This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 10 failed / 86 succeeded
Started2020-01-15 15:25
Elapsed4h12m
Revision
Buildergke-prow-default-pool-cf4891d4-m8gw
links{u'resultstore': {u'url': u'https://source.cloud.google.com/results/invocations/6203691d-3b3e-4ca6-a21c-a2184e43f46c/targets/test'}}
pod102c98d7-37ab-11ea-8603-da2f7a5855b4
resultstorehttps://source.cloud.google.com/results/invocations/6203691d-3b3e-4ca6-a21c-a2184e43f46c/targets/test
infra-commitff065cbb3
job-versionv1.18.0-alpha.1.755+05209312b74eac
pod102c98d7-37ab-11ea-8603-da2f7a5855b4
repok8s.io/kubernetes
repo-commit05209312b74eac15f0a80577cbbdcb11bce9b8ad
repos{u'k8s.io/kubernetes': u'master', u'github.com/containerd/cri': u'master'}
revisionv1.18.0-alpha.1.755+05209312b74eac

Test Failures


E2eNode Suite [k8s.io] CriticalPod [Serial] [Disruptive] [NodeFeature:CriticalPod] when we need to admit a critical pod should be able to create and delete a critical pod 0.12s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sCriticalPod\s\[Serial\]\s\[Disruptive\]\s\[NodeFeature\:CriticalPod\]\swhen\swe\sneed\sto\sadmit\sa\scritical\spod\sshould\sbe\sable\sto\screate\sand\sdelete\sa\scritical\spod$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/critical_pod_test.go:54
failed to create PriorityClasses with an error: PriorityClass.scheduling.k8s.io "critical-pod-test-high-priority" is invalid: value: Forbidden: maximum allowed value of a user defined priority is 1000000000
Expected
    <bool>: false
to equal
    <bool>: true
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/critical_pod_test.go:88
				
				Click to see stdout/stderrfrom junit_cos-stable_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] CriticalPod [Serial] [Disruptive] [NodeFeature:CriticalPod] when we need to admit a critical pod should be able to create and delete a critical pod 0.29s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sCriticalPod\s\[Serial\]\s\[Disruptive\]\s\[NodeFeature\:CriticalPod\]\swhen\swe\sneed\sto\sadmit\sa\scritical\spod\sshould\sbe\sable\sto\screate\sand\sdelete\sa\scritical\spod$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/critical_pod_test.go:54
failed to create PriorityClasses with an error: PriorityClass.scheduling.k8s.io "critical-pod-test-high-priority" is invalid: value: Forbidden: maximum allowed value of a user defined priority is 1000000000
Expected
    <bool>: false
to equal
    <bool>: true
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/critical_pod_test.go:88
				
				Click to see stdout/stderrfrom junit_ubuntu_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Device Plugin [Feature:DevicePluginProbe][NodeFeature:DevicePluginProbe][Serial] DevicePlugin Verifies the Kubelet device plugin functionality. 35s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sDevice\sPlugin\s\[Feature\:DevicePluginProbe\]\[NodeFeature\:DevicePluginProbe\]\[Serial\]\sDevicePlugin\sVerifies\sthe\sKubelet\sdevice\splugin\sfunctionality\.$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/device_plugin_test.go:115
Expected
    <int>: 3
to equal
    <int>: 2
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/device_plugin_test.go:158
				
				Click to see stdout/stderrfrom junit_ubuntu_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] Device Plugin [Feature:DevicePluginProbe][NodeFeature:DevicePluginProbe][Serial] DevicePlugin Verifies the Kubelet device plugin functionality. 36s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sDevice\sPlugin\s\[Feature\:DevicePluginProbe\]\[NodeFeature\:DevicePluginProbe\]\[Serial\]\sDevicePlugin\sVerifies\sthe\sKubelet\sdevice\splugin\sfunctionality\.$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/device_plugin_test.go:115
Expected
    <int>: 3
to equal
    <int>: 2
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/device_plugin_test.go:158
				
				Click to see stdout/stderrfrom junit_cos-stable_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] InodeEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 16m26s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sInodeEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:468
priority 0 pod: innocent-pod failed
Expected
    <v1.PodPhase>: Failed
not to equal
    <v1.PodPhase>: Failed
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:631
				
				Click to see stdout/stderrfrom junit_ubuntu_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 14m22s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:468
Timed out after 600.000s.
Expected
    <*errors.errorString | 0xc0017aa4a0>: {
        s: "pods that should be evicted are still running",
    }
to be nil
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:490
				
				Click to see stdout/stderrfrom junit_cos-stable_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageSoftEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 11m16s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageSoftEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:468
priority 0 pod: innocent-pod failed
Expected
    <v1.PodPhase>: Failed
not to equal
    <v1.PodPhase>: Failed
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:631
				
				Click to see stdout/stderrfrom junit_ubuntu_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageSoftEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 14m34s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageSoftEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:468
Timed out after 600.000s.
Expected
    <*errors.errorString | 0xc0017db3a0>: {
        s: "pods that should be evicted are still running",
    }
to be nil
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:490