This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 5 failed / 48 succeeded
Started2020-01-14 02:00
Elapsed3h59m
Revision
Buildergke-prow-default-pool-cf4891d4-2j9c
links{u'resultstore': {u'url': u'https://source.cloud.google.com/results/invocations/28ae8e95-de30-4183-b625-19d5e2c297f4/targets/test'}}
pod8111791c-3671-11ea-9f20-3687633bf296
resultstorehttps://source.cloud.google.com/results/invocations/28ae8e95-de30-4183-b625-19d5e2c297f4/targets/test
infra-commite65b3228e
job-versionv1.15.8-beta.1.30+14ede42c4fe699
pod8111791c-3671-11ea-9f20-3687633bf296
repok8s.io/kubernetes
repo-commit14ede42c4fe699a7078b566d89abc160f26857a2
repos{u'k8s.io/kubernetes': u'release-1.15'}
revisionv1.15.8-beta.1.30+14ede42c4fe699

Test Failures


E2eNode Suite [k8s.io] Downward API [Serial] [Disruptive] [NodeFeature:EphemeralStorage] Downward API tests for local ephemeral storage should provide default limits.ephemeral-storage from node allocatable 5m8s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sDownward\sAPI\s\[Serial\]\s\[Disruptive\]\s\[NodeFeature\:EphemeralStorage\]\sDownward\sAPI\stests\sfor\slocal\sephemeral\sstorage\sshould\sprovide\sdefault\slimits\.ephemeral\-storage\sfrom\snode\sallocatable$'
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/common/downward_api.go:271
Unexpected error:
    <*errors.errorString | 0xc00031a5c0>: {
        s: "expected pod \"downward-api-9abda5f3-18ec-4d3b-b72c-77c1c6ed57cc\" success: Gave up after waiting 5m0s for pod \"downward-api-9abda5f3-18ec-4d3b-b72c-77c1c6ed57cc\" to be \"success or failure\"",
    }
    expected pod "downward-api-9abda5f3-18ec-4d3b-b72c-77c1c6ed57cc" success: Gave up after waiting 5m0s for pod "downward-api-9abda5f3-18ec-4d3b-b72c-77c1c6ed57cc" to be "success or failure"
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/util.go:2342
				
				Click to see stdout/stderrfrom junit_ubuntu-gke-1804-d1809-0-v20200110_01.xml

Find downward-api-9abda5f3-18ec-4d3b-b72c-77c1c6ed57cc mentions in log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageCapacityIsolationEviction [Slow] [Serial] [Disruptive] [Feature:LocalStorageCapacityIsolation][NodeFeature:Eviction] when we run containers that should cause evictions due to pod local storage violations should eventually evict all of the correct pods 6m34s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageCapacityIsolationEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\s\[Feature\:LocalStorageCapacityIsolation\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sevictions\sdue\sto\spod\slocal\sstorage\sviolations\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:455
Unexpected error:
    <*errors.errorString | 0xc000216500>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/pods.go:113
				
				Click to see stdout/stderrfrom junit_ubuntu-gke-1804-d1809-0-v20200110_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] LocalStorageEviction [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 6m37s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sLocalStorageEviction\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:455
Unexpected error:
    <*errors.errorString | 0xc000216500>: {
        s: "timed out waiting for the condition",
    }
    timed out waiting for the condition
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/pods.go:113
				
				Click to see stdout/stderrfrom junit_ubuntu-gke-1804-d1809-0-v20200110_01.xml

Filter through log files | View test history on testgrid


E2eNode Suite [k8s.io] PriorityLocalStorageEvictionOrdering [Slow] [Serial] [Disruptive][NodeFeature:Eviction] when we run containers that should cause DiskPressure should eventually evict all of the correct pods 4m57s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=E2eNode\sSuite\s\[k8s\.io\]\sPriorityLocalStorageEvictionOrdering\s\[Slow\]\s\[Serial\]\s\[Disruptive\]\[NodeFeature\:Eviction\]\swhen\swe\srun\scontainers\sthat\sshould\scause\sDiskPressure\s\sshould\seventually\sevict\sall\sof\sthe\scorrect\spods$'
_output/local/go/src/k8s.io/kubernetes/test/e2e_node/eviction_test.go:523
Unexpected error:
    <*errors.errorString | 0xc000415240>: {
        s: "pod ran to completion",
    }
    pod ran to completion
occurred
/go/src/k8s.io/kubernetes/_output/local/go/src/k8s.io/kubernetes/test/e2e/framework/pods.go:113
				
				Click to see stdout/stderrfrom junit_ubuntu-gke-1804-d1809-0-v20200110_01.xml

Find ran mentions in log files | View test history on testgrid


Node Tests 3h58m

error during go run /go/src/k8s.io/kubernetes/test/e2e_node/runner/remote/run_remote.go --cleanup --logtostderr --vmodule=*=4 --ssh-env=gce --results-dir=/workspace/_artifacts --project=ubuntu-image-validation --zone=us-west1-b --ssh-user=prow --ssh-key=/workspace/.ssh/google_compute_engine --ginkgo-flags=--nodes=1 --focus="\[Serial\]" --skip="\[Flaky\]|\[Benchmark\]|\[NodeAlphaFeature:.+\]" --test_args=--feature-gates=DynamicKubeletConfig=true --test-timeout=5h0m0s --images=ubuntu-gke-1804-d1809-0-v20200110 --image-project=ubuntu-os-gke-cloud-devel: exit status 1
				from junit_runner.xml

Filter through log files | View test history on testgrid


Show 48 Passed Tests

Show 264 Skipped Tests