This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 17 failed / 14 succeeded
Started2020-10-20 11:13
Elapsed5h11m
Revision
Builder334eecbc-12c5-11eb-a3cf-6ef374375bd7
infra-commit237fbef29
job-versionv1.20.0-alpha.2.191+6d3ccd8e6c409a
master_os_imagecos-85-13310-1041-9
node_os_imagecos-85-13310-1041-9
revisionv1.20.0-alpha.2.191+6d3ccd8e6c409a

Test Failures


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] Shouldn't perform scale up operation and should list unhealthy status if most of the cluster is broken[Feature:ClusterSizeAutoscalingScaleUp] 17m54s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sShouldn\'t\sperform\sscale\sup\soperation\sand\sshould\slist\sunhealthy\sstatus\sif\smost\sof\sthe\scluster\sis\sbroken\[Feature\:ClusterSizeAutoscalingScaleUp\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:881
Oct 20 12:23:03.787: Unexpected error:
    <*errors.StatusError | 0xc000f5cb40>: {
        ErrStatus: {
            TypeMeta: {Kind: "", APIVersion: ""},
            ListMeta: {
                SelfLink: "",
                ResourceVersion: "",
                Continue: "",
                RemainingItemCount: nil,
            },
            Status: "Failure",
            Message: "configmaps \"cluster-autoscaler-status\" not found",
            Reason: "NotFound",
            Details: {
                Name: "cluster-autoscaler-status",
                Group: "",
                Kind: "configmaps",
                UID: "",
                Causes: nil,
                RetryAfterSeconds: 0,
            },
            Code: 404,
        },
    }
    configmaps "cluster-autoscaler-status" not found
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:930
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should add node to the particular mig [Feature:ClusterSizeAutoscalingScaleUp] 38m41s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sadd\snode\sto\sthe\sparticular\smig\s\[Feature\:ClusterSizeAutoscalingScaleUp\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:538
Oct 20 12:07:41.848: Unexpected error:
    <*errors.errorString | 0xc0008c8020>: {
        s: "timeout waiting 5m0s for appropriate cluster size",
    }
    timeout waiting 5m0s for appropriate cluster size
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:583
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should be able to scale down by draining multiple pods one by one as dictated by pdb[Feature:ClusterSizeAutoscalingScaleDown] 24m36s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sbe\sable\sto\sscale\sdown\sby\sdraining\smultiple\spods\sone\sby\sone\sas\sdictated\sby\spdb\[Feature\:ClusterSizeAutoscalingScaleDown\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:733
Oct 20 15:11:06.603: Unexpected error:
    <*errors.errorString | 0xc0001556f0>: {
        s: "timeout waiting 20m0s for appropriate cluster size",
    }
    timeout waiting 20m0s for appropriate cluster size
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:736
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should be able to scale down by draining system pods with pdb[Feature:ClusterSizeAutoscalingScaleDown] 24m37s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sbe\sable\sto\sscale\sdown\sby\sdraining\ssystem\spods\swith\spdb\[Feature\:ClusterSizeAutoscalingScaleDown\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:741
Oct 20 13:18:34.093: Unexpected error:
    <*errors.errorString | 0xc002bd4060>: {
        s: "timeout waiting 20m0s for appropriate cluster size",
    }
    timeout waiting 20m0s for appropriate cluster size
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:744
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should correctly scale down after a node is not needed [Feature:ClusterSizeAutoscalingScaleDown] 24m3s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\scorrectly\sscale\sdown\safter\sa\snode\sis\snot\sneeded\s\[Feature\:ClusterSizeAutoscalingScaleDown\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:684
Oct 20 14:07:06.175: Unexpected error:
    <*errors.errorString | 0xc001b2a090>: {
        s: "timeout waiting 20m0s for appropriate cluster size",
    }
    timeout waiting 20m0s for appropriate cluster size
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:680
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should correctly scale down after a node is not needed and one node is broken [Feature:ClusterSizeAutoscalingScaleDown] 24m45s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\scorrectly\sscale\sdown\safter\sa\snode\sis\snot\sneeded\sand\sone\snode\sis\sbroken\s\[Feature\:ClusterSizeAutoscalingScaleDown\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:687
Oct 20 12:48:20.984: Unexpected error:
    <*errors.errorString | 0xc00160c320>: {
        s: "timeout waiting 20m0s for appropriate cluster size",
    }
    timeout waiting 20m0s for appropriate cluster size
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:680
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pending pods are small and one node is broken [Feature:ClusterSizeAutoscalingScaleUp] 7m53s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sincrease\scluster\ssize\sif\spending\spods\sare\ssmall\sand\sone\snode\sis\sbroken\s\[Feature\:ClusterSizeAutoscalingScaleUp\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:325
Oct 20 15:52:07.666: Unexpected error:
    <*errors.errorString | 0xc0039d0240>: {
        s: "timeout waiting 5m0s for appropriate cluster size",
    }
    timeout waiting 5m0s for appropriate cluster size
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:199
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pod requesting EmptyDir volume is pending [Feature:ClusterSizeAutoscalingScaleUp] 5m41s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sincrease\scluster\ssize\sif\spod\srequesting\sEmptyDir\svolume\sis\spending\s\[Feature\:ClusterSizeAutoscalingScaleUp\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:445
Oct 20 12:56:00.838: Unexpected error:
    <*errors.errorString | 0xc0009b29d0>: {
        s: "Only 0 pods started out of 1",
    }
    Only 0 pods started out of 1
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:459
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pod requesting volume is pending [Feature:ClusterSizeAutoscalingScaleUp] 10m39s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sincrease\scluster\ssize\sif\spod\srequesting\svolume\sis\spending\s\[Feature\:ClusterSizeAutoscalingScaleUp\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:466
Oct 20 15:32:58.227: Unexpected error:
    <*errors.errorString | 0xc0037f6410>: {
        s: "Only 0 pods started out of 1",
    }
    Only 0 pods started out of 1
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:528
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pods are pending due to host port conflict [Feature:ClusterSizeAutoscalingScaleUp] 8m36s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sincrease\scluster\ssize\sif\spods\sare\spending\sdue\sto\shost\sport\sconflict\s\[Feature\:ClusterSizeAutoscalingScaleUp\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:417
Oct 20 14:42:47.787: Unexpected error:
    <*errors.errorString | 0xc003f68780>: {
        s: "timeout waiting 5m0s for appropriate cluster size",
    }
    timeout waiting 5m0s for appropriate cluster size
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:421
				
				Click to see stdout/stderrfrom junit_01.xml

Filter through log files | View test history on testgrid


Kubernetes e2e suite [sig-autoscaling] Cluster size autoscaling [Slow] should increase cluster size if pods are pending due to pod anti-affinity [Feature:ClusterSizeAutoscalingScaleUp] 5m39s

go run hack/e2e.go -v --test --test_args='--ginkgo.focus=Kubernetes\se2e\ssuite\s\[sig\-autoscaling\]\sCluster\ssize\sautoscaling\s\[Slow\]\sshould\sincrease\scluster\ssize\sif\spods\sare\spending\sdue\sto\spod\santi\-affinity\s\[Feature\:ClusterSizeAutoscalingScaleUp\]$'
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:426
Oct 20 14:48:33.844: Unexpected error:
    <*errors.errorString | 0xc0030704a0>: {
        s: "Only 0 pods started out of 2",
    }
    Only 0 pods started out of 2
occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/autoscaling/cluster_size_autoscaling.go:438