PR | andyxning: do not create StatefulSet pods when PVC is being deleted |
Result | FAILURE |
Tests | 0 failed / 95 succeeded |
Started | |
Elapsed | 12m44s |
Revision | 677073960427075d3da7d791bed84224e8d1898b |
Refs |
98732 |
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdCompletion
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdCompletion/shell_not_expected
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdCompletion/unsupported_shell_type
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitAPIPort
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitAPIPort/accept_a_valid_port_number
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitAPIPort/fail_on_negative_port_number
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitAPIPort/fail_on_non-string_port
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitAPIPort/fail_on_too_large_port_number
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitCertPhaseCSR
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitCertPhaseCSR/fails_on_CSR
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitCertPhaseCSR/fails_on_all
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitCertPhaseCSR/generate_CSR
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/can't_load_old_component_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/can't_load_old_v1alpha1_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/can't_load_old_v1alpha2_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/can't_load_old_v1alpha3_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/can_load_current_component_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/can_load_v1beta1_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/can_load_v1beta2_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/don't_allow_mixed_arguments_v1beta1
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/don't_allow_mixed_arguments_v1beta2
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitConfig/fail_on_non_existing_path
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitFeatureGates
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitFeatureGates/feature_gate_IPv6DualStack=true
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitFeatureGates/feature_gate_PublicKeysECDSA=true
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitFeatureGates/no_feature_gates_passed
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitKubernetesVersion
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitKubernetesVersion/invalid_semantic_version_string_is_detected
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitKubernetesVersion/valid_version_is_accepted
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitToken
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitToken/invalid_token_non-lowercase
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitToken/invalid_token_size
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdInitToken/valid_token_is_accepted
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinArgsMixed
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinArgsMixed/discovery-token_and_config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinBadArgs
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinBadArgs/discovery-token_and_discovery-file_can't_both_be_set
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinBadArgs/discovery-token_or_discovery-file_must_be_set
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinConfig
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinConfig/config
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinConfig/config_path
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinDiscoveryFile
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinDiscoveryFile/invalid_discovery_file
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinDiscoveryFile/valid_discovery_file
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinDiscoveryToken
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinDiscoveryToken/valid_discovery_token
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinDiscoveryToken/valid_discovery_token_url
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinNodeName
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinNodeName/valid_node_name
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinTLSBootstrapToken
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinTLSBootstrapToken/valid_bootstrap_token
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinTLSBootstrapToken/valid_bootstrap_token_url
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinToken
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinToken/valid_token
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdJoinToken/valid_token_url
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdTokenDelete
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdTokenDelete/invalid_token
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdTokenDelete/no_token_provided
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdTokenGenerate
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdTokenGenerateTypoError
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdVersion
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdVersion/default_output
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdVersion/invalid_output_option
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdVersion/short_output
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdVersionOutputJsonOrYaml
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdVersionOutputJsonOrYaml/json_output
k8s.io/kubernetes/cmd/kubeadm/test/cmd TestCmdVersionOutputJsonOrYaml/yaml_output
test-cmd run_RESTMapper_evaluation_tests
test-cmd run_assert_categories_tests
test-cmd run_assert_short_name_tests
test-cmd run_clusterroles_tests
test-cmd run_create_secret_tests
test-cmd run_kubectl_apply_deployments_tests
test-cmd run_kubectl_apply_tests
test-cmd run_kubectl_config_set_cluster_tests
test-cmd run_kubectl_config_set_credentials_tests
test-cmd run_kubectl_config_set_tests
test-cmd run_kubectl_create_error_tests
test-cmd run_kubectl_create_filter_tests
test-cmd run_kubectl_create_kustomization_directory_tests
test-cmd run_kubectl_delete_allnamespaces_tests
test-cmd run_kubectl_diff_same_names
test-cmd run_kubectl_diff_tests
test-cmd run_kubectl_exec_pod_tests
test-cmd run_kubectl_exec_resource_name_tests
test-cmd run_kubectl_get_tests
test-cmd run_kubectl_local_proxy_tests
test-cmd run_kubectl_request_timeout_tests
test-cmd run_kubectl_run_tests
test-cmd run_kubectl_server_side_apply_tests
test-cmd run_kubectl_version_tests
test-cmd run_pod_tests
test-cmd run_role_tests
test-cmd run_save_config_tests
... skipping 61 lines ... Recording: record_command_canary Running command: record_command_canary +++ Running case: test-cmd.record_command_canary +++ working dir: /home/prow/go/src/k8s.io/kubernetes +++ command: record_command_canary /home/prow/go/src/k8s.io/kubernetes/test/cmd/legacy-script.sh: line 155: bogus-expected-to-fail: command not found !!! [0408 07:15:01] Call tree: !!! [0408 07:15:01] 1: /home/prow/go/src/k8s.io/kubernetes/test/cmd/../../third_party/forked/shell2junit/sh2ju.sh:47 record_command_canary(...) !!! [0408 07:15:01] 2: /home/prow/go/src/k8s.io/kubernetes/test/cmd/../../third_party/forked/shell2junit/sh2ju.sh:112 eVal(...) !!! [0408 07:15:01] 3: /home/prow/go/src/k8s.io/kubernetes/test/cmd/legacy-script.sh:131 juLog(...) !!! [0408 07:15:01] 4: /home/prow/go/src/k8s.io/kubernetes/test/cmd/legacy-script.sh:159 record_command(...) !!! [0408 07:15:01] 5: hack/make-rules/test-cmd.sh:35 source(...) +++ exit code: 1 +++ error: 1 +++ [0408 07:15:01] Running kubeadm tests +++ [0408 07:15:07] Building go targets for linux/amd64: cmd/kubeadm +++ [0408 07:15:53] Running tests without code coverage {"Time":"2021-04-08T07:17:23.229588833Z","Action":"output","Package":"k8s.io/kubernetes/cmd/kubeadm/test/cmd","Output":"ok \tk8s.io/kubernetes/cmd/kubeadm/test/cmd\t52.196s\n"} ✓ cmd/kubeadm/test/cmd (52.199s) ... skipping 320 lines ... +++ [0408 07:19:12] Building kube-controller-manager +++ [0408 07:19:17] Building go targets for linux/amd64: cmd/kube-controller-manager +++ [0408 07:19:43] Starting controller-manager Flag --port has been deprecated, see --secure-port instead. I0408 07:19:44.905093 58006 serving.go:331] Generated self-signed cert in-memory W0408 07:19:45.295985 58006 authentication.go:368] failed to read in-cluster kubeconfig for delegated authentication: open /var/run/secrets/kubernetes.io/serviceaccount/token: no such file or directory W0408 07:19:45.296054 58006 authentication.go:265] No authentication-kubeconfig provided in order to lookup client-ca-file in configmap/extension-apiserver-authentication in kube-system, so client certificate authentication won't work. W0408 07:19:45.296065 58006 authentication.go:289] No authentication-kubeconfig provided in order to lookup requestheader-client-ca-file in configmap/extension-apiserver-authentication in kube-system, so request-header client certificate authentication won't work. W0408 07:19:45.296132 58006 authorization.go:187] failed to read in-cluster kubeconfig for delegated authorization: open /var/run/secrets/kubernetes.io/serviceaccount/token: no such file or directory W0408 07:19:45.296158 58006 authorization.go:156] No authorization-kubeconfig provided, so SubjectAccessReview of authorization tokens won't work. I0408 07:19:45.296184 58006 controllermanager.go:175] Version: v1.19.10-rc.0.47+df994f89b54ffc I0408 07:19:45.297671 58006 secure_serving.go:197] Serving securely on [::]:10257 I0408 07:19:45.297795 58006 tlsconfig.go:240] Starting DynamicServingCertificateController I0408 07:19:45.298512 58006 deprecated_insecure_serving.go:53] Serving insecurely on [::]:10252 I0408 07:19:45.298641 58006 leaderelection.go:243] attempting to acquire leader lease kube-system/kube-controller-manager... ... skipping 15 lines ... I0408 07:19:45.682876 58006 shared_informer.go:240] Waiting for caches to sync for ReplicaSet I0408 07:19:45.683221 58006 controllermanager.go:549] Started "cronjob" W0408 07:19:45.683251 58006 controllermanager.go:541] Skipping "csrsigning" I0408 07:19:45.683380 58006 cronjob_controller.go:96] Starting CronJob Manager W0408 07:19:45.683543 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. W0408 07:19:45.683619 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. E0408 07:19:45.683651 58006 core.go:90] Failed to start service controller: WARNING: no cloud provider provided, services of type LoadBalancer will fail W0408 07:19:45.683671 58006 controllermanager.go:541] Skipping "service" W0408 07:19:45.684115 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. I0408 07:19:45.684181 58006 controllermanager.go:549] Started "endpoint" I0408 07:19:45.684288 58006 endpoints_controller.go:184] Starting endpoint controller I0408 07:19:45.684308 58006 shared_informer.go:240] Waiting for caches to sync for endpoint W0408 07:19:45.684707 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. I0408 07:19:45.684747 58006 controllermanager.go:549] Started "horizontalpodautoscaling" W0408 07:19:45.684756 58006 controllermanager.go:528] "tokencleaner" is disabled I0408 07:19:45.684774 58006 horizontal.go:169] Starting HPA controller I0408 07:19:45.684784 58006 shared_informer.go:240] Waiting for caches to sync for HPA I0408 07:19:45.684988 58006 node_lifecycle_controller.go:77] Sending events to api server E0408 07:19:45.685050 58006 core.go:230] failed to start cloud node lifecycle controller: no cloud provider provided W0408 07:19:45.685070 58006 controllermanager.go:541] Skipping "cloud-node-lifecycle" W0408 07:19:45.685303 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. W0408 07:19:45.685331 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. W0408 07:19:45.685343 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. I0408 07:19:45.685441 58006 controllermanager.go:549] Started "statefulset" I0408 07:19:45.685742 58006 stateful_set.go:146] Starting stateful set controller ... skipping 120 lines ... I0408 07:19:46.063176 58006 controllermanager.go:549] Started "ttl" I0408 07:19:46.064218 58006 ttl_controller.go:118] Starting TTL controller I0408 07:19:46.064235 58006 shared_informer.go:240] Waiting for caches to sync for TTL I0408 07:19:46.083417 58006 shared_informer.go:247] Caches are synced for service account I0408 07:19:46.090966 58006 shared_informer.go:247] Caches are synced for ClusterRoleAggregator I0408 07:19:46.091872 54487 controller.go:609] quota admission added evaluator for: serviceaccounts E0408 07:19:46.110895 58006 clusterroleaggregation_controller.go:181] view failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "view": the object has been modified; please apply your changes to the latest version and try again E0408 07:19:46.111302 58006 clusterroleaggregation_controller.go:181] edit failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "edit": the object has been modified; please apply your changes to the latest version and try again E0408 07:19:46.123526 58006 clusterroleaggregation_controller.go:181] edit failed with : Operation cannot be fulfilled on clusterroles.rbac.authorization.k8s.io "edit": the object has been modified; please apply your changes to the latest version and try again I0408 07:19:46.145021 58006 shared_informer.go:247] Caches are synced for PVC protection I0408 07:19:46.145708 58006 shared_informer.go:247] Caches are synced for PV protection I0408 07:19:46.148696 58006 shared_informer.go:247] Caches are synced for deployment I0408 07:19:46.148762 58006 shared_informer.go:247] Caches are synced for expand I0408 07:19:46.148773 58006 shared_informer.go:247] Caches are synced for certificate-csrapproving I0408 07:19:46.148799 58006 shared_informer.go:247] Caches are synced for disruption ... skipping 9 lines ... I0408 07:19:46.188210 58006 shared_informer.go:247] Caches are synced for persistent volume I0408 07:19:46.188864 58006 shared_informer.go:247] Caches are synced for daemon sets I0408 07:19:46.249591 58006 shared_informer.go:247] Caches are synced for GC I0408 07:19:46.264366 58006 shared_informer.go:247] Caches are synced for TTL I0408 07:19:46.285105 58006 shared_informer.go:247] Caches are synced for HPA node/127.0.0.1 created W0408 07:19:46.323133 58006 actual_state_of_world.go:534] Failed to update statusUpdateNeeded field in actual state of world: Failed to set statusUpdateNeeded to needed true, because nodeName="127.0.0.1" does not exist +++ [0408 07:19:46] Checking kubectl version I0408 07:19:46.386378 58006 shared_informer.go:247] Caches are synced for endpoint_slice_mirroring I0408 07:19:46.388545 58006 shared_informer.go:247] Caches are synced for endpoint_slice Client Version: version.Info{Major:"1", Minor:"19+", GitVersion:"v1.19.10-rc.0.47+df994f89b54ffc", GitCommit:"df994f89b54ffcdcadfb365e573788a2545cb8e6", GitTreeState:"clean", BuildDate:"2021-04-08T06:19:44Z", GoVersion:"go1.15.10", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"19+", GitVersion:"v1.19.10-rc.0.47+df994f89b54ffc", GitCommit:"df994f89b54ffcdcadfb365e573788a2545cb8e6", GitTreeState:"clean", BuildDate:"2021-04-08T06:19:44Z", GoVersion:"go1.15.10", Compiler:"gc", Platform:"linux/amd64"} I0408 07:19:46.444141 58006 shared_informer.go:247] Caches are synced for resource quota ... skipping 127 lines ... +++ working dir: /home/prow/go/src/k8s.io/kubernetes +++ command: run_RESTMapper_evaluation_tests +++ [0408 07:19:52] Creating namespace namespace-1617866392-13550 namespace/namespace-1617866392-13550 created Context "test" modified. +++ [0408 07:19:52] Testing RESTMapper +++ [0408 07:19:53] "kubectl get unknownresourcetype" returns error as expected: error: the server doesn't have a resource type "unknownresourcetype" +++ exit code: 0 NAME SHORTNAMES APIGROUP NAMESPACED KIND bindings true Binding componentstatuses cs false ComponentStatus configmaps cm true ConfigMap endpoints ep true Endpoints ... skipping 59 lines ... namespace/namespace-1617866398-22348 created Context "test" modified. +++ [0408 07:19:58] Testing clusterroles [32mrbac.sh:29: Successful get clusterroles/cluster-admin {{.metadata.name}}: cluster-admin (B[m[32mrbac.sh:30: Successful get clusterrolebindings/cluster-admin {{.metadata.name}}: cluster-admin (B[mSuccessful message:Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pod-admin" not found has:clusterroles.rbac.authorization.k8s.io "pod-admin" not found clusterrole.rbac.authorization.k8s.io/pod-admin created (dry run) clusterrole.rbac.authorization.k8s.io/pod-admin created (server dry run) Successful message:Error from server (NotFound): clusterroles.rbac.authorization.k8s.io "pod-admin" not found has:clusterroles.rbac.authorization.k8s.io "pod-admin" not found clusterrole.rbac.authorization.k8s.io/pod-admin created [32mrbac.sh:42: Successful get clusterrole/pod-admin {{range.rules}}{{range.verbs}}{{.}}:{{end}}{{end}}: *: (B[mSuccessful message:warning: deleting cluster-scoped resources, not scoped to the provided namespace clusterrole.rbac.authorization.k8s.io "pod-admin" deleted ... skipping 18 lines ... (B[mclusterrole.rbac.authorization.k8s.io/url-reader created [32mrbac.sh:61: Successful get clusterrole/url-reader {{range.rules}}{{range.verbs}}{{.}}:{{end}}{{end}}: get: (B[m[32mrbac.sh:62: Successful get clusterrole/url-reader {{range.rules}}{{range.nonResourceURLs}}{{.}}:{{end}}{{end}}: /logs/*:/healthz/*: (B[mclusterrole.rbac.authorization.k8s.io/aggregation-reader created [32mrbac.sh:64: Successful get clusterrole/aggregation-reader {{.metadata.name}}: aggregation-reader (B[mSuccessful message:Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "super-admin" not found has:clusterrolebindings.rbac.authorization.k8s.io "super-admin" not found clusterrolebinding.rbac.authorization.k8s.io/super-admin created (dry run) clusterrolebinding.rbac.authorization.k8s.io/super-admin created (server dry run) Successful message:Error from server (NotFound): clusterrolebindings.rbac.authorization.k8s.io "super-admin" not found has:clusterrolebindings.rbac.authorization.k8s.io "super-admin" not found clusterrolebinding.rbac.authorization.k8s.io/super-admin created [32mrbac.sh:77: Successful get clusterrolebinding/super-admin {{range.subjects}}{{.name}}:{{end}}: super-admin: (B[mclusterrolebinding.rbac.authorization.k8s.io/super-admin subjects updated (dry run) clusterrolebinding.rbac.authorization.k8s.io/super-admin subjects updated (server dry run) [32mrbac.sh:80: Successful get clusterrolebinding/super-admin {{range.subjects}}{{.name}}:{{end}}: super-admin: ... skipping 59 lines ... [32mrbac.sh:102: Successful get clusterrolebinding/super-admin {{range.subjects}}{{.name}}:{{end}}: super-admin:foo:test-all-user: (B[m[32mrbac.sh:103: Successful get clusterrolebinding/super-group {{range.subjects}}{{.name}}:{{end}}: the-group:foo:test-all-user: (B[m[32mrbac.sh:104: Successful get clusterrolebinding/super-sa {{range.subjects}}{{.name}}:{{end}}: sa-name:foo:test-all-user: (B[mrolebinding.rbac.authorization.k8s.io/admin created (dry run) rolebinding.rbac.authorization.k8s.io/admin created (server dry run) Successful message:Error from server (NotFound): rolebindings.rbac.authorization.k8s.io "admin" not found has: not found rolebinding.rbac.authorization.k8s.io/admin created [32mrbac.sh:113: Successful get rolebinding/admin {{.roleRef.kind}}: ClusterRole (B[m[32mrbac.sh:114: Successful get rolebinding/admin {{range.subjects}}{{.name}}:{{end}}: default-admin: (B[mrolebinding.rbac.authorization.k8s.io/admin subjects updated [32mrbac.sh:116: Successful get rolebinding/admin {{range.subjects}}{{.name}}:{{end}}: default-admin:foo: ... skipping 29 lines ... message:Warning: rbac.authorization.k8s.io/v1beta1 Role is deprecated in v1.17+, unavailable in v1.22+; use rbac.authorization.k8s.io/v1 Role No resources found in namespace-1617866407-30296 namespace. has:Role is deprecated Successful message:Warning: rbac.authorization.k8s.io/v1beta1 Role is deprecated in v1.17+, unavailable in v1.22+; use rbac.authorization.k8s.io/v1 Role No resources found in namespace-1617866407-30296 namespace. Error: 1 warning received has:Role is deprecated Successful message:Warning: rbac.authorization.k8s.io/v1beta1 Role is deprecated in v1.17+, unavailable in v1.22+; use rbac.authorization.k8s.io/v1 Role No resources found in namespace-1617866407-30296 namespace. Error: 1 warning received has:Error: 1 warning received role.rbac.authorization.k8s.io/pod-admin created (dry run) role.rbac.authorization.k8s.io/pod-admin created (server dry run) Successful message:Error from server (NotFound): roles.rbac.authorization.k8s.io "pod-admin" not found has: not found role.rbac.authorization.k8s.io/pod-admin created [32mrbac.sh:163: Successful get role/pod-admin {{range.rules}}{{range.verbs}}{{.}}:{{end}}{{end}}: *: (B[m[32mrbac.sh:164: Successful get role/pod-admin {{range.rules}}{{range.resources}}{{.}}:{{end}}{{end}}: pods: (B[m[32mrbac.sh:165: Successful get role/pod-admin {{range.rules}}{{range.apiGroups}}{{.}}:{{end}}{{end}}: : (B[mSuccessful ... skipping 464 lines ... has:valid-pod Successful message:NAME READY STATUS RESTARTS AGE valid-pod 0/1 Pending 0 0s has:valid-pod [32mcore.sh:190: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: valid-pod: (B[merror: resource(s) were provided, but no name, label selector, or --all flag specified [32mcore.sh:194: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: valid-pod: (B[m[32mcore.sh:198: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: valid-pod: (B[merror: setting 'all' parameter but found a non empty selector. [32mcore.sh:202: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: valid-pod: (B[m[32mcore.sh:206: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: valid-pod: (B[mwarning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. pod "valid-pod" force deleted [32mcore.sh:210: Successful get pods -l'name in (valid-pod)' {{range.items}}{{.metadata.name}}:{{end}}: (B[m[32mcore.sh:215: Successful get namespaces {{range.items}}{{ if eq .metadata.name \"test-kubectl-describe-pod\" }}found{{end}}{{end}}:: : ... skipping 19 lines ... (B[mpoddisruptionbudget.policy/test-pdb-2 created [32mcore.sh:259: Successful get pdb/test-pdb-2 --namespace=test-kubectl-describe-pod {{.spec.minAvailable}}: 50% (B[mpoddisruptionbudget.policy/test-pdb-3 created [32mcore.sh:265: Successful get pdb/test-pdb-3 --namespace=test-kubectl-describe-pod {{.spec.maxUnavailable}}: 2 (B[mpoddisruptionbudget.policy/test-pdb-4 created [32mcore.sh:269: Successful get pdb/test-pdb-4 --namespace=test-kubectl-describe-pod {{.spec.maxUnavailable}}: 50% (B[merror: min-available and max-unavailable cannot be both specified [32mcore.sh:275: Successful get pods --namespace=test-kubectl-describe-pod {{range.items}}{{.metadata.name}}:{{end}}: (B[mpod/env-test-pod created matched TEST_CMD_1 matched <set to the key 'key-1' in secret 'test-secret'> matched TEST_CMD_2 matched <set to the key 'key-2' of config map 'test-configmap'> ... skipping 221 lines ... [32mcore.sh:534: Successful get pods {{range.items}}{{(index .spec.containers 0).image}}:{{end}}: k8s.gcr.io/pause:3.2: (B[mSuccessful message:kubectl-create kubectl-patch has:kubectl-patch pod/valid-pod patched [32mcore.sh:554: Successful get pods {{range.items}}{{(index .spec.containers 0).image}}:{{end}}: nginx: (B[m+++ [0408 07:20:40] "kubectl patch with resourceVersion 562" returns error as expected: Error from server (Conflict): Operation cannot be fulfilled on pods "valid-pod": the object has been modified; please apply your changes to the latest version and try again pod "valid-pod" deleted pod/valid-pod replaced [32mcore.sh:578: Successful get pod valid-pod {{(index .spec.containers 0).name}}: replaced-k8s-serve-hostname (B[mSuccessful message:kubectl-create kubectl-patch kubectl-replace has:kubectl-replace Successful message:error: --grace-period must have --force specified has:\-\-grace-period must have \-\-force specified Successful message:error: --timeout must have --force specified has:\-\-timeout must have \-\-force specified node/node-v1-test created W0408 07:20:41.825499 58006 actual_state_of_world.go:534] Failed to update statusUpdateNeeded field in actual state of world: Failed to set statusUpdateNeeded to needed true, because nodeName="node-v1-test" does not exist [32mcore.sh:606: Successful get node node-v1-test {{range.items}}{{if .metadata.annotations.a}}found{{end}}{{end}}:: : (B[mnode/node-v1-test replaced (server dry run) node/node-v1-test replaced (dry run) [32mcore.sh:631: Successful get node node-v1-test {{range.items}}{{if .metadata.annotations.a}}found{{end}}{{end}}:: : (B[mnode/node-v1-test replaced [32mcore.sh:647: Successful get node node-v1-test {{.metadata.annotations.a}}: b ... skipping 29 lines ... spec: containers: - image: k8s.gcr.io/pause:2.0 name: kubernetes-pause has:localonlyvalue [32mcore.sh:683: Successful get pod valid-pod {{.metadata.labels.name}}: valid-pod (B[merror: 'name' already has a value (valid-pod), and --overwrite is false [32mcore.sh:687: Successful get pod valid-pod {{.metadata.labels.name}}: valid-pod (B[m[32mcore.sh:691: Successful get pod valid-pod {{.metadata.labels.name}}: valid-pod (B[mpod/valid-pod labeled [32mcore.sh:695: Successful get pod valid-pod {{.metadata.labels.name}}: valid-pod-super-sayan (B[m[32mcore.sh:699: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: valid-pod: (B[mwarning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely. ... skipping 83 lines ... +++ Running case: test-cmd.run_kubectl_create_error_tests +++ working dir: /home/prow/go/src/k8s.io/kubernetes +++ command: run_kubectl_create_error_tests +++ [0408 07:20:53] Creating namespace namespace-1617866453-23322 namespace/namespace-1617866453-23322 created Context "test" modified. +++ [0408 07:20:53] Testing kubectl create with error Error: must specify one of -f and -k Create a resource from a file or from stdin. JSON and YAML formats are accepted. Examples: ... skipping 42 lines ... Usage: kubectl create -f FILENAME [options] Use "kubectl <command> --help" for more information about a given command. Use "kubectl options" for a list of global command-line options (applies to all commands). +++ [0408 07:20:54] "kubectl create with empty string list returns error as expected: error: error validating "hack/testdata/invalid-rc-with-empty-args.yaml": error validating data: ValidationError(ReplicationController.spec.template.spec.containers[0].args): unknown object type "nil" in ReplicationController.spec.template.spec.containers[0].args[0]; if you choose to ignore these errors, turn validation off with --validate=false kubectl convert is DEPRECATED and will be removed in a future version. In order to convert, kubectl apply the object to the cluster, then kubectl get at the desired version. +++ exit code: 0 Recording: run_kubectl_apply_tests Running command: run_kubectl_apply_tests ... skipping 34 lines ... I0408 07:20:57.372410 58006 event.go:291] "Event occurred" object="namespace-1617866454-29923/test-deployment-retainkeys-8695b756f8" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: test-deployment-retainkeys-8695b756f8-c8jhm" deployment.apps "test-deployment-retainkeys" deleted [32mapply.sh:88: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mpod/selector-test-pod created [32mapply.sh:92: Successful get pods selector-test-pod {{.metadata.labels.name}}: selector-test-pod (B[mSuccessful message:Error from server (NotFound): pods "selector-test-pod-dont-apply" not found has:pods "selector-test-pod-dont-apply" not found pod "selector-test-pod" deleted [32mapply.sh:101: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mW0408 07:20:58.564913 66435 helpers.go:567] --dry-run=true is deprecated (boolean value) and can be replaced with --dry-run=client. pod/test-pod created (dry run) pod/test-pod created (dry run) ... skipping 14 lines ... has:resources.mygroup.example.com Warning: apiextensions.k8s.io/v1beta1 CustomResourceDefinition is deprecated in v1.16+, unavailable in v1.22+; use apiextensions.k8s.io/v1 CustomResourceDefinition I0408 07:21:01.637750 54487 client.go:360] parsed scheme: "endpoint" I0408 07:21:01.637798 54487 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{http://127.0.0.1:2379 <nil> 0 <nil>}] I0408 07:21:01.646125 54487 controller.go:609] quota admission added evaluator for: resources.mygroup.example.com kind.mygroup.example.com/myobj created (server dry run) Error from server (NotFound): resources.mygroup.example.com "myobj" not found customresourcedefinition.apiextensions.k8s.io "resources.mygroup.example.com" deleted namespace/nsb created [32mapply.sh:170: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mpod/a created [32mapply.sh:173: Successful get pods a -n nsb {{.metadata.name}}: a (B[mpod/b created pod/a pruned Warning: extensions/v1beta1 Ingress is deprecated in v1.14+, unavailable in v1.22+; use networking.k8s.io/v1 Ingress [32mapply.sh:177: Successful get pods b -n nsb {{.metadata.name}}: b (B[mSuccessful message:Error from server (NotFound): pods "a" not found has:pods "a" not found pod "b" deleted [32mapply.sh:187: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mpod/a created [32mapply.sh:192: Successful get pods a {{.metadata.name}}: a (B[mSuccessful message:Error from server (NotFound): pods "b" not found has:pods "b" not found pod/b created [32mapply.sh:200: Successful get pods a {{.metadata.name}}: a (B[m[32mapply.sh:201: Successful get pods b -n nsb {{.metadata.name}}: b (B[mpod "a" deleted pod "b" deleted Successful message:error: all resources selected for prune without explicitly passing --all. To prune all resources, pass the --all flag. If you did not mean to prune all resources, specify a label selector has:all resources selected for prune without explicitly passing --all pod/a created pod/b created service/prune-svc created Warning: extensions/v1beta1 Ingress is deprecated in v1.14+, unavailable in v1.22+; use networking.k8s.io/v1 Ingress I0408 07:21:07.946971 58006 horizontal.go:354] Horizontal Pod Autoscaler frontend has been deleted in namespace-1617866450-7591 ... skipping 41 lines ... (B[mpod/b created [32mapply.sh:254: Successful get pods b -n nsb {{.metadata.name}}: b (B[mpod/b unchanged pod/a pruned Warning: extensions/v1beta1 Ingress is deprecated in v1.14+, unavailable in v1.22+; use networking.k8s.io/v1 Ingress Successful message:Error from server (NotFound): pods "a" not found has:pods "a" not found [32mapply.sh:261: Successful get pods b -n nsb {{.metadata.name}}: b (B[mnamespace "nsb" deleted Successful message:error: the namespace from the provided object "nsb" does not match the namespace "foo". You must pass '--namespace=nsb' to perform this operation. has:the namespace from the provided object "nsb" does not match the namespace "foo". [32mapply.sh:272: Successful get services {{range.items}}{{.metadata.name}}:{{end}}: (B[mservice/a created [32mapply.sh:276: Successful get services a {{.metadata.name}}: a (B[mSuccessful message:The Service "a" is invalid: spec.clusterIP: Invalid value: "10.0.0.12": field is immutable ... skipping 28 lines ... (B[m[32mapply.sh:298: Successful get deployment test-the-deployment {{.metadata.name}}: test-the-deployment (B[m[32mapply.sh:299: Successful get service test-the-service {{.metadata.name}}: test-the-service (B[mconfigmap "test-the-map" deleted service "test-the-service" deleted deployment.apps "test-the-deployment" deleted Successful message:Error from server (NotFound): namespaces "multi-resource-ns" not found has:namespaces "multi-resource-ns" not found [32mapply.sh:307: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:namespace/multi-resource-ns created Error from server (NotFound): error when creating "hack/testdata/multi-resource-1.yaml": namespaces "multi-resource-ns" not found has:namespaces "multi-resource-ns" not found Successful message:Error from server (NotFound): pods "test-pod" not found has:pods "test-pod" not found pod/test-pod created namespace/multi-resource-ns unchanged [32mapply.sh:315: Successful get pods test-pod -n multi-resource-ns {{.metadata.name}}: test-pod (B[mpod "test-pod" deleted namespace "multi-resource-ns" deleted I0408 07:21:35.871937 58006 namespace_controller.go:185] Namespace has been deleted nsb [32mapply.sh:321: Successful get configmaps {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:configmap/foo created error: unable to recognize "hack/testdata/multi-resource-2.yaml": no matches for kind "Bogus" in version "example.com/v1" has:no matches for kind "Bogus" in version "example.com/v1" [32mapply.sh:327: Successful get configmaps foo {{.metadata.name}}: foo (B[mconfigmap "foo" deleted [32mapply.sh:333: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:pod/pod-a created ... skipping 6 lines ... pod "pod-c" deleted [32mapply.sh:341: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[m[32mapply.sh:345: Successful get crds {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:Warning: apiextensions.k8s.io/v1beta1 CustomResourceDefinition is deprecated in v1.16+, unavailable in v1.22+; use apiextensions.k8s.io/v1 CustomResourceDefinition customresourcedefinition.apiextensions.k8s.io/widgets.example.com created error: unable to recognize "hack/testdata/multi-resource-4.yaml": no matches for kind "Widget" in version "example.com/v1" has:no matches for kind "Widget" in version "example.com/v1" I0408 07:21:43.089558 54487 client.go:360] parsed scheme: "endpoint" I0408 07:21:43.089617 54487 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{http://127.0.0.1:2379 <nil> 0 <nil>}] Successful message:Error from server (NotFound): widgets.example.com "foo" not found has:widgets.example.com "foo" not found [32mapply.sh:351: Successful get crds widgets.example.com {{.metadata.name}}: widgets.example.com (B[mI0408 07:21:43.429197 54487 controller.go:609] quota admission added evaluator for: widgets.example.com widget.example.com/foo created Warning: apiextensions.k8s.io/v1beta1 CustomResourceDefinition is deprecated in v1.16+, unavailable in v1.22+; use apiextensions.k8s.io/v1 CustomResourceDefinition customresourcedefinition.apiextensions.k8s.io/widgets.example.com unchanged ... skipping 34 lines ... message:868 has:868 pod "test-pod" deleted [32mapply.sh:410: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[m+++ [0408 07:21:46] Testing upgrade kubectl client-side apply to server-side apply pod/test-pod created error: Apply failed with 1 conflict: conflict with "kubectl-client-side-apply" using v1: .metadata.labels.name Please review the fields above--they currently have other managers. Here are the ways you can resolve this warning: * If you intend to manage all of these fields, please re-run the apply command with the `--force-conflicts` flag. * If you do not intend to manage all of the fields, please edit your manifest to remove references to the fields that should keep their ... skipping 55 lines ... message:resources.mygroup.example.com has:resources.mygroup.example.com Warning: apiextensions.k8s.io/v1beta1 CustomResourceDefinition is deprecated in v1.16+, unavailable in v1.22+; use apiextensions.k8s.io/v1 CustomResourceDefinition I0408 07:21:48.912890 54487 client.go:360] parsed scheme: "endpoint" I0408 07:21:48.912939 54487 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{http://127.0.0.1:2379 <nil> 0 <nil>}] kind.mygroup.example.com/myobj serverside-applied (server dry run) Error from server (NotFound): resources.mygroup.example.com "myobj" not found customresourcedefinition.apiextensions.k8s.io "resources.mygroup.example.com" deleted +++ exit code: 0 Recording: run_kubectl_run_tests Running command: run_kubectl_run_tests +++ Running case: test-cmd.run_kubectl_run_tests +++ working dir: /home/prow/go/src/k8s.io/kubernetes +++ command: run_kubectl_run_tests +++ [0408 07:21:49] Creating namespace namespace-1617866509-1739 namespace/namespace-1617866509-1739 created W0408 07:21:49.350088 58006 mutation_detector.go:53] Mutation detector is enabled, this will result in memory leakage. I0408 07:21:49.350152 58006 resource_quota_monitor.go:228] QuotaMonitor created object count evaluator for resources.mygroup.example.com I0408 07:21:49.350232 58006 shared_informer.go:240] Waiting for caches to sync for resource quota E0408 07:21:49.351531 58006 reflector.go:127] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource Context "test" modified. +++ [0408 07:21:49] Testing kubectl run pod/nginx-extensions created (dry run) pod/nginx-extensions created (server dry run) [32mrun.sh:32: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[m[32mrun.sh:35: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: ... skipping 2 lines ... (B[mpod "nginx-extensions" deleted Successful message:pod/test1 created has:pod/test1 created pod "test1" deleted Successful message:error: Invalid image name "InvalidImageName": invalid reference format has:error: Invalid image name "InvalidImageName": invalid reference format +++ exit code: 0 Recording: run_kubectl_create_filter_tests Running command: run_kubectl_create_filter_tests +++ Running case: test-cmd.run_kubectl_create_filter_tests +++ working dir: /home/prow/go/src/k8s.io/kubernetes +++ command: run_kubectl_create_filter_tests E0408 07:21:50.650303 58006 reflector.go:127] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource +++ [0408 07:21:50] Creating namespace namespace-1617866510-23232 namespace/namespace-1617866510-23232 created Context "test" modified. +++ [0408 07:21:50] Testing kubectl create filter [32mcreate.sh:50: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mpod/selector-test-pod created [32mcreate.sh:54: Successful get pods selector-test-pod {{.metadata.labels.name}}: selector-test-pod (B[mSuccessful message:Error from server (NotFound): pods "selector-test-pod-dont-apply" not found has:pods "selector-test-pod-dont-apply" not found pod "selector-test-pod" deleted +++ exit code: 0 Recording: run_kubectl_apply_deployments_tests Running command: run_kubectl_apply_deployments_tests ... skipping 7 lines ... [32mapps.sh:119: Successful get deployments {{range.items}}{{.metadata.name}}:{{end}}: (B[m[32mapps.sh:120: Successful get replicasets {{range.items}}{{.metadata.name}}:{{end}}: (B[m[32mapps.sh:121: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mdeployment.apps/my-depl created I0408 07:21:52.296262 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/my-depl" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set my-depl-84fb47b469 to 1" I0408 07:21:52.308607 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/my-depl-84fb47b469" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: my-depl-84fb47b469-pfn2s" E0408 07:21:52.394866 58006 reflector.go:127] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource [32mapps.sh:125: Successful get deployments my-depl {{.metadata.name}}: my-depl (B[m[32mapps.sh:127: Successful get deployments my-depl {{.spec.template.metadata.labels.l1}}: l1 (B[m[32mapps.sh:128: Successful get deployments my-depl {{.spec.selector.matchLabels.l1}}: l1 (B[m[32mapps.sh:129: Successful get deployments my-depl {{.metadata.labels.l1}}: l1 (B[mdeployment.apps/my-depl configured [32mapps.sh:134: Successful get deployments my-depl {{.spec.template.metadata.labels.l1}}: l1 ... skipping 9 lines ... (B[mdeployment.apps/nginx created I0408 07:21:54.327601 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set nginx-9bb9c4878 to 3" I0408 07:21:54.332656 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-9bb9c4878" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-9bb9c4878-jfrjl" I0408 07:21:54.339173 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-9bb9c4878" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-9bb9c4878-pv5br" I0408 07:21:54.342569 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-9bb9c4878" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-9bb9c4878-79s2f" [32mapps.sh:152: Successful get deployment nginx {{.metadata.name}}: nginx (B[mE0408 07:21:55.842545 58006 reflector.go:127] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource Successful message:Error from server (Conflict): error when applying patch: {"metadata":{"annotations":{"kubectl.kubernetes.io/last-applied-configuration":"{\"apiVersion\":\"apps/v1\",\"kind\":\"Deployment\",\"metadata\":{\"annotations\":{},\"labels\":{\"name\":\"nginx\"},\"name\":\"nginx\",\"namespace\":\"namespace-1617866511-17197\",\"resourceVersion\":\"99\"},\"spec\":{\"replicas\":3,\"selector\":{\"matchLabels\":{\"name\":\"nginx2\"}},\"template\":{\"metadata\":{\"labels\":{\"name\":\"nginx2\"}},\"spec\":{\"containers\":[{\"image\":\"k8s.gcr.io/nginx:test-cmd\",\"name\":\"nginx\",\"ports\":[{\"containerPort\":80}]}]}}}}\n"},"resourceVersion":"99"},"spec":{"selector":{"matchLabels":{"name":"nginx2"}},"template":{"metadata":{"labels":{"name":"nginx2"}}}}} to: Resource: "apps/v1, Resource=deployments", GroupVersionKind: "apps/v1, Kind=Deployment" Name: "nginx", Namespace: "namespace-1617866511-17197" for: "hack/testdata/deployment-label-change2.yaml": Operation cannot be fulfilled on deployments.apps "nginx": the object has been modified; please apply your changes to the latest version and try again has:Error from server (Conflict) deployment.apps/nginx configured I0408 07:22:04.035903 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set nginx-6dd6cfdb57 to 3" I0408 07:22:04.040442 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-6dd6cfdb57" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-6dd6cfdb57-4p67p" I0408 07:22:04.047854 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-6dd6cfdb57" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-6dd6cfdb57-rppch" I0408 07:22:04.047964 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-6dd6cfdb57" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-6dd6cfdb57-rxscb" Successful message: "name": "nginx2" "name": "nginx2" has:"name": "nginx2" E0408 07:22:05.539829 58006 reflector.go:127] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource E0408 07:22:08.511562 58006 replica_set.go:532] sync "namespace-1617866511-17197/nginx-6dd6cfdb57" failed with replicasets.apps "nginx-6dd6cfdb57" not found I0408 07:22:09.468438 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx" kind="Deployment" apiVersion="apps/v1" type="Normal" reason="ScalingReplicaSet" message="Scaled up replica set nginx-6dd6cfdb57 to 3" Successful message:The Deployment "nginx" is invalid: spec.template.metadata.labels: Invalid value: map[string]string{"name":"nginx3"}: `selector` does not match template `labels` has:Invalid value I0408 07:22:09.477527 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-6dd6cfdb57" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-6dd6cfdb57-6h7q2" I0408 07:22:09.483559 58006 event.go:291] "Event occurred" object="namespace-1617866511-17197/nginx-6dd6cfdb57" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: nginx-6dd6cfdb57-6hqpg" ... skipping 302 lines ... +++ [0408 07:22:14] Creating namespace namespace-1617866534-9824 namespace/namespace-1617866534-9824 created Context "test" modified. +++ [0408 07:22:14] Testing kubectl get [32mget.sh:29: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:Error from server (NotFound): pods "abc" not found has:pods "abc" not found [32mget.sh:37: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:Error from server (NotFound): pods "abc" not found has:pods "abc" not found [32mget.sh:45: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:{ "apiVersion": "v1", "items": [], ... skipping 26 lines ... I0408 07:22:15.730424 54487 clientconn.go:948] ClientConn switching balancer to "pick_first" Successful message:NAME has not:No resources found [32mget.sh:73: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:error: the server doesn't have a resource type "foobar" has not:No resources found Successful message:No resources found in namespace-1617866534-9824 namespace. has:No resources found Successful message: has not:No resources found Successful message:No resources found in namespace-1617866534-9824 namespace. has:No resources found [32mget.sh:93: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mSuccessful message:Error from server (NotFound): pods "abc" not found has:pods "abc" not found Successful message:Error from server (NotFound): pods "abc" not found has not:List Successful message:I0408 07:22:16.747309 69951 loader.go:375] Config loaded from file: /tmp/tmp.tKtnRtl63A/.kube/config I0408 07:22:16.748944 69951 round_trippers.go:444] GET http://127.0.0.1:8080/version?timeout=32s 200 OK in 1 milliseconds I0408 07:22:16.778036 69951 round_trippers.go:444] GET http://127.0.0.1:8080/api/v1/namespaces/default/pods 200 OK in 2 milliseconds I0408 07:22:16.779875 69951 round_trippers.go:444] GET http://127.0.0.1:8080/api/v1/namespaces/default/replicationcontrollers 200 OK in 1 milliseconds ... skipping 629 lines ... } [32mget.sh:158: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: valid-pod: (B[m<no value>Successful message:valid-pod: has:valid-pod: Successful message:error: error executing jsonpath "{.missing}": Error executing template: missing is not found. Printing more information for debugging the template: template was: {.missing} object given to jsonpath engine was: map[string]interface {}{"apiVersion":"v1", "kind":"Pod", "metadata":map[string]interface {}{"creationTimestamp":"2021-04-08T07:22:24Z", "labels":map[string]interface {}{"name":"valid-pod"}, "managedFields":[]interface {}{map[string]interface {}{"apiVersion":"v1", "fieldsType":"FieldsV1", "fieldsV1":map[string]interface {}{"f:metadata":map[string]interface {}{"f:labels":map[string]interface {}{".":map[string]interface {}{}, "f:name":map[string]interface {}{}}}, "f:spec":map[string]interface {}{"f:containers":map[string]interface {}{"k:{\"name\":\"kubernetes-serve-hostname\"}":map[string]interface {}{".":map[string]interface {}{}, "f:image":map[string]interface {}{}, "f:imagePullPolicy":map[string]interface {}{}, "f:name":map[string]interface {}{}, "f:resources":map[string]interface {}{".":map[string]interface {}{}, "f:limits":map[string]interface {}{".":map[string]interface {}{}, "f:cpu":map[string]interface {}{}, "f:memory":map[string]interface {}{}}, "f:requests":map[string]interface {}{".":map[string]interface {}{}, "f:cpu":map[string]interface {}{}, "f:memory":map[string]interface {}{}}}, "f:terminationMessagePath":map[string]interface {}{}, "f:terminationMessagePolicy":map[string]interface {}{}}}, "f:dnsPolicy":map[string]interface {}{}, "f:enableServiceLinks":map[string]interface {}{}, "f:restartPolicy":map[string]interface {}{}, "f:schedulerName":map[string]interface {}{}, "f:securityContext":map[string]interface {}{}, "f:terminationGracePeriodSeconds":map[string]interface {}{}}}, "manager":"kubectl-create", "operation":"Update", "time":"2021-04-08T07:22:24Z"}}, "name":"valid-pod", "namespace":"namespace-1617866544-7927", "resourceVersion":"1068", "selfLink":"/api/v1/namespaces/namespace-1617866544-7927/pods/valid-pod", "uid":"8560cf4f-f045-402f-be23-fc33f1f2bde9"}, "spec":map[string]interface {}{"containers":[]interface {}{map[string]interface {}{"image":"k8s.gcr.io/serve_hostname", "imagePullPolicy":"Always", "name":"kubernetes-serve-hostname", "resources":map[string]interface {}{"limits":map[string]interface {}{"cpu":"1", "memory":"512Mi"}, "requests":map[string]interface {}{"cpu":"1", "memory":"512Mi"}}, "terminationMessagePath":"/dev/termination-log", "terminationMessagePolicy":"File"}}, "dnsPolicy":"ClusterFirst", "enableServiceLinks":true, "preemptionPolicy":"PreemptLowerPriority", "priority":0, "restartPolicy":"Always", "schedulerName":"default-scheduler", "securityContext":map[string]interface {}{}, "terminationGracePeriodSeconds":30}, "status":map[string]interface {}{"phase":"Pending", "qosClass":"Guaranteed"}} has:missing is not found error: error executing template "{{.missing}}": template: output:1:2: executing "output" at <.missing>: map has no entry for key "missing" Successful message:Error executing template: template: output:1:2: executing "output" at <.missing>: map has no entry for key "missing". Printing more information for debugging the template: template was: {{.missing}} raw data was: {"apiVersion":"v1","kind":"Pod","metadata":{"creationTimestamp":"2021-04-08T07:22:24Z","labels":{"name":"valid-pod"},"managedFields":[{"apiVersion":"v1","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"kubernetes-serve-hostname\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{".":{},"f:limits":{".":{},"f:cpu":{},"f:memory":{}},"f:requests":{".":{},"f:cpu":{},"f:memory":{}}},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}},"manager":"kubectl-create","operation":"Update","time":"2021-04-08T07:22:24Z"}],"name":"valid-pod","namespace":"namespace-1617866544-7927","resourceVersion":"1068","selfLink":"/api/v1/namespaces/namespace-1617866544-7927/pods/valid-pod","uid":"8560cf4f-f045-402f-be23-fc33f1f2bde9"},"spec":{"containers":[{"image":"k8s.gcr.io/serve_hostname","imagePullPolicy":"Always","name":"kubernetes-serve-hostname","resources":{"limits":{"cpu":"1","memory":"512Mi"},"requests":{"cpu":"1","memory":"512Mi"}},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File"}],"dnsPolicy":"ClusterFirst","enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority","priority":0,"restartPolicy":"Always","schedulerName":"default-scheduler","securityContext":{},"terminationGracePeriodSeconds":30},"status":{"phase":"Pending","qosClass":"Guaranteed"}} object given to template engine was: map[apiVersion:v1 kind:Pod metadata:map[creationTimestamp:2021-04-08T07:22:24Z labels:map[name:valid-pod] managedFields:[map[apiVersion:v1 fieldsType:FieldsV1 fieldsV1:map[f:metadata:map[f:labels:map[.:map[] f:name:map[]]] f:spec:map[f:containers:map[k:{"name":"kubernetes-serve-hostname"}:map[.:map[] f:image:map[] f:imagePullPolicy:map[] f:name:map[] f:resources:map[.:map[] f:limits:map[.:map[] f:cpu:map[] f:memory:map[]] f:requests:map[.:map[] f:cpu:map[] f:memory:map[]]] f:terminationMessagePath:map[] f:terminationMessagePolicy:map[]]] f:dnsPolicy:map[] f:enableServiceLinks:map[] f:restartPolicy:map[] f:schedulerName:map[] f:securityContext:map[] f:terminationGracePeriodSeconds:map[]]] manager:kubectl-create operation:Update time:2021-04-08T07:22:24Z]] name:valid-pod namespace:namespace-1617866544-7927 resourceVersion:1068 selfLink:/api/v1/namespaces/namespace-1617866544-7927/pods/valid-pod uid:8560cf4f-f045-402f-be23-fc33f1f2bde9] spec:map[containers:[map[image:k8s.gcr.io/serve_hostname imagePullPolicy:Always name:kubernetes-serve-hostname resources:map[limits:map[cpu:1 memory:512Mi] requests:map[cpu:1 memory:512Mi]] terminationMessagePath:/dev/termination-log terminationMessagePolicy:File]] dnsPolicy:ClusterFirst enableServiceLinks:true preemptionPolicy:PreemptLowerPriority priority:0 restartPolicy:Always schedulerName:default-scheduler securityContext:map[] terminationGracePeriodSeconds:30] status:map[phase:Pending qosClass:Guaranteed]] ... skipping 9 lines ... Successful message:pod/valid-pod has not:STATUS Successful message:pod/valid-pod has:pod/valid-pod E0408 07:22:28.502442 58006 reflector.go:127] k8s.io/client-go/metadata/metadatainformer/informer.go:90: Failed to watch *v1.PartialObjectMetadata: failed to list *v1.PartialObjectMetadata: the server could not find the requested resource Successful message:apiVersion: v1 kind: Pod metadata: creationTimestamp: "2021-04-08T07:22:24Z" labels: ... skipping 137 lines ... terminationGracePeriodSeconds: 30 status: phase: Pending qosClass: Guaranteed has:name: valid-pod Successful message:Error from server (NotFound): pods "invalid-pod" not found has:"invalid-pod" not found pod "valid-pod" deleted [32mget.sh:196: Successful get pods {{range.items}}{{.metadata.name}}:{{end}}: (B[mpod/redis-master created pod/valid-pod created Successful ... skipping 36 lines ... +++ [0408 07:22:30] Creating namespace namespace-1617866550-1219 namespace/namespace-1617866550-1219 created Context "test" modified. +++ [0408 07:22:30] Testing kubectl exec POD COMMAND Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (NotFound): pods "abc" not found has:pods "abc" not found pod/test-pod created Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (BadRequest): pod test-pod does not have a host assigned has not:pods "test-pod" not found Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (BadRequest): pod test-pod does not have a host assigned has not:pod or type/name must be specified pod "test-pod" deleted +++ exit code: 0 Recording: run_kubectl_exec_resource_name_tests Running command: run_kubectl_exec_resource_name_tests ... skipping 3 lines ... +++ [0408 07:22:31] Creating namespace namespace-1617866551-15713 namespace/namespace-1617866551-15713 created Context "test" modified. +++ [0408 07:22:31] Testing kubectl exec TYPE/NAME COMMAND Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. error: the server doesn't have a resource type "foo" has:error: Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (NotFound): deployments.apps "bar" not found has:"bar" not found pod/test-pod created replicaset.apps/frontend created I0408 07:22:32.497147 58006 event.go:291] "Event occurred" object="namespace-1617866551-15713/frontend" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: frontend-q48l8" I0408 07:22:32.502658 58006 event.go:291] "Event occurred" object="namespace-1617866551-15713/frontend" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: frontend-2t64m" I0408 07:22:32.506028 58006 event.go:291] "Event occurred" object="namespace-1617866551-15713/frontend" kind="ReplicaSet" apiVersion="apps/v1" type="Normal" reason="SuccessfulCreate" message="Created pod: frontend-w2sbm" configmap/test-set-env-config created Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. error: cannot attach to *v1.ConfigMap: selector for *v1.ConfigMap not implemented has:not implemented Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (BadRequest): pod test-pod does not have a host assigned has not:not found Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (BadRequest): pod test-pod does not have a host assigned has not:pod, type/name or --filename must be specified Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (BadRequest): pod frontend-2t64m does not have a host assigned has not:not found Successful message:kubectl exec [POD] [COMMAND] is DEPRECATED and will be removed in a future version. Use kubectl exec [POD] -- [COMMAND] instead. Error from server (BadRequest): pod frontend-2t64m does not have a host assigned has not:pod, type/name or --filename must be specified pod "test-pod" deleted replicaset.apps "frontend" deleted configmap "test-set-env-config" deleted +++ exit code: 0 Recording: run_create_secret_tests Running command: run_create_secret_tests +++ Running case: test-cmd.run_create_secret_tests +++ working dir: /home/prow/go/src/k8s.io/kubernetes +++ command: run_create_secret_tests Successful message:Error from server (NotFound): secrets "mysecret" not found has:secrets "mysecret" not found Successful message:user-specified has:user-specified Successful message:Error from server (NotFound): secrets "mysecret" not found has:secrets "mysecret" not found Successful {"kind":"ConfigMap","apiVersion":"v1","metadata":{"name":"tester-update-cm","namespace":"default","selfLink":"/api/v1/namespaces/default/configmaps/tester-update-cm","uid":"9d9c8ccf-7b14-4340-b732-b66429111f23","resourceVersion":"1151","creationTimestamp":"2021-04-08T07:22:33Z"}} Successful message:{"kind":"ConfigMap","apiVersion":"v1","metadata":{"name":"tester-update-cm","namespace":"default","selfLink":"/api/v1/namespaces/default/configmaps/tester-update-cm","uid":"9d9c8ccf-7b14-4340-b732-b66429111f23","resourceVersion":"1152","creationTimestamp":"2021-04-08T07:22:33Z"},"data":{"key1":"config1"}} has:uid Successful message:{"kind":"ConfigMap","apiVersion":"v1","metadata":{"name":"tester-update-cm","namespace":"default","selfLink":"/api/v1/namespaces/default/configmaps/tester-update-cm","uid":"9d9c8ccf-7b14-4340-b732-b66429111f23","resourceVersion":"1152","creationTimestamp":"2021-04-08T07:22:33Z"},"data":{"key1":"config1"}} has:config1 {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Success","details":{"name":"tester-update-cm","kind":"configmaps","uid":"9d9c8ccf-7b14-4340-b732-b66429111f23"}} Successful message:Error from server (NotFound): configmaps "tester-update-cm" not found has:configmaps "tester-update-cm" not found +++ exit code: 0 Recording: run_kubectl_create_kustomization_directory_tests Running command: run_kubectl_create_kustomization_directory_tests +++ Running case: test-cmd.run_kubectl_create_kustomization_directory_tests ... skipping 173 lines ... has:Timeout Successful message:NAME READY STATUS RESTARTS AGE valid-pod 0/1 Pending 0 2s has:valid-pod Successful message:error: Invalid timeout value. Timeout must be a single integer in seconds, or an integer followed by a corresponding time unit (e.g. 1s | 2m | 3h) has:Invalid timeout value pod "valid-pod" deleted +++ exit code: 0 Recording: run_crd_tests Running command: run_crd_tests ... skipping 244 lines ... foo.company.com/test patched [32mcrd.sh:236: Successful get foos/test {{.patched}}: value1 (B[mfoo.company.com/test patched [32mcrd.sh:238: Successful get foos/test {{.patched}}: value2 (B[mfoo.company.com/test patched [32mcrd.sh:240: Successful get foos/test {{.patched}}: <no value> (B[m+++ [0408 07:22:46] "kubectl patch --local" returns error as expected for CustomResource: error: cannot apply strategic merge patch for company.com/v1, Kind=Foo locally, try --type merge { "apiVersion": "company.com/v1", "kind": "Foo", "metadata": { "annotations": { "kubernetes.io/change-cause": "kubectl patch foos/test --server=http://127.0.0.1:8080 --match-server-version=true --patch={\"patched\":null} --type=merge --record=true" ... skipping 304 lines ... W0408 07:22:49.803986 58006 shared_informer.go:494] resyncPeriod 67589840627519 is smaller than resyncCheckPeriod 68970300535873 and the informer has already started. Changing it to 68970300535873 I0408 07:22:49.804005 58006 resource_quota_monitor.go:228] QuotaMonitor created object count evaluator for foos.company.com I0408 07:22:49.804051 58006 shared_informer.go:240] Waiting for caches to sync for resource quota I0408 07:22:57.425045 54487 client.go:360] parsed scheme: "passthrough" I0408 07:22:57.425112 54487 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{http://127.0.0.1:2379 <nil> 0 <nil>}] <nil> <nil>} I0408 07:22:57.425127 54487 clientconn.go:948] ClientConn switching balancer to "pick_first" {"component":"entrypoint","file":"prow/entrypoint/run.go:169","func":"k8s.io/test-infra/prow/entrypoint.Options.ExecuteProcess","level":"error","msg":"Entrypoint received interrupt: terminated","severity":"error","time":"2021-04-08T07:23:04Z"}