This job view page is being replaced by Spyglass soon. Check out the new job view.
ResultFAILURE
Tests 2 failed / 605 succeeded
Started2019-01-11 18:51
Elapsed29m33s
Revision
Buildergke-prow-containerd-pool-99179761-5z9f
poddc30e33a-15d1-11e9-9d8c-0a580a6c019e
infra-commita2ccf6d48
poddc30e33a-15d1-11e9-9d8c-0a580a6c019e
repok8s.io/kubernetes
repo-commitd2d9a3aba471e56e1a753a05061dc6cb60137711
repos{u'k8s.io/kubernetes': u'master'}

Test Failures


k8s.io/kubernetes/test/integration/scheduler TestPreemptionRaces 21s

go test -v k8s.io/kubernetes/test/integration/scheduler -run TestPreemptionRaces$
I0111 19:13:59.612428  120459 services.go:33] Network range for service cluster IPs is unspecified. Defaulting to {10.0.0.0 ffffff00}.
I0111 19:13:59.612455  120459 services.go:45] Setting service IP to "10.0.0.1" (read-write).
I0111 19:13:59.612469  120459 master.go:273] Node port range unspecified. Defaulting to 30000-32767.
I0111 19:13:59.612479  120459 master.go:229] Using reconciler: 
I0111 19:13:59.613996  120459 storage_factory.go:285] storing podtemplates in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.614113  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.614134  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.614168  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.614218  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.614569  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.614665  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.614739  120459 store.go:1414] Monitoring podtemplates count at <storage-prefix>//podtemplates
I0111 19:13:59.614778  120459 reflector.go:169] Listing and watching *core.PodTemplate from storage/cacher.go:/podtemplates
I0111 19:13:59.614791  120459 storage_factory.go:285] storing events in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.615007  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.615069  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.615115  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.615178  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.615446  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.615505  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.615587  120459 store.go:1414] Monitoring events count at <storage-prefix>//events
I0111 19:13:59.615652  120459 storage_factory.go:285] storing limitranges in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.615759  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.615785  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.615832  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.615907  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.616177  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.616256  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.616286  120459 store.go:1414] Monitoring limitranges count at <storage-prefix>//limitranges
I0111 19:13:59.616316  120459 storage_factory.go:285] storing resourcequotas in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.616340  120459 reflector.go:169] Listing and watching *core.LimitRange from storage/cacher.go:/limitranges
I0111 19:13:59.616395  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.616414  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.616445  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.616574  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.616803  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.616882  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.616930  120459 store.go:1414] Monitoring resourcequotas count at <storage-prefix>//resourcequotas
I0111 19:13:59.616957  120459 reflector.go:169] Listing and watching *core.ResourceQuota from storage/cacher.go:/resourcequotas
I0111 19:13:59.617399  120459 storage_factory.go:285] storing secrets in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.617719  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.617768  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.617811  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.617892  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.618182  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.618302  120459 store.go:1414] Monitoring secrets count at <storage-prefix>//secrets
I0111 19:13:59.618487  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.618542  120459 reflector.go:169] Listing and watching *core.Secret from storage/cacher.go:/secrets
I0111 19:13:59.618543  120459 storage_factory.go:285] storing persistentvolumes in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.619376  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.619414  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.619445  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.619488  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.619787  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.619838  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.619902  120459 store.go:1414] Monitoring persistentvolumes count at <storage-prefix>//persistentvolumes
I0111 19:13:59.619949  120459 reflector.go:169] Listing and watching *core.PersistentVolume from storage/cacher.go:/persistentvolumes
I0111 19:13:59.620084  120459 storage_factory.go:285] storing persistentvolumeclaims in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.620168  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.620186  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.620219  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.620273  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.620500  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.620547  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.620594  120459 store.go:1414] Monitoring persistentvolumeclaims count at <storage-prefix>//persistentvolumeclaims
I0111 19:13:59.620650  120459 reflector.go:169] Listing and watching *core.PersistentVolumeClaim from storage/cacher.go:/persistentvolumeclaims
I0111 19:13:59.621069  120459 storage_factory.go:285] storing configmaps in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.621171  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.621192  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.621229  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.621295  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.621562  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.621718  120459 store.go:1414] Monitoring configmaps count at <storage-prefix>//configmaps
I0111 19:13:59.621859  120459 storage_factory.go:285] storing namespaces in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.621928  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.621940  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.621972  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.622012  120459 reflector.go:169] Listing and watching *core.ConfigMap from storage/cacher.go:/configmaps
I0111 19:13:59.622064  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.622164  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.622652  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.622720  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.622771  120459 store.go:1414] Monitoring namespaces count at <storage-prefix>//namespaces
I0111 19:13:59.622829  120459 reflector.go:169] Listing and watching *core.Namespace from storage/cacher.go:/namespaces
I0111 19:13:59.622928  120459 storage_factory.go:285] storing endpoints in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.623051  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.623084  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.623126  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.623171  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.623437  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.623523  120459 store.go:1414] Monitoring endpoints count at <storage-prefix>//endpoints
I0111 19:13:59.623570  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.623660  120459 reflector.go:169] Listing and watching *core.Endpoints from storage/cacher.go:/endpoints
I0111 19:13:59.623704  120459 storage_factory.go:285] storing nodes in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.623800  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.623817  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.623843  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.623880  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.624106  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.624168  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.624223  120459 store.go:1414] Monitoring nodes count at <storage-prefix>//nodes
I0111 19:13:59.624294  120459 reflector.go:169] Listing and watching *core.Node from storage/cacher.go:/nodes
I0111 19:13:59.624460  120459 storage_factory.go:285] storing pods in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.624551  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.624572  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.624599  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.624690  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.624922  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.624969  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.625058  120459 store.go:1414] Monitoring pods count at <storage-prefix>//pods
I0111 19:13:59.625210  120459 reflector.go:169] Listing and watching *core.Pod from storage/cacher.go:/pods
I0111 19:13:59.625225  120459 storage_factory.go:285] storing serviceaccounts in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.625301  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.625318  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.625346  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.625439  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.625741  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.625771  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.625856  120459 store.go:1414] Monitoring serviceaccounts count at <storage-prefix>//serviceaccounts
I0111 19:13:59.625897  120459 reflector.go:169] Listing and watching *core.ServiceAccount from storage/cacher.go:/serviceaccounts
I0111 19:13:59.626300  120459 storage_factory.go:285] storing services in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.626759  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.626830  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.626905  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.627144  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.627396  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.627464  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.627518  120459 store.go:1414] Monitoring services count at <storage-prefix>//services
I0111 19:13:59.627548  120459 reflector.go:169] Listing and watching *core.Service from storage/cacher.go:/services
I0111 19:13:59.627549  120459 storage_factory.go:285] storing services in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.627632  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.627645  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.627716  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.628162  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.628458  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.628490  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.628573  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.628595  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.628634  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.628730  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.629009  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.629200  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.629191  120459 storage_factory.go:285] storing replicationcontrollers in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.629271  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.629291  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.629317  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.629359  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.629731  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.629768  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.629832  120459 store.go:1414] Monitoring replicationcontrollers count at <storage-prefix>//replicationcontrollers
I0111 19:13:59.629872  120459 reflector.go:169] Listing and watching *core.ReplicationController from storage/cacher.go:/replicationcontrollers
I0111 19:13:59.642745  120459 master.go:408] Skipping disabled API group "auditregistration.k8s.io".
I0111 19:13:59.642790  120459 master.go:416] Enabling API group "authentication.k8s.io".
I0111 19:13:59.642816  120459 master.go:416] Enabling API group "authorization.k8s.io".
I0111 19:13:59.643011  120459 storage_factory.go:285] storing horizontalpodautoscalers.autoscaling in autoscaling/v1, reading as autoscaling/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.643259  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.643278  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.643350  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.643430  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.643929  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.644040  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.644143  120459 store.go:1414] Monitoring horizontalpodautoscalers.autoscaling count at <storage-prefix>//horizontalpodautoscalers
I0111 19:13:59.644311  120459 reflector.go:169] Listing and watching *autoscaling.HorizontalPodAutoscaler from storage/cacher.go:/horizontalpodautoscalers
I0111 19:13:59.644375  120459 storage_factory.go:285] storing horizontalpodautoscalers.autoscaling in autoscaling/v1, reading as autoscaling/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.644470  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.644497  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.644563  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.644623  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.644895  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.645006  120459 store.go:1414] Monitoring horizontalpodautoscalers.autoscaling count at <storage-prefix>//horizontalpodautoscalers
I0111 19:13:59.645036  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.645126  120459 reflector.go:169] Listing and watching *autoscaling.HorizontalPodAutoscaler from storage/cacher.go:/horizontalpodautoscalers
I0111 19:13:59.645325  120459 storage_factory.go:285] storing horizontalpodautoscalers.autoscaling in autoscaling/v1, reading as autoscaling/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.645437  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.645462  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.645540  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.645602  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.645970  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.646085  120459 store.go:1414] Monitoring horizontalpodautoscalers.autoscaling count at <storage-prefix>//horizontalpodautoscalers
I0111 19:13:59.646110  120459 master.go:416] Enabling API group "autoscaling".
I0111 19:13:59.646348  120459 storage_factory.go:285] storing jobs.batch in batch/v1, reading as batch/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.646517  120459 reflector.go:169] Listing and watching *autoscaling.HorizontalPodAutoscaler from storage/cacher.go:/horizontalpodautoscalers
I0111 19:13:59.646544  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.646752  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.646420  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.646868  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.646933  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.647333  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.647387  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.647485  120459 store.go:1414] Monitoring jobs.batch count at <storage-prefix>//jobs
I0111 19:13:59.647521  120459 reflector.go:169] Listing and watching *batch.Job from storage/cacher.go:/jobs
I0111 19:13:59.647707  120459 storage_factory.go:285] storing cronjobs.batch in batch/v1beta1, reading as batch/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.648456  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.648495  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.648526  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.648609  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.648874  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.648950  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.649082  120459 store.go:1414] Monitoring cronjobs.batch count at <storage-prefix>//cronjobs
I0111 19:13:59.649139  120459 master.go:416] Enabling API group "batch".
I0111 19:13:59.649156  120459 reflector.go:169] Listing and watching *batch.CronJob from storage/cacher.go:/cronjobs
I0111 19:13:59.649566  120459 storage_factory.go:285] storing certificatesigningrequests.certificates.k8s.io in certificates.k8s.io/v1beta1, reading as certificates.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.649662  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.649707  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.649761  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.649820  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.650152  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.650271  120459 store.go:1414] Monitoring certificatesigningrequests.certificates.k8s.io count at <storage-prefix>//certificatesigningrequests
I0111 19:13:59.650288  120459 master.go:416] Enabling API group "certificates.k8s.io".
I0111 19:13:59.650471  120459 storage_factory.go:285] storing leases.coordination.k8s.io in coordination.k8s.io/v1beta1, reading as coordination.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.650497  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.650544  120459 reflector.go:169] Listing and watching *certificates.CertificateSigningRequest from storage/cacher.go:/certificatesigningrequests
I0111 19:13:59.650554  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.650567  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.650621  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.650713  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.650992  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.651084  120459 store.go:1414] Monitoring leases.coordination.k8s.io count at <storage-prefix>//leases
I0111 19:13:59.651128  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.651155  120459 reflector.go:169] Listing and watching *coordination.Lease from storage/cacher.go:/leases
I0111 19:13:59.651266  120459 storage_factory.go:285] storing leases.coordination.k8s.io in coordination.k8s.io/v1beta1, reading as coordination.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.651613  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.651627  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.651654  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.651715  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.651967  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.652091  120459 store.go:1414] Monitoring leases.coordination.k8s.io count at <storage-prefix>//leases
I0111 19:13:59.652118  120459 master.go:416] Enabling API group "coordination.k8s.io".
I0111 19:13:59.652161  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.652225  120459 reflector.go:169] Listing and watching *coordination.Lease from storage/cacher.go:/leases
I0111 19:13:59.652322  120459 storage_factory.go:285] storing replicationcontrollers in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.652402  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.652417  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.652447  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.652482  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.652702  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.652750  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.653303  120459 reflector.go:169] Listing and watching *core.ReplicationController from storage/cacher.go:/replicationcontrollers
I0111 19:13:59.658418  120459 store.go:1414] Monitoring replicationcontrollers count at <storage-prefix>//replicationcontrollers
I0111 19:13:59.658667  120459 storage_factory.go:285] storing daemonsets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.658803  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.658818  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.658859  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.661001  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.661968  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.662276  120459 store.go:1414] Monitoring daemonsets.apps count at <storage-prefix>//daemonsets
I0111 19:13:59.662539  120459 storage_factory.go:285] storing deployments.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.662732  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.662774  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.662822  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.662959  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.663015  120459 reflector.go:169] Listing and watching *apps.DaemonSet from storage/cacher.go:/daemonsets
I0111 19:13:59.663316  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.664761  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.664997  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.665399  120459 store.go:1414] Monitoring deployments.apps count at <storage-prefix>//deployments
I0111 19:13:59.665630  120459 storage_factory.go:285] storing ingresses.extensions in extensions/v1beta1, reading as extensions/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.665794  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.665849  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.665946  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.666026  120459 reflector.go:169] Listing and watching *apps.Deployment from storage/cacher.go:/deployments
I0111 19:13:59.666327  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.666857  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.666976  120459 store.go:1414] Monitoring ingresses.extensions count at <storage-prefix>//ingresses
I0111 19:13:59.667246  120459 storage_factory.go:285] storing podsecuritypolicies.policy in policy/v1beta1, reading as policy/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.667325  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.667337  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.667372  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.667450  120459 reflector.go:169] Listing and watching *extensions.Ingress from storage/cacher.go:/ingresses
I0111 19:13:59.667450  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.667602  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.668083  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.668204  120459 store.go:1414] Monitoring podsecuritypolicies.policy count at <storage-prefix>//podsecuritypolicies
I0111 19:13:59.668362  120459 storage_factory.go:285] storing replicasets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.668693  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.668705  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.668734  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.668863  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.668888  120459 reflector.go:169] Listing and watching *policy.PodSecurityPolicy from storage/cacher.go:/podsecuritypolicies
I0111 19:13:59.668997  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.669204  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.669341  120459 store.go:1414] Monitoring replicasets.apps count at <storage-prefix>//replicasets
I0111 19:13:59.669479  120459 storage_factory.go:285] storing networkpolicies.networking.k8s.io in networking.k8s.io/v1, reading as networking.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.669529  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.669583  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.669594  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.669617  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.669714  120459 reflector.go:169] Listing and watching *apps.ReplicaSet from storage/cacher.go:/replicasets
I0111 19:13:59.670066  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.670604  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.670769  120459 store.go:1414] Monitoring networkpolicies.networking.k8s.io count at <storage-prefix>//networkpolicies
I0111 19:13:59.670808  120459 master.go:416] Enabling API group "extensions".
I0111 19:13:59.670833  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.671266  120459 storage_factory.go:285] storing networkpolicies.networking.k8s.io in networking.k8s.io/v1, reading as networking.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.671376  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.671409  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.671461  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.670887  120459 reflector.go:169] Listing and watching *networking.NetworkPolicy from storage/cacher.go:/networkpolicies
I0111 19:13:59.671548  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.671850  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.671937  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.671985  120459 store.go:1414] Monitoring networkpolicies.networking.k8s.io count at <storage-prefix>//networkpolicies
I0111 19:13:59.672020  120459 master.go:416] Enabling API group "networking.k8s.io".
I0111 19:13:59.672118  120459 reflector.go:169] Listing and watching *networking.NetworkPolicy from storage/cacher.go:/networkpolicies
I0111 19:13:59.672201  120459 storage_factory.go:285] storing poddisruptionbudgets.policy in policy/v1beta1, reading as policy/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.672315  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.672332  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.672357  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.672406  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.672666  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.672719  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.672768  120459 store.go:1414] Monitoring poddisruptionbudgets.policy count at <storage-prefix>//poddisruptionbudgets
I0111 19:13:59.673325  120459 storage_factory.go:285] storing podsecuritypolicies.policy in policy/v1beta1, reading as policy/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.673415  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.673434  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.673601  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.673669  120459 reflector.go:169] Listing and watching *policy.PodDisruptionBudget from storage/cacher.go:/poddisruptionbudgets
I0111 19:13:59.673897  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.674794  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.674846  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.674980  120459 store.go:1414] Monitoring podsecuritypolicies.policy count at <storage-prefix>//podsecuritypolicies
I0111 19:13:59.675059  120459 master.go:416] Enabling API group "policy".
I0111 19:13:59.675111  120459 reflector.go:169] Listing and watching *policy.PodSecurityPolicy from storage/cacher.go:/podsecuritypolicies
I0111 19:13:59.675133  120459 storage_factory.go:285] storing roles.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.675245  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.675272  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.675457  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.675523  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.675959  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.676046  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.676134  120459 store.go:1414] Monitoring roles.rbac.authorization.k8s.io count at <storage-prefix>//roles
I0111 19:13:59.676341  120459 reflector.go:169] Listing and watching *rbac.Role from storage/cacher.go:/roles
I0111 19:13:59.676375  120459 storage_factory.go:285] storing rolebindings.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.676468  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.676490  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.676585  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.676634  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.676970  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.677099  120459 store.go:1414] Monitoring rolebindings.rbac.authorization.k8s.io count at <storage-prefix>//rolebindings
I0111 19:13:59.677144  120459 storage_factory.go:285] storing clusterroles.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.677215  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.677243  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.677268  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.677332  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.677359  120459 reflector.go:169] Listing and watching *rbac.RoleBinding from storage/cacher.go:/rolebindings
I0111 19:13:59.677492  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.677791  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.677884  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.678036  120459 store.go:1414] Monitoring clusterroles.rbac.authorization.k8s.io count at <storage-prefix>//clusterroles
I0111 19:13:59.678319  120459 reflector.go:169] Listing and watching *rbac.ClusterRole from storage/cacher.go:/clusterroles
I0111 19:13:59.678620  120459 storage_factory.go:285] storing clusterrolebindings.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.678832  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.678857  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.678889  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.678962  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.679440  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.679653  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.679781  120459 store.go:1414] Monitoring clusterrolebindings.rbac.authorization.k8s.io count at <storage-prefix>//clusterrolebindings
I0111 19:13:59.679863  120459 reflector.go:169] Listing and watching *rbac.ClusterRoleBinding from storage/cacher.go:/clusterrolebindings
I0111 19:13:59.679867  120459 storage_factory.go:285] storing roles.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.679983  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.680006  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.680044  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.680121  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.680373  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.680446  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.680487  120459 store.go:1414] Monitoring roles.rbac.authorization.k8s.io count at <storage-prefix>//roles
I0111 19:13:59.680565  120459 reflector.go:169] Listing and watching *rbac.Role from storage/cacher.go:/roles
I0111 19:13:59.680696  120459 storage_factory.go:285] storing rolebindings.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.680771  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.680802  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.680879  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.680965  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.681389  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.681494  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.681502  120459 store.go:1414] Monitoring rolebindings.rbac.authorization.k8s.io count at <storage-prefix>//rolebindings
I0111 19:13:59.681566  120459 reflector.go:169] Listing and watching *rbac.RoleBinding from storage/cacher.go:/rolebindings
I0111 19:13:59.681564  120459 storage_factory.go:285] storing clusterroles.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.681717  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.681791  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.681841  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.681919  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.682162  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.682289  120459 store.go:1414] Monitoring clusterroles.rbac.authorization.k8s.io count at <storage-prefix>//clusterroles
I0111 19:13:59.682336  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.682375  120459 reflector.go:169] Listing and watching *rbac.ClusterRole from storage/cacher.go:/clusterroles
I0111 19:13:59.682529  120459 storage_factory.go:285] storing clusterrolebindings.rbac.authorization.k8s.io in rbac.authorization.k8s.io/v1, reading as rbac.authorization.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.682830  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.682862  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.682959  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.683099  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.683993  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.684068  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.684083  120459 store.go:1414] Monitoring clusterrolebindings.rbac.authorization.k8s.io count at <storage-prefix>//clusterrolebindings
I0111 19:13:59.684114  120459 master.go:416] Enabling API group "rbac.authorization.k8s.io".
I0111 19:13:59.684113  120459 reflector.go:169] Listing and watching *rbac.ClusterRoleBinding from storage/cacher.go:/clusterrolebindings
I0111 19:13:59.686652  120459 storage_factory.go:285] storing priorityclasses.scheduling.k8s.io in scheduling.k8s.io/v1beta1, reading as scheduling.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.686800  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.686820  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.686865  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.686927  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.687177  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.687286  120459 store.go:1414] Monitoring priorityclasses.scheduling.k8s.io count at <storage-prefix>//priorityclasses
I0111 19:13:59.687304  120459 master.go:416] Enabling API group "scheduling.k8s.io".
I0111 19:13:59.687326  120459 master.go:408] Skipping disabled API group "settings.k8s.io".
I0111 19:13:59.687492  120459 storage_factory.go:285] storing storageclasses.storage.k8s.io in storage.k8s.io/v1beta1, reading as storage.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.687591  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.687617  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.687649  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.687787  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.687828  120459 reflector.go:169] Listing and watching *scheduling.PriorityClass from storage/cacher.go:/priorityclasses
I0111 19:13:59.688011  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.688299  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.688473  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.688587  120459 store.go:1414] Monitoring storageclasses.storage.k8s.io count at <storage-prefix>//storageclasses
I0111 19:13:59.689421  120459 reflector.go:169] Listing and watching *storage.StorageClass from storage/cacher.go:/storageclasses
I0111 19:13:59.689468  120459 storage_factory.go:285] storing volumeattachments.storage.k8s.io in storage.k8s.io/v1beta1, reading as storage.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.689745  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.689830  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.689907  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.690158  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.693386  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.694158  120459 store.go:1414] Monitoring volumeattachments.storage.k8s.io count at <storage-prefix>//volumeattachments
I0111 19:13:59.695422  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.696024  120459 storage_factory.go:285] storing storageclasses.storage.k8s.io in storage.k8s.io/v1beta1, reading as storage.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.696409  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.696425  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.696721  120459 reflector.go:169] Listing and watching *storage.VolumeAttachment from storage/cacher.go:/volumeattachments
I0111 19:13:59.697062  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.697737  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.710400  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.711010  120459 store.go:1414] Monitoring storageclasses.storage.k8s.io count at <storage-prefix>//storageclasses
I0111 19:13:59.711267  120459 storage_factory.go:285] storing volumeattachments.storage.k8s.io in storage.k8s.io/v1beta1, reading as storage.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.711605  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.711655  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.711852  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.712297  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.712687  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.713748  120459 reflector.go:169] Listing and watching *storage.StorageClass from storage/cacher.go:/storageclasses
I0111 19:13:59.715132  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.715700  120459 store.go:1414] Monitoring volumeattachments.storage.k8s.io count at <storage-prefix>//volumeattachments
I0111 19:13:59.715728  120459 master.go:416] Enabling API group "storage.k8s.io".
I0111 19:13:59.716654  120459 storage_factory.go:285] storing deployments.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.716957  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.716981  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.721905  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.722173  120459 reflector.go:169] Listing and watching *storage.VolumeAttachment from storage/cacher.go:/volumeattachments
I0111 19:13:59.725738  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.725965  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.730297  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.731078  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.731467  120459 store.go:1414] Monitoring deployments.apps count at <storage-prefix>//deployments
I0111 19:13:59.732291  120459 reflector.go:169] Listing and watching *apps.Deployment from storage/cacher.go:/deployments
I0111 19:13:59.732878  120459 storage_factory.go:285] storing statefulsets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.733154  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.733179  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.733389  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.735876  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.736623  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.736889  120459 store.go:1414] Monitoring statefulsets.apps count at <storage-prefix>//statefulsets
I0111 19:13:59.737417  120459 storage_factory.go:285] storing controllerrevisions.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.737650  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.737704  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.737754  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.737934  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.737993  120459 reflector.go:169] Listing and watching *apps.StatefulSet from storage/cacher.go:/statefulsets
I0111 19:13:59.738392  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.738706  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.738856  120459 store.go:1414] Monitoring controllerrevisions.apps count at <storage-prefix>//controllerrevisions
I0111 19:13:59.739129  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.739552  120459 reflector.go:169] Listing and watching *apps.ControllerRevision from storage/cacher.go:/controllerrevisions
I0111 19:13:59.739603  120459 storage_factory.go:285] storing deployments.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.739766  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.739791  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.739830  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.739881  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.740154  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.740359  120459 store.go:1414] Monitoring deployments.apps count at <storage-prefix>//deployments
I0111 19:13:59.740944  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.741520  120459 reflector.go:169] Listing and watching *apps.Deployment from storage/cacher.go:/deployments
I0111 19:13:59.743323  120459 storage_factory.go:285] storing statefulsets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.743455  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.743477  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.743520  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.743642  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.745050  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.745290  120459 store.go:1414] Monitoring statefulsets.apps count at <storage-prefix>//statefulsets
I0111 19:13:59.745657  120459 storage_factory.go:285] storing daemonsets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.745782  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.745802  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.745871  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.745969  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.746032  120459 reflector.go:169] Listing and watching *apps.StatefulSet from storage/cacher.go:/statefulsets
I0111 19:13:59.746423  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.747121  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.747431  120459 store.go:1414] Monitoring daemonsets.apps count at <storage-prefix>//daemonsets
I0111 19:13:59.747595  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.749169  120459 reflector.go:169] Listing and watching *apps.DaemonSet from storage/cacher.go:/daemonsets
I0111 19:13:59.749931  120459 storage_factory.go:285] storing replicasets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.750326  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.750351  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.752616  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.752710  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.753437  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.753909  120459 store.go:1414] Monitoring replicasets.apps count at <storage-prefix>//replicasets
I0111 19:13:59.754616  120459 storage_factory.go:285] storing controllerrevisions.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.754790  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.754814  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.754861  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.755147  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.755839  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.757044  120459 reflector.go:169] Listing and watching *apps.ReplicaSet from storage/cacher.go:/replicasets
I0111 19:13:59.758064  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.758257  120459 store.go:1414] Monitoring controllerrevisions.apps count at <storage-prefix>//controllerrevisions
I0111 19:13:59.758570  120459 storage_factory.go:285] storing deployments.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.758862  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.758890  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.758929  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.758998  120459 reflector.go:169] Listing and watching *apps.ControllerRevision from storage/cacher.go:/controllerrevisions
I0111 19:13:59.759216  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.759252  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.760833  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.761016  120459 store.go:1414] Monitoring deployments.apps count at <storage-prefix>//deployments
I0111 19:13:59.761389  120459 storage_factory.go:285] storing statefulsets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.761550  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.761565  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.761620  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.761803  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.762053  120459 reflector.go:169] Listing and watching *apps.Deployment from storage/cacher.go:/deployments
I0111 19:13:59.762362  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.763979  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.764651  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.764825  120459 store.go:1414] Monitoring statefulsets.apps count at <storage-prefix>//statefulsets
I0111 19:13:59.765329  120459 reflector.go:169] Listing and watching *apps.StatefulSet from storage/cacher.go:/statefulsets
I0111 19:13:59.765899  120459 storage_factory.go:285] storing daemonsets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.766269  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.766286  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.766533  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.767219  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.768968  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.769169  120459 store.go:1414] Monitoring daemonsets.apps count at <storage-prefix>//daemonsets
I0111 19:13:59.769480  120459 storage_factory.go:285] storing replicasets.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.769566  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.769586  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.769622  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.769780  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.769928  120459 reflector.go:169] Listing and watching *apps.DaemonSet from storage/cacher.go:/daemonsets
I0111 19:13:59.770424  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.773303  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.773471  120459 store.go:1414] Monitoring replicasets.apps count at <storage-prefix>//replicasets
I0111 19:13:59.773880  120459 storage_factory.go:285] storing controllerrevisions.apps in apps/v1, reading as apps/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.774069  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.774087  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.774160  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.774306  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.774382  120459 reflector.go:169] Listing and watching *apps.ReplicaSet from storage/cacher.go:/replicasets
I0111 19:13:59.774620  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.774889  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.775004  120459 store.go:1414] Monitoring controllerrevisions.apps count at <storage-prefix>//controllerrevisions
I0111 19:13:59.775021  120459 master.go:416] Enabling API group "apps".
I0111 19:13:59.775096  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.775118  120459 storage_factory.go:285] storing validatingwebhookconfigurations.admissionregistration.k8s.io in admissionregistration.k8s.io/v1beta1, reading as admissionregistration.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.775214  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.775248  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.775301  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.775698  120459 reflector.go:169] Listing and watching *apps.ControllerRevision from storage/cacher.go:/controllerrevisions
I0111 19:13:59.776736  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.780141  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.780338  120459 store.go:1414] Monitoring validatingwebhookconfigurations.admissionregistration.k8s.io count at <storage-prefix>//validatingwebhookconfigurations
I0111 19:13:59.780410  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.780412  120459 storage_factory.go:285] storing mutatingwebhookconfigurations.admissionregistration.k8s.io in admissionregistration.k8s.io/v1beta1, reading as admissionregistration.k8s.io/__internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.780512  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.780523  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.780562  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.780653  120459 reflector.go:169] Listing and watching *admissionregistration.ValidatingWebhookConfiguration from storage/cacher.go:/validatingwebhookconfigurations
I0111 19:13:59.780666  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.791841  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.791998  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.800025  120459 store.go:1414] Monitoring mutatingwebhookconfigurations.admissionregistration.k8s.io count at <storage-prefix>//mutatingwebhookconfigurations
I0111 19:13:59.800069  120459 master.go:416] Enabling API group "admissionregistration.k8s.io".
I0111 19:13:59.800083  120459 reflector.go:169] Listing and watching *admissionregistration.MutatingWebhookConfiguration from storage/cacher.go:/mutatingwebhookconfigurations
I0111 19:13:59.800115  120459 storage_factory.go:285] storing events in v1, reading as __internal from storagebackend.Config{Type:"", Prefix:"0c88a306-f2a2-40a8-93c1-472b7f2fb7b6", Transport:storagebackend.TransportConfig{ServerList:[]string{"http://127.0.0.1:2379"}, KeyFile:"", CertFile:"", CAFile:""}, Quorum:false, Paging:true, Codec:runtime.Codec(nil), Transformer:value.Transformer(nil), CompactionInterval:300000000000, CountMetricPollPeriod:60000000000}
I0111 19:13:59.800462  120459 clientconn.go:551] parsed scheme: ""
I0111 19:13:59.800481  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:13:59.800539  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:13:59.800601  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.801711  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:13:59.801786  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:13:59.801830  120459 store.go:1414] Monitoring events count at <storage-prefix>//events
I0111 19:13:59.801860  120459 master.go:416] Enabling API group "events.k8s.io".
W0111 19:13:59.807019  120459 genericapiserver.go:334] Skipping API batch/v2alpha1 because it has no resources.
W0111 19:13:59.821020  120459 genericapiserver.go:334] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources.
W0111 19:13:59.821862  120459 genericapiserver.go:334] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources.
W0111 19:13:59.824229  120459 genericapiserver.go:334] Skipping API storage.k8s.io/v1alpha1 because it has no resources.
W0111 19:13:59.838108  120459 genericapiserver.go:334] Skipping API admissionregistration.k8s.io/v1alpha1 because it has no resources.
I0111 19:13:59.840607  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:13:59.840640  120459 healthz.go:170] healthz check poststarthook/bootstrap-controller failed: not finished
I0111 19:13:59.840649  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:13:59.840658  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:13:59.840665  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:13:59.840841  120459 wrap.go:47] GET /healthz: (372.66µs) 500
goroutine 27549 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00fd861c0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00fd861c0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00ff52480, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc01006e048, 0xc0061bc000, 0x18a, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc01006e048, 0xc00fe18300)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc01006e048, 0xc00fe18300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00ff955c0, 0xc00f15bda0, 0x604c620, 0xc01006e048, 0xc00fe18300)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[-]poststarthook/bootstrap-controller failed: reason withheld\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58612]
I0111 19:13:59.842346  120459 wrap.go:47] GET /api/v1/services: (1.009764ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.846087  120459 wrap.go:47] GET /api/v1/services: (1.064941ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.848847  120459 wrap.go:47] GET /api/v1/namespaces/default: (1.074407ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.850496  120459 wrap.go:47] POST /api/v1/namespaces: (1.298022ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.851969  120459 wrap.go:47] GET /api/v1/namespaces/default/services/kubernetes: (1.057299ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.855607  120459 wrap.go:47] POST /api/v1/namespaces/default/services: (3.185145ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.857091  120459 wrap.go:47] GET /api/v1/namespaces/default/endpoints/kubernetes: (1.116479ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.859175  120459 wrap.go:47] POST /api/v1/namespaces/default/endpoints: (1.741898ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.860832  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.09387ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.861617  120459 wrap.go:47] GET /api/v1/namespaces/default: (1.431907ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:13:59.861980  120459 wrap.go:47] GET /api/v1/services: (1.0261ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:13:59.862711  120459 wrap.go:47] GET /api/v1/services: (1.31408ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58618]
I0111 19:13:59.862802  120459 wrap.go:47] POST /api/v1/namespaces: (1.638046ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58612]
I0111 19:13:59.863816  120459 wrap.go:47] GET /api/v1/namespaces/default/services/kubernetes: (1.653725ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:13:59.863941  120459 wrap.go:47] GET /api/v1/namespaces/kube-public: (745.915µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58618]
I0111 19:13:59.865154  120459 wrap.go:47] GET /api/v1/namespaces/default/endpoints/kubernetes: (1.036269ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:13:59.865984  120459 wrap.go:47] POST /api/v1/namespaces: (1.660413ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:13:59.867221  120459 wrap.go:47] GET /api/v1/namespaces/kube-node-lease: (851.013µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:13:59.869009  120459 wrap.go:47] POST /api/v1/namespaces: (1.390372ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:13:59.941710  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:13:59.941753  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:13:59.941765  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:13:59.941773  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:13:59.941942  120459 wrap.go:47] GET /healthz: (371.244µs) 500
goroutine 27581 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc010d9d6c0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc010d9d6c0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc0098a09a0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff642b8, 0xc0033fe180, 0x175, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353e00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff642b8, 0xc003353c00)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff642b8, 0xc003353c00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00ca7f740, 0xc00f15bda0, 0x604c620, 0xc00ff642b8, 0xc003353c00)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.041704  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:14:00.041748  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.041758  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.041765  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.041922  120459 wrap.go:47] GET /healthz: (368.357µs) 500
goroutine 27583 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc010d9d7a0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc010d9d7a0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc0098a0a40, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff642c0, 0xc0033fea80, 0x175, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38200)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff642c0, 0xc003d38200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00ca7f8c0, 0xc00f15bda0, 0x604c620, 0xc00ff642c0, 0xc003d38200)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.141624  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:14:00.141662  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.141691  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.141698  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.141836  120459 wrap.go:47] GET /healthz: (339.748µs) 500
goroutine 27628 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00026c5b0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00026c5b0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc009877d20, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40a268, 0xc0034cf500, 0x175, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b000)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40a268, 0xc003d9b000)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00c5dd4a0, 0xc00f15bda0, 0x604c620, 0xc00f40a268, 0xc003d9b000)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.241621  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:14:00.241660  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.241694  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.241702  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.241852  120459 wrap.go:47] GET /healthz: (340.569µs) 500
goroutine 27454 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc0002941c0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc0002941c0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00988d760, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc006404288, 0xc0098d0180, 0x175, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc006404288, 0xc0020c9400)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc006404288, 0xc0020c9400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00c62ff20, 0xc00f15bda0, 0x604c620, 0xc006404288, 0xc0020c9400)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.341621  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:14:00.341654  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.341664  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.341686  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.341819  120459 wrap.go:47] GET /healthz: (318.336µs) 500
goroutine 27630 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00026c7e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00026c7e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc009877fa0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40a290, 0xc0034cfb00, 0x175, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b700)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40a290, 0xc003d9b700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00c4d8780, 0xc00f15bda0, 0x604c620, 0xc00f40a290, 0xc003d9b700)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.443095  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:14:00.443138  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.443148  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.443155  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.443312  120459 wrap.go:47] GET /healthz: (354.068µs) 500
goroutine 27585 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc010d9d880, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc010d9d880, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc0098a0ae0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff642c8, 0xc0033ff080, 0x175, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38800)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff642c8, 0xc003d38800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00ca7f980, 0xc00f15bda0, 0x604c620, 0xc00ff642c8, 0xc003d38800)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.541620  120459 healthz.go:170] healthz check etcd failed: etcd client connection not yet established
I0111 19:14:00.541656  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.541665  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.541693  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.541848  120459 wrap.go:47] GET /healthz: (362.005µs) 500
goroutine 27632 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00026c930, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00026c930, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc009852220, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40a2b8, 0xc003b9a480, 0x175, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9be00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9bd00)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40a2b8, 0xc003d9bd00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00c468300, 0xc00f15bda0, 0x604c620, 0xc00f40a2b8, 0xc003d9bd00)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[-]etcd failed: reason withheld\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.613649  120459 clientconn.go:551] parsed scheme: ""
I0111 19:14:00.613703  120459 clientconn.go:557] scheme "" not registered, fallback to default scheme
I0111 19:14:00.613756  120459 resolver_conn_wrapper.go:116] ccResolverWrapper: sending new addresses to cc: [{127.0.0.1:2379 0  <nil>}]
I0111 19:14:00.613852  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:14:00.614296  120459 balancer_v1_wrapper.go:245] clientv3/balancer: pin "127.0.0.1:2379"
I0111 19:14:00.614405  120459 balancer_v1_wrapper.go:125] balancerWrapper: got update addr from Notify: [{127.0.0.1:2379 <nil>}]
I0111 19:14:00.642633  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.642664  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.642692  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.642850  120459 wrap.go:47] GET /healthz: (1.281923ms) 500
goroutine 27653 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc000229110, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc000229110, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc0098623a0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc0090589f8, 0xc0020c4580, 0x160, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3eb00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc0090589f8, 0xc002c3e900)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc0090589f8, 0xc002c3e900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00c3175c0, 0xc00f15bda0, 0x604c620, 0xc0090589f8, 0xc002c3e900)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.742367  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.742398  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.742406  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.742562  120459 wrap.go:47] GET /healthz: (1.063648ms) 500
goroutine 27456 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc0002945b0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc0002945b0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00988db40, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc006404290, 0xc00a402580, 0x160, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc006404290, 0xc0020c9800)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc006404290, 0xc0020c9800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00c58e6c0, 0xc00f15bda0, 0x604c620, 0xc006404290, 0xc0020c9800)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.842621  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.761177ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.842817  120459 wrap.go:47] GET /apis/scheduling.k8s.io/v1beta1/priorityclasses/system-node-critical: (1.746438ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58642]
I0111 19:14:00.843326  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (2.477395ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:00.844113  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.844131  120459 healthz.go:170] healthz check poststarthook/scheduling/bootstrap-system-priority-classes failed: not finished
I0111 19:14:00.844138  120459 healthz.go:170] healthz check poststarthook/ca-registration failed: not finished
I0111 19:14:00.844162  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.137226ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.844290  120459 wrap.go:47] GET /healthz: (1.683947ms) 500
goroutine 27673 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00026caf0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00026caf0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc009852660, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40a328, 0xc002a84c60, 0x160, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40a328, 0xc003a53200)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40a328, 0xc003a53200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00c1a5da0, 0xc00f15bda0, 0x604c620, 0xc00f40a328, 0xc003a53200)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld\n[-]poststarthook/ca-registration failed: reason withheld\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58644]
I0111 19:14:00.844635  120459 wrap.go:47] POST /apis/scheduling.k8s.io/v1beta1/priorityclasses: (1.451481ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58642]
I0111 19:14:00.844818  120459 storage_scheduling.go:91] created PriorityClass system-node-critical with value 2000001000
I0111 19:14:00.845382  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:aggregate-to-admin: (812.613µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58644]
I0111 19:14:00.845757  120459 wrap.go:47] GET /apis/scheduling.k8s.io/v1beta1/priorityclasses/system-cluster-critical: (757.494µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58642]
I0111 19:14:00.845905  120459 wrap.go:47] GET /api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication: (2.270373ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:00.847543  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin: (1.226308ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58644]
I0111 19:14:00.847925  120459 wrap.go:47] POST /api/v1/namespaces/kube-system/configmaps: (1.404548ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:00.847957  120459 wrap.go:47] POST /apis/scheduling.k8s.io/v1beta1/priorityclasses: (1.492732ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.848084  120459 storage_scheduling.go:91] created PriorityClass system-cluster-critical with value 2000000000
I0111 19:14:00.848096  120459 storage_scheduling.go:100] all system priority classes are created successfully or already exist.
I0111 19:14:00.848633  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:aggregate-to-edit: (759.464µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58644]
I0111 19:14:00.849750  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit: (724.501µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.850851  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:aggregate-to-view: (792.853µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.851970  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view: (799.472µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.853035  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/cluster-admin: (738.325µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.854812  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.364093ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.855058  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/cluster-admin
I0111 19:14:00.856050  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:discovery: (765.424µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.857699  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.21775ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.857854  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:discovery
I0111 19:14:00.858782  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:basic-user: (749.208µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.860321  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.182536ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.860554  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:basic-user
I0111 19:14:00.861447  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/admin: (701.356µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.863195  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.250148ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.863366  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/admin
I0111 19:14:00.864268  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/edit: (712.226µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.865957  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.348427ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.866149  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/edit
I0111 19:14:00.867096  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/view: (765.716µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.868754  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.24014ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.868954  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/view
I0111 19:14:00.869853  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:aggregate-to-admin: (717.977µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.871505  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.29541ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.871727  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:aggregate-to-admin
I0111 19:14:00.872592  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:aggregate-to-edit: (720.362µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.874513  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.512841ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.874808  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:aggregate-to-edit
I0111 19:14:00.875799  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:aggregate-to-view: (748.363µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.877642  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.461723ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.877867  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:aggregate-to-view
I0111 19:14:00.878863  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:heapster: (839.774µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.880517  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.336986ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.880738  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:heapster
I0111 19:14:00.881708  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:node: (779.813µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.884346  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (2.215991ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.884627  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:node
I0111 19:14:00.885707  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:node-problem-detector: (861.111µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.887352  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.300056ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.887609  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:node-problem-detector
I0111 19:14:00.888939  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:node-proxier: (814.15µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.890474  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.20503ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.890640  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:node-proxier
I0111 19:14:00.891569  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:kubelet-api-admin: (721.723µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.893208  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.257036ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.893388  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:kubelet-api-admin
I0111 19:14:00.894390  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:node-bootstrapper: (813.116µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.895792  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.110942ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.896017  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:node-bootstrapper
I0111 19:14:00.896938  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:auth-delegator: (754.959µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.898398  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.168511ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.898633  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:auth-delegator
I0111 19:14:00.899643  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:kube-aggregator: (817.465µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.901198  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.156505ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.901372  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:kube-aggregator
I0111 19:14:00.902473  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:kube-controller-manager: (871.457µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.904292  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.400856ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.904475  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:kube-controller-manager
I0111 19:14:00.905334  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:kube-scheduler: (705.496µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.907266  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.491089ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.907467  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:kube-scheduler
I0111 19:14:00.908362  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:kube-dns: (730.446µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.909832  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.117054ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.910024  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:kube-dns
I0111 19:14:00.910968  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:persistent-volume-provisioner: (763.669µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.912776  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.413859ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.912969  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:persistent-volume-provisioner
I0111 19:14:00.913925  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:csi-external-attacher: (777.243µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.915653  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.366157ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.915865  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:csi-external-attacher
I0111 19:14:00.916845  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:aws-cloud-provider: (774.882µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.918545  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.303688ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.918779  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:aws-cloud-provider
I0111 19:14:00.919794  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:certificates.k8s.io:certificatesigningrequests:nodeclient: (825.213µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.921475  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.319347ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.921656  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:certificates.k8s.io:certificatesigningrequests:nodeclient
I0111 19:14:00.922616  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient: (758.704µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.924178  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.202371ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.924391  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:certificates.k8s.io:certificatesigningrequests:selfnodeclient
I0111 19:14:00.925395  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:volume-scheduler: (793.729µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.926936  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.179039ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.927115  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:volume-scheduler
I0111 19:14:00.928121  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:csi-external-provisioner: (823.188µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.929804  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.344668ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.930050  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:csi-external-provisioner
I0111 19:14:00.931071  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:attachdetach-controller: (838.318µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.932656  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.285285ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.932864  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:attachdetach-controller
I0111 19:14:00.933802  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:clusterrole-aggregation-controller: (746.045µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.935326  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.126191ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.935523  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:clusterrole-aggregation-controller
I0111 19:14:00.936360  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:cronjob-controller: (683.243µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.938025  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.275518ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.938274  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:cronjob-controller
I0111 19:14:00.939230  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:daemon-set-controller: (764.774µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.941004  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.378412ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.941242  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:daemon-set-controller
I0111 19:14:00.942067  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:00.942123  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:deployment-controller: (713.687µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:00.942219  120459 wrap.go:47] GET /healthz: (856.216µs) 500
goroutine 27483 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00036f420, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00036f420, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00920f500, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc000d485e8, 0xc000076780, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc000d485e8, 0xc00636f700)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc000d485e8, 0xc00636f700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0056d0600, 0xc00f15bda0, 0x604c620, 0xc000d485e8, 0xc00636f700)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:00.943980  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.44442ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.944187  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:deployment-controller
I0111 19:14:00.945173  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:disruption-controller: (791.768µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.946944  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.426508ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.947185  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:disruption-controller
I0111 19:14:00.948283  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:endpoint-controller: (897.154µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.950059  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.395255ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.950253  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:endpoint-controller
I0111 19:14:00.951209  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:expand-controller: (765.408µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.952780  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.209419ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.953017  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:expand-controller
I0111 19:14:00.954065  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:generic-garbage-collector: (878.533µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.955756  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.255014ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.955969  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:generic-garbage-collector
I0111 19:14:00.956905  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:horizontal-pod-autoscaler: (757.455µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.958501  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.231803ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.958729  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:horizontal-pod-autoscaler
I0111 19:14:00.959599  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:job-controller: (697.374µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.961279  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.285113ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.961502  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:job-controller
I0111 19:14:00.962451  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:namespace-controller: (772.015µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.964137  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.278433ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.964383  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:namespace-controller
I0111 19:14:00.965330  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:node-controller: (763.113µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.966903  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.192744ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.967094  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:node-controller
I0111 19:14:00.968034  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:persistent-volume-binder: (765.864µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.969700  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.28061ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.969943  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:persistent-volume-binder
I0111 19:14:00.970885  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:pod-garbage-collector: (731.817µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.972598  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.32446ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.972799  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:pod-garbage-collector
I0111 19:14:00.973745  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:replicaset-controller: (756.741µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.975769  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.636449ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.975993  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:replicaset-controller
I0111 19:14:00.976939  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:replication-controller: (774.518µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.978735  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.417069ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.978981  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:replication-controller
I0111 19:14:00.979968  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:resourcequota-controller: (804.072µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.981496  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.183118ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.981739  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:resourcequota-controller
I0111 19:14:00.982844  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:route-controller: (912.747µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.984509  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.331245ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.984757  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:route-controller
I0111 19:14:00.985753  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:service-account-controller: (799.457µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.987250  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.087536ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.987479  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:service-account-controller
I0111 19:14:00.988617  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:service-controller: (879.31µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.990215  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.252651ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.990454  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:service-controller
I0111 19:14:00.991425  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:statefulset-controller: (782.627µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.992964  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.151985ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:00.993195  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:statefulset-controller
I0111 19:14:00.994197  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:ttl-controller: (794.873µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.002659  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (1.682322ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.002946  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:ttl-controller
I0111 19:14:01.022703  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:certificate-controller: (1.109803ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.042340  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.042593  120459 wrap.go:47] GET /healthz: (1.166968ms) 500
goroutine 27928 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007e28380, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007e28380, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc005287500, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff65490, 0xc004a20640, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656da00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff65490, 0xc00656d900)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff65490, 0xc00656d900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc005695920, 0xc00f15bda0, 0x604c620, 0xc00ff65490, 0xc00656d900)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.043132  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (2.059954ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.043374  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:certificate-controller
I0111 19:14:01.062562  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:pvc-protection-controller: (1.532544ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.084641  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (3.503174ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.085017  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:pvc-protection-controller
I0111 19:14:01.102470  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterroles/system:controller:pv-protection-controller: (1.41521ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.123147  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterroles: (2.059496ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.123382  120459 storage_rbac.go:187] created clusterrole.rbac.authorization.k8s.io/system:controller:pv-protection-controller
I0111 19:14:01.142180  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.142293  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/cluster-admin: (1.27129ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.142346  120459 wrap.go:47] GET /healthz: (956.54µs) 500
goroutine 27930 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007e28540, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007e28540, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc005287ba0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff654e0, 0xc004a20a00, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8100)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff654e0, 0xc006eb8100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0056b5320, 0xc00f15bda0, 0x604c620, 0xc00ff654e0, 0xc006eb8100)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.163054  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.038112ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.163390  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/cluster-admin
I0111 19:14:01.182489  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:discovery: (1.463527ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.203466  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.33865ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.203738  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:discovery
I0111 19:14:01.222285  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:basic-user: (1.147318ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.242256  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.242434  120459 wrap.go:47] GET /healthz: (1.074622ms) 500
goroutine 27988 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007f72770, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007f72770, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc0029e9e40, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc000d48da8, 0xc004a20dc0, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc000d48da8, 0xc006faf800)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc000d48da8, 0xc006faf800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc005571440, 0xc00f15bda0, 0x604c620, 0xc000d48da8, 0xc006faf800)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.242887  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.802496ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.243094  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:basic-user
I0111 19:14:01.262572  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:node-proxier: (1.454222ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.283116  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.000283ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.283372  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:node-proxier
I0111 19:14:01.302351  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:kube-controller-manager: (1.315866ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.323003  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.968485ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.323321  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:kube-controller-manager
I0111 19:14:01.342164  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.342332  120459 wrap.go:47] GET /healthz: (903.818µs) 500
goroutine 28005 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007f64a80, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007f64a80, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00285f100, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc002ab1310, 0xc001de2500, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fade00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc002ab1310, 0xc006fadd00)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc002ab1310, 0xc006fadd00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc004c1b6e0, 0xc00f15bda0, 0x604c620, 0xc002ab1310, 0xc006fadd00)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.342340  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:kube-dns: (1.285503ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.363634  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.558818ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.363946  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:kube-dns
I0111 19:14:01.382373  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:kube-scheduler: (1.290129ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.403374  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.28714ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.403651  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:kube-scheduler
I0111 19:14:01.422414  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:aws-cloud-provider: (1.220534ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.442213  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.442394  120459 wrap.go:47] GET /healthz: (979.101µs) 500
goroutine 28019 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007e29650, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007e29650, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002817300, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff65778, 0xc004a21400, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e400)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff65778, 0xc004d2e400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc005191140, 0xc00f15bda0, 0x604c620, 0xc00ff65778, 0xc004d2e400)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.443029  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.959429ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.443254  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:aws-cloud-provider
I0111 19:14:01.462421  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:volume-scheduler: (1.406799ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.484140  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (3.039794ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.484426  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:volume-scheduler
I0111 19:14:01.502477  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:node: (1.39165ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.523103  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.992056ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.523356  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:node
I0111 19:14:01.542296  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.542559  120459 wrap.go:47] GET /healthz: (1.061056ms) 500
goroutine 27982 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007e35f10, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007e35f10, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002604b40, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f77d7b8, 0xc00f972280, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9a00)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f77d7b8, 0xc0073b9a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0053d7a40, 0xc00f15bda0, 0x604c620, 0xc00f77d7b8, 0xc0073b9a00)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.542904  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:attachdetach-controller: (1.884534ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.563078  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.089532ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.563317  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:attachdetach-controller
I0111 19:14:01.582488  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:clusterrole-aggregation-controller: (1.44312ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.603114  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.015172ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.603393  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:clusterrole-aggregation-controller
I0111 19:14:01.623137  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:cronjob-controller: (1.707837ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.642279  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.642444  120459 wrap.go:47] GET /healthz: (1.012212ms) 500
goroutine 27963 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007ffec40, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007ffec40, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002491020, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40afc0, 0xc004a217c0, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40afc0, 0xc001081300)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40afc0, 0xc001081300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0057954a0, 0xc00f15bda0, 0x604c620, 0xc00f40afc0, 0xc001081300)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.642987  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.923622ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.643178  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:cronjob-controller
I0111 19:14:01.662430  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:daemon-set-controller: (1.326255ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.683728  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.664493ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.684028  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:daemon-set-controller
I0111 19:14:01.702486  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:deployment-controller: (1.356415ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.723386  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.296821ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.723645  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:deployment-controller
I0111 19:14:01.742310  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.742516  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:disruption-controller: (1.472919ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.742518  120459 wrap.go:47] GET /healthz: (1.069692ms) 500
goroutine 27965 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007ffefc0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007ffefc0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00221aa00, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40b090, 0xc004eec280, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40b090, 0xc007b08a00)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40b090, 0xc007b08a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0036b8060, 0xc00f15bda0, 0x604c620, 0xc00f40b090, 0xc007b08a00)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:01.763086  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.063725ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.763324  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:disruption-controller
I0111 19:14:01.782653  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:endpoint-controller: (1.532781ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.803420  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.278641ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.804363  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:endpoint-controller
I0111 19:14:01.822564  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:expand-controller: (1.44783ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.843364  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.843547  120459 wrap.go:47] GET /healthz: (2.097952ms) 500
goroutine 28057 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007d915e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007d915e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00215bda0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc0060cef08, 0xc0049faa00, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebf00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc0060cef08, 0xc003aebe00)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc0060cef08, 0xc003aebe00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc004bef020, 0xc00f15bda0, 0x604c620, 0xc0060cef08, 0xc003aebe00)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:01.844039  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.62765ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.844282  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:expand-controller
I0111 19:14:01.862329  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:generic-garbage-collector: (1.342792ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.883152  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.127628ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.883415  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:generic-garbage-collector
I0111 19:14:01.902397  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:horizontal-pod-autoscaler: (1.298694ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.923367  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.319032ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.923599  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:horizontal-pod-autoscaler
I0111 19:14:01.942140  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:01.942301  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:job-controller: (1.301017ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:01.942316  120459 wrap.go:47] GET /healthz: (889.101µs) 500
goroutine 28072 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc007fffc00, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc007fffc00, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc00123ca60, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40b2d8, 0xc0049fadc0, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1200)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40b2d8, 0xc00a4b1200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc004bf1560, 0xc00f15bda0, 0x604c620, 0xc00f40b2d8, 0xc00a4b1200)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:01.963015  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.974959ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:01.963287  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:job-controller
I0111 19:14:01.982714  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:namespace-controller: (1.59319ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.003208  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.057201ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.003481  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:namespace-controller
I0111 19:14:02.022976  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:node-controller: (1.586173ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.042383  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.042563  120459 wrap.go:47] GET /healthz: (1.161987ms) 500
goroutine 28077 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00a9782a0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00a9782a0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc001129800, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40b3a8, 0xc004a21e00, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434400)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40b3a8, 0xc00b434400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc006218660, 0xc00f15bda0, 0x604c620, 0xc00f40b3a8, 0xc00b434400)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:02.043104  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.9591ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.043340  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:node-controller
I0111 19:14:02.062451  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:persistent-volume-binder: (1.401194ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.085308  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.86479ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.085934  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:persistent-volume-binder
I0111 19:14:02.102491  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:pod-garbage-collector: (1.440372ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.123386  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.252309ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.123809  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:pod-garbage-collector
I0111 19:14:02.142282  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.142658  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:replicaset-controller: (1.471808ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.142946  120459 wrap.go:47] GET /healthz: (1.582096ms) 500
goroutine 28128 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00aa640e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00aa640e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc0009df880, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff65d30, 0xc001de2dc0, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5b00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5a00)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff65d30, 0xc00b3f5a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0023440c0, 0xc00f15bda0, 0x604c620, 0xc00ff65d30, 0xc00b3f5a00)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:02.163302  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.225799ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.163626  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:replicaset-controller
I0111 19:14:02.182826  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:replication-controller: (1.747267ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.214960  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (4.10639ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.215228  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:replication-controller
I0111 19:14:02.222053  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:resourcequota-controller: (1.030352ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.242399  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.242577  120459 wrap.go:47] GET /healthz: (1.181618ms) 500
goroutine 28182 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00aa64460, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00aa64460, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc001c06240, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00ff65d78, 0xc0049fb180, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2000)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00ff65d78, 0xc00b8d2000)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0023450e0, 0xc00f15bda0, 0x604c620, 0xc00ff65d78, 0xc00b8d2000)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:02.242931  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.882212ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.243162  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:resourcequota-controller
I0111 19:14:02.262396  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:route-controller: (1.336321ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.283693  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.553523ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.284665  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:route-controller
I0111 19:14:02.302281  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:service-account-controller: (1.233124ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.323015  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.963062ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.323317  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:service-account-controller
I0111 19:14:02.342275  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.342444  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:service-controller: (1.351628ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.342440  120459 wrap.go:47] GET /healthz: (982.294µs) 500
goroutine 28168 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00a4a9b90, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00a4a9b90, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002d14300, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc006405100, 0xc000076dc0, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc006405100, 0xc00b5b3400)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc006405100, 0xc00b5b3400)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0064e1620, 0xc00f15bda0, 0x604c620, 0xc006405100, 0xc00b5b3400)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:02.363141  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (1.965207ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.363435  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:service-controller
I0111 19:14:02.382499  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:statefulset-controller: (1.403789ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.403315  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.243877ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.403597  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:statefulset-controller
I0111 19:14:02.422067  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:ttl-controller: (1.052286ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.443324  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.249289ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:02.443523  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:ttl-controller
I0111 19:14:02.444217  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.444402  120459 wrap.go:47] GET /healthz: (1.112338ms) 500
goroutine 28202 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00b4e03f0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00b4e03f0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002d0de40, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f77dd38, 0xc0069b0640, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f77dd38, 0xc007836500)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f77dd38, 0xc007836500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc006ebe9c0, 0xc00f15bda0, 0x604c620, 0xc00f77dd38, 0xc007836500)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58614]
I0111 19:14:02.463121  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:certificate-controller: (2.106471ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.506337  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (25.189535ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.506628  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:certificate-controller
I0111 19:14:02.514313  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:pvc-protection-controller: (7.454208ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.523581  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.58499ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.523851  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:pvc-protection-controller
I0111 19:14:02.542884  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.543075  120459 wrap.go:47] GET /healthz: (1.67727ms) 500
goroutine 28214 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00b3455e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00b3455e0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002d9b720, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc0060cf470, 0xc0069b0c80, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de800)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc0060cf470, 0xc0066de700)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc0060cf470, 0xc0066de700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc006ef22a0, 0xc00f15bda0, 0x604c620, 0xc0060cf470, 0xc0066de700)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:02.543246  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:controller:pv-protection-controller: (2.251108ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.563475  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/clusterrolebindings: (2.460415ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.563780  120459 storage_rbac.go:215] created clusterrolebinding.rbac.authorization.k8s.io/system:controller:pv-protection-controller
I0111 19:14:02.582488  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles/extension-apiserver-authentication-reader: (1.493804ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.584354  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.383212ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.603314  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles: (2.196845ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.603667  120459 storage_rbac.go:246] created role.rbac.authorization.k8s.io/extension-apiserver-authentication-reader in kube-system
I0111 19:14:02.623786  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles/system:controller:bootstrap-signer: (2.780699ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.632552  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (8.338265ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.642147  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.642345  120459 wrap.go:47] GET /healthz: (943.724µs) 500
goroutine 28170 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00b6c2150, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00b6c2150, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002de65e0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc006405180, 0xc000077400, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3a00)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc006405180, 0xc00b5b3900)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc006405180, 0xc00b5b3900)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0064e1ce0, 0xc00f15bda0, 0x604c620, 0xc006405180, 0xc00b5b3900)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:02.642763  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles: (1.738251ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.642986  120459 storage_rbac.go:246] created role.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-system
I0111 19:14:02.662392  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles/system:controller:cloud-provider: (1.296884ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.664135  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.275176ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.683222  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles: (1.862695ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.683518  120459 storage_rbac.go:246] created role.rbac.authorization.k8s.io/system:controller:cloud-provider in kube-system
I0111 19:14:02.702317  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles/system:controller:token-cleaner: (1.262186ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.704125  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.357518ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.723500  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles: (2.425408ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.723849  120459 storage_rbac.go:246] created role.rbac.authorization.k8s.io/system:controller:token-cleaner in kube-system
I0111 19:14:02.742182  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.742330  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles/system::leader-locking-kube-controller-manager: (1.240485ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.742345  120459 wrap.go:47] GET /healthz: (920.633µs) 500
goroutine 28209 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00b4e0fc0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00b4e0fc0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002e3c5c0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f77deb8, 0xc001de3180, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837700)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f77deb8, 0xc007837600)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f77deb8, 0xc007837600)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0071cca20, 0xc00f15bda0, 0x604c620, 0xc00f77deb8, 0xc007837600)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:02.744052  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.235173ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.763298  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles: (2.280656ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.763567  120459 storage_rbac.go:246] created role.rbac.authorization.k8s.io/system::leader-locking-kube-controller-manager in kube-system
I0111 19:14:02.782398  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles/system::leader-locking-kube-scheduler: (1.302236ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.784530  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.640542ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.803752  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles: (2.668879ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.804045  120459 storage_rbac.go:246] created role.rbac.authorization.k8s.io/system::leader-locking-kube-scheduler in kube-system
I0111 19:14:02.822438  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/roles/system:controller:bootstrap-signer: (1.356551ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.824340  120459 wrap.go:47] GET /api/v1/namespaces/kube-public: (1.339494ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.842223  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.842400  120459 wrap.go:47] GET /healthz: (986.18µs) 500
goroutine 28237 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00b5693b0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00b5693b0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002eecda0, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc00f40bb88, 0xc00ef2a780, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd200)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd100)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc00f40bb88, 0xc0063dd100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc0074de3c0, 0xc00f15bda0, 0x604c620, 0xc00f40bb88, 0xc0063dd100)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:02.843152  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/roles: (2.075298ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.843400  120459 storage_rbac.go:246] created role.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-public
I0111 19:14:02.862512  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system::leader-locking-kube-controller-manager: (1.439063ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.864463  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.451903ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.916070  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings: (33.913782ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.916381  120459 storage_rbac.go:276] created rolebinding.rbac.authorization.k8s.io/system::leader-locking-kube-controller-manager in kube-system
I0111 19:14:02.917626  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system::leader-locking-kube-scheduler: (994.524µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.919869  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.796398ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.930983  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings: (10.087112ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.931258  120459 storage_rbac.go:276] created rolebinding.rbac.authorization.k8s.io/system::leader-locking-kube-scheduler in kube-system
I0111 19:14:02.942309  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system:controller:bootstrap-signer: (1.304878ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.942481  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:02.942628  120459 wrap.go:47] GET /healthz: (1.158307ms) 500
goroutine 28301 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00b4e1ea0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00b4e1ea0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002f70e40, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc002508498, 0xc001de3680, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c500)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc002508498, 0xc00ad5c300)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc002508498, 0xc00ad5c300)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00733b080, 0xc00f15bda0, 0x604c620, 0xc002508498, 0xc00ad5c300)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:02.944043  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.363507ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.963030  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings: (1.958329ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.963297  120459 storage_rbac.go:276] created rolebinding.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-system
I0111 19:14:02.982449  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system:controller:cloud-provider: (1.362837ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:02.984377  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.369983ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:03.003184  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings: (2.113393ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:03.003497  120459 storage_rbac.go:276] created rolebinding.rbac.authorization.k8s.io/system:controller:cloud-provider in kube-system
I0111 19:14:03.022409  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system:controller:token-cleaner: (1.334195ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:03.024339  120459 wrap.go:47] GET /api/v1/namespaces/kube-system: (1.358103ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:03.043172  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings: (2.146424ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:03.043628  120459 storage_rbac.go:276] created rolebinding.rbac.authorization.k8s.io/system:controller:token-cleaner in kube-system
I0111 19:14:03.044285  120459 healthz.go:170] healthz check poststarthook/rbac/bootstrap-roles failed: not finished
I0111 19:14:03.044458  120459 wrap.go:47] GET /healthz: (943.449µs) 500
goroutine 28338 [running]:
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).recordStatus(0xc00b9989a0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/httplog.(*respLogger).WriteHeader(0xc00b9989a0, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*baseTimeoutWriter).WriteHeader(0xc002fa8b00, 0x1f4)
net/http.Error(0x7f2aec1899c0, 0xc002508588, 0xc00f972780, 0x136, 0x1f4)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/healthz.handleRootHealthz.func1(0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
net/http.HandlerFunc.ServeHTTP(0xc009939d40, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*pathHandler).ServeHTTP(0xc00c74f680, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/mux.(*PathRecorderMux).ServeHTTP(0xc00dcd0ee0, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server.director.ServeHTTP(0x40e851a, 0xe, 0xc00fc9f7a0, 0xc00dcd0ee0, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthorization.func1(0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e40, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.WithMaxInFlightLimit.func1(0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
net/http.HandlerFunc.ServeHTTP(0xc0101280c0, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithImpersonation.func1(0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
net/http.HandlerFunc.ServeHTTP(0xc00fef0e80, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d100)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/endpoints/filters.WithAuthentication.func1(0x7f2aec1899c0, 0xc002508588, 0xc00ad5d000)
net/http.HandlerFunc.ServeHTTP(0xc00dcd6870, 0x7f2aec1899c0, 0xc002508588, 0xc00ad5d000)
k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP.func1(0xc00733b7a0, 0xc00f15bda0, 0x604c620, 0xc002508588, 0xc00ad5d000)
created by k8s.io/kubernetes/vendor/k8s.io/apiserver/pkg/server/filters.(*timeoutHandler).ServeHTTP

logging error output: "[+]ping ok\n[+]log ok\n[+]etcd ok\n[+]poststarthook/generic-apiserver-start-informers ok\n[+]poststarthook/bootstrap-controller ok\n[-]poststarthook/rbac/bootstrap-roles failed: reason withheld\n[+]poststarthook/scheduling/bootstrap-system-priority-classes ok\n[+]poststarthook/ca-registration ok\nhealthz check failed\n"
 [Go-http-client/1.1 127.0.0.1:58616]
I0111 19:14:03.062783  120459 wrap.go:47] GET /apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/rolebindings/system:controller:bootstrap-signer: (1.678261ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:03.064903  120459 wrap.go:47] GET /api/v1/namespaces/kube-public: (1.460213ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:03.083264  120459 wrap.go:47] POST /apis/rbac.authorization.k8s.io/v1/namespaces/kube-public/rolebindings: (2.149185ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:03.083512  120459 storage_rbac.go:276] created rolebinding.rbac.authorization.k8s.io/system:controller:bootstrap-signer in kube-public
I0111 19:14:03.142586  120459 wrap.go:47] GET /healthz: (1.018842ms) 200 [Go-http-client/1.1 127.0.0.1:58616]
W0111 19:14:03.143381  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143447  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143482  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143506  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143519  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143532  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143543  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143558  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143575  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
W0111 19:14:03.143588  120459 mutation_detector.go:48] Mutation detector is enabled, this will result in memory leakage.
I0111 19:14:03.143758  120459 factory.go:745] Creating scheduler from algorithm provider 'DefaultProvider'
I0111 19:14:03.143776  120459 factory.go:826] Creating scheduler with fit predicates 'map[MaxCSIVolumeCountPred:{} NoDiskConflict:{} CheckNodeMemoryPressure:{} CheckVolumeBinding:{} MaxGCEPDVolumeCount:{} CheckNodeDiskPressure:{} PodToleratesNodeTaints:{} CheckNodePIDPressure:{} CheckNodeCondition:{} MaxEBSVolumeCount:{} MaxAzureDiskVolumeCount:{} GeneralPredicates:{} NoVolumeZoneConflict:{} MatchInterPodAffinity:{}]' and priority functions 'map[ImageLocalityPriority:{} SelectorSpreadPriority:{} InterPodAffinityPriority:{} LeastRequestedPriority:{} BalancedResourceAllocation:{} NodePreferAvoidPodsPriority:{} NodeAffinityPriority:{} TaintTolerationPriority:{}]'
I0111 19:14:03.143893  120459 controller_utils.go:1021] Waiting for caches to sync for scheduler controller
I0111 19:14:03.144191  120459 reflector.go:131] Starting reflector *v1.Pod (12h0m0s) from k8s.io/kubernetes/test/integration/scheduler/util.go:194
I0111 19:14:03.144218  120459 reflector.go:169] Listing and watching *v1.Pod from k8s.io/kubernetes/test/integration/scheduler/util.go:194
I0111 19:14:03.145142  120459 wrap.go:47] GET /api/v1/pods?fieldSelector=status.phase%21%3DFailed%2Cstatus.phase%21%3DSucceeded&limit=500&resourceVersion=0: (637.926µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58616]
I0111 19:14:03.145982  120459 get.go:251] Starting watch for /api/v1/pods, rv=18528 labels= fields=status.phase!=Failed,status.phase!=Succeeded timeout=9m59s
I0111 19:14:03.246492  120459 shared_informer.go:123] caches populated
I0111 19:14:03.246522  120459 controller_utils.go:1028] Caches are synced for scheduler controller
I0111 19:14:03.246928  120459 reflector.go:131] Starting reflector *v1.PersistentVolume (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.246947  120459 reflector.go:169] Listing and watching *v1.PersistentVolume from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.246955  120459 reflector.go:131] Starting reflector *v1.Node (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.246976  120459 reflector.go:169] Listing and watching *v1.Node from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247321  120459 reflector.go:131] Starting reflector *v1.PersistentVolumeClaim (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247332  120459 reflector.go:169] Listing and watching *v1.PersistentVolumeClaim from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247414  120459 reflector.go:131] Starting reflector *v1.StatefulSet (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247426  120459 reflector.go:169] Listing and watching *v1.StatefulSet from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247731  120459 reflector.go:131] Starting reflector *v1.Service (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247743  120459 reflector.go:169] Listing and watching *v1.Service from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247837  120459 reflector.go:131] Starting reflector *v1beta1.PodDisruptionBudget (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.247849  120459 reflector.go:169] Listing and watching *v1beta1.PodDisruptionBudget from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.248049  120459 reflector.go:131] Starting reflector *v1.ReplicationController (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.248059  120459 reflector.go:169] Listing and watching *v1.ReplicationController from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.248188  120459 reflector.go:131] Starting reflector *v1.StorageClass (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.248204  120459 reflector.go:169] Listing and watching *v1.StorageClass from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.248845  120459 wrap.go:47] GET /api/v1/nodes?limit=500&resourceVersion=0: (648.819µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58988]
I0111 19:14:03.248960  120459 wrap.go:47] GET /api/v1/persistentvolumes?limit=500&resourceVersion=0: (492.195µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58614]
I0111 19:14:03.249285  120459 wrap.go:47] GET /apis/storage.k8s.io/v1/storageclasses?limit=500&resourceVersion=0: (360.165µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59000]
I0111 19:14:03.249306  120459 wrap.go:47] GET /api/v1/persistentvolumeclaims?limit=500&resourceVersion=0: (359.02µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58990]
I0111 19:14:03.249722  120459 wrap.go:47] GET /apis/apps/v1/statefulsets?limit=500&resourceVersion=0: (331.183µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58992]
I0111 19:14:03.249807  120459 wrap.go:47] GET /apis/policy/v1beta1/poddisruptionbudgets?limit=500&resourceVersion=0: (353.987µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58996]
I0111 19:14:03.250205  120459 wrap.go:47] GET /api/v1/services?limit=500&resourceVersion=0: (382.725µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58994]
I0111 19:14:03.250504  120459 get.go:251] Starting watch for /api/v1/nodes, rv=18528 labels= fields= timeout=8m17s
I0111 19:14:03.248388  120459 reflector.go:131] Starting reflector *v1.ReplicaSet (1s) from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.250623  120459 reflector.go:169] Listing and watching *v1.ReplicaSet from k8s.io/client-go/informers/factory.go:132
I0111 19:14:03.250921  120459 wrap.go:47] GET /api/v1/replicationcontrollers?limit=500&resourceVersion=0: (1.54333ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:03.251061  120459 get.go:251] Starting watch for /api/v1/persistentvolumeclaims, rv=18528 labels= fields= timeout=6m7s
I0111 19:14:03.251201  120459 get.go:251] Starting watch for /apis/storage.k8s.io/v1/storageclasses, rv=18532 labels= fields= timeout=9m59s
I0111 19:14:03.251389  120459 get.go:251] Starting watch for /apis/apps/v1/statefulsets, rv=18533 labels= fields= timeout=7m59s
I0111 19:14:03.251495  120459 get.go:251] Starting watch for /apis/policy/v1beta1/poddisruptionbudgets, rv=18530 labels= fields= timeout=7m18s
I0111 19:14:03.252062  120459 get.go:251] Starting watch for /api/v1/replicationcontrollers, rv=18528 labels= fields= timeout=7m45s
I0111 19:14:03.252078  120459 get.go:251] Starting watch for /api/v1/persistentvolumes, rv=18528 labels= fields= timeout=6m52s
I0111 19:14:03.251986  120459 get.go:251] Starting watch for /api/v1/services, rv=18540 labels= fields= timeout=6m55s
I0111 19:14:03.253871  120459 wrap.go:47] GET /apis/apps/v1/replicasets?limit=500&resourceVersion=0: (2.361378ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58994]
I0111 19:14:03.254461  120459 get.go:251] Starting watch for /apis/apps/v1/replicasets, rv=18533 labels= fields= timeout=9m43s
I0111 19:14:03.346864  120459 shared_informer.go:123] caches populated
I0111 19:14:03.447112  120459 shared_informer.go:123] caches populated
I0111 19:14:03.547326  120459 shared_informer.go:123] caches populated
I0111 19:14:03.647580  120459 shared_informer.go:123] caches populated
I0111 19:14:03.747785  120459 shared_informer.go:123] caches populated
I0111 19:14:03.848029  120459 shared_informer.go:123] caches populated
I0111 19:14:03.948258  120459 shared_informer.go:123] caches populated
I0111 19:14:04.048465  120459 shared_informer.go:123] caches populated
I0111 19:14:04.148716  120459 shared_informer.go:123] caches populated
I0111 19:14:04.248970  120459 shared_informer.go:123] caches populated
I0111 19:14:04.249424  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:04.249542  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:04.250520  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:04.250613  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:04.251406  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:04.251518  120459 wrap.go:47] POST /api/v1/nodes: (2.072611ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.254088  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.045663ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.254402  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:04.254415  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:04.254559  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1"
I0111 19:14:04.254587  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1": all PVCs bound and nothing to do
I0111 19:14:04.254634  120459 factory.go:1166] Attempting to bind rpod-0 to node1
I0111 19:14:04.256368  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.865956ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.256415  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:04.256427  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:04.256538  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1"
I0111 19:14:04.256566  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1": all PVCs bound and nothing to do
I0111 19:14:04.256608  120459 factory.go:1166] Attempting to bind rpod-1 to node1
I0111 19:14:04.256731  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0/binding: (1.613376ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.256870  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:04.258669  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.512233ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.258880  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1/binding: (2.05009ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.259072  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:04.260813  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.462086ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.359050  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (1.92833ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.461625  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (1.745934ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.462002  120459 preemption_test.go:561] Creating the preemptor pod...
I0111 19:14:04.464872  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.545782ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.465204  120459 preemption_test.go:567] Creating additional pods...
I0111 19:14:04.465248  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:04.465261  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:04.465387  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.465454  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.467060  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.296886ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.468209  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.993044ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59058]
I0111 19:14:04.468397  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (2.181576ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59056]
I0111 19:14:04.468468  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.033712ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:58998]
I0111 19:14:04.469807  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.092237ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59058]
I0111 19:14:04.470080  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.471821  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.920211ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.472569  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (2.076953ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59058]
I0111 19:14:04.473786  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.596153ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.476002  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.860217ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.477873  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (4.960891ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59058]
I0111 19:14:04.478206  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:04.478222  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:04.478360  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.478400  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.478625  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.120602ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.480555  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (1.728124ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59060]
I0111 19:14:04.480657  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.302753ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59064]
I0111 19:14:04.481090  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.245653ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.481229  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.880448ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59058]
I0111 19:14:04.483546  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.663937ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59060]
I0111 19:14:04.483875  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.486423  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (4.532431ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59062]
I0111 19:14:04.487340  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/preemptor-pod.1578e12a74c0e852: (4.808556ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59058]
I0111 19:14:04.488070  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (3.805232ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59060]
I0111 19:14:04.488427  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:04.488448  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:04.488589  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1"
I0111 19:14:04.488608  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1": all PVCs bound and nothing to do
I0111 19:14:04.488719  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:04.488735  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:04.488826  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.488875  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.489499  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.652778ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59062]
I0111 19:14:04.491128  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.513381ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59066]
I0111 19:14:04.491422  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.806324ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.491996  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.47633ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59062]
I0111 19:14:04.492295  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6/status: (2.151623ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59058]
I0111 19:14:04.493307  120459 factory.go:1166] Attempting to bind preemptor-pod to node1
I0111 19:14:04.493838  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.073392ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.494093  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.494273  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:04.494294  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:04.494414  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.494469  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.495166  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.709186ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59062]
I0111 19:14:04.495649  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/binding: (2.147471ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59066]
I0111 19:14:04.495793  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:04.498005  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.95116ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59070]
I0111 19:14:04.498061  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.642559ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59062]
I0111 19:14:04.498307  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.546577ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59072]
I0111 19:14:04.498461  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1/status: (1.494478ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.499940  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.457711ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59068]
I0111 19:14:04.500124  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.256761ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59018]
I0111 19:14:04.500386  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.500584  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:04.500607  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:04.500730  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.500778  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.501197  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.561125ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59072]
I0111 19:14:04.502229  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.010902ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59068]
I0111 19:14:04.502575  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.298491ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59074]
I0111 19:14:04.503002  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (1.81477ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59066]
I0111 19:14:04.504726  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.719717ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59068]
I0111 19:14:04.505627  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (2.285956ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59066]
I0111 19:14:04.506553  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.506718  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:04.506734  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:04.506824  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.506876  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.508579  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.979699ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59068]
I0111 19:14:04.510730  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10/status: (3.382969ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59072]
I0111 19:14:04.510834  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (3.426493ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59076]
I0111 19:14:04.510847  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.053838ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59078]
I0111 19:14:04.512327  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.243635ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59076]
I0111 19:14:04.512582  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.512791  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:04.512813  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:04.512933  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.512973  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.092888ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59068]
I0111 19:14:04.512979  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.515093  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.611176ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59072]
I0111 19:14:04.515414  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12/status: (2.146518ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59078]
I0111 19:14:04.515506  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.840449ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59082]
I0111 19:14:04.516841  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.944332ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59080]
I0111 19:14:04.517508  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.490752ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59082]
I0111 19:14:04.517542  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.447008ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59072]
I0111 19:14:04.517966  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.518167  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:04.518189  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:04.518397  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.518451  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.519948  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.083603ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59086]
I0111 19:14:04.520726  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.357743ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59072]
I0111 19:14:04.520789  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.427561ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59088]
I0111 19:14:04.522388  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14/status: (3.724321ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59080]
I0111 19:14:04.523535  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.712978ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59088]
I0111 19:14:04.525889  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (2.856262ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59080]
I0111 19:14:04.526128  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.526569  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.711457ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59088]
I0111 19:14:04.527089  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:04.527104  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:04.527224  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.527279  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.529924  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.782304ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59080]
I0111 19:14:04.529936  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.848773ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59090]
I0111 19:14:04.530339  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.189352ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.530816  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16/status: (3.315108ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59086]
I0111 19:14:04.532284  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.085552ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.532285  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.941455ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59090]
I0111 19:14:04.532555  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.532727  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:04.532738  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:04.532821  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.532874  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.534814  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.504578ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59094]
I0111 19:14:04.534921  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.221155ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59096]
I0111 19:14:04.535140  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18/status: (2.030285ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59080]
I0111 19:14:04.535626  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.670481ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.537078  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.453822ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59096]
I0111 19:14:04.537430  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.354223ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.537813  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.538023  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:04.538042  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:04.538167  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.538218  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.540345  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.251159ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.540403  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.42035ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.540781  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (2.133321ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59098]
I0111 19:14:04.542015  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20/status: (3.563343ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59094]
I0111 19:14:04.542866  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.01436ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.544155  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.669162ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59098]
I0111 19:14:04.544425  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.544591  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:04.544603  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:04.544697  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.544736  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.546057  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.027803ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.546214  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.687057ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.546730  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23/status: (1.771464ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59098]
I0111 19:14:04.548175  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.096422ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59098]
I0111 19:14:04.548519  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.548530  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.90282ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59102]
I0111 19:14:04.548742  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:04.548766  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:04.548862  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.199524ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59092]
I0111 19:14:04.548912  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.548986  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.551079  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.810128ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59098]
I0111 19:14:04.551722  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25/status: (2.380041ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.551823  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.299486ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59104]
I0111 19:14:04.551997  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.142065ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59106]
I0111 19:14:04.553971  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.379713ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.554038  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.451655ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59104]
I0111 19:14:04.554214  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.554383  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:04.554403  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:04.554493  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.554554  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.555954  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.463631ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.557021  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.817352ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59110]
I0111 19:14:04.557154  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (2.06198ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59108]
I0111 19:14:04.557499  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (2.689675ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59098]
I0111 19:14:04.558116  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.597511ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.559014  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.122345ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59108]
I0111 19:14:04.559322  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.559484  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:04.559497  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:04.559604  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.559640  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.560272  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.711664ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.582216  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (21.779674ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59112]
I0111 19:14:04.592989  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (32.576878ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59110]
I0111 19:14:04.594046  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (32.773645ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.594059  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29/status: (33.142045ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59108]
I0111 19:14:04.596289  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.189703ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59110]
I0111 19:14:04.597219  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.597394  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.950455ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.597428  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:04.597446  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:04.597591  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.597630  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.600386  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.857011ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.600639  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.329662ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59112]
I0111 19:14:04.601283  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31/status: (2.254881ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59110]
I0111 19:14:04.602103  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.843253ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59154]
I0111 19:14:04.604747  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.605446ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59112]
I0111 19:14:04.607804  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (3.461596ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59110]
I0111 19:14:04.608076  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.608294  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:04.608310  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:04.608516  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.608621  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.610467  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.099246ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59170]
I0111 19:14:04.610788  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.671596ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.611999  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (2.982284ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59110]
I0111 19:14:04.615051  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (2.591155ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.615386  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (10.172296ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59154]
I0111 19:14:04.617191  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.617428  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:04.617452  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:04.617604  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.617712  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.620513  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (4.192072ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59154]
I0111 19:14:04.621516  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36/status: (2.089527ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59170]
I0111 19:14:04.621606  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (2.109978ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.623084  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.158795ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59170]
I0111 19:14:04.623635  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.45568ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59186]
I0111 19:14:04.624387  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.624601  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:04.624645  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:04.624770  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.624838  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.626147  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.095587ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.627204  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.126913ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59196]
I0111 19:14:04.628437  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.692262ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.629649  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37/status: (1.975768ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59182]
I0111 19:14:04.630979  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (9.124547ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59154]
I0111 19:14:04.631202  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.303727ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59100]
I0111 19:14:04.632453  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.740523ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59182]
I0111 19:14:04.632725  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.632770  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.216945ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59154]
I0111 19:14:04.632898  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:04.632928  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:04.633023  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.633071  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.633289  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.616057ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59196]
I0111 19:14:04.634469  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.073614ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59182]
I0111 19:14:04.635085  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.362109ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59196]
I0111 19:14:04.635758  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39/status: (2.387685ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59154]
I0111 19:14:04.635778  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.988912ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59204]
I0111 19:14:04.637723  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.524711ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59196]
I0111 19:14:04.637757  120459 cacher.go:598] cacher (*core.Pod): 1 objects queued in incoming channel.
I0111 19:14:04.638861  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.660585ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59182]
I0111 19:14:04.639444  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.640521  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:04.640707  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:04.640934  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.641012  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.641440  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.150384ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59182]
I0111 19:14:04.643410  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.343218ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59218]
I0111 19:14:04.643435  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.474994ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.644013  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42/status: (2.34834ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59196]
I0111 19:14:04.646150  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.124852ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59182]
I0111 19:14:04.648349  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.519996ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.648782  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.884329ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59218]
I0111 19:14:04.649385  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.649569  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:04.649585  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:04.649668  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.649776  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.652561  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.786684ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.652814  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (2.697751ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59218]
I0111 19:14:04.653017  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.511807ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59232]
I0111 19:14:04.653074  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (2.560523ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59230]
I0111 19:14:04.654795  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.36868ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.655131  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.655302  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:04.655320  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:04.655399  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.655505  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.657359  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.437486ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.660855  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48/status: (4.610339ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59218]
I0111 19:14:04.661072  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.924019ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59242]
I0111 19:14:04.662736  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.461796ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59218]
I0111 19:14:04.663112  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.663363  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:04.663383  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:04.663469  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.663510  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.672444  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (7.712698ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.673647  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (8.217815ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.674301  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49/status: (10.523212ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59242]
I0111 19:14:04.677646  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.481641ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.678003  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.678186  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:04.678203  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:04.678308  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.678386  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.682613  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48/status: (3.337368ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.683069  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (3.734399ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.686797  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.81599ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.687151  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.687383  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:04.687423  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:04.687642  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.687752  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.701332  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-48.1578e12a8014d32b: (21.321446ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59252]
I0111 19:14:04.701423  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (13.225045ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.709745  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-49.1578e12a808f0673: (2.806644ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.711388  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49/status: (22.762924ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59216]
I0111 19:14:04.713150  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.272811ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.713605  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.713773  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:04.713790  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:04.713863  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.713906  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.717021  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (2.838474ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.717766  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-45.1578e12a7fbd5d8f: (2.655213ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.719148  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (4.331521ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59252]
I0111 19:14:04.719495  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.879502ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59250]
I0111 19:14:04.719828  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.720078  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:04.720119  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:04.720257  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.720326  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.722144  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.083879ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.722953  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.860544ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.723909  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47/status: (3.241275ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59252]
I0111 19:14:04.727368  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.616223ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.727861  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.728079  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:04.728117  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:04.728246  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.728315  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.731669  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (2.024668ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.731741  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42/status: (1.928434ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.749179  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (16.986741ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.749498  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.749694  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:04.749714  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:04.749853  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.749902  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.751788  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.588886ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.752247  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47/status: (2.034845ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.753880  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.267207ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.754168  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.754355  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:04.754368  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:04.754471  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.754509  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.754544  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-42.1578e12a7f37594e: (24.266985ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59260]
I0111 19:14:04.756031  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.239544ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59260]
I0111 19:14:04.756935  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (2.180681ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.757978  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-47.1578e12a83f1eb3b: (2.636248ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59262]
I0111 19:14:04.758848  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.308046ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.759115  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.759282  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:04.759292  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:04.759359  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.759393  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.760812  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.09184ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59260]
I0111 19:14:04.762113  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44/status: (2.34916ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.762348  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.336485ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.762980  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.463741ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59262]
I0111 19:14:04.763389  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (953.725µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.763610  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.764796  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:04.764812  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:04.764903  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.764940  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.766626  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.106821ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.766884  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.604608ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59260]
I0111 19:14:04.767137  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (1.871362ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.768587  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (996.323µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.768875  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.769141  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:04.769153  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:04.769255  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.769273  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-46.1578e12a85fb90e3: (1.92192ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59260]
I0111 19:14:04.769290  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.771169  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44/status: (1.617866ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.771283  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.054085ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.773493  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-44.1578e12a864618ad: (2.446659ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59264]
I0111 19:14:04.773788  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (2.300007ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59258]
I0111 19:14:04.774078  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.774221  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:04.774249  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:04.774325  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.774403  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.779426  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.280225ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.779751  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39/status: (1.598231ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59264]
I0111 19:14:04.781010  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-39.1578e12a7ebe8f16: (2.086295ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59266]
I0111 19:14:04.781245  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.076229ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59264]
I0111 19:14:04.781524  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.781710  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:04.781726  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:04.781826  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.781897  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.783435  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.1152ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.783883  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.350362ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59268]
I0111 19:14:04.784826  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43/status: (2.528278ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59266]
I0111 19:14:04.786498  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.138351ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59268]
I0111 19:14:04.786782  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.786972  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:04.786985  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:04.787096  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.787149  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.788366  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (954.292µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.789101  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.448423ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59270]
I0111 19:14:04.789365  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41/status: (1.937225ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59268]
I0111 19:14:04.790910  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.083084ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59270]
I0111 19:14:04.791221  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.791407  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:04.791437  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:04.791529  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.791595  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.793323  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.141688ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.793700  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43/status: (1.435599ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59270]
I0111 19:14:04.794817  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-43.1578e12a879d72da: (2.44754ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59272]
I0111 19:14:04.795219  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.037833ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59270]
I0111 19:14:04.795520  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.795668  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:04.795704  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:04.795796  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.795880  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.797123  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.026314ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59272]
I0111 19:14:04.798307  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41/status: (1.378669ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59274]
I0111 19:14:04.799456  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-41.1578e12a87ed98bd: (2.999067ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.799891  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (996.741µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59274]
I0111 19:14:04.800205  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.800442  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:04.800460  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:04.800540  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.800578  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.801985  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.076776ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59272]
I0111 19:14:04.802759  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37/status: (1.944607ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.804358  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.019558ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.804362  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-37.1578e12a7e40e7f2: (2.970932ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59276]
I0111 19:14:04.804631  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.805112  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:04.805148  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:04.805347  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.805440  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.806964  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.190171ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.807910  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40/status: (2.14007ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59272]
I0111 19:14:04.809160  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.342349ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59276]
I0111 19:14:04.809568  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.307528ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59272]
I0111 19:14:04.809856  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.809992  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:04.810005  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:04.810090  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.810163  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.811554  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.162747ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59276]
I0111 19:14:04.812061  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36/status: (1.661894ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.813865  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-36.1578e12a7dd41ec8: (2.634333ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59278]
I0111 19:14:04.814063  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.544172ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59256]
I0111 19:14:04.814372  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.814533  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:04.814554  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:04.814663  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.814725  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.816880  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.128995ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59276]
I0111 19:14:04.817325  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40/status: (2.32783ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59278]
I0111 19:14:04.818362  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-40.1578e12a8904a69d: (2.89006ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.819947  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.313284ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59278]
I0111 19:14:04.820282  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.820438  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:04.820455  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:04.820548  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.820591  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.822256  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.063946ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59276]
I0111 19:14:04.823083  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38/status: (1.743178ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.827560  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.591253ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.828092  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.828588  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.440161ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59276]
I0111 19:14:04.828971  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:04.828993  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:04.829302  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.829353  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.832138  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (2.183296ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.832259  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (1.884362ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59282]
I0111 19:14:04.832767  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-33.1578e12a7d496b81: (2.412829ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59284]
I0111 19:14:04.833986  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.217333ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59282]
I0111 19:14:04.834345  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.834504  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:04.834525  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:04.834616  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.834659  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.836524  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.607513ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.837960  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38/status: (3.050753ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59284]
I0111 19:14:04.838621  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-38.1578e12a89ebe428: (2.857193ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59286]
I0111 19:14:04.840168  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.335599ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59284]
I0111 19:14:04.840465  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.840695  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:04.840712  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:04.840809  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.840883  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.842327  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.124968ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.843321  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31/status: (2.143346ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59286]
I0111 19:14:04.845106  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-31.1578e12a7ca1c640: (3.541073ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59288]
I0111 19:14:04.848731  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.738959ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59286]
I0111 19:14:04.849080  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.849306  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:04.849347  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:04.849474  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.849612  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.852034  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.735861ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.852520  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35/status: (2.107107ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59290]
I0111 19:14:04.852978  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (3.072346ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59286]
I0111 19:14:04.854439  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.34035ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59290]
I0111 19:14:04.854923  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.855187  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:04.855283  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:04.855398  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.855473  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.857096  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.221093ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59286]
I0111 19:14:04.858147  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34/status: (1.970175ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.860158  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.437287ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.860282  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.780194ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59286]
I0111 19:14:04.860425  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.860602  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:04.860635  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:04.860775  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.860841  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.862789  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35/status: (1.701967ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.862931  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.359363ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59292]
I0111 19:14:04.864102  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-35.1578e12a8ba6936c: (2.010215ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59294]
I0111 19:14:04.864657  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.549009ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59280]
I0111 19:14:04.864982  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.747452ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59292]
I0111 19:14:04.865069  120459 preemption_test.go:583] Check unschedulable pods still exists and were never scheduled...
I0111 19:14:04.865392  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.865586  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:04.865604  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:04.865730  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.865773  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.868451  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (2.14723ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59298]
I0111 19:14:04.869034  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34/status: (3.015994ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.869465  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-34.1578e12a8c001e64: (2.879945ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59300]
I0111 19:14:04.869892  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (4.135284ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59294]
I0111 19:14:04.870599  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.199044ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59298]
I0111 19:14:04.870884  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.871051  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:04.871068  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:04.871157  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.871196  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.872228  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.647333ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59300]
I0111 19:14:04.874138  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (2.26972ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.874320  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.274113ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59300]
I0111 19:14:04.874738  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29/status: (2.860756ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59298]
I0111 19:14:04.874865  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-29.1578e12a7a5e16c5: (2.688068ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59302]
I0111 19:14:04.875775  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.019722ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59300]
I0111 19:14:04.876701  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.442126ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59298]
I0111 19:14:04.876919  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.877142  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:04.877183  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:04.877245  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.058778ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59300]
I0111 19:14:04.877343  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.877393  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.878850  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.013944ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59298]
I0111 19:14:04.879383  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.486885ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.879617  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (1.792597ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.881000  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (875.783µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.881012  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.714429ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59298]
I0111 19:14:04.881319  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.881388  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.774634ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.881442  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:04.881460  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:04.881524  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.881571  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.882781  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.338115ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.883117  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.384649ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.883738  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (1.929275ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.884513  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.1889ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.886148  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (2.067294ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.887390  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.887601  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:04.887619  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:04.887629  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.881696ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.887723  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.887781  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.887901  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-26.1578e12a7a102a17: (3.787411ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.889467  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.176643ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.891224  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-32.1578e12a8d4ea135: (2.399677ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59310]
I0111 19:14:04.891389  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (2.886676ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59296]
I0111 19:14:04.892055  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (3.839735ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.892809  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.070201ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59310]
I0111 19:14:04.893554  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.020099ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.893992  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.894082  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (988.888µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59310]
I0111 19:14:04.894458  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:04.894477  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:04.894608  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.895427  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (977.483µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.896247  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.347443ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.894871  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.896775  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (972.666µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.896852  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.413422ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59312]
I0111 19:14:04.898016  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (936.245µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59312]
I0111 19:14:04.898477  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (1.632814ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.899361  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (988.448µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59312]
I0111 19:14:04.900757  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.749356ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.901035  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.901110  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.262578ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59312]
I0111 19:14:04.901209  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:04.901225  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:04.901314  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.901425  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.902534  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.035189ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.903212  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.644448ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.903711  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (852.991µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.904135  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28/status: (1.86252ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59316]
I0111 19:14:04.905191  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.220277ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.905859  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.750112ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59314]
I0111 19:14:04.906576  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (2.069492ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59316]
I0111 19:14:04.906826  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.109803ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59304]
I0111 19:14:04.906867  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.907011  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:04.907069  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:04.907258  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.907358  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.908391  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.123808ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59314]
I0111 19:14:04.910138  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.33163ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59314]
I0111 19:14:04.910143  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (1.937074ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59318]
I0111 19:14:04.910688  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (2.521136ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.911093  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-30.1578e12a8e55fdf5: (2.873567ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59320]
I0111 19:14:04.912014  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.209465ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59314]
I0111 19:14:04.912191  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.111681ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59318]
I0111 19:14:04.912458  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.912600  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:04.912614  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:04.912706  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.912752  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.914776  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (2.288678ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59320]
I0111 19:14:04.915275  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28/status: (2.329078ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.915448  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (2.309163ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59322]
I0111 19:14:04.915765  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-28.1578e12a8ebd43e5: (2.285451ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59324]
I0111 19:14:04.916375  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.26263ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59320]
I0111 19:14:04.917317  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.120186ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.917582  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.917786  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:04.917804  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:04.917818  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.032181ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59320]
I0111 19:14:04.917882  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.917924  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.919491  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.092769ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59326]
I0111 19:14:04.919657  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.469142ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59322]
I0111 19:14:04.919786  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25/status: (1.592655ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.921271  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (998.18µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59322]
I0111 19:14:04.922359  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-25.1578e12a79bb2730: (3.715205ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59328]
I0111 19:14:04.922500  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (2.40448ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.922721  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.922843  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:04.922874  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:04.922942  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.922981  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.923406  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.297216ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59322]
I0111 19:14:04.925640  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.001042ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59330]
I0111 19:14:04.926770  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27/status: (3.438102ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.926859  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (3.587231ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59328]
I0111 19:14:04.926861  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (3.09719ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59322]
I0111 19:14:04.928085  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (966.452µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.928339  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.928465  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.237853ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59328]
I0111 19:14:04.928498  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:04.928541  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:04.928636  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.928925  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.929934  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (973.509µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59330]
I0111 19:14:04.930174  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.215488ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.930845  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23/status: (1.45447ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59332]
I0111 19:14:04.931610  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.080239ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.932593  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-23.1578e12a797aa96f: (2.309248ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59330]
I0111 19:14:04.933095  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.586124ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59332]
I0111 19:14:04.933403  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.933617  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:04.933652  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:04.933761  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.933828  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.933966  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.061474ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.935575  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.236973ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59334]
I0111 19:14:04.935725  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.24185ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59336]
I0111 19:14:04.936059  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.869783ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.936959  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24/status: (2.892686ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59330]
I0111 19:14:04.937529  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.498423ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59336]
I0111 19:14:04.938382  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (992.648µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.938992  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.939153  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.161971ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59336]
I0111 19:14:04.939827  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:04.939884  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:04.940132  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.940419  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.942154  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.306864ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.942329  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.26181ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.942407  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.451705ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59336]
I0111 19:14:04.944792  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (2.167845ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.945225  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22/status: (4.134814ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59334]
I0111 19:14:04.946263  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.193602ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.946923  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.027701ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59334]
I0111 19:14:04.947230  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.947396  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:04.947412  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:04.947481  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.947521  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.947952  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.103706ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.950081  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (979.757µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.950248  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-24.1578e12a90abb736: (2.077508ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.950431  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24/status: (2.646445ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59334]
I0111 19:14:04.950768  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (2.385593ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.951858  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.015211ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.952137  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.952351  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:04.952371  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:04.952453  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.952495  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.953934  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (2.780491ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.954488  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22/status: (1.745958ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.954603  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.73678ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.956104  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.074044ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.956186  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-22.1578e12a910cfd59: (3.055875ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59342]
I0111 19:14:04.956393  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.956557  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:04.956595  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:04.956740  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.956809  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.958181  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (2.210561ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.959005  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18/status: (1.898198ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.960010  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.83802ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:04.961152  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-18.1578e12a78c5a7de: (3.603395ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:04.961210  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.209079ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.961264  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (2.137238ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59338]
I0111 19:14:04.961559  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:04.961739  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:04.961780  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:04.962121  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:04.962337  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:04.962881  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.261481ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.964639  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.009885ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.965909  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (905.515µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:04.966105  120459 preemption_test.go:598] Cleaning up all pods...
I0111 19:14:04.980573  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21/status: (17.257067ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59346]
I0111 19:14:05.000475  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (37.758202ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.000960  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (19.823703ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59346]
I0111 19:14:05.011431  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.011650  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (11.406768ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59306]
I0111 19:14:05.012396  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:05.012415  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:05.012551  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.012603  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.013531  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (47.294177ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:05.016725  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.214316ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59348]
I0111 19:14:05.017877  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (3.972485ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.018999  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19/status: (4.766378ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59346]
I0111 19:14:05.019959  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (5.460041ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:05.021831  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.861796ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.022090  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.022629  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:05.022646  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:05.022768  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.022814  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.025293  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21/status: (1.735179ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.025746  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (2.331721ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59348]
I0111 19:14:05.027145  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (6.74051ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59340]
I0111 19:14:05.027187  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-21.1578e12a925d26af: (3.327947ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59350]
I0111 19:14:05.027215  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.552438ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.027625  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.027948  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:05.027974  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:05.028077  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.028121  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.029978  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14/status: (1.649408ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.030226  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.85528ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59348]
I0111 19:14:05.031856  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.461118ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.031998  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (4.506185ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59350]
I0111 19:14:05.032081  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.032286  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:05.032304  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:05.032321  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-14.1578e12a77e99367: (2.795192ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59352]
I0111 19:14:05.032391  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.032433  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.034443  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.684935ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59348]
I0111 19:14:05.034694  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12/status: (1.822372ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59352]
I0111 19:14:05.035880  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-12.1578e12a77961839: (2.795993ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59354]
I0111 19:14:05.036867  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.252349ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59352]
I0111 19:14:05.037245  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.037412  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:05.037425  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:05.037546  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.037614  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.037967  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (5.484391ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.039848  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.469872ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.040025  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.91069ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59348]
I0111 19:14:05.040151  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13/status: (2.133538ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59354]
I0111 19:14:05.042604  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.565494ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.042971  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.043155  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:05.043179  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:05.043230  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (4.837955ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59358]
I0111 19:14:05.043272  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.043305  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.045710  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (2.18415ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59360]
I0111 19:14:05.047293  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (2.167247ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.047963  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.216431ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59360]
I0111 19:14:05.048188  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.048513  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-8.1578e12a76dbf06f: (4.57861ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59364]
I0111 19:14:05.048703  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:05.048718  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:05.048807  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.048841  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.048901  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (4.950811ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59362]
I0111 19:14:05.050642  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.125421ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59366]
I0111 19:14:05.051125  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (2.134352ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59360]
I0111 19:14:05.052764  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17/status: (3.693207ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59344]
I0111 19:14:05.053830  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (4.686292ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59362]
I0111 19:14:05.054746  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.622378ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59360]
I0111 19:14:05.054945  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.055073  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.055087  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.055152  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.055189  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.057047  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.539206ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59366]
I0111 19:14:05.057575  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.487118ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59370]
I0111 19:14:05.058176  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15/status: (2.518306ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59360]
I0111 19:14:05.058890  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (4.438514ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59362]
I0111 19:14:05.060373  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.576979ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59370]
I0111 19:14:05.060589  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.060730  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:05.060743  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:05.060967  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.061019  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.062312  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.135196ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59370]
I0111 19:14:05.062665  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11/status: (1.426054ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59366]
I0111 19:14:05.063226  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (4.04011ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59362]
I0111 19:14:05.063463  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.933123ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59372]
I0111 19:14:05.064959  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.967426ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59366]
I0111 19:14:05.065205  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.065346  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:05.065363  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:05.065424  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.065463  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.067127  120459 store.go:355] GuaranteedUpdate of /0c88a306-f2a2-40a8-93c1-472b7f2fb7b6/pods/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 failed because of a conflict, going to retry
I0111 19:14:05.067306  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.141929ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59370]
I0111 19:14:05.067350  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10/status: (1.486261ms) 409 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59372]
I0111 19:14:05.068738  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (5.272436ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59362]
I0111 19:14:05.069907  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.593309ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59370]
I0111 19:14:05.070033  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-10.1578e12a7738fa76: (3.710097ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
E0111 19:14:05.070570  120459 scheduler.go:292] Error getting the updated preemptor pod object: pods "ppod-10" not found
I0111 19:14:05.070739  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.070773  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.070899  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.070946  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.072134  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (952.835µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59372]
I0111 19:14:05.072821  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15/status: (1.654128ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.073874  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-15.1578e12a97e796e8: (2.161271ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59378]
I0111 19:14:05.075276  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (6.158373ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59362]
I0111 19:14:05.076129  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.476162ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.076458  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.076587  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:05.076603  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:05.076665  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.076721  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.079436  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (2.342689ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59372]
I0111 19:14:05.080949  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17/status: (3.770206ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.081893  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-17.1578e12a9786b977: (3.964423ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59380]
I0111 19:14:05.082469  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (6.509516ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59378]
I0111 19:14:05.087922  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (6.150594ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.088143  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.088296  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:05.088314  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:05.088448  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.088497  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.099417  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.295513ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.101335  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-19.1578e12a955dbcb1: (6.046537ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59372]
I0111 19:14:05.101615  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (14.567149ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59380]
I0111 19:14:05.106330  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (4.288837ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59372]
I0111 19:14:05.108452  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19/status: (7.659957ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.117011  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (8.149028ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.117386  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.117625  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.117689  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.117906  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (11.252976ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59372]
I0111 19:14:05.119873  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.489207ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.122488  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (4.103439ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.122532  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:05.122634  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:05.124558  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.659439ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.126209  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:05.126259  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:05.128220  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.683769ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.128267  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (5.399418ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.131160  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:05.131196  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:05.136330  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (7.698148ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.137878  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.071253ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.139329  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:05.139396  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:05.140212  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (3.492069ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.141844  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.112635ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.143388  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:05.143451  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:05.151877  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (7.521647ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.152270  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (11.767109ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.155487  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:05.155558  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:05.157464  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (4.829116ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.157482  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.677108ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.160283  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:05.160326  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:05.161592  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (3.762007ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.161902  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.333565ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.166767  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:05.166808  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:05.168627  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.53441ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.169380  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (7.466934ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.172725  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:05.173052  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:05.173789  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (4.046248ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.175457  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.339951ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.177148  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:05.177189  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:05.178561  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (4.424693ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.179273  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.542663ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.181467  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:05.181500  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:05.182724  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (3.675927ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.183188  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.458049ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.185601  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:05.185637  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:05.186990  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (3.922015ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.187151  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.282656ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.189487  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:05.189553  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:05.191180  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (3.898735ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.193119  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.30565ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.193756  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:05.193791  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:05.195652  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.493573ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.196186  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (4.704173ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.199378  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:05.199415  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:05.200882  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (4.383405ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.202803  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.029855ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.204953  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:05.204995  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:05.205982  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (4.704016ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.211436  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:05.211477  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:05.212901  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (6.474376ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.214815  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (7.165271ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.216800  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:05.216837  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:05.217054  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.764092ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.219392  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.828255ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.220052  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (6.163064ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.223173  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:05.223222  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:05.225704  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (5.189305ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.228047  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.215785ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.229383  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:05.229420  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:05.230908  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (4.528891ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.232399  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.931902ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.234629  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:05.234663  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:05.235753  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (4.56585ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.235985  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.108348ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.238140  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:05.238180  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:05.239321  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (3.31872ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.239849  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.410708ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.242036  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:05.242069  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:05.242974  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (3.382179ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.247003  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:05.247050  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:05.247383  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.226797ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.248447  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (4.42249ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.249425  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.561498ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.249571  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:05.249641  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:05.250697  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:05.250716  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:05.251145  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:05.251173  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:05.251572  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:05.252260  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (3.377583ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.252847  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.456448ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.254947  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:05.254979  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:05.256398  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (3.748792ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.256999  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.796224ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.259210  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:05.259260  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:05.261221  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (4.046749ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.261863  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.15613ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.264797  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:05.264884  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:05.266630  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (4.978196ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.268104  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.986929ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.270322  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:05.270362  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:05.271193  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (3.700183ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.271992  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.213183ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.273874  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:05.273904  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:05.275553  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (4.110697ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.277255  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.004302ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.278529  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:05.278570  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:05.279729  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (3.908053ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.280352  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.184263ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.282452  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:05.282484  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:05.283919  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (3.902027ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.284584  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.620955ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.289310  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:05.289402  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:05.290708  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (5.920446ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.291319  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.561197ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.294183  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:05.294595  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:05.296987  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (5.700322ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.297491  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.847798ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.301900  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (4.499799ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.304392  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (1.991864ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.311198  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (5.532171ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.314087  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.350731ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.316742  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (982.096µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.319628  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.271754ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.322577  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.24499ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.328527  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.337501ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.331450  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.339093ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.334275  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.179291ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.337014  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.126808ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.339719  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.096816ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.342387  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.124157ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.347306  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (3.25474ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.350103  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.086607ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.352980  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.019313ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.355818  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.28143ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.358552  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.19453ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.361243  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.050185ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.364924  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (2.086369ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.367753  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.033198ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.370384  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.060296ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.373205  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.037827ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.375754  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (944.659µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.378329  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.007455ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.381150  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.255887ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.383836  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.087652ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.387405  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.097941ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.390000  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.03747ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.394448  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (2.782992ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.397814  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.200727ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.400475  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.105778ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.405058  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (2.963495ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.408227  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.405282ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.411614  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.628345ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.414485  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.222845ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.417809  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.524648ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.421147  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.576206ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.424186  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.41846ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.427766  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.178915ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.431059  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.477756ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.434255  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.49989ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.437353  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.352034ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.440871  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.908831ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.443608  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.167106ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.447790  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.462702ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.450938  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.408483ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.454226  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.545367ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.457217  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.326669ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.460931  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.75518ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.464245  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.554056ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.468542  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.576083ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.471642  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.422064ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.474644  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (1.329664ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.477369  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (1.096883ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.479993  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.019934ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.484091  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.624629ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.484651  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:05.484727  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:05.484887  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1"
I0111 19:14:05.484927  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1": all PVCs bound and nothing to do
I0111 19:14:05.484979  120459 factory.go:1166] Attempting to bind rpod-0 to node1
I0111 19:14:05.488405  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.83647ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.488730  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0/binding: (3.412638ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.489419  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:05.489653  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:05.489710  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:05.489827  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1"
I0111 19:14:05.489850  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1": all PVCs bound and nothing to do
I0111 19:14:05.489929  120459 factory.go:1166] Attempting to bind rpod-1 to node1
I0111 19:14:05.492722  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1/binding: (2.521624ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.493095  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:05.493365  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.314804ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.495651  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.680839ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.618314  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (27.819993ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.721261  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (2.002688ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.721791  120459 preemption_test.go:561] Creating the preemptor pod...
I0111 19:14:05.725392  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.34604ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.725618  120459 preemption_test.go:567] Creating additional pods...
I0111 19:14:05.725973  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:05.725985  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:05.726098  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.726135  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.729907  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (2.15446ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.730312  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.302061ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59388]
I0111 19:14:05.731369  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.820467ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.733352  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.353216ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.733597  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.734495  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.076135ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.736855  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (1.981109ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.737763  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.579207ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.742113  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (4.585465ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59382]
I0111 19:14:05.743199  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:05.743213  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:05.743336  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.743388  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.745634  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.090858ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.748925  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/preemptor-pod.1578e12abfe56684: (2.747438ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.749604  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (14.011659ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.750806  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (6.055931ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59388]
I0111 19:14:05.751066  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (6.648793ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59390]
I0111 19:14:05.753410  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.311188ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59390]
I0111 19:14:05.753648  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.757187  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (3.175008ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59390]
I0111 19:14:05.757574  120459 cacher.go:598] cacher (*core.Pod): 2 objects queued in incoming channel.
I0111 19:14:05.758115  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.903688ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.760312  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (1.153076ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
E0111 19:14:05.761223  120459 scheduler.go:319] Error preempting pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1: pods "rpod-1" not found
I0111 19:14:05.761344  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:05.761393  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:05.761553  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.761622  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.766401  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.110847ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59392]
I0111 19:14:05.767965  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (8.251974ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.768587  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1/status: (6.578661ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.770376  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.769909ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59376]
I0111 19:14:05.770485  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.55231ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.771520  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (4.770899ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59392]
I0111 19:14:05.771771  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.771936  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:05.771984  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:05.772079  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.772144  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.773655  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.491956ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.775335  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.028608ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59396]
I0111 19:14:05.776228  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (3.383424ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59394]
I0111 19:14:05.776753  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3/status: (3.901149ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59392]
I0111 19:14:05.779123  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.121868ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59396]
I0111 19:14:05.779745  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.779958  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:05.779994  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:05.780128  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.780219  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.785862  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (4.765803ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.788276  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (7.837575ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59396]
I0111 19:14:05.788606  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.416286ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59394]
I0111 19:14:05.788656  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.126034ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.791764  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.760914ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59394]
I0111 19:14:05.791803  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5/status: (2.579626ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59396]
I0111 19:14:05.794471  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.31152ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59394]
I0111 19:14:05.796713  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.814454ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59394]
I0111 19:14:05.799203  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (6.270106ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59396]
I0111 19:14:05.799750  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.639686ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59394]
I0111 19:14:05.800881  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.801087  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:05.801121  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:05.801222  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.801587  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.804652  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.126501ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59398]
I0111 19:14:05.805985  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (3.015819ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.806503  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6/status: (3.236025ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59396]
I0111 19:14:05.806503  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (5.416381ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59394]
I0111 19:14:05.809178  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (2.196246ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.809447  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.809589  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:05.809598  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:05.809731  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.809763  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.316057ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59398]
I0111 19:14:05.809773  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.811510  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.207674ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.812949  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.130484ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59398]
I0111 19:14:05.813168  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.080857ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59402]
I0111 19:14:05.813370  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12/status: (3.052054ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59400]
I0111 19:14:05.815499  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.144641ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59400]
I0111 19:14:05.815661  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.8753ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59402]
I0111 19:14:05.816099  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.816318  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:05.816367  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:05.816547  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.816618  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.818349  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.163354ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59402]
I0111 19:14:05.819819  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13/status: (2.915359ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.820333  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (2.587652ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59404]
I0111 19:14:05.821749  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.175367ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.823387  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.156094ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59402]
I0111 19:14:05.824600  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (3.031313ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.824836  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.826814  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.826836  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:05.826957  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.827013  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.828998  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.790235ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.837985  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (8.524664ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.839467  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (11.261277ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59408]
I0111 19:14:05.841107  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.463498ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.841568  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (3.683668ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.842332  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15/status: (3.65165ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59404]
I0111 19:14:05.843553  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.060749ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.844007  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.205529ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.844258  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.844425  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:05.844445  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:05.844551  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.844603  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.846941  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.988184ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.847249  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.275041ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.847781  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18/status: (1.888557ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59408]
I0111 19:14:05.847960  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.56954ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59416]
I0111 19:14:05.848714  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.364815ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.849317  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.150642ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59408]
I0111 19:14:05.849533  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.849721  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:05.849738  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:05.849852  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.849901  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.851167  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.04789ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.851701  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.519956ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.852375  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22/status: (2.191268ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59408]
I0111 19:14:05.852385  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.820467ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59418]
I0111 19:14:05.853369  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.571771ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.854486  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.205605ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59418]
I0111 19:14:05.854865  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.855097  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:05.855112  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:05.855196  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.855273  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.855598  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.643178ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.857647  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.451144ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59422]
I0111 19:14:05.857897  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (2.136925ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.858156  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24/status: (2.373ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59418]
I0111 19:14:05.859840  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.56921ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.859903  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.224955ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59386]
I0111 19:14:05.860250  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.860483  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.176161ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59422]
I0111 19:14:05.860586  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:05.860602  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:05.860718  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.860767  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.862509  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.600067ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.862963  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.704351ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59424]
I0111 19:14:05.863442  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.132398ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59426]
I0111 19:14:05.863532  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (2.355549ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59420]
I0111 19:14:05.865286  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.88139ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59424]
I0111 19:14:05.865301  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.207356ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59426]
I0111 19:14:05.865606  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.865958  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:05.866280  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:05.866382  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.866440  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.869194  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29/status: (2.184341ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.869634  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.670774ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59424]
I0111 19:14:05.869639  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (2.565915ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.871310  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.092168ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.871548  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.871740  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:05.871797  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:05.872102  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.872132  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.826045ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.872174  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.872631  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.364779ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59430]
I0111 19:14:05.873799  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.339874ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.874191  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (1.655474ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59406]
I0111 19:14:05.875744  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.480011ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59432]
I0111 19:14:05.876327  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.618781ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.876739  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.877085  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:05.878170  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:05.878264  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-26.1578e12ac7ebadb9: (3.232587ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59430]
I0111 19:14:05.878100  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.04024ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59432]
I0111 19:14:05.878475  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.878517  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.882256  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (3.099037ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.882656  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (3.150508ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59430]
I0111 19:14:05.882968  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.848805ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59436]
I0111 19:14:05.883285  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.701772ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59432]
I0111 19:14:05.885814  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.772616ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.886099  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.886315  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:05.886335  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:05.886438  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.886494  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.682174ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59432]
I0111 19:14:05.886493  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.889084  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.421776ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59432]
I0111 19:14:05.889172  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.082907ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.889495  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (2.614183ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59434]
I0111 19:14:05.889943  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (3.147691ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.891840  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.722996ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59434]
I0111 19:14:05.892206  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.373793ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59432]
I0111 19:14:05.892745  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.892889  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:05.892908  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:05.892982  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.893133  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.894578  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.668315ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.895366  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (1.709956ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.895419  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.546961ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59440]
I0111 19:14:05.897079  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.953735ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.898504  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-32.1578e12ac8fa9191: (3.929955ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59442]
I0111 19:14:05.899180  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.577153ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59428]
I0111 19:14:05.901271  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.612266ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59442]
I0111 19:14:05.903524  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.476574ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.903590  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.859842ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59442]
I0111 19:14:05.903824  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.903983  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:05.903999  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:05.904080  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.904137  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.905875  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.807003ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.906385  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.363441ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59446]
I0111 19:14:05.906594  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.537231ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59444]
I0111 19:14:05.907729  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39/status: (2.718812ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59440]
I0111 19:14:05.908578  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.166857ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.909400  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.165739ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59446]
I0111 19:14:05.909634  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.909833  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:05.909852  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:05.909929  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.909969  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.911279  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.079244ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.912069  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.480012ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59448]
I0111 19:14:05.912540  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (2.298018ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59444]
I0111 19:14:05.913113  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44/status: (2.846105ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59446]
I0111 19:14:05.914295  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.17438ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.914850  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.267317ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59444]
I0111 19:14:05.915065  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.915214  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:05.915228  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:05.915330  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.915428  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.916412  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.554088ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.917668  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.702318ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59448]
I0111 19:14:05.917739  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.778378ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59444]
I0111 19:14:05.918948  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (1.731256ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59438]
I0111 19:14:05.920552  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.08433ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59444]
I0111 19:14:05.920918  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.921146  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:05.921162  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:05.921252  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.921306  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.923215  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48/status: (1.688423ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59444]
I0111 19:14:05.923912  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.780715ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.924261  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (2.02211ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59448]
I0111 19:14:05.925059  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.478748ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59444]
I0111 19:14:05.925403  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.925708  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:05.925728  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:05.925824  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.925882  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.928199  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.329773ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.928767  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (2.63659ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59448]
I0111 19:14:05.929326  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-46.1578e12acb2dbc71: (2.466562ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59452]
I0111 19:14:05.930565  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.158556ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59448]
I0111 19:14:05.930908  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.931089  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:05.931099  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:05.931201  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.931258  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.933001  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.508477ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.936131  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48/status: (3.892973ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59452]
I0111 19:14:05.936210  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-48.1578e12acb875f9f: (4.011573ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59458]
I0111 19:14:05.951630  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (14.422037ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59452]
I0111 19:14:05.952034  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.952268  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:05.952290  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:05.952456  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.952516  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.955013  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.741607ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59464]
I0111 19:14:05.955778  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (2.908571ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.956472  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49/status: (3.472208ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59458]
I0111 19:14:05.958853  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.540815ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.959115  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.959380  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:05.959398  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:05.959482  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.959526  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.961283  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.229777ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59464]
I0111 19:14:05.961933  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44/status: (2.133161ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.963840  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.268151ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.964309  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.964530  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:05.964547  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:05.964645  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.964716  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.965067  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-44.1578e12acada7db0: (2.503146ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59464]
I0111 19:14:05.966296  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.30275ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.968788  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-49.1578e12acd63ac75: (2.707552ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:05.968999  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49/status: (3.214985ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59464]
I0111 19:14:05.970944  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.389859ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:05.971293  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.971470  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:05.971489  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:05.971569  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.971614  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.973591  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.723829ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:05.973760  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.627178ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.976494  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47/status: (2.160146ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59468]
I0111 19:14:05.978365  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.298664ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.978645  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.978922  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:05.978938  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:05.979037  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.979082  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.981180  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39/status: (1.822941ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.981636  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.653232ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:05.982194  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-39.1578e12aca814812: (2.13301ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59470]
I0111 19:14:05.982968  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.288503ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59450]
I0111 19:14:05.983717  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.983904  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:05.983921  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:05.984043  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.984122  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.986996  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.538722ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:05.987317  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47/status: (2.870609ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59470]
I0111 19:14:05.987417  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-47.1578e12ace871491: (2.470941ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59472]
I0111 19:14:05.988974  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.15099ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59470]
I0111 19:14:05.989219  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.989486  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:05.989502  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:05.989714  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.989770  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.991463  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.373726ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:05.991943  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (1.954024ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59470]
I0111 19:14:05.992423  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.82846ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59474]
I0111 19:14:05.993529  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.044444ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59470]
I0111 19:14:05.993876  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.994099  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:05.994116  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:05.994211  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.994282  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:05.995570  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.046623ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:05.996342  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43/status: (1.824648ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59474]
I0111 19:14:05.996847  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.621643ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59476]
I0111 19:14:05.998004  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.147179ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59474]
I0111 19:14:05.998336  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:05.998506  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:05.998528  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:05.998688  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:05.998739  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.000067  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.071937ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:06.001023  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (2.055751ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59476]
I0111 19:14:06.002171  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-45.1578e12acf9c0caa: (2.457957ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.002550  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.033936ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59476]
I0111 19:14:06.002885  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.003078  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:06.003098  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:06.003211  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.003272  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.005351  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.542787ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.005702  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43/status: (1.894083ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:06.007202  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.148234ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:06.007425  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.007569  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:06.007586  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:06.007699  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.007742  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.009124  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.040223ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.009269  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-43.1578e12acfe10042: (5.299756ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59480]
I0111 19:14:06.010366  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42/status: (2.362004ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:06.012277  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.481499ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:06.012468  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.650715ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59480]
I0111 19:14:06.012656  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.013059  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:06.013117  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:06.013269  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.013318  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.015473  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41/status: (1.927009ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:06.015803  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (2.197175ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.016214  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.204915ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59482]
I0111 19:14:06.017423  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.157152ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.017749  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.018029  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0
I0111 19:14:06.018121  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0
I0111 19:14:06.018456  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.018553  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.018615  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.455264ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59482]
I0111 19:14:06.020274  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.16592ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.021247  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.857542ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59482]
I0111 19:14:06.021597  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0/status: (2.4484ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59466]
I0111 19:14:06.023283  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.224031ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.023644  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.023863  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:06.023883  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:06.024045  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.024105  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.026726  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (2.215971ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59478]
I0111 19:14:06.026798  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.948233ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.027033  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4/status: (2.442593ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59484]
I0111 19:14:06.028839  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.319492ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59484]
I0111 19:14:06.029127  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.029330  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:06.029348  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:06.029478  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.029536  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.031047  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.208336ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59484]
I0111 19:14:06.032444  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5/status: (2.597923ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.032631  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-5.1578e12ac31ea51a: (2.432959ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59488]
I0111 19:14:06.034128  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.231764ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.034378  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.034543  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:06.034560  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:06.034620  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.034652  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.036778  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17/status: (1.901838ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.036857  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.964681ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59484]
I0111 19:14:06.037440  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.984235ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.038980  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.360088ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59484]
I0111 19:14:06.039329  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.039540  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:06.039558  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:06.039652  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.039717  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.041284  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.151759ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.042375  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40/status: (2.242773ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.043463  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.713673ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.043950  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.126154ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.044273  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.044463  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:06.044481  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:06.044582  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.044629  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.045996  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.135858ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.046992  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36/status: (1.906682ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.047625  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.192414ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59496]
I0111 19:14:06.048662  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.236349ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59486]
I0111 19:14:06.049003  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.049174  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:06.049191  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:06.049336  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.049386  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.050795  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.132901ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59496]
I0111 19:14:06.051359  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.358676ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59498]
I0111 19:14:06.051557  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37/status: (1.940626ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.053158  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.104737ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.053400  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.053585  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:06.053599  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:06.053767  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.053826  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.055667  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.371265ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.057211  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13/status: (2.923392ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59496]
I0111 19:14:06.057357  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-13.1578e12ac54a084c: (2.519093ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59502]
I0111 19:14:06.059143  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.139959ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59502]
I0111 19:14:06.059422  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.059606  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:06.059620  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:06.059722  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.059763  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.061477  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.416022ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59502]
I0111 19:14:06.062539  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (2.425029ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.063248  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-33.1578e12ac9741c67: (2.3948ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59504]
I0111 19:14:06.064349  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.070835ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59490]
I0111 19:14:06.064728  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.065138  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:06.065156  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:06.065316  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.065367  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.066992  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.337435ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59502]
I0111 19:14:06.067551  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38/status: (1.93151ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59504]
I0111 19:14:06.068559  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.986051ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59508]
I0111 19:14:06.069003  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.062162ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59504]
I0111 19:14:06.069306  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.069504  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:06.069524  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:06.069627  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.069668  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.071469  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.51661ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59502]
I0111 19:14:06.072078  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.656511ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59510]
I0111 19:14:06.072542  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9/status: (2.580849ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59508]
I0111 19:14:06.074295  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.241935ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59510]
I0111 19:14:06.074576  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.074813  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:06.074828  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:06.074905  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.074962  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.076569  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.159479ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59510]
I0111 19:14:06.076877  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.469978ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59502]
I0111 19:14:06.077386  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19/status: (1.770362ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59512]
I0111 19:14:06.078843  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.09928ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59512]
I0111 19:14:06.079123  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.079312  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:06.079325  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:06.079415  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.079456  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.080972  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.247916ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59510]
I0111 19:14:06.081888  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.682829ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59514]
I0111 19:14:06.082059  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20/status: (2.274411ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59502]
I0111 19:14:06.083870  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.322943ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59514]
I0111 19:14:06.084210  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.084430  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:06.084448  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:06.084536  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.084621  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.088879  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1/status: (3.946208ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59514]
I0111 19:14:06.089001  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (3.906934ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59510]
I0111 19:14:06.090530  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-1.1578e12ac202d7b9: (4.905226ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.090958  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.399533ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59510]
I0111 19:14:06.091342  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.091607  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:06.091622  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:06.091736  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.091795  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.093451  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.343863ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.094581  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.405207ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59518]
I0111 19:14:06.094822  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10/status: (2.716168ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59514]
I0111 19:14:06.096531  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.198194ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59518]
I0111 19:14:06.096796  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.097011  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:06.097025  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:06.097095  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.097131  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.098617  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.122828ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.099807  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21/status: (2.410385ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59518]
I0111 19:14:06.099971  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.159311ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59520]
I0111 19:14:06.101660  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.267316ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59518]
I0111 19:14:06.101985  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.102194  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:06.102210  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:06.102321  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.102380  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.103886  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.218776ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.104923  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23/status: (2.272027ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59520]
I0111 19:14:06.105083  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.092927ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59522]
I0111 19:14:06.107001  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.247964ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59520]
I0111 19:14:06.107316  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.107629  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:06.107647  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:06.107789  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.107839  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.109303  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.101269ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.109937  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11/status: (1.868943ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59520]
I0111 19:14:06.110819  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.450286ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59524]
I0111 19:14:06.111702  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.249291ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59520]
I0111 19:14:06.112029  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.112272  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:06.112329  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:06.112452  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.112503  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.114227  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.358471ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.114744  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18/status: (1.9797ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59524]
I0111 19:14:06.115806  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-18.1578e12ac6f51473: (2.48372ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59526]
I0111 19:14:06.116300  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.172626ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59524]
I0111 19:14:06.116623  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.116899  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:06.116916  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:06.117096  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.117145  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.119123  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.253403ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.119306  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.448429ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59528]
I0111 19:14:06.119463  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25/status: (2.068694ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59526]
I0111 19:14:06.129426  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (9.868873ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59528]
I0111 19:14:06.130428  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.573022ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59526]
I0111 19:14:06.130797  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.130982  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2
I0111 19:14:06.130998  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2
I0111 19:14:06.131100  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.131154  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.132780  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.258293ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.133778  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2/status: (2.335617ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59528]
I0111 19:14:06.134066  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.826037ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59530]
I0111 19:14:06.135465  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.187787ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59528]
I0111 19:14:06.135786  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.135992  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:06.136055  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:06.136190  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.136244  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.137523  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.016685ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.138566  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.837491ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59532]
I0111 19:14:06.138749  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7/status: (2.237893ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59530]
I0111 19:14:06.140326  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.094662ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59532]
I0111 19:14:06.140547  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.140709  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:06.140726  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:06.140914  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.140956  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.142476  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.252014ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.143326  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6/status: (2.159182ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59532]
I0111 19:14:06.144563  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-6.1578e12ac4603ade: (2.73556ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59534]
I0111 19:14:06.144807  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.093263ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59532]
I0111 19:14:06.145115  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.145354  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:06.145371  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:06.145484  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.145531  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.147991  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22/status: (2.228907ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59534]
I0111 19:14:06.148528  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-22.1578e12ac745e7db: (2.396184ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.148651  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (2.216292ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59536]
I0111 19:14:06.149436  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.012687ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59534]
I0111 19:14:06.149755  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.149935  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:06.149956  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:06.150066  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.150120  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.151604  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.235073ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.151986  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.306153ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59538]
I0111 19:14:06.152550  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27/status: (2.193712ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59536]
I0111 19:14:06.154208  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.206921ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59538]
I0111 19:14:06.154474  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.154720  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:06.154741  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:06.154870  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.154923  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.156433  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.145404ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.157226  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.712361ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59540]
I0111 19:14:06.158915  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14/status: (3.509473ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59538]
I0111 19:14:06.160785  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.398564ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59540]
I0111 19:14:06.161064  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.161272  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:06.161292  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:06.161405  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.161478  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.163168  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.248163ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.164204  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27/status: (2.424847ms) 409 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59540]
I0111 19:14:06.165287  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-27.1578e12ad92adf8d: (2.363989ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59542]
I0111 19:14:06.166092  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.146988ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59540]
I0111 19:14:06.166459  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.166629  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:06.166647  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:06.166783  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.166862  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.168631  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.497031ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59542]
I0111 19:14:06.170361  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14/status: (1.316773ms) 409 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59542]
I0111 19:14:06.170986  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-14.1578e12ad97428bf: (3.587545ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.172025  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.181424ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59542]
I0111 19:14:06.172496  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.172743  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:06.172766  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:06.172897  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.172960  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.174479  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.289913ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.175051  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35/status: (1.820967ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59544]
I0111 19:14:06.175064  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.502413ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59546]
I0111 19:14:06.176496  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.056461ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59544]
I0111 19:14:06.176782  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.177033  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:06.177048  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:06.177197  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.177264  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.178857  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.260813ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.179479  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.555372ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59548]
I0111 19:14:06.179762  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34/status: (2.273827ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59544]
I0111 19:14:06.208440  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (27.920589ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59548]
I0111 19:14:06.209060  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.216170  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:06.216276  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:06.216549  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.216652  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.219693  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (2.297153ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.223128  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24/status: (5.646913ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59548]
I0111 19:14:06.224726  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-24.1578e12ac797da0c: (6.227973ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59550]
I0111 19:14:06.225361  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.573735ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59548]
I0111 19:14:06.225815  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.230747  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.074155ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59550]
I0111 19:14:06.240782  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:06.240810  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:06.241053  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.241128  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.245093  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (3.037932ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.245136  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28/status: (3.345826ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59550]
I0111 19:14:06.246818  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.266321ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59552]
I0111 19:14:06.247437  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.701008ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59550]
I0111 19:14:06.247918  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.248149  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:06.248162  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:06.248321  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.248379  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.252904  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:06.253717  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:06.254170  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.785153ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59554]
I0111 19:14:06.254171  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (5.165599ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.265290  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:06.265404  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:06.268281  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31/status: (19.343517ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59552]
I0111 19:14:06.270786  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.819382ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59554]
I0111 19:14:06.271180  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.271484  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:06.271530  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:06.271893  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.272040  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.275619  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:06.276259  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.896725ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.278219  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (5.757886ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59554]
I0111 19:14:06.282150  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (7.760875ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.282244  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.608106ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59554]
I0111 19:14:06.282777  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.283028  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:06.283050  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:06.283508  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.283558  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.285616  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.397958ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.287921  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12/status: (3.694782ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.288784  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-12.1578e12ac4e18be3: (2.436285ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.291584  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (2.796071ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.291907  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.292059  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:06.292074  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:06.292154  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.292197  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.293744  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.072814ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.294647  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.784356ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59558]
I0111 19:14:06.297006  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (4.565583ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59516]
I0111 19:14:06.298841  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.350066ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59558]
I0111 19:14:06.299262  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.299474  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:06.299489  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:06.299612  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.299654  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.301472  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.174369ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.301902  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.544123ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59560]
I0111 19:14:06.302491  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16/status: (2.558363ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59558]
I0111 19:14:06.304062  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.083284ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59560]
I0111 19:14:06.304381  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.304583  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:06.304601  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:06.304711  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.304756  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.306421  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16/status: (1.443588ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59560]
I0111 19:14:06.306934  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (916.093µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59562]
I0111 19:14:06.307961  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-16.1578e12ae21495de: (2.653655ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.308398  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.440159ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59560]
I0111 19:14:06.308779  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.308966  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:06.308982  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:06.309093  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.309142  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.310520  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.118009ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.311558  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (2.1352ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59562]
I0111 19:14:06.313537  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-8.1578e12ae06e0ae9: (3.685097ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.313732  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.740215ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59562]
I0111 19:14:06.314250  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.314437  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:06.314453  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:06.314551  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.314639  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.316130  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.271709ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.316660  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6/status: (1.780598ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.318812  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-6.1578e12ac4603ade: (3.512403ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59566]
I0111 19:14:06.318900  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.632723ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59556]
I0111 19:14:06.319182  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.319393  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:06.319414  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:06.319548  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.319599  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.320897  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.307101ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59566]
I0111 19:14:06.322294  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11/status: (1.89374ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.322574  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (2.017583ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59568]
I0111 19:14:06.324320  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.219756ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.324749  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.324822  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-11.1578e12ad6a5b9dc: (3.140976ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59570]
I0111 19:14:06.324987  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:06.325005  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:06.325220  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.325278  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.326541  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.009545ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.327497  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21/status: (1.955528ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59566]
I0111 19:14:06.328314  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-21.1578e12ad6024631: (2.350999ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59572]
I0111 19:14:06.329229  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.038575ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59566]
I0111 19:14:06.329639  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.329871  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:06.329888  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:06.329997  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.330056  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.331515  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.190366ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.333150  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-42.1578e12ad0ae63e7: (2.325812ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59574]
I0111 19:14:06.334467  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42/status: (4.140218ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59572]
I0111 19:14:06.336112  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.127028ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59574]
I0111 19:14:06.336387  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.336610  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0
I0111 19:14:06.336627  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0
I0111 19:14:06.336760  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.336812  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.338130  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.044688ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.339007  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0/status: (1.885059ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59574]
I0111 19:14:06.339984  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-0.1578e12ad15327cc: (2.50871ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59576]
I0111 19:14:06.341086  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.09804ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59574]
I0111 19:14:06.341422  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.341700  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:06.341750  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:06.341857  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.341902  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.343760  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.551248ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.344049  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41/status: (1.844356ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59576]
I0111 19:14:06.344817  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-41.1578e12ad10373b6: (2.113919ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59578]
I0111 19:14:06.345597  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.122103ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59576]
I0111 19:14:06.346038  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.346199  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:06.346215  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:06.346395  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.346445  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.347995  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.081219ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.349108  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36/status: (2.220374ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59578]
I0111 19:14:06.350599  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-36.1578e12ad2e138eb: (2.846214ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59580]
I0111 19:14:06.350980  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.412579ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59578]
I0111 19:14:06.351317  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.351476  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:06.351492  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:06.351589  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.351645  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.353490  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.585233ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.354104  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9/status: (2.200449ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59580]
I0111 19:14:06.355181  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-9.1578e12ad45f4f91: (2.676686ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.356106  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.206897ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59580]
I0111 19:14:06.356469  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.356668  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:06.356712  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:06.356836  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.356888  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.359572  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.409045ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.359970  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17/status: (1.811286ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.361461  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.06078ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.361900  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.362109  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:06.362164  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:06.362305  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.362356  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.362434  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-17.1578e12ad2490e92: (2.829998ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59584]
I0111 19:14:06.363743  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.09346ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.366392  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-37.1578e12ad329ce47: (3.086025ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59584]
I0111 19:14:06.366462  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37/status: (3.761289ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59564]
I0111 19:14:06.368137  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.156375ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59584]
I0111 19:14:06.368473  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.368686  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:06.368709  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:06.368834  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.368887  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.370463  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.312507ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.371525  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40/status: (2.373043ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59584]
I0111 19:14:06.372522  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-40.1578e12ad2964969: (2.669492ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59586]
I0111 19:14:06.373205  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.203396ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59584]
I0111 19:14:06.373502  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.373828  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2
I0111 19:14:06.373853  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2
I0111 19:14:06.373958  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.374078  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.376540  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2/status: (1.735614ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59586]
I0111 19:14:06.377561  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (2.781661ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.377578  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-2.1578e12ad809623f: (2.642366ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59588]
I0111 19:14:06.378339  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.141445ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59586]
I0111 19:14:06.378607  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.378811  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:06.378830  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:06.378936  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.379046  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.380959  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.605534ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.382105  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-28.1578e12ade97531a: (2.289185ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59590]
I0111 19:14:06.382197  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28/status: (2.833137ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59588]
I0111 19:14:06.383786  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.143051ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59590]
I0111 19:14:06.384080  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.384294  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:06.384319  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:06.384458  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.384513  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.386096  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.277785ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.386697  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34/status: (1.858807ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59590]
I0111 19:14:06.388647  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.093447ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59590]
I0111 19:14:06.388994  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.389183  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:06.389205  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:06.389345  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.389398  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.389403  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-34.1578e12adac916d5: (3.143563ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59592]
I0111 19:14:06.390608  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (962.843µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59592]
I0111 19:14:06.391300  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22/status: (1.688479ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59590]
I0111 19:14:06.392336  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-22.1578e12ac745e7db: (2.312169ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.392939  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.009696ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59590]
I0111 19:14:06.393397  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.393604  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:06.393628  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:06.393832  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.393920  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.395431  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.151207ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59592]
I0111 19:14:06.396040  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23/status: (1.828475ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.397097  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-23.1578e12ad65259c1: (2.432381ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59594]
I0111 19:14:06.398045  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.244408ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59582]
I0111 19:14:06.398355  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.398565  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:06.398586  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:06.398761  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.398808  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.400322  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.252465ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59592]
I0111 19:14:06.401187  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19/status: (2.127342ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59594]
I0111 19:14:06.402447  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-19.1578e12ad4b00ef3: (2.969141ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.402720  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.126462ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59594]
I0111 19:14:06.402992  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.403166  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:06.403185  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:06.403294  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.403339  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.404939  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.381955ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59592]
I0111 19:14:06.405352  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38/status: (1.773293ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.407051  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.310648ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.407436  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.407596  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-38.1578e12ad41da6d6: (2.996247ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59598]
I0111 19:14:06.407752  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:06.407779  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:06.407911  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.407965  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.409810  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.553667ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.410162  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4/status: (1.942272ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59592]
I0111 19:14:06.411585  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-4.1578e12ad1a8116c: (2.807535ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59600]
I0111 19:14:06.412005  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.390912ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59592]
I0111 19:14:06.412360  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.412642  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:06.412662  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:06.412809  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.412858  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.414593  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.128486ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.415297  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7/status: (2.157011ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59600]
I0111 19:14:06.416441  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-7.1578e12ad8570697: (2.587865ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59602]
I0111 19:14:06.417519  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.232589ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59600]
I0111 19:14:06.417842  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.418053  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:06.418069  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:06.418191  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.418253  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.419897  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.217568ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.421297  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.153992ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59606]
I0111 19:14:06.421612  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (2.887375ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59602]
I0111 19:14:06.421940  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-30.1578e12ae1a2d08d: (2.817078ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59604]
I0111 19:14:06.423525  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.494661ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59606]
I0111 19:14:06.423882  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.424099  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:06.424120  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:06.424261  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.424316  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.425738  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.187099ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59604]
I0111 19:14:06.426806  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35/status: (2.26021ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.428304  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-35.1578e12ada876301: (2.959158ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59608]
I0111 19:14:06.429021  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.471639ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59596]
I0111 19:14:06.429338  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.429531  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:06.429548  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:06.429644  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.429711  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.432843  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20/status: (2.827468ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59608]
I0111 19:14:06.433555  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (2.315472ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59604]
I0111 19:14:06.434217  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-20.1578e12ad4f4a794: (3.625793ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59610]
I0111 19:14:06.435380  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.271775ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59608]
I0111 19:14:06.435761  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.436000  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:06.436021  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:06.436162  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.436221  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.439912  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (3.335141ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59610]
I0111 19:14:06.440285  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-10.1578e12ad5b0e8b1: (2.415775ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59612]
I0111 19:14:06.440927  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10/status: (3.837158ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59604]
I0111 19:14:06.444511  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.503179ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59612]
I0111 19:14:06.444869  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.445071  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:06.445088  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:06.445198  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.445264  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.448150  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31/status: (2.624324ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59612]
I0111 19:14:06.448487  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-31.1578e12adf0616ae: (2.424748ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59614]
I0111 19:14:06.449362  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (3.823976ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59610]
I0111 19:14:06.450061  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.430882ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59612]
I0111 19:14:06.450356  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.450550  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:06.450574  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:06.450842  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:06.450922  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:06.453047  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.421273ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59610]
I0111 19:14:06.453878  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25/status: (2.270399ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59614]
I0111 19:14:06.454365  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-25.1578e12ad733bbcf: (2.701654ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:06.455516  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.117513ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59614]
I0111 19:14:06.455882  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:06.525409  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.033885ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:06.621566  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.806904ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:06.721801  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.989468ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:06.821804  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.022526ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:06.921692  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.87798ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.021548  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.86206ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.121627  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.827755ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.144524  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:07.144559  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:07.144767  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1"
I0111 19:14:07.144790  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1": all PVCs bound and nothing to do
I0111 19:14:07.144830  120459 factory.go:1166] Attempting to bind preemptor-pod to node1
I0111 19:14:07.145154  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:07.145165  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:07.145297  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.145340  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.148667  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/binding: (3.455382ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.149059  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:07.149475  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3/status: (3.19375ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59610]
I0111 19:14:07.149595  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (2.727501ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59624]
I0111 19:14:07.150018  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.152015  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-3.1578e12ac2a36ab9: (5.566659ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59626]
I0111 19:14:07.152168  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (2.261643ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59610]
I0111 19:14:07.152453  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.152810  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:07.152823  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:07.152913  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.152950  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.155183  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.6091ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59628]
I0111 19:14:07.155445  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.155645  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29/status: (2.426853ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.155810  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.355736ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59626]
I0111 19:14:07.157193  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.149228ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.157562  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.157763  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:07.157784  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:07.157974  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.158062  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.158807  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-29.1578e12ac84249ee: (2.354132ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59626]
I0111 19:14:07.159309  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (963.799µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59628]
I0111 19:14:07.159657  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.161463  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48/status: (3.123885ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.162199  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-48.1578e12acb875f9f: (2.824872ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59626]
I0111 19:14:07.163181  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.210291ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59616]
I0111 19:14:07.163525  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.163714  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:07.163735  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:07.163865  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.163917  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.165250  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.037048ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59628]
I0111 19:14:07.165886  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.166776  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-44.1578e12acada7db0: (2.066497ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59630]
I0111 19:14:07.166850  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44/status: (2.643403ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59626]
I0111 19:14:07.168752  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.377562ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59630]
I0111 19:14:07.168993  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.169168  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:07.169183  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:07.169266  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.169325  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.171759  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.703123ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59628]
I0111 19:14:07.171896  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (2.348007ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59630]
I0111 19:14:07.173580  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-26.1578e12ac7ebadb9: (3.369119ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.175420  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (3.072652ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59628]
I0111 19:14:07.175710  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.175888  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:07.175928  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:07.176074  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.176140  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.178588  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49/status: (2.139832ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.180855  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-49.1578e12acd63ac75: (3.924861ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59630]
I0111 19:14:07.181500  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.891191ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59636]
I0111 19:14:07.181781  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.181933  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:07.181947  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:07.182014  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.182060  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.185613  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (6.167756ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.185941  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.187061  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (2.794512ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59630]
I0111 19:14:07.187155  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39/status: (4.392194ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59636]
I0111 19:14:07.187319  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.187661  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-39.1578e12aca814812: (3.817031ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59638]
I0111 19:14:07.189351  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.295354ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59630]
I0111 19:14:07.189766  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.189970  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:07.189989  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:07.190178  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.190251  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.193184  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (2.682937ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59638]
I0111 19:14:07.193396  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (2.007055ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.194119  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-45.1578e12acf9c0caa: (2.977377ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59640]
I0111 19:14:07.195735  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.623271ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59638]
I0111 19:14:07.196074  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.196218  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:07.196248  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:07.196321  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.196365  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.198863  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.683173ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.199385  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.200321  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-47.1578e12ace871491: (2.959389ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59642]
I0111 19:14:07.200592  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47/status: (3.926144ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59640]
I0111 19:14:07.202464  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.276214ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59642]
I0111 19:14:07.202777  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.203493  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:07.203511  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:07.203727  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.203802  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.205396  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.306538ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.205709  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.207113  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15/status: (3.03033ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59642]
I0111 19:14:07.207493  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-15.1578e12ac5e8884f: (2.915582ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59644]
I0111 19:14:07.208811  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.159908ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59642]
I0111 19:14:07.209128  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.209323  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:07.209339  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:07.209416  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.209461  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.212225  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (2.481128ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59644]
I0111 19:14:07.213887  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (3.559429ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.214731  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (2.10593ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59644]
I0111 19:14:07.214765  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-32.1578e12ac8fa9191: (4.423076ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.215636  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.215840  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:07.215860  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:07.216002  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.216064  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.218422  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (2.099637ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.219198  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.543679ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59648]
I0111 19:14:07.220054  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.15917ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.220334  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.220473  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:07.220489  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:07.220579  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.220627  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.220810  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-46.1578e12acb2dbc71: (4.234399ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.223174  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (2.296347ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.223884  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (3.989095ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.224066  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-33.1578e12ac9741c67: (2.585657ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.224168  120459 preemption_test.go:583] Check unschedulable pods still exists and were never scheduled...
I0111 19:14:07.222821  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.851514ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59648]
I0111 19:14:07.224719  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.116631ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.224937  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.225114  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:07.225168  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:07.225313  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.225562  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.225796  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.464545ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.227044  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.263517ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.227303  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.227826  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.35552ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59652]
I0111 19:14:07.228418  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-18.1578e12ac6f51473: (2.246912ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.229276  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.013091ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59652]
I0111 19:14:07.231211  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18/status: (3.903535ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.231256  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.438335ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.232991  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.063579ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.233247  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.233274  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.16738ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59632]
I0111 19:14:07.233406  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:07.233522  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:07.233647  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.233751  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.235252  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.286214ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.235277  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.572487ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.236314  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.237770  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1/status: (3.479854ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59654]
I0111 19:14:07.238060  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-1.1578e12ac202d7b9: (2.560634ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59656]
I0111 19:14:07.238149  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.153991ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59646]
I0111 19:14:07.239861  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.198798ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.240269  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.627383ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59654]
I0111 19:14:07.240530  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.240723  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:07.240745  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:07.240848  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.240896  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.242565  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.653877ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.242640  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.409282ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59654]
I0111 19:14:07.243430  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.243277  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43/status: (1.851501ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59658]
I0111 19:14:07.244740  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.756433ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59654]
I0111 19:14:07.246075  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.974417ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59658]
I0111 19:14:07.246295  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-43.1578e12acfe10042: (4.751913ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59660]
I0111 19:14:07.246427  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.246580  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:07.246605  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:07.246701  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.246748  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.249494  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (3.77962ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59654]
I0111 19:14:07.249846  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-13.1578e12ac54a084c: (2.259396ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59662]
I0111 19:14:07.249947  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13/status: (2.970244ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59658]
I0111 19:14:07.250647  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (3.120098ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.251316  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.456323ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59654]
I0111 19:14:07.251437  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.010509ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59658]
I0111 19:14:07.251755  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.251944  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:07.251981  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:07.252145  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.252216  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.253221  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:07.253444  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.760679ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.254112  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:07.254834  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (2.199377ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59662]
I0111 19:14:07.255073  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5/status: (2.170231ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59664]
I0111 19:14:07.255961  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.106279ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59650]
I0111 19:14:07.257777  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.393909ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59662]
I0111 19:14:07.257816  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-5.1578e12ac31ea51a: (4.863056ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59666]
I0111 19:14:07.258246  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.846429ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59664]
I0111 19:14:07.258469  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.258662  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:07.258700  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:07.258781  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.258859  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.259289  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.119074ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59666]
I0111 19:14:07.261308  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.232977ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59670]
I0111 19:14:07.263087  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-13.1578e12ac54a084c: (2.902867ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59666]
I0111 19:14:07.263946  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (4.381484ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59662]
I0111 19:14:07.263973  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.837981ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59670]
I0111 19:14:07.264344  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.264514  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13/status: (4.743898ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59664]
I0111 19:14:07.265531  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:07.265603  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:07.265870  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.254819ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59666]
I0111 19:14:07.265880  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (998.309µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59664]
I0111 19:14:07.266391  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.266560  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:07.266598  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:07.266756  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.266827  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.267470  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.096795ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59666]
I0111 19:14:07.269071  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.701713ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59672]
I0111 19:14:07.269368  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.269704  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5/status: (2.144305ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59668]
I0111 19:14:07.271041  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (925.593µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59668]
I0111 19:14:07.271405  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.838172ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59666]
I0111 19:14:07.271527  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.271723  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:07.271766  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:07.271787  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-5.1578e12ac31ea51a: (2.748768ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59674]
I0111 19:14:07.272089  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.272141  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.274269  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (2.240271ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59668]
I0111 19:14:07.274464  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (2.093533ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59672]
I0111 19:14:07.274734  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.275972  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-45.1578e12acf9c0caa: (3.139184ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59676]
I0111 19:14:07.276007  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (3.66481ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59674]
I0111 19:14:07.276079  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.300858ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59672]
I0111 19:14:07.276109  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:07.277819  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.166705ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59668]
I0111 19:14:07.277984  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.337163ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59676]
I0111 19:14:07.278212  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.278394  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:07.278411  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:07.278483  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.278525  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.279490  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.206886ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59676]
I0111 19:14:07.280721  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.449748ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59678]
I0111 19:14:07.280990  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.282082  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.564209ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59682]
I0111 19:14:07.282548  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (3.317049ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59668]
I0111 19:14:07.283383  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-33.1578e12ac9741c67: (2.858279ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59676]
I0111 19:14:07.283976  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (954.898µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59668]
I0111 19:14:07.284004  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.363225ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59680]
I0111 19:14:07.284218  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.284424  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:07.284444  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:07.284530  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.284570  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.286763  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (2.353866ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59676]
I0111 19:14:07.287062  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (2.015509ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59678]
I0111 19:14:07.287451  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (2.505917ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59684]
I0111 19:14:07.287522  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-46.1578e12acb2dbc71: (2.319257ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59686]
I0111 19:14:07.288876  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.274157ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59676]
I0111 19:14:07.288892  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.257636ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59678]
I0111 19:14:07.288979  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.289174  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.289321  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:07.289344  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:07.289424  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.289464  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.290996  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.368704ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59684]
I0111 19:14:07.291658  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.292255  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (1.993153ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59688]
I0111 19:14:07.292571  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (3.135668ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59686]
I0111 19:14:07.293446  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-26.1578e12ac7ebadb9: (2.376548ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59690]
I0111 19:14:07.293853  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.080532ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59688]
I0111 19:14:07.294060  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.072823ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59686]
I0111 19:14:07.294171  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.294459  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:07.294508  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:07.294634  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:07.294738  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:07.296125  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.418921ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59690]
I0111 19:14:07.296758  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.148303ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.297011  120459 backoff_utils.go:79] Backing off 2s
I0111 19:14:07.297252  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (2.243155ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59684]
I0111 19:14:07.298698  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (2.077442ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59690]
I0111 19:14:07.299081  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-32.1578e12ac8fa9191: (2.557947ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.300721  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.129082ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59684]
I0111 19:14:07.301022  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:07.301080  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.94316ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59690]
I0111 19:14:07.302496  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.05865ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.304023  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.156278ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.305650  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.265183ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.307256  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.163246ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.308800  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.142135ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.310328  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.181955ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.311756  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.080187ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.313461  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.252778ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.315744  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.714719ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.317428  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.186962ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.318908  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.005ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.320209  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (999.786µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.321695  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.074521ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.323391  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.273612ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.324940  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.087063ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.326591  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.12455ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.326867  120459 preemption_test.go:598] Cleaning up all pods...
I0111 19:14:07.329950  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0
I0111 19:14:07.330035  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-0
I0111 19:14:07.331983  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.663679ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.332637  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (5.496981ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.335469  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:07.335550  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-1
I0111 19:14:07.336813  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (3.845299ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.337414  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.527257ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.340053  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2
I0111 19:14:07.340094  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-2
I0111 19:14:07.341127  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (3.757825ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.342593  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.952664ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.344337  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:07.344371  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:07.346146  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (4.171829ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.346946  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.270291ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.353248  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:07.353740  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:07.355877  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (9.115297ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.355885  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.658016ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.359406  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:07.359447  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:07.361090  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (4.81972ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.361604  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.485245ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.365406  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:07.365496  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:07.367722  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (5.749033ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.370118  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.197447ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.371094  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:07.371889  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:07.373432  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (5.012961ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.375581  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.552398ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.376711  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:07.376744  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:07.378401  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (4.475341ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.381046  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.008361ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.382370  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:07.382408  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:07.384069  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.427529ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.384132  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (5.360528ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.387434  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:07.387544  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:07.388931  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (4.466894ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.389580  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.674457ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.391987  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:07.392104  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:07.393266  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (3.912075ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.393874  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.450241ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.396401  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:07.396786  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:07.397968  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (4.359925ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.398577  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.297441ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.401954  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:07.402047  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:07.402625  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (4.068663ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.405211  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.804457ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.406206  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:07.406311  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:07.407895  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (4.964239ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.409730  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.104398ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.412107  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:07.412221  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:07.416523  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.559659ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.416659  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (7.50991ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.420909  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:07.421009  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:07.423611  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (6.46401ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.423750  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.659291ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.426547  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:07.426586  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:07.428340  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.409094ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.429336  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (5.171469ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.433054  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:07.433636  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:07.434899  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (4.685077ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.437535  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.670748ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.438827  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:07.438919  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:07.440465  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (5.236106ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.440994  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.775854ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.443753  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:07.443790  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:07.445100  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (4.231562ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.445485  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.497832ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.448665  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:07.448722  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:07.450961  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (4.977671ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.452430  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.430776ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.455176  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:07.455221  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:07.455793  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (4.109467ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.456937  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.315959ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.458895  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:07.458933  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:07.460654  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (4.507249ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.461015  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.844716ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.463739  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:07.463808  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:07.465529  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (4.484748ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.466364  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.052366ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.469436  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:07.469540  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:07.470529  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (4.002527ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.471219  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.366029ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.474197  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:07.474270  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:07.474876  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (4.002479ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.476203  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.585425ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.477956  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:07.477988  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:07.479337  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (3.982511ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.480053  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.71826ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.483588  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:07.483767  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:07.485258  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (5.28611ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.486397  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.285681ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.488933  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:07.488993  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:07.490455  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (4.080115ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.490932  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.631334ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.493360  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:07.493395  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:07.494641  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (3.699318ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.495080  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.333697ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.497377  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:07.497411  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:07.498994  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (3.916414ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.499339  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.605885ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.501764  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:07.501801  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:07.503173  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (3.777293ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.504561  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.014286ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.505852  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:07.506015  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:07.507569  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (3.888831ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.508292  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.349895ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.511749  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:07.511785  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:07.514218  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (5.247633ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.515205  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.859994ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.517132  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:07.517169  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:07.519598  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (4.926145ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.519995  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.434623ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.522764  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:07.522801  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:07.524576  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (4.496014ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.525631  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.582058ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.528371  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:07.528443  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:07.530483  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (4.975804ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.530895  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.672739ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.533843  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:07.533888  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:07.534896  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (3.885891ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.536297  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.836254ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.537558  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:07.537611  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:07.539305  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (4.136364ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.539792  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.859337ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.542654  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:07.542708  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:07.544585  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (4.240058ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.544722  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.77135ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.548038  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:07.548122  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:07.548980  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (3.925775ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.550203  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.729652ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.553559  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:07.553715  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (4.219095ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.553722  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:07.555914  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.700491ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.556956  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:07.556994  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:07.558490  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (4.425647ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.559013  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.56488ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.561601  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:07.561634  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:07.563932  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.818176ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.584379  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (25.188113ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.593893  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:07.593940  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:07.596756  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (6.379767ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.597432  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.836635ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.600075  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:07.600106  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:07.601831  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (4.559897ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.601980  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.626987ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.605548  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:07.605666  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:07.611086  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.48556ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.612502  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (10.344026ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.616275  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:07.616309  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:07.618536  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.952351ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.620771  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (7.864214ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.624294  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:07.624339  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:07.626990  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.18878ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.629044  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (7.898333ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.637076  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (7.666493ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.652110  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (14.127588ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.659259  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (6.320295ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.662187  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.303897ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.690145  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (26.169211ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.693268  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.358606ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.696020  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.12444ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.699692  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.444483ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.706651  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (4.433348ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.709601  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.186199ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.716113  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (2.182329ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.720411  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.748023ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.723320  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.267758ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.725935  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (946.378µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.735469  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (4.57988ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.739390  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.148464ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.746539  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (3.768597ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.751772  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.1057ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.757101  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (3.732776ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.763536  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (4.783388ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.767314  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.289033ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.770074  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.122279ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.772957  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.211726ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.775628  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.121251ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.778613  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (883.935µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.781922  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.822753ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.784446  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (979.564µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.788507  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (2.57564ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.791960  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.963957ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.794500  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (943.055µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.797666  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.709099ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.800303  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.066195ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.802586  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (763.563µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.805129  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.044901ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.807767  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (975.119µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.810449  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.107551ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.812933  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (927.319µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.815324  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (912.598µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.818887  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (929.551µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.821521  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (955.761µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.824024  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (990.368µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.827063  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.487263ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.830015  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.372064ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.833074  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.405313ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.836229  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.563626ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.839057  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.260311ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.841850  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.105206ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.844556  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.02414ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.851925  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.560341ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.855073  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.515626ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.858077  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.397514ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.860634  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.020234ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.863583  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.438798ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.866908  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (1.62248ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.870687  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (2.110631ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.874718  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.52549ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.878477  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:07.878558  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:07.878734  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1"
I0111 19:14:07.878772  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1": all PVCs bound and nothing to do
I0111 19:14:07.878821  120459 factory.go:1166] Attempting to bind rpod-0 to node1
I0111 19:14:07.878494  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.192354ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.881350  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.005947ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.881788  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:07.881807  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:07.881915  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1"
I0111 19:14:07.881931  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1": all PVCs bound and nothing to do
I0111 19:14:07.881965  120459 factory.go:1166] Attempting to bind rpod-1 to node1
I0111 19:14:07.883739  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0/binding: (4.580834ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:07.884183  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1/binding: (2.000726ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.884248  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:07.884339  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:07.886451  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.931783ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.893859  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (6.981587ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:07.984942  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (2.705514ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.120054  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (31.276286ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.120404  120459 preemption_test.go:561] Creating the preemptor pod...
I0111 19:14:08.123436  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.81612ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.123694  120459 preemption_test.go:567] Creating additional pods...
I0111 19:14:08.123807  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:08.123822  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:08.123932  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.123970  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.126118  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.220923ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.126630  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (1.802615ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:08.127060  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.162068ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59846]
I0111 19:14:08.127089  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.30446ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.128514  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.036733ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.128716  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.29815ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.129247  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.130759  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.657473ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.131830  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (2.26056ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.132965  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.659734ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.135365  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.072453ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.137497  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (4.492383ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.137916  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:08.137969  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:08.138154  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1"
I0111 19:14:08.138186  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1": all PVCs bound and nothing to do
I0111 19:14:08.138281  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:08.138310  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4
I0111 19:14:08.138322  120459 factory.go:1166] Attempting to bind preemptor-pod to node1
I0111 19:14:08.138435  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.139377  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-4 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.139519  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.320931ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.140767  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.749616ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59850]
I0111 19:14:08.140832  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/binding: (2.298459ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59692]
I0111 19:14:08.140995  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:08.141098  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (5.350724ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59694]
I0111 19:14:08.142151  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.840906ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59854]
I0111 19:14:08.142223  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4/status: (2.16162ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.143869  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.290971ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59852]
I0111 19:14:08.144188  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.321757ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.144390  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.900827ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59850]
I0111 19:14:08.144478  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.144632  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:08.144662  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:08.144793  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.144858  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.146520  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.655319ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.147066  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3/status: (1.765772ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59852]
I0111 19:14:08.148200  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.209582ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59858]
I0111 19:14:08.148914  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.912065ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59848]
I0111 19:14:08.149162  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.744018ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59852]
I0111 19:14:08.149499  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.149701  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:08.149714  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5
I0111 19:14:08.149821  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.149900  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-5 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.152875  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (2.388986ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59858]
I0111 19:14:08.153347  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5/status: (2.877146ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59856]
I0111 19:14:08.154154  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.370621ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59862]
I0111 19:14:08.149930  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.768572ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59860]
I0111 19:14:08.154838  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.092586ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59856]
I0111 19:14:08.155086  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.155256  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:08.155275  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:08.155383  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.155427  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.157335  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.051386ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59862]
I0111 19:14:08.157353  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.300016ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59858]
I0111 19:14:08.158261  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (2.140218ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59856]
I0111 19:14:08.158606  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.012356ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59864]
I0111 19:14:08.159637  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (980.884µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59856]
I0111 19:14:08.159868  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.992486ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59862]
I0111 19:14:08.159872  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.160123  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:08.160143  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3
I0111 19:14:08.160244  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.160285  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-3 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.163144  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3/status: (2.311204ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59858]
I0111 19:14:08.163737  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.405944ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59856]
I0111 19:14:08.166178  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.031926ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59856]
I0111 19:14:08.166724  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (3.10178ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59858]
I0111 19:14:08.167710  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-3.1578e12b5010252a: (6.621718ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59868]
I0111 19:14:08.168871  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.170788  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:08.170841  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:08.170986  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.171064  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.171969  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.184828ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59858]
I0111 19:14:08.173929  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (13.16623ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59866]
I0111 19:14:08.174878  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.734026ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59872]
I0111 19:14:08.176519  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.367091ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59874]
I0111 19:14:08.176550  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (3.952282ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59858]
I0111 19:14:08.176714  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10/status: (4.970757ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59868]
I0111 19:14:08.178971  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.693252ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59870]
I0111 19:14:08.179375  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.156134ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59866]
I0111 19:14:08.179387  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.179738  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:08.179755  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:08.179850  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.179887  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.181975  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.472826ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59878]
I0111 19:14:08.182009  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.466648ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59876]
I0111 19:14:08.183397  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12/status: (2.912756ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59870]
I0111 19:14:08.183757  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.382027ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59866]
I0111 19:14:08.185844  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.123282ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59876]
I0111 19:14:08.186220  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.186409  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:08.186430  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:08.186522  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.186568  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.187993  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.00291ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59866]
I0111 19:14:08.189300  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.929721ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59880]
I0111 19:14:08.189375  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14/status: (2.200363ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59876]
I0111 19:14:08.189614  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (2.7036ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59878]
I0111 19:14:08.191249  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.973666ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59866]
I0111 19:14:08.191303  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.565947ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59876]
I0111 19:14:08.192148  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.192355  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:08.192393  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:08.192498  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.192559  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.195615  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.246745ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59878]
I0111 19:14:08.197260  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.666401ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59884]
I0111 19:14:08.198892  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (4.688022ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59882]
I0111 19:14:08.200141  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.983507ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59878]
I0111 19:14:08.202816  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.985372ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59882]
I0111 19:14:08.204522  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16/status: (11.009905ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59880]
I0111 19:14:08.205789  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.447048ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59882]
I0111 19:14:08.206883  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.168856ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59880]
I0111 19:14:08.207470  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.207737  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:08.207906  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:08.208151  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.208227  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.209813  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.168484ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59882]
I0111 19:14:08.210425  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.542551ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59886]
I0111 19:14:08.211207  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17/status: (2.590857ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59880]
I0111 19:14:08.213200  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.580595ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59886]
I0111 19:14:08.213476  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.213485  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.080593ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59882]
I0111 19:14:08.213638  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:08.213652  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:08.213763  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.213799  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.215997  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.667434ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59886]
I0111 19:14:08.216450  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16/status: (2.132738ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59880]
I0111 19:14:08.216825  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.909476ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59888]
I0111 19:14:08.218654  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-16.1578e12b52e802ea: (4.117443ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59890]
I0111 19:14:08.219159  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (7.404449ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59884]
I0111 19:14:08.220062  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.83986ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59880]
I0111 19:14:08.220360  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.220524  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:08.220540  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:08.220629  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.220698  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.220734  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.870472ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59886]
I0111 19:14:08.222614  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17/status: (1.622906ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59890]
I0111 19:14:08.223737  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-17.1578e12b53d7102c: (2.550261ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59886]
I0111 19:14:08.224483  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.312569ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59890]
I0111 19:14:08.224735  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.238012ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59892]
I0111 19:14:08.224809  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.224949  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:08.224984  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:08.225080  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.225147  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.226216  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.423127ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59888]
I0111 19:14:08.228632  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.649949ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59896]
I0111 19:14:08.229338  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (3.22655ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59894]
I0111 19:14:08.229622  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24/status: (3.761684ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59886]
I0111 19:14:08.230717  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (5.500572ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59890]
I0111 19:14:08.231334  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.316286ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59896]
I0111 19:14:08.231591  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.231800  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:08.231835  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:08.231939  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.232000  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.233414  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.150416ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59890]
I0111 19:14:08.234538  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.981055ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59888]
I0111 19:14:08.235540  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.513649ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59898]
I0111 19:14:08.235890  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23/status: (3.317128ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59896]
I0111 19:14:08.235944  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.921414ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59890]
I0111 19:14:08.237951  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.600421ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59888]
I0111 19:14:08.238396  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.054843ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59898]
I0111 19:14:08.238655  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.238859  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:08.238914  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:08.239113  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.239208  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.240665  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.802945ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59888]
I0111 19:14:08.241107  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.213706ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.241909  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28/status: (1.679774ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59902]
I0111 19:14:08.243003  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (3.551282ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59898]
I0111 19:14:08.243308  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.039626ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59902]
I0111 19:14:08.243594  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.360265ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59888]
I0111 19:14:08.243894  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.244118  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:08.244176  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:08.244321  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.244398  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.245941  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.251644ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.246766  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.692812ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59904]
I0111 19:14:08.249467  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31/status: (2.088938ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59904]
I0111 19:14:08.251099  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.157898ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59904]
I0111 19:14:08.251491  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.252113  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (8.116278ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59898]
I0111 19:14:08.252861  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:08.252883  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:08.253018  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.253071  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.254080  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:08.254973  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.367639ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59904]
I0111 19:14:08.255073  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:08.257702  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.999018ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59906]
I0111 19:14:08.258391  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.804097ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59904]
I0111 19:14:08.258926  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28/status: (4.151896ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.260775  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.493622ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.261079  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.261334  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:08.261358  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:08.261757  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.261895  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.262146  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-28.1578e12b55afc19b: (7.65228ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59908]
I0111 19:14:08.262210  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.278277ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59904]
I0111 19:14:08.265727  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:08.265866  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.270868ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59906]
I0111 19:14:08.266353  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34/status: (2.835751ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59912]
I0111 19:14:08.266391  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (3.916228ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.267384  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:08.267914  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.881855ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59910]
I0111 19:14:08.269935  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (3.184447ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59912]
I0111 19:14:08.270520  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.270765  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:08.270784  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:08.270787  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (4.027397ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59906]
I0111 19:14:08.270911  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.270973  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.273452  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (2.260268ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59910]
I0111 19:14:08.273886  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.20682ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59916]
I0111 19:14:08.274082  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36/status: (2.545058ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.274759  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.142424ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59914]
I0111 19:14:08.276527  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:08.277104  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.181984ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59916]
I0111 19:14:08.277584  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (2.90232ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.277991  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.278159  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:08.278222  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:08.278322  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.278365  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.279464  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.906967ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59914]
I0111 19:14:08.281335  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39/status: (2.206013ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.281844  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.33789ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59918]
I0111 19:14:08.282645  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (3.539378ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59910]
I0111 19:14:08.283772  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.539084ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59918]
I0111 19:14:08.284155  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.284316  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.881138ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59914]
I0111 19:14:08.284324  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:08.284340  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:08.284419  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.284480  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.286461  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.238345ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59922]
I0111 19:14:08.287205  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.78529ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59920]
I0111 19:14:08.287289  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.502106ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59900]
I0111 19:14:08.289074  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41/status: (4.139538ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59910]
I0111 19:14:08.290046  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.334208ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59920]
I0111 19:14:08.292042  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (2.111096ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59910]
I0111 19:14:08.292192  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.732154ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59920]
I0111 19:14:08.292661  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.292832  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:08.292873  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:08.293010  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.293067  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.294870  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.114272ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59926]
I0111 19:14:08.295335  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.714019ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59924]
I0111 19:14:08.295853  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.1848ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59910]
I0111 19:14:08.298081  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43/status: (4.449033ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59922]
I0111 19:14:08.298126  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.70421ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59924]
I0111 19:14:08.300983  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.630158ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59922]
I0111 19:14:08.301176  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.461414ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59926]
I0111 19:14:08.301227  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.301390  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:08.301437  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:08.301647  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.301741  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.303205  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.136517ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59922]
I0111 19:14:08.303877  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.489003ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.305347  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (3.315818ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59926]
I0111 19:14:08.308108  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.24763ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.308707  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.308917  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:08.308960  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:08.309089  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.309165  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.311195  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.765312ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.312529  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49/status: (2.749391ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59922]
I0111 19:14:08.312980  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.866967ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59930]
I0111 19:14:08.314708  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.613331ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59922]
I0111 19:14:08.315095  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.315312  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:08.315366  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:08.315515  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.315570  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.317048  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.167461ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.318305  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46/status: (2.436327ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59930]
I0111 19:14:08.320027  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-46.1578e12b5969f53b: (3.639518ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.320815  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.774029ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59930]
I0111 19:14:08.321155  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.321364  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:08.321386  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:08.321568  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.321727  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.323431  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.459306ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.324382  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49/status: (2.255275ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.326089  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.300904ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.326336  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.326482  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:08.326503  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:08.326587  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.326639  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.328010  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (986.562µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.328966  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48/status: (2.068183ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.330534  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.13795ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.330781  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.330921  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:08.330938  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:08.331031  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.331073  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.332614  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.222721ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.334525  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-49.1578e12b59db33b4: (11.202757ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59934]
I0111 19:14:08.334591  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47/status: (3.292019ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59928]
I0111 19:14:08.336507  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.413595ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.336640  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.500633ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59934]
I0111 19:14:08.337117  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.337375  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:08.337392  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:08.337509  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.338046  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.338598  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.589303ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59934]
I0111 19:14:08.339381  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (1.452906ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.340049  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48/status: (1.501548ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59936]
I0111 19:14:08.341373  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (964.759µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59936]
I0111 19:14:08.341880  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.342396  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-48.1578e12b5ae5f2d3: (2.542412ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59934]
I0111 19:14:08.342663  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:08.342720  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:08.342840  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.342908  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.344406  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.132928ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59936]
I0111 19:14:08.345665  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47/status: (2.391182ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.346265  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-47.1578e12b5b299ee2: (2.53106ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59938]
I0111 19:14:08.347203  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.091189ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59932]
I0111 19:14:08.347489  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.347665  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:08.347704  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:08.347783  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.347825  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.349980  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.790038ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59936]
I0111 19:14:08.350441  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43/status: (2.398094ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59938]
I0111 19:14:08.351014  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-43.1578e12b58e5accf: (2.402149ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59940]
I0111 19:14:08.352463  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (1.582342ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59938]
I0111 19:14:08.352764  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.352965  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:08.353002  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:08.353142  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.353195  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.354497  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.066944ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59936]
I0111 19:14:08.355246  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.432585ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59942]
I0111 19:14:08.355790  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (2.319822ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59940]
I0111 19:14:08.357492  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.188326ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59942]
I0111 19:14:08.357819  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.357988  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:08.358025  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:08.358171  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.358255  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.359806  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.010672ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59942]
I0111 19:14:08.360875  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41/status: (2.354944ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59936]
I0111 19:14:08.362305  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-41.1578e12b5862a646: (2.418241ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59944]
I0111 19:14:08.362908  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.580122ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59936]
I0111 19:14:08.363773  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.363942  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:08.363958  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:08.364044  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.364090  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.365982  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45/status: (1.658599ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59944]
I0111 19:14:08.366039  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.314384ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59942]
I0111 19:14:08.367564  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.132053ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59944]
I0111 19:14:08.367825  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.367962  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:08.367977  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:08.368053  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.368095  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.368110  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-45.1578e12b5c7b24f8: (3.062593ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.370402  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.745613ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59942]
I0111 19:14:08.370595  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44/status: (2.183357ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59944]
I0111 19:14:08.371348  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (2.694558ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.372222  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.071178ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59944]
I0111 19:14:08.372574  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.372787  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:08.372829  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:08.373062  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.373143  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.374691  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.262437ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.375942  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39/status: (2.43935ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59942]
I0111 19:14:08.376202  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-39.1578e12b5805573f: (2.165993ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59948]
I0111 19:14:08.377837  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.468978ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59942]
I0111 19:14:08.378158  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.378419  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:08.378439  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:08.378537  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.378597  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.380521  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.15266ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.381168  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44/status: (1.774486ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59948]
I0111 19:14:08.382966  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-44.1578e12b5d5e81b9: (2.055682ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59950]
I0111 19:14:08.383457  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.48126ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59948]
I0111 19:14:08.383776  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.383962  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:08.383982  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:08.384150  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.384206  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.386062  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.263585ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59952]
I0111 19:14:08.386470  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (2.024005ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.386958  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42/status: (2.508421ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59950]
I0111 19:14:08.388521  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.140851ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.388846  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.389061  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:08.389098  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:08.389219  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.389303  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.390973  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.31839ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.391648  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40/status: (2.047902ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59952]
I0111 19:14:08.391780  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.245015ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59954]
I0111 19:14:08.393224  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.129473ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59952]
I0111 19:14:08.393474  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.393625  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:08.393641  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:08.393746  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.393788  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.396334  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.475297ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.396564  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42/status: (2.557832ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59952]
I0111 19:14:08.396881  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-42.1578e12b5e545483: (2.260681ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59956]
I0111 19:14:08.403962  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (7.052537ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59952]
I0111 19:14:08.404298  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.418627ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59956]
I0111 19:14:08.404458  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.404663  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:08.404719  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:08.404846  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.404917  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.407418  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.562185ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.411319  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-40.1578e12b5ea210f5: (3.258704ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.413422  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40/status: (2.967459ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59952]
I0111 19:14:08.415791  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.647865ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.416079  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.416289  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:08.416329  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:08.416496  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.416579  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.418873  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36/status: (1.958708ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.421080  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-36.1578e12b5794893d: (3.958257ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59958]
I0111 19:14:08.421153  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.564375ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59960]
I0111 19:14:08.422262  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (1.915785ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59946]
I0111 19:14:08.422597  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.422903  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:08.422921  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:08.423067  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.423124  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.424614  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.209568ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59958]
I0111 19:14:08.425136  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.347991ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.426393  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38/status: (3.036011ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59960]
I0111 19:14:08.428493  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.545705ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.428959  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.429185  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:08.429226  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:08.429498  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.429602  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.431918  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.848378ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.432417  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34/status: (1.929968ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59958]
I0111 19:14:08.434125  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-34.1578e12b5709f69e: (3.621811ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59964]
I0111 19:14:08.434505  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.651691ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59958]
I0111 19:14:08.434863  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.435092  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:08.435129  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:08.435284  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.435358  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.436814  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (1.177024ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.440564  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-38.1578e12b60a62c50: (4.051201ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59966]
I0111 19:14:08.440641  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38/status: (4.965274ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59964]
I0111 19:14:08.445752  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (3.186528ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59964]
I0111 19:14:08.446068  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.446285  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:08.446359  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:08.446487  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.446544  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.449355  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.400689ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.449509  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37/status: (2.271161ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59968]
I0111 19:14:08.449437  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (2.511234ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59964]
I0111 19:14:08.451712  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.515206ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59964]
I0111 19:14:08.452057  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.452213  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:08.452228  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:08.452395  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.452464  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.454865  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.654515ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59970]
I0111 19:14:08.454968  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (2.241353ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.455215  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35/status: (2.50726ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59964]
I0111 19:14:08.457067  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.348318ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.457378  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.457554  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:08.457579  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:08.457720  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.457821  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.460890  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (2.470962ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.460983  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-37.1578e12b620b86b7: (2.625757ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59970]
I0111 19:14:08.463262  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37/status: (1.99658ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59970]
I0111 19:14:08.465321  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.478577ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59972]
I0111 19:14:08.465644  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.465815  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:08.465832  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:08.465911  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.465958  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.468555  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.289832ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.469227  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35/status: (1.604201ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59974]
I0111 19:14:08.469571  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-35.1578e12b6265cd0f: (3.062606ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59972]
I0111 19:14:08.471086  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.175075ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59974]
I0111 19:14:08.471527  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.471715  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:08.471735  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:08.471813  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.471859  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.474430  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31/status: (1.836715ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59972]
I0111 19:14:08.474998  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.526567ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.478279  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (3.085457ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59972]
I0111 19:14:08.478520  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.478704  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:08.478751  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:08.478885  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.478950  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.479896  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-31.1578e12b55fefce0: (2.691749ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59976]
I0111 19:14:08.480588  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.137408ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.481487  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (2.06309ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59972]
I0111 19:14:08.492275  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (11.842354ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59976]
I0111 19:14:08.492375  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (4.437348ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59972]
I0111 19:14:08.492721  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.492884  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:08.492901  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:08.493015  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.493079  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.494658  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.284731ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.495294  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (1.93483ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59976]
I0111 19:14:08.496468  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.369841ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.497047  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.281055ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59978]
I0111 19:14:08.497954  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.498151  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:08.498190  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:08.498366  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.498444  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.501187  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (1.867042ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59980]
I0111 19:14:08.502069  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-33.1578e12b63fa00fe: (3.099828ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59976]
I0111 19:14:08.502339  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (3.623348ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.503867  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.427429ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59980]
I0111 19:14:08.504213  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.504388  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:08.504435  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:08.504533  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.504579  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.507348  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (2.49288ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.507989  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (2.175389ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59976]
I0111 19:14:08.508746  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-32.1578e12b64d19a1c: (3.225581ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59982]
I0111 19:14:08.508981  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (1.452789ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59984]
I0111 19:14:08.509197  120459 preemption_test.go:583] Check unschedulable pods still exists and were never scheduled...
I0111 19:14:08.509394  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.344848ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.509648  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.509851  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:08.509873  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:08.509997  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.510055  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.510556  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.19568ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59984]
I0111 19:14:08.512468  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.579565ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59984]
I0111 19:14:08.513326  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (2.415802ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.513069  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (2.740158ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59976]
I0111 19:14:08.514453  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.574532ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59986]
I0111 19:14:08.515283  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.193185ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.516049  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.516312  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:08.516330  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:08.516445  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.516499  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.519080  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (3.941141ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59986]
I0111 19:14:08.519360  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (2.621936ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.520305  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-23.1578e12b5541d655: (2.775105ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59988]
I0111 19:14:08.521068  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23/status: (4.254874ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59984]
I0111 19:14:08.521697  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.922604ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59986]
I0111 19:14:08.522766  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (1.18799ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59988]
I0111 19:14:08.523354  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.523463  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (1.294418ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59986]
I0111 19:14:08.523592  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:08.523610  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:08.523730  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.523788  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.525380  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.075659ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59990]
I0111 19:14:08.526365  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (1.386984ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.527025  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-30.1578e12b65d4a0e4: (2.536315ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59992]
I0111 19:14:08.527368  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (3.333812ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59988]
I0111 19:14:08.528864  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.921345ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.529168  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.187669ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59992]
I0111 19:14:08.529400  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.529568  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:08.529587  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:08.529698  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.529746  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.531918  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.516896ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59994]
I0111 19:14:08.532335  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.869313ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59990]
I0111 19:14:08.532645  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29/status: (2.677616ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59992]
I0111 19:14:08.532799  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (3.45551ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.534403  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.378253ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59990]
I0111 19:14:08.534765  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.534922  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:08.534939  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:08.535011  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.535102  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.536634  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (2.064907ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.537434  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.863485ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59990]
I0111 19:14:08.538528  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24/status: (1.979831ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59996]
I0111 19:14:08.539065  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-24.1578e12b54d940ed: (3.269746ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59994]
I0111 19:14:08.540516  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (1.430611ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59996]
I0111 19:14:08.541131  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (4.036065ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59962]
I0111 19:14:08.541164  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.541382  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:08.541401  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:08.541483  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.541529  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.543370  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.549123ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59994]
I0111 19:14:08.544254  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.667047ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59998]
I0111 19:14:08.545217  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.062527ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59994]
I0111 19:14:08.545930  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27/status: (1.912848ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60000]
I0111 19:14:08.546717  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (5.041402ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59990]
I0111 19:14:08.546926  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.336946ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59994]
I0111 19:14:08.548408  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.123305ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60000]
I0111 19:14:08.548658  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.548786  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.320518ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59998]
I0111 19:14:08.548918  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:08.548937  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:08.549023  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.549078  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.550442  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.094082ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59998]
I0111 19:14:08.551202  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.480961ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.553257  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (3.884468ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60000]
I0111 19:14:08.553316  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (2.081299ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60006]
I0111 19:14:08.553577  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (2.364335ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59998]
I0111 19:14:08.554926  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.247202ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.555330  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.379454ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59998]
I0111 19:14:08.555534  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.555736  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:08.555754  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:08.555856  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.555899  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.559614  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (4.191622ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.559842  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27/status: (2.907349ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59998]
I0111 19:14:08.559889  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-27.1578e12b67b4e47b: (2.836515ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60004]
I0111 19:14:08.561836  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (2.648107ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60008]
I0111 19:14:08.561890  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.658164ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.562267  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.562876  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:08.562898  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:08.563329  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.563420  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.565071  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.152412ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60008]
I0111 19:14:08.565572  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26/status: (1.784541ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.565695  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (5.497318ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:59998]
I0111 19:14:08.567181  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.145805ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.567739  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-26.1578e12b6828152d: (3.092538ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60010]
I0111 19:14:08.567801  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.700453ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60008]
I0111 19:14:08.567982  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.568130  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:08.568148  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:08.568217  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.568275  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.569663  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.237448ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60010]
I0111 19:14:08.570163  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.331331ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60014]
I0111 19:14:08.570913  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25/status: (2.185916ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60012]
I0111 19:14:08.571938  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (3.593642ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.572622  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.290054ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60014]
I0111 19:14:08.572950  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.573177  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:08.573196  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:08.573287  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.573377  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.575078  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.40319ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60014]
I0111 19:14:08.575503  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (3.111805ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.577230  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.392384ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.577545  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22/status: (3.845824ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60010]
I0111 19:14:08.578418  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (4.398571ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60016]
I0111 19:14:08.580154  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (2.1287ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60010]
I0111 19:14:08.580982  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.580641  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (3.09609ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60002]
I0111 19:14:08.588667  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:08.588783  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:08.588971  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.589064  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.590130  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (8.676166ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60016]
I0111 19:14:08.592864  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25/status: (2.116534ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.594365  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (4.947886ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60014]
I0111 19:14:08.594605  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (3.047444ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60020]
I0111 19:14:08.595095  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.201167ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.595461  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.596350  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:08.596407  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:08.596556  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.596627  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.596650  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.542079ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60014]
I0111 19:14:08.597205  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-25.1578e12b694d00c9: (5.380553ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60016]
I0111 19:14:08.599352  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (2.23406ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60022]
I0111 19:14:08.600342  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-22.1578e12b699ad3b6: (2.251107ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.601053  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22/status: (4.065054ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.601253  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (3.704298ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60016]
I0111 19:14:08.601256  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.204671ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60022]
I0111 19:14:08.602935  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (1.394443ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.603184  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.603430  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:08.603456  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:08.603515  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (1.80129ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60016]
I0111 19:14:08.603610  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.603662  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.606337  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.967563ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.608027  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (2.34949ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60028]
I0111 19:14:08.608449  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.927594ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60016]
I0111 19:14:08.608886  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21/status: (4.152508ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.610892  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.158246ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.611277  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.611638  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:08.611652  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:08.611807  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.611859  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.612187  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (3.285637ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60026]
I0111 19:14:08.614059  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.385074ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60026]
I0111 19:14:08.614324  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20/status: (2.186907ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.615333  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (2.293102ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.616015  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.57316ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.616341  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.605596ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60026]
I0111 19:14:08.616609  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.69056ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.616897  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.617323  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:08.617353  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:08.617474  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.617532  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.619750  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21/status: (1.981336ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.619817  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (3.030354ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.620180  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (2.363847ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.621297  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.116683ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.621588  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.621796  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:08.621807  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:08.621833  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.468889ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.621882  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.621920  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.622566  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-21.1578e12b6b68f6fe: (4.270878ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60034]
I0111 19:14:08.629059  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19/status: (6.768961ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.629622  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (6.651282ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.630206  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (6.827197ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.631861  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (2.007865ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.632118  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.632306  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:08.632316  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:08.632407  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.632447  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.632530  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (1.451242ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.634108  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.306453ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.635721  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (12.335697ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60034]
I0111 19:14:08.636381  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (2.583684ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60018]
I0111 19:14:08.636751  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18/status: (3.820066ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.637741  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.51933ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60034]
I0111 19:14:08.638888  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (1.607984ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.639434  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (2.211119ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.639811  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.640560  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.225426ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.640569  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:08.640583  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:08.640708  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.640749  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.642843  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.452134ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.643922  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19/status: (2.483455ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.645109  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-19.1578e12b6c7f90c8: (3.215291ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60036]
I0111 19:14:08.645522  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.159142ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60032]
I0111 19:14:08.645984  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.646293  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:08.646314  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:08.646419  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.646514  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.648278  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.491781ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60036]
I0111 19:14:08.649085  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (7.732254ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60034]
I0111 19:14:08.651020  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.568715ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60034]
I0111 19:14:08.652663  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-18.1578e12b6d202eae: (5.170532ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.653420  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (2.008167ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60034]
I0111 19:14:08.655503  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (1.646006ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.655920  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18/status: (7.299615ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60036]
I0111 19:14:08.657186  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.314261ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.659088  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (1.262512ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.659850  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (1.708908ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60036]
I0111 19:14:08.660095  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.660250  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:08.660261  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:08.660350  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.660387  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.661800  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (1.978194ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.663012  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12/status: (1.855788ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60036]
I0111 19:14:08.663057  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (2.208671ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60038]
I0111 19:14:08.665004  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-12.1578e12b5226ad41: (2.677699ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.665630  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (3.002363ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60040]
I0111 19:14:08.665963  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (2.341239ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60038]
I0111 19:14:08.666289  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.666722  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:08.666747  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:08.666883  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.666951  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.667942  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (1.280945ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.668159  120459 preemption_test.go:598] Cleaning up all pods...
I0111 19:14:08.670129  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.60382ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60046]
I0111 19:14:08.670718  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15/status: (2.474629ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.670954  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (3.274727ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60042]
I0111 19:14:08.672474  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.343311ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60024]
I0111 19:14:08.672891  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.673352  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (4.806791ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60048]
I0111 19:14:08.673563  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:08.673580  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:08.673709  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.673757  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.676163  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (2.207253ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60046]
I0111 19:14:08.676910  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10/status: (2.948517ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60042]
I0111 19:14:08.677517  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-10.1578e12b519ffc86: (3.095036ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.679321  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.22789ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.679569  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.679801  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:08.679818  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:08.679931  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.679968  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.680004  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (5.278208ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.682756  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15/status: (2.485167ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.682872  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (2.608108ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.684512  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.093106ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.685114  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.685303  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:08.685320  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:08.685403  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.685443  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.686362  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (5.747808ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60046]
I0111 19:14:08.688714  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (2.924545ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.689010  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11/status: (3.243513ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.691577  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-15.1578e12b6f2e697c: (10.059917ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60054]
I0111 19:14:08.692613  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (2.63503ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.692883  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.693076  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:08.693093  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:08.693186  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.693252  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.694247  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.192804ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60054]
I0111 19:14:08.695016  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (8.211704ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60046]
I0111 19:14:08.696607  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (2.720166ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.696610  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (2.700896ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.697589  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-8.1578e12b50b17592: (2.745216ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60054]
I0111 19:14:08.699890  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (2.135587ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60050]
I0111 19:14:08.700330  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.700493  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:08.700510  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:08.700632  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.700703  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.702157  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (6.838623ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60046]
I0111 19:14:08.703738  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (2.300226ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.704318  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11/status: (2.590385ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60054]
I0111 19:14:08.705120  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-11.1578e12b7048d7de: (2.263453ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60046]
I0111 19:14:08.706139  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.396691ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60054]
I0111 19:14:08.706378  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.706536  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:08.706550  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6
I0111 19:14:08.706640  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.706694  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-6 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.709360  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6/status: (2.446749ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.710096  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.748855ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60060]
I0111 19:14:08.710508  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (3.646324ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60056]
I0111 19:14:08.711062  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (8.072968ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60058]
I0111 19:14:08.712340  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (2.545908ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.712577  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.712755  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:08.712777  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:08.712868  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.712947  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.717588  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (6.219814ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60056]
I0111 19:14:08.721471  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (2.587016ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60060]
I0111 19:14:08.721766  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7/status: (3.041641ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.723106  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.737083ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60062]
I0111 19:14:08.723408  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (1.294876ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60052]
I0111 19:14:08.723827  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.723975  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:08.723992  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:08.724072  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.724113  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.726223  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.602277ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60060]
I0111 19:14:08.726801  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.945021ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60064]
I0111 19:14:08.727041  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9/status: (2.642072ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60062]
I0111 19:14:08.729063  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.589122ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60064]
I0111 19:14:08.729357  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.729584  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:08.729617  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (8.668281ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60056]
I0111 19:14:08.729651  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-7
I0111 19:14:08.729719  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:08.729753  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:08.729950  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.730017  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.731599  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (993.456µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.732824  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.772419ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60060]
I0111 19:14:08.733135  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13/status: (2.468428ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60066]
I0111 19:14:08.734724  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.525834ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60060]
I0111 19:14:08.735042  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.409699ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60066]
I0111 19:14:08.735308  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.735513  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:08.735534  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:08.735662  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (5.686003ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60064]
I0111 19:14:08.735656  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:08.735902  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:08.737147  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (966.72µs) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.738496  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13/status: (1.841591ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.739771  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-13.1578e12b72f0fc35: (2.319379ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.740733  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (4.575704ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60060]
I0111 19:14:08.742889  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.212982ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.743196  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:08.744285  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:08.744321  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:08.745730  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (4.44176ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.746537  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.980432ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.749055  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:08.749126  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-11
I0111 19:14:08.751275  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (5.033414ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.751991  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.40162ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.754779  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:08.754814  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-12
I0111 19:14:08.756005  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (4.347436ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.757539  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.950664ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.759124  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:08.759213  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-13
I0111 19:14:08.761309  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (4.904173ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.767423  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:08.767500  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:08.767783  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (8.146248ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.769492  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (7.379114ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.770406  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.502156ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.773269  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:08.774738  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:08.775510  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (5.275519ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60068]
I0111 19:14:08.776106  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.986148ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.780356  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:08.780718  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:08.782069  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (5.476878ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.783176  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.955564ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.785652  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:08.785710  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:08.787185  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (4.522798ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.787741  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.77066ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.790958  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:08.791031  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-18
I0111 19:14:08.792717  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (4.847864ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.793143  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.550238ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.795773  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:08.795807  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:08.797480  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (4.441499ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.797987  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.90164ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.800780  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:08.801152  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-20
I0111 19:14:08.803128  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (5.281553ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.804014  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.341723ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.807163  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:08.807262  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:08.808526  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (4.285609ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.809120  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.570115ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.811501  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:08.811543  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-22
I0111 19:14:08.813360  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.504221ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.814827  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (5.941741ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.817701  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:08.817739  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-23
I0111 19:14:08.819647  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (4.477871ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.819778  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.720452ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.823161  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:08.823273  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-24
I0111 19:14:08.824448  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (4.442788ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.826418  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.756334ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.827258  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:08.827293  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:08.829571  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.784544ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.831749  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (6.969853ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.836262  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:08.836307  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-26
I0111 19:14:08.838933  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (6.71784ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.839606  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.560218ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.842498  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:08.842535  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:08.844113  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (4.694754ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.844996  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.175686ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.847796  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:08.847835  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:08.849856  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (5.331953ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.849924  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.827209ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.853355  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:08.853804  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-29
I0111 19:14:08.855369  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (5.098254ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.856599  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.143999ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.858847  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:08.858886  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:08.860780  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (4.927647ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.861366  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.214609ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.867814  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:08.867849  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-31
I0111 19:14:08.869273  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (7.155852ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.871906  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.110748ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.873127  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:08.873171  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:08.874792  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (5.144193ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.876314  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.879013ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.880225  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:08.880276  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:08.883316  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (7.986207ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.885875  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (5.144038ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.887901  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:08.887942  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:08.890343  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.172433ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.890498  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (5.652015ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.894703  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:08.894732  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:08.901550  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (5.954377ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.902540  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (11.671341ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.907399  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:08.907440  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-36
I0111 19:14:08.911532  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (7.792529ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.912534  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (4.818506ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.917129  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:08.917205  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:08.919891  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (7.925143ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.921558  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.78182ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.926010  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:08.926061  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-38
I0111 19:14:08.928523  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.696167ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.930982  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (10.745776ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.935942  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:08.935984  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:08.938327  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.648455ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.938964  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (7.653514ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.946009  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (6.711608ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.949109  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:08.949185  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:08.952797  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.284842ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.956326  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:08.956362  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-41
I0111 19:14:08.959976  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.317386ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.964504  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (16.002776ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.969491  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:08.970048  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-42
I0111 19:14:08.973210  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (8.180761ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.973879  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.903902ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.979992  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:08.980045  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-43
I0111 19:14:08.982607  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.048217ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.985969  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (12.234953ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.990385  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:08.990429  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-44
I0111 19:14:08.992285  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (5.773571ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:08.992360  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.550707ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.996776  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:08.996878  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-45
I0111 19:14:08.998661  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.446174ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:08.999151  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (6.283704ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.002569  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:09.002606  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-46
I0111 19:14:09.005956  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (6.192049ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.005977  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.681327ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.010130  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:09.010177  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-47
I0111 19:14:09.012123  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.621709ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.012959  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (6.347611ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.018609  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:09.018719  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-48
I0111 19:14:09.019756  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (5.560753ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.021364  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.274452ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.023152  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:09.023226  120459 scheduler.go:450] Skip schedule deleting pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-49
I0111 19:14:09.025089  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.521624ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.025548  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (5.404889ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.032392  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (6.493074ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.034040  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (1.303874ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.039874  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (5.494356ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.043310  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-0: (1.411961ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.047258  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-1: (1.246526ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.050067  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-2: (1.161123ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.052619  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-3: (1.023662ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.055558  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-4: (932.443µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.058024  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-5: (950.884µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.060813  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-6: (1.013066ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.063276  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-7: (807.811µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.076622  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.299472ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.079363  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.170512ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.092308  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (11.376533ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.095288  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-11: (1.299687ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.098382  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-12: (1.466937ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.101205  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-13: (1.212066ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.103828  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.025305ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.106911  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.263413ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.109453  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (983.117µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.111914  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (907.602µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.114375  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-18: (836.925µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.116730  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (866.923µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.119852  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-20: (1.040449ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.122467  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.034769ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.124951  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-22: (915.945µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.127833  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-23: (862.05µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.130603  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-24: (830.36µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.133333  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.120091ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.135955  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-26: (1.036108ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.138544  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.04386ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.141153  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.105426ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.145955  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-29: (3.382617ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.148506  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.041699ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.151080  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-31: (1.037326ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.153654  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.047865ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.156097  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (943.936µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.158385  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (815.567µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.161384  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.355143ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.163940  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-36: (937.447µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.166280  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (844.246µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.168591  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-38: (780.293µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.171134  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-39: (968.963µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.173750  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (823.466µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.176729  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-41: (1.328168ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.179552  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-42: (1.224278ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.182088  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-43: (996.16µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.184785  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-44: (830.765µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.187276  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-45: (1.0028ms) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.189704  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-46: (889.256µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.192171  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-47: (934.514µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.194566  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-48: (815.88µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.197038  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-49: (943.059µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.199559  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (880.255µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.202102  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (927.545µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.204509  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (853.408µs) 404 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.207382  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.394756ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.207640  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:09.207653  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0
I0111 19:14:09.207795  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1"
I0111 19:14:09.207808  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0", node "node1": all PVCs bound and nothing to do
I0111 19:14:09.207846  120459 factory.go:1166] Attempting to bind rpod-0 to node1
I0111 19:14:09.209812  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0/binding: (1.716547ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.210198  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.46057ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.210782  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-0 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:09.212762  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.782041ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.254279  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:09.255243  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:09.266089  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:09.267528  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:09.276724  120459 reflector.go:215] k8s.io/client-go/informers/factory.go:132: forcing resync
I0111 19:14:09.289299  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:09.289326  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1
I0111 19:14:09.289472  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1"
I0111 19:14:09.289486  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1", node "node1": all PVCs bound and nothing to do
I0111 19:14:09.289527  120459 factory.go:1166] Attempting to bind rpod-1 to node1
I0111 19:14:09.292786  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1/binding: (2.721943ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.293002  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/rpod-1 is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:09.295195  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.802855ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.313167  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-0: (1.846817ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.420843  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (1.971791ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.421275  120459 preemption_test.go:561] Creating the preemptor pod...
I0111 19:14:09.424894  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.244404ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.425136  120459 preemption_test.go:567] Creating additional pods...
I0111 19:14:09.425480  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:09.425492  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:09.425592  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.425625  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.429058  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.554699ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60086]
I0111 19:14:09.429262  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.484596ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.432318  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.38396ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.432972  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (6.481582ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.437304  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.317515ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60072]
I0111 19:14:09.437741  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (3.107635ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.438373  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.441408  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (2.510719ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60086]
I0111 19:14:09.444573  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (16.641437ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60088]
I0111 19:14:09.446529  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (7.912644ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.449504  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.62609ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.450457  120459 wrap.go:47] DELETE /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/rpod-1: (8.620935ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60086]
I0111 19:14:09.450842  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:09.450858  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:09.450965  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.451012  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.452934  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.808235ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.453402  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.171077ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60088]
I0111 19:14:09.455356  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.71734ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.455556  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (2.94075ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60090]
I0111 19:14:09.457597  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (6.559294ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60086]
I0111 19:14:09.458230  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.924707ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.458265  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod: (2.313499ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60090]
I0111 19:14:09.458826  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.460811  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.905888ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.461553  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/status: (2.243448ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60088]
I0111 19:14:09.464553  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:09.464577  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod
I0111 19:14:09.464866  120459 scheduler_binder.go:211] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1"
I0111 19:14:09.464897  120459 scheduler_binder.go:221] AssumePodVolumes for pod "preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod", node "node1": all PVCs bound and nothing to do
I0111 19:14:09.465005  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:09.465015  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9
I0111 19:14:09.465135  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.465176  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-9 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.465251  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (4.026853ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.465490  120459 factory.go:1166] Attempting to bind preemptor-pod to node1
I0111 19:14:09.466731  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.217549ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.467907  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9/status: (2.458375ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60088]
I0111 19:14:09.467913  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/preemptor-pod/binding: (1.958731ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60094]
I0111 19:14:09.468455  120459 scheduler.go:569] pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/preemptor-pod is bound successfully on node node1, 1 nodes evaluated, 1 nodes were found feasible
I0111 19:14:09.469819  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/preemptor-pod.1578e12b9c67269d: (9.820919ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60086]
I0111 19:14:09.470160  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-9: (1.751722ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60088]
I0111 19:14:09.470393  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.191615ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60092]
I0111 19:14:09.470438  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.470572  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:09.470590  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:09.470712  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.470754  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.472842  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.606425ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60086]
I0111 19:14:09.472954  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (2.004746ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.474319  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.945389ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.474387  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.555304ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60088]
I0111 19:14:09.475159  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.70951ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.475256  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.977732ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60086]
I0111 19:14:09.475565  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.477728  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.894751ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.477957  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.981261ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.478293  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:09.478333  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10
I0111 19:14:09.478472  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.478540  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-10 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.486389  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10/status: (7.58559ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.486914  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (6.274144ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60070]
I0111 19:14:09.487738  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (7.625484ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60100]
I0111 19:14:09.488012  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (6.747056ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60102]
I0111 19:14:09.490951  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.244953ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60100]
I0111 19:14:09.491451  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-10: (1.933446ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.492243  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.492445  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:09.492467  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8
I0111 19:14:09.492558  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.492640  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-8 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.495085  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.450301ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.496513  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-8.1578e12b9f17c2e0: (2.463023ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60106]
I0111 19:14:09.496755  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8/status: (3.162962ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60102]
I0111 19:14:09.497197  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (3.592799ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60104]
I0111 19:14:09.499075  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-8: (1.300447ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60104]
I0111 19:14:09.499429  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.580726ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.500391  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.500649  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:09.500670  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14
I0111 19:14:09.500783  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.500838  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-14 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.504448  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.809504ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.505361  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14/status: (2.038707ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60106]
I0111 19:14:09.505772  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (2.809831ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60108]
I0111 19:14:09.509534  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.180023ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.509914  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-14: (1.334862ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60106]
I0111 19:14:09.510154  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.550152ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60108]
I0111 19:14:09.510186  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.510375  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:09.510395  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16
I0111 19:14:09.510502  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.510547  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-16 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.512440  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.836655ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.513577  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.17367ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60116]
I0111 19:14:09.513714  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16/status: (2.339594ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60112]
I0111 19:14:09.514125  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (2.190137ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60114]
I0111 19:14:09.516578  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-16: (1.650189ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60112]
I0111 19:14:09.516909  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.92123ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.516943  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.517126  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:09.517143  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15
I0111 19:14:09.517252  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.517307  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-15 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.519772  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15/status: (2.209297ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60114]
I0111 19:14:09.520051  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.990762ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60118]
I0111 19:14:09.520191  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.740613ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60116]
I0111 19:14:09.521572  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-15: (1.187145ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60118]
I0111 19:14:09.522576  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.522762  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:09.522790  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17
I0111 19:14:09.522836  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.205983ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60098]
I0111 19:14:09.522940  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.523057  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-17 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.525840  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17/status: (2.18455ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60114]
I0111 19:14:09.525868  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.911286ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60120]
I0111 19:14:09.525878  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.250038ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60110]
I0111 19:14:09.525994  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.896543ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.528029  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-17: (1.360156ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.528274  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.703543ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60120]
I0111 19:14:09.528339  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.528744  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:09.528766  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19
I0111 19:14:09.528893  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.528953  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-19 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.531011  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19/status: (1.825301ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60110]
I0111 19:14:09.531316  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.541641ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.532585  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (1.97251ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60126]
I0111 19:14:09.534289  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.47112ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.535548  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-19: (3.341388ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60110]
I0111 19:14:09.535935  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.536104  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:09.536147  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21
I0111 19:14:09.536291  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.536314  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.665502ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60124]
I0111 19:14:09.536377  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-21 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.538550  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.439985ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.538576  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21/status: (1.757665ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60110]
I0111 19:14:09.538595  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.333609ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60126]
I0111 19:14:09.539060  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.991337ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60128]
I0111 19:14:09.540975  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.539888ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60126]
I0111 19:14:09.541557  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-21: (1.223493ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60128]
I0111 19:14:09.541878  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.542052  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:09.542070  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25
I0111 19:14:09.542140  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.542181  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-25 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.544047  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.249578ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.544499  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25/status: (2.056693ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60128]
I0111 19:14:09.544838  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.632574ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60130]
I0111 19:14:09.547887  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-25: (1.435644ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60130]
I0111 19:14:09.548363  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.548753  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (7.268458ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60126]
I0111 19:14:09.549385  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:09.549446  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27
I0111 19:14:09.549581  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.549638  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-27 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.552202  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.534702ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60130]
I0111 19:14:09.553542  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (2.900128ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60132]
I0111 19:14:09.553733  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27/status: (3.634871ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.555007  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.008443ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60130]
I0111 19:14:09.556351  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-27: (1.993259ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.556762  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (6.257603ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60134]
I0111 19:14:09.556957  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.557215  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:09.557259  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28
I0111 19:14:09.557353  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.557466  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-28 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.558145  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.56161ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60130]
I0111 19:14:09.559356  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28/status: (1.600778ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.559919  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.597267ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60136]
I0111 19:14:09.560937  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (2.663625ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60132]
I0111 19:14:09.561171  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.385795ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60130]
I0111 19:14:09.561401  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-28: (1.328642ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.561736  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.562627  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:09.562643  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:09.562754  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.562796  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.565230  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.519493ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60138]
I0111 19:14:09.565780  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (2.211855ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60136]
I0111 19:14:09.565947  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (2.918897ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60122]
I0111 19:14:09.567246  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.286357ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60132]
I0111 19:14:09.567378  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.053323ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60136]
I0111 19:14:09.567853  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.568010  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:09.568027  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32
I0111 19:14:09.568155  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.568272  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-32 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.570145  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.295218ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60132]
I0111 19:14:09.571209  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.070588ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60142]
I0111 19:14:09.571287  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32/status: (2.753324ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60138]
I0111 19:14:09.572139  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.204913ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60140]
I0111 19:14:09.572982  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.956252ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60132]
I0111 19:14:09.573012  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-32: (1.233815ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60142]
I0111 19:14:09.573645  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.574019  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:09.574050  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33
I0111 19:14:09.574155  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.574204  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-33 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.575506  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.793878ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60140]
I0111 19:14:09.576694  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.925514ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60144]
I0111 19:14:09.577040  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.666217ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60146]
I0111 19:14:09.577208  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33/status: (2.780152ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60138]
I0111 19:14:09.578993  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-33: (1.249498ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60140]
I0111 19:14:09.579322  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.579473  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.866492ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60144]
I0111 19:14:09.579557  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:09.579574  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35
I0111 19:14:09.579724  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.579761  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-35 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.583159  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35/status: (2.885264ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60140]
I0111 19:14:09.583594  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (3.382547ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60148]
I0111 19:14:09.583982  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (4.079116ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60144]
I0111 19:14:09.584404  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (3.683856ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60150]
I0111 19:14:09.586724  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-35: (1.444522ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60150]
I0111 19:14:09.586953  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.865768ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60148]
I0111 19:14:09.587001  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.587291  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:09.587308  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34
I0111 19:14:09.587443  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.587483  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-34 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.589496  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.465646ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60154]
I0111 19:14:09.589819  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34/status: (1.996646ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60150]
I0111 19:14:09.590510  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.976104ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60148]
I0111 19:14:09.590649  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (2.65718ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60152]
I0111 19:14:09.591539  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-34: (1.13651ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60150]
I0111 19:14:09.592040  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.592257  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:09.592270  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30
I0111 19:14:09.592527  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.592575  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-30 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.593292  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.246497ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60148]
I0111 19:14:09.595718  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30/status: (2.244986ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60150]
I0111 19:14:09.595939  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (2.278648ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60148]
I0111 19:14:09.596375  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (2.949447ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60154]
I0111 19:14:09.596551  120459 wrap.go:47] PATCH /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events/ppod-30.1578e12ba49420ec: (3.010204ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60156]
I0111 19:14:09.597422  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-30: (1.239016ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60150]
I0111 19:14:09.597718  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.598503  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.894379ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60148]
I0111 19:14:09.598948  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:09.598962  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37
I0111 19:14:09.599056  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.599090  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-37 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.602402  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (2.164823ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60160]
I0111 19:14:09.602469  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37/status: (2.701673ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60154]
I0111 19:14:09.602835  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (3.860014ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60156]
I0111 19:14:09.603455  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (3.693021ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60158]
I0111 19:14:09.603815  120459 cacher.go:598] cacher (*core.Pod): 3 objects queued in incoming channel.
I0111 19:14:09.603837  120459 cacher.go:598] cacher (*core.Pod): 4 objects queued in incoming channel.
I0111 19:14:09.605258  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.67241ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60156]
I0111 19:14:09.605745  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-37: (2.619227ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60154]
I0111 19:14:09.605944  120459 cacher.go:598] cacher (*core.Pod): 5 objects queued in incoming channel.
I0111 19:14:09.606223  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.608267  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:09.608283  120459 scheduler.go:454] Attempting to schedule pod: preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40
I0111 19:14:09.608625  120459 factory.go:1070] Unable to schedule preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40: no fit: 0/1 nodes are available: 1 Insufficient cpu, 1 Insufficient memory.; waiting
I0111 19:14:09.608696  120459 factory.go:1175] Updating pod condition for preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-40 to (PodScheduled==False, Reason=Unschedulable)
I0111 19:14:09.610906  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (4.48553ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60158]
I0111 19:14:09.612891  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/events: (1.943863ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60164]
I0111 19:14:09.612938  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.503438ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60158]
I0111 19:14:09.613341  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (2.971218ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60162]
I0111 19:14:09.613822  120459 wrap.go:47] PUT /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40/status: (3.487885ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60160]
I0111 19:14:09.615275  120459 wrap.go:47] POST /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods: (1.748505ms) 201 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60158]
I0111 19:14:09.616013  120459 wrap.go:47] GET /api/v1/namespaces/preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/pods/ppod-40: (1.057987ms) 200 [scheduler.test/v0.0.0 (linux/amd64) kubernetes/$Format 127.0.0.1:60162]
I0111 19:14:09.616300  120459 generic_scheduler.go:1108] Node node1 is a potential node for preemption.
I0111 19:14:09.616534  120459 scheduling_queue.go:821] About to try and schedule pod preemption-race0ec39b98-15d5-11e9-8dcb-0242ac110002/ppod-39
I0111 19:14:09.616565  120459 scheduler.go