Created
July 10, 2018 02:51
-
-
Save yifan-gu/a1cc0a8af097166f316fa90e6652f3db to your computer and use it in GitHub Desktop.
conformance test
This file has been truncated, but you can view the full file.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
[INFO] [19:08:41-0700] Running tests against existing cluster... | |
[INFO] [19:08:41-0700] Running parallel tests N=<default> | |
I0709 19:08:41.741853 10764 test.go:86] Extended test version v3.10.0-alpha.0+e63afaa-1228-dirty | |
Running Suite: Extended | |
======================= | |
Random Seed: 1531188522 - Will randomize all specs | |
Will run 447 specs | |
Running in parallel across 5 nodes | |
I0709 19:08:43.830656 11717 e2e.go:56] The --provider flag is not set. Treating as a conformance test. Some tests may not be run. | |
Jul 9 19:08:43.830: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:08:43.832: INFO: Waiting up to 4h0m0s for all (but 0) nodes to be schedulable | |
Jul 9 19:08:44.246: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready | |
Jul 9 19:08:44.634: INFO: 20 / 20 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) | |
Jul 9 19:08:44.634: INFO: expected 7 pod replicas in namespace 'kube-system', 7 are Running and Ready. | |
Jul 9 19:08:44.692: INFO: Waiting for pods to enter Success, but no pods in "kube-system" match label map[name:e2e-image-puller] | |
Jul 9 19:08:44.692: INFO: Dumping network health container logs from all nodes... | |
Jul 9 19:08:44.761: INFO: e2e test version: v1.10.0+b81c8f8 | |
Jul 9 19:08:44.840: INFO: kube-apiserver version: v1.11.0+d4cacc0 | |
I0709 19:08:44.840549 11717 e2e.go:56] The --provider flag is not set. Treating as a conformance test. Some tests may not be run. | |
SSS | |
------------------------------ | |
I0709 19:08:44.845674 11716 e2e.go:56] The --provider flag is not set. Treating as a conformance test. Some tests may not be run. | |
I0709 19:08:44.850342 11714 e2e.go:56] The --provider flag is not set. Treating as a conformance test. Some tests may not be run. | |
S | |
------------------------------ | |
I0709 19:08:44.859308 11713 e2e.go:56] The --provider flag is not set. Treating as a conformance test. Some tests may not be run. | |
I0709 19:08:44.859324 11748 e2e.go:56] The --provider flag is not set. Treating as a conformance test. Some tests may not be run. | |
SSSSSS | |
------------------------------ | |
[sig-storage] HostPath | |
should give a volume the correct mode [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:08:44.852: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:08:47.067: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
Jul 9 19:08:47.860: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. | |
Jul 9 19:08:48.099: INFO: Found ClusterRoles; assuming RBAC is enabled. | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-hostpath-mrzt2 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:37 | |
[It] should give a volume the correct mode [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test hostPath mode | |
Jul 9 19:08:48.437: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "e2e-tests-hostpath-mrzt2" to be "success or failure" | |
Jul 9 19:08:48.465: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 27.784025ms | |
Jul 9 19:08:50.532: INFO: Pod "pod-host-path-test": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.095115151s | |
STEP: Saw pod success | |
Jul 9 19:08:50.532: INFO: Pod "pod-host-path-test" satisfied condition "success or failure" | |
Jul 9 19:08:50.602: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-host-path-test container test-container-1: <nil> | |
STEP: delete the pod | |
Jul 9 19:08:50.751: INFO: Waiting for pod pod-host-path-test to disappear | |
Jul 9 19:08:50.819: INFO: Pod pod-host-path-test no longer exists | |
[AfterEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:08:50.819: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-hostpath-mrzt2" for this suite. | |
Jul 9 19:08:57.018: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:01.300: INFO: namespace: e2e-tests-hostpath-mrzt2, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:02.072: INFO: namespace e2e-tests-hostpath-mrzt2 deletion completed in 11.171547155s | |
• [SLOW TEST:17.221 seconds] | |
[sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:34 | |
should give a volume the correct mode [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Projected | |
should provide node allocatable (cpu) as default cpu limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:08:44.861: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:08:48.672: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
Jul 9 19:08:49.441: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. | |
Jul 9 19:08:49.693: INFO: Found ClusterRoles; assuming RBAC is enabled. | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-hz5j2 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:08:51.143: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276" in namespace "e2e-tests-projected-hz5j2" to be "success or failure" | |
Jul 9 19:08:51.191: INFO: Pod "downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 48.433563ms | |
Jul 9 19:08:53.252: INFO: Pod "downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.109326818s | |
Jul 9 19:08:55.331: INFO: Pod "downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.188430871s | |
STEP: Saw pod success | |
Jul 9 19:08:55.331: INFO: Pod "downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:08:55.414: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:08:55.588: INFO: Waiting for pod downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:08:55.656: INFO: Pod downwardapi-volume-2fa0cddb-83e6-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:08:55.656: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-hz5j2" for this suite. | |
Jul 9 19:09:01.872: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:04.923: INFO: namespace: e2e-tests-projected-hz5j2, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:07.026: INFO: namespace e2e-tests-projected-hz5j2 deletion completed in 11.313794635s | |
• [SLOW TEST:22.165 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should provide node allocatable (cpu) as default cpu limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Secrets | |
should be able to mount in a volume regardless of a different secret existing with same name in different namespace [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:86 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:08:44.863: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:08:47.936: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
Jul 9 19:08:48.568: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. | |
Jul 9 19:08:48.738: INFO: Found ClusterRoles; assuming RBAC is enabled. | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-z59rs | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:86 | |
Jul 9 19:08:49.815: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secret-namespace-txhdr | |
STEP: Creating secret with name secret-test-2f0ac750-83e6-11e8-bd2e-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:08:51.010: INFO: Waiting up to 5m0s for pod "pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276" in namespace "e2e-tests-secrets-z59rs" to be "success or failure" | |
Jul 9 19:08:51.046: INFO: Pod "pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 36.426674ms | |
Jul 9 19:08:53.109: INFO: Pod "pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.099545743s | |
Jul 9 19:08:55.156: INFO: Pod "pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.146331244s | |
STEP: Saw pod success | |
Jul 9 19:08:55.156: INFO: Pod "pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:08:55.217: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276 container secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:08:55.370: INFO: Waiting for pod pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:08:55.447: INFO: Pod pod-secrets-3037ca1c-83e6-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:08:55.447: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-z59rs" for this suite. | |
Jul 9 19:09:01.708: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:04.411: INFO: namespace: e2e-tests-secrets-z59rs, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:06.079: INFO: namespace e2e-tests-secrets-z59rs deletion completed in 10.554564859s | |
STEP: Destroying namespace "e2e-tests-secret-namespace-txhdr" for this suite. | |
Jul 9 19:09:12.187: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:15.304: INFO: namespace: e2e-tests-secret-namespace-txhdr, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:15.666: INFO: namespace e2e-tests-secret-namespace-txhdr deletion completed in 9.58619834s | |
• [SLOW TEST:30.803 seconds] | |
[sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 | |
should be able to mount in a volume regardless of a different secret existing with same name in different namespace [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:86 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable from pods in volume with mappings [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:02.074: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:04.463: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-fpcz7 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable from pods in volume with mappings [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating projection with secret that has name projected-secret-test-map-38a11c87-83e6-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:09:05.132: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-38a63b92-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-projected-fpcz7" to be "success or failure" | |
Jul 9 19:09:05.162: INFO: Pod "pod-projected-secrets-38a63b92-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.038412ms | |
Jul 9 19:09:07.190: INFO: Pod "pod-projected-secrets-38a63b92-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.058319067s | |
STEP: Saw pod success | |
Jul 9 19:09:07.190: INFO: Pod "pod-projected-secrets-38a63b92-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:09:07.220: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-secrets-38a63b92-83e6-11e8-8401-28d244b00276 container projected-secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:09:07.293: INFO: Waiting for pod pod-projected-secrets-38a63b92-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:09:07.320: INFO: Pod pod-projected-secrets-38a63b92-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:07.320: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-fpcz7" for this suite. | |
Jul 9 19:09:13.443: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:16.255: INFO: namespace: e2e-tests-projected-fpcz7, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:16.782: INFO: namespace e2e-tests-projected-fpcz7 deletion completed in 9.431912033s | |
• [SLOW TEST:14.708 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable from pods in volume with mappings [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-api-machinery] Downward API | |
should provide default limits.cpu/memory from node allocatable [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:15.667: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:17.263: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-94pzx | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should provide default limits.cpu/memory from node allocatable [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward api env vars | |
Jul 9 19:09:18.065: INFO: Waiting up to 5m0s for pod "downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276" in namespace "e2e-tests-downward-api-94pzx" to be "success or failure" | |
Jul 9 19:09:18.094: INFO: Pod "downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 29.033879ms | |
Jul 9 19:09:20.138: INFO: Pod "downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.072997211s | |
Jul 9 19:09:22.168: INFO: Pod "downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.103381227s | |
STEP: Saw pod success | |
Jul 9 19:09:22.168: INFO: Pod "downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:09:22.206: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276 container dapi-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:09:22.277: INFO: Waiting for pod downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:09:22.308: INFO: Pod downward-api-405b7c12-83e6-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:22.308: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-94pzx" for this suite. | |
Jul 9 19:09:28.434: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:30.064: INFO: namespace: e2e-tests-downward-api-94pzx, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:31.863: INFO: namespace e2e-tests-downward-api-94pzx deletion completed in 9.523567764s | |
• [SLOW TEST:16.195 seconds] | |
[sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downward_api.go:37 | |
should provide default limits.cpu/memory from node allocatable [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] ConfigMap | |
binary data should be reflected in volume [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:187 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:07.027: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:08.916: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-wcrw4 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] binary data should be reflected in volume [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:187 | |
Jul 9 19:09:09.671: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node | |
STEP: Creating configMap with name configmap-test-upd-3b60bc1b-83e6-11e8-992b-28d244b00276 | |
STEP: Creating the pod | |
STEP: Waiting for pod with text data | |
STEP: Waiting for pod with binary data | |
[AfterEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:13.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-configmap-wcrw4" for this suite. | |
Jul 9 19:09:36.135: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:37.910: INFO: namespace: e2e-tests-configmap-wcrw4, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:40.347: INFO: namespace e2e-tests-configmap-wcrw4 deletion completed in 26.324645331s | |
• [SLOW TEST:33.320 seconds] | |
[sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 | |
binary data should be reflected in volume [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:187 | |
------------------------------ | |
SS | |
------------------------------ | |
[sig-storage] EmptyDir volumes when FSGroup is specified | |
files with FSGroup ownership should support (root,0644,tmpfs) [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:57 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:31.864: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:33.500: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-q75bf | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] files with FSGroup ownership should support (root,0644,tmpfs) [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:57 | |
STEP: Creating a pod to test emptydir 0644 on tmpfs | |
Jul 9 19:09:34.110: INFO: Waiting up to 5m0s for pod "pod-49ebef96-83e6-11e8-bd2e-28d244b00276" in namespace "e2e-tests-emptydir-q75bf" to be "success or failure" | |
Jul 9 19:09:34.141: INFO: Pod "pod-49ebef96-83e6-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 31.003318ms | |
Jul 9 19:09:36.172: INFO: Pod "pod-49ebef96-83e6-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.062287969s | |
STEP: Saw pod success | |
Jul 9 19:09:36.172: INFO: Pod "pod-49ebef96-83e6-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:09:36.206: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-49ebef96-83e6-11e8-bd2e-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:09:36.273: INFO: Waiting for pod pod-49ebef96-83e6-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:09:36.303: INFO: Pod pod-49ebef96-83e6-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:36.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-q75bf" for this suite. | |
Jul 9 19:09:42.463: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:44.883: INFO: namespace: e2e-tests-emptydir-q75bf, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:45.978: INFO: namespace e2e-tests-emptydir-q75bf deletion completed in 9.628539544s | |
• [SLOW TEST:14.114 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
when FSGroup is specified | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:44 | |
files with FSGroup ownership should support (root,0644,tmpfs) [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:57 | |
------------------------------ | |
[k8s.io] Pods | |
should be updated [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:16.783: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:18.406: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-vvlc8 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/pods.go:127 | |
[It] should be updated [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: creating the pod | |
STEP: submitting the pod to kubernetes | |
STEP: verifying the pod is in kubernetes | |
STEP: updating the pod | |
Jul 9 19:09:21.782: INFO: Successfully updated pod "pod-update-40eff97e-83e6-11e8-8401-28d244b00276" | |
STEP: verifying the updated pod is in kubernetes | |
Jul 9 19:09:21.840: INFO: Pod update OK | |
[AfterEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:21.840: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pods-vvlc8" for this suite. | |
Jul 9 19:09:44.044: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:09:46.796: INFO: namespace: e2e-tests-pods-vvlc8, resource: bindings, ignored listing per whitelist | |
Jul 9 19:09:47.464: INFO: namespace e2e-tests-pods-vvlc8 deletion completed in 25.509105364s | |
• [SLOW TEST:30.681 seconds] | |
[k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should be updated [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:Builds][Conformance] oc new-app | |
should fail with a --name longer than 58 characters [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:66 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:40.350: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:09:42.725: INFO: configPath is now "/tmp/e2e-test-new-app-dk5fm-user.kubeconfig" | |
Jul 9 19:09:42.725: INFO: The user is now "e2e-test-new-app-dk5fm-user" | |
Jul 9 19:09:42.725: INFO: Creating project "e2e-test-new-app-dk5fm" | |
Jul 9 19:09:42.839: INFO: Waiting on permissions in project "e2e-test-new-app-dk5fm" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:26 | |
Jul 9 19:09:42.921: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:30 | |
STEP: waiting for builder service account | |
STEP: waiting for openshift namespace imagestreams | |
Jul 9 19:09:43.049: INFO: Running scan #0 | |
Jul 9 19:09:43.049: INFO: Checking language ruby | |
Jul 9 19:09:43.100: INFO: Checking tag 2.0 | |
Jul 9 19:09:43.100: INFO: Checking tag 2.2 | |
Jul 9 19:09:43.100: INFO: Checking tag 2.3 | |
Jul 9 19:09:43.100: INFO: Checking tag 2.4 | |
Jul 9 19:09:43.100: INFO: Checking tag 2.5 | |
Jul 9 19:09:43.100: INFO: Checking tag latest | |
Jul 9 19:09:43.100: INFO: Checking language nodejs | |
Jul 9 19:09:43.138: INFO: Checking tag 0.10 | |
Jul 9 19:09:43.138: INFO: Checking tag 4 | |
Jul 9 19:09:43.138: INFO: Checking tag 6 | |
Jul 9 19:09:43.138: INFO: Checking tag 8 | |
Jul 9 19:09:43.138: INFO: Checking tag latest | |
Jul 9 19:09:43.138: INFO: Checking language perl | |
Jul 9 19:09:43.171: INFO: Checking tag 5.16 | |
Jul 9 19:09:43.171: INFO: Checking tag 5.20 | |
Jul 9 19:09:43.171: INFO: Checking tag 5.24 | |
Jul 9 19:09:43.171: INFO: Checking tag latest | |
Jul 9 19:09:43.171: INFO: Checking language php | |
Jul 9 19:09:43.204: INFO: Checking tag 5.6 | |
Jul 9 19:09:43.204: INFO: Checking tag 7.0 | |
Jul 9 19:09:43.204: INFO: Checking tag 7.1 | |
Jul 9 19:09:43.204: INFO: Checking tag latest | |
Jul 9 19:09:43.204: INFO: Checking tag 5.5 | |
Jul 9 19:09:43.204: INFO: Checking language python | |
Jul 9 19:09:43.238: INFO: Checking tag latest | |
Jul 9 19:09:43.238: INFO: Checking tag 2.7 | |
Jul 9 19:09:43.238: INFO: Checking tag 3.3 | |
Jul 9 19:09:43.238: INFO: Checking tag 3.4 | |
Jul 9 19:09:43.238: INFO: Checking tag 3.5 | |
Jul 9 19:09:43.238: INFO: Checking tag 3.6 | |
Jul 9 19:09:43.238: INFO: Checking language wildfly | |
Jul 9 19:09:43.272: INFO: Checking tag latest | |
Jul 9 19:09:43.272: INFO: Checking tag 10.0 | |
Jul 9 19:09:43.272: INFO: Checking tag 10.1 | |
Jul 9 19:09:43.272: INFO: Checking tag 11.0 | |
Jul 9 19:09:43.272: INFO: Checking tag 12.0 | |
Jul 9 19:09:43.272: INFO: Checking tag 8.1 | |
Jul 9 19:09:43.272: INFO: Checking tag 9.0 | |
Jul 9 19:09:43.272: INFO: Checking language mysql | |
Jul 9 19:09:43.303: INFO: Checking tag 5.5 | |
Jul 9 19:09:43.303: INFO: Checking tag 5.6 | |
Jul 9 19:09:43.303: INFO: Checking tag 5.7 | |
Jul 9 19:09:43.303: INFO: Checking tag latest | |
Jul 9 19:09:43.303: INFO: Checking language postgresql | |
Jul 9 19:09:43.341: INFO: Checking tag 9.5 | |
Jul 9 19:09:43.341: INFO: Checking tag 9.6 | |
Jul 9 19:09:43.341: INFO: Checking tag latest | |
Jul 9 19:09:43.341: INFO: Checking tag 9.2 | |
Jul 9 19:09:43.341: INFO: Checking tag 9.4 | |
Jul 9 19:09:43.341: INFO: Checking language mongodb | |
Jul 9 19:09:43.382: INFO: Checking tag 3.4 | |
Jul 9 19:09:43.382: INFO: Checking tag latest | |
Jul 9 19:09:43.382: INFO: Checking tag 2.4 | |
Jul 9 19:09:43.382: INFO: Checking tag 2.6 | |
Jul 9 19:09:43.382: INFO: Checking tag 3.2 | |
Jul 9 19:09:43.382: INFO: Checking language jenkins | |
Jul 9 19:09:43.417: INFO: Checking tag 1 | |
Jul 9 19:09:43.417: INFO: Checking tag 2 | |
Jul 9 19:09:43.417: INFO: Checking tag latest | |
Jul 9 19:09:43.417: INFO: Success! | |
[It] should fail with a --name longer than 58 characters [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:66 | |
STEP: calling oc new-app | |
Jul 9 19:09:43.417: INFO: Running 'oc new-app --config=/tmp/e2e-test-new-app-dk5fm-user.kubeconfig --namespace=e2e-test-new-app-dk5fm https://github.com/openshift/nodejs-ex --name a2345678901234567890123456789012345678901234567890123456789' | |
Jul 9 19:09:46.048: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc new-app --config=/tmp/e2e-test-new-app-dk5fm-user.kubeconfig --namespace=e2e-test-new-app-dk5fm https://github.com/openshift/nodejs-ex --name a2345678901234567890123456789012345678901234567890123456789] [] error: invalid name: a2345678901234567890123456789012345678901234567890123456789. Must be an a lower case alphanumeric (a-z, and 0-9) string with a maximum length of 58 characters, where the first character is a letter (a-z), and the '-' character is allowed anywhere except the first or last character. | |
error: invalid name: a2345678901234567890123456789012345678901234567890123456789. Must be an a lower case alphanumeric (a-z, and 0-9) string with a maximum length of 58 characters, where the first character is a letter (a-z), and the '-' character is allowed anywhere except the first or last character. | |
[] <nil> 0xc42105f740 exit status 1 <nil> <nil> true [0xc4200dc888 0xc4200dc8f8 0xc4200dc8f8] [0xc4200dc888 0xc4200dc8f8] [0xc4200dc890 0xc4200dc8f0] [0x916090 0x916190] 0xc4214ecd80 <nil>}: | |
error: invalid name: a2345678901234567890123456789012345678901234567890123456789. Must be an a lower case alphanumeric (a-z, and 0-9) string with a maximum length of 58 characters, where the first character is a letter (a-z), and the '-' character is allowed anywhere except the first or last character. | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:40 | |
[AfterEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:09:46.239: INFO: namespace : e2e-test-new-app-dk5fm api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:52.283: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:11.979 seconds] | |
[Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:16 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:24 | |
should fail with a --name longer than 58 characters [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:66 | |
------------------------------ | |
[sig-storage] ConfigMap | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:54 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:47.465: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:48.936: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-llm69 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:54 | |
STEP: Creating configMap with name configmap-test-volume-532239eb-83e6-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:09:49.596: INFO: Waiting up to 5m0s for pod "pod-configmaps-5326fb5a-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-configmap-llm69" to be "success or failure" | |
Jul 9 19:09:49.630: INFO: Pod "pod-configmaps-5326fb5a-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 33.792324ms | |
Jul 9 19:09:51.659: INFO: Pod "pod-configmaps-5326fb5a-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.062907306s | |
STEP: Saw pod success | |
Jul 9 19:09:51.659: INFO: Pod "pod-configmaps-5326fb5a-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:09:51.698: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-configmaps-5326fb5a-83e6-11e8-8401-28d244b00276 container configmap-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:09:51.762: INFO: Waiting for pod pod-configmaps-5326fb5a-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:09:51.799: INFO: Pod pod-configmaps-5326fb5a-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:51.800: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-configmap-llm69" for this suite. | |
Jul 9 19:09:57.927: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:00.793: INFO: namespace: e2e-tests-configmap-llm69, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:01.310: INFO: namespace e2e-tests-configmap-llm69 deletion completed in 9.47277623s | |
• [SLOW TEST:13.845 seconds] | |
[sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:54 | |
------------------------------ | |
S | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using one of the plugins 'redhat/openshift-ovs-multitenant, redhat/openshift-ovs-networkpolicy' | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:444 | |
Jul 9 19:10:01.576: INFO: Could not check network plugin name: exit status 1. Assuming a non-OpenShift plugin | |
Jul 9 19:10:01.576: INFO: Not using one of the specified plugins | |
[AfterEach] when using one of the plugins 'redhat/openshift-ovs-multitenant, redhat/openshift-ovs-networkpolicy' | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
[AfterEach] when using one of the plugins 'redhat/openshift-ovs-multitenant, redhat/openshift-ovs-networkpolicy' | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:01.576: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.265 seconds] | |
[Area:Networking] multicast | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/multicast.go:21 | |
when using one of the plugins 'redhat/openshift-ovs-multitenant, redhat/openshift-ovs-networkpolicy' | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:442 | |
should allow multicast traffic in namespaces where it is enabled [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/multicast.go:45 | |
Jul 9 19:10:01.576: Not using one of the specified plugins | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
[k8s.io] InitContainer | |
should not start app containers if init containers fail on a RestartAlways pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:166 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:08:44.845: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:08:47.331: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
Jul 9 19:08:48.100: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. | |
Jul 9 19:08:48.285: INFO: Found ClusterRoles; assuming RBAC is enabled. | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-vc5jv | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:40 | |
[It] should not start app containers if init containers fail on a RestartAlways pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:166 | |
STEP: creating the pod | |
Jul 9 19:08:48.568: INFO: PodSpec: initContainers in spec.initContainers | |
Jul 9 19:09:43.506: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-2ec7c3dc-83e6-11e8-8fe2-28d244b00276", GenerateName:"", Namespace:"e2e-tests-init-container-vc5jv", SelfLink:"/api/v1/namespaces/e2e-tests-init-container-vc5jv/pods/pod-init-2ec7c3dc-83e6-11e8-8fe2-28d244b00276", UID:"2edb82bb-83e6-11e8-84c6-0af96768d57e", ResourceVersion:"69577", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666785328, loc:(*time.Location)(0x6b11480)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"536063916"}, Annotations:map[string]string{"openshift.io/scc":"anyuid"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-2dvtq", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc4211d0e00), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"busybox", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-2dvtq", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"Always", SecurityContext:(*v1.SecurityContext)(0xc4211d0f80), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"busybox", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-2dvtq", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"Always", SecurityContext:(*v1.SecurityContext)(0xc4211d1000), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"k8s.gcr.io/pause-amd64:3.1", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:31457280, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"31457280", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}, "memory":resource.Quantity{i:resource.int64Amount{value:31457280, scale:0}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"31457280", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-2dvtq", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc4211d0e80), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc4214a4a48), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"ip-10-0-130-54.us-west-2.compute.internal", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc4211d0ec0), ImagePullSecrets:[]v1.LocalObjectReference{v1.LocalObjectReference{Name:"default-dockercfg-7gp5s"}}, Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/memory-pressure", Operator:"Exists", Value:"", Effect:"NoSchedule", TolerationSeconds:(*int64)(nil)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(nil), DNSConfig:(*v1.PodDNSConfig)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666785328, loc:(*time.Location)(0x6b11480)}}, Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666785328, loc:(*time.Location)(0x6b11480)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666785328, loc:(*time.Location)(0x6b11480)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.0.130.54", PodIP:"10.2.2.191", StartTime:(*v1.Time)(0xc4217b30e0), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc42045c700)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc42045c770)}, Ready:false, RestartCount:3, Image:"busybox:latest", ImageID:"docker-pullable://busybox@sha256:141c253bc4c3fd0a201d32dc1f493bcf3fff003b6df416dea4f41046e0f37d47", ContainerID:"docker://aad60e5897b9c2cfbf55c49ff779c85feaef0893f60edb22c46ab15ad8fae41a"}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc4217b3120), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"busybox", ImageID:"", ContainerID:""}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc4217b3100), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"k8s.gcr.io/pause-amd64:3.1", ImageID:"", ContainerID:""}}, QOSClass:"Guaranteed"}} | |
[AfterEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:09:43.507: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-init-container-vc5jv" for this suite. | |
Jul 9 19:10:05.682: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:08.083: INFO: namespace: e2e-tests-init-container-vc5jv, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:09.484: INFO: namespace e2e-tests-init-container-vc5jv deletion completed in 25.913101336s | |
• [SLOW TEST:84.639 seconds] | |
[k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should not start app containers if init containers fail on a RestartAlways pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:166 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:01.581: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:10:03.428: INFO: configPath is now "/tmp/e2e-test-router-stress-p27nt-user.kubeconfig" | |
Jul 9 19:10:03.428: INFO: The user is now "e2e-test-router-stress-p27nt-user" | |
Jul 9 19:10:03.428: INFO: Creating project "e2e-test-router-stress-p27nt" | |
Jul 9 19:10:03.577: INFO: Waiting on permissions in project "e2e-test-router-stress-p27nt" ... | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:45 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:32 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:10:03.719: INFO: namespace : e2e-test-router-stress-p27nt api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:09.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [8.212 seconds] | |
[Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:21 | |
The HAProxy router [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:68 | |
should respond with 503 to unrecognized hosts [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:69 | |
no router installed on the cluster | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:48 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-api-machinery] ConfigMap | |
should be consumable via the environment [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-api-machinery] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:09.795: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:10:11.209: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-m2zlr | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable via the environment [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating configMap e2e-tests-configmap-m2zlr/configmap-test-606abb83-83e6-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:10:11.880: INFO: Waiting up to 5m0s for pod "pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-configmap-m2zlr" to be "success or failure" | |
Jul 9 19:10:11.909: INFO: Pod "pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 29.02209ms | |
Jul 9 19:10:13.938: INFO: Pod "pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.058136545s | |
Jul 9 19:10:15.968: INFO: Pod "pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 4.087813715s | |
Jul 9 19:10:17.997: INFO: Pod "pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.117318537s | |
STEP: Saw pod success | |
Jul 9 19:10:17.997: INFO: Pod "pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:10:18.025: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276 container env-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:10:18.098: INFO: Waiting for pod pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:10:18.127: INFO: Pod pod-configmaps-606f137b-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-api-machinery] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:18.127: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-configmap-m2zlr" for this suite. | |
Jul 9 19:10:24.264: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:26.021: INFO: namespace: e2e-tests-configmap-m2zlr, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:27.719: INFO: namespace e2e-tests-configmap-m2zlr deletion completed in 9.559471834s | |
• [SLOW TEST:17.924 seconds] | |
[sig-api-machinery] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap.go:29 | |
should be consumable via the environment [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-network] Networking Granular Checks: Pods | |
should function for node-pod communication: udp [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:52.331: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:54.259: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-2hdld | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should function for node-pod communication: udp [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-2hdld | |
STEP: creating a selector | |
STEP: Creating the service pods in kubernetes | |
Jul 9 19:09:55.154: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable | |
STEP: Creating test pods | |
Jul 9 19:10:11.775: INFO: ExecWithOptions {Command:[/bin/sh -c echo 'hostName' | timeout -t 2 nc -w 1 -u 10.2.2.207 8081 | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-2hdld PodName:host-test-container-pod ContainerName:hostexec Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} | |
Jul 9 19:10:11.775: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:10:13.217: INFO: Found all expected endpoints: [netserver-0] | |
[AfterEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:13.217: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pod-network-test-2hdld" for this suite. | |
Jul 9 19:10:35.383: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:39.766: INFO: namespace: e2e-tests-pod-network-test-2hdld, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:39.884: INFO: namespace e2e-tests-pod-network-test-2hdld deletion completed in 26.620108146s | |
• [SLOW TEST:47.553 seconds] | |
[sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:25 | |
Granular Checks: Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:28 | |
should function for node-pod communication: udp [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SSSS | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:422 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:27.721: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:10:29.267: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-h97lg | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:422 | |
STEP: Creating configMap with name projected-configmap-test-volume-6b2667d6-83e6-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:10:29.896: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6b2b5fb4-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-projected-h97lg" to be "success or failure" | |
Jul 9 19:10:29.927: INFO: Pod "pod-projected-configmaps-6b2b5fb4-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.894207ms | |
Jul 9 19:10:31.958: INFO: Pod "pod-projected-configmaps-6b2b5fb4-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.06175169s | |
STEP: Saw pod success | |
Jul 9 19:10:31.958: INFO: Pod "pod-projected-configmaps-6b2b5fb4-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:10:31.984: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-configmaps-6b2b5fb4-83e6-11e8-8401-28d244b00276 container projected-configmap-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:10:32.059: INFO: Waiting for pod pod-projected-configmaps-6b2b5fb4-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:10:32.089: INFO: Pod pod-projected-configmaps-6b2b5fb4-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:32.089: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-h97lg" for this suite. | |
Jul 9 19:10:38.220: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:41.402: INFO: namespace: e2e-tests-projected-h97lg, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:41.732: INFO: namespace e2e-tests-projected-h97lg deletion completed in 9.608230757s | |
• [SLOW TEST:14.011 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:422 | |
------------------------------ | |
[sig-storage] Projected | |
updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:08:44.846: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:08:48.929: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
Jul 9 19:08:49.775: INFO: Found PodSecurityPolicies; assuming PodSecurityPolicy is enabled. | |
Jul 9 19:08:50.089: INFO: Found ClusterRoles; assuming RBAC is enabled. | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-b9w9s | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
Jul 9 19:08:50.558: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node | |
STEP: Creating projection with configMap that has name projected-configmap-test-upd-2ffc6e90-83e6-11e8-881a-28d244b00276 | |
STEP: Creating the pod | |
STEP: Updating configmap projected-configmap-test-upd-2ffc6e90-83e6-11e8-881a-28d244b00276 | |
STEP: waiting to observe update in volume | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:21.631: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-b9w9s" for this suite. | |
Jul 9 19:10:43.925: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:46.903: INFO: namespace: e2e-tests-projected-b9w9s, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:48.827: INFO: namespace e2e-tests-projected-b9w9s deletion completed in 27.040637991s | |
• [SLOW TEST:123.981 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable in multiple volumes in the same pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:39.890: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:10:42.001: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-dw7b2 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable in multiple volumes in the same pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating configMap with name projected-configmap-test-volume-72d724e5-83e6-11e8-992b-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:10:42.805: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-72dcf3bd-83e6-11e8-992b-28d244b00276" in namespace "e2e-tests-projected-dw7b2" to be "success or failure" | |
Jul 9 19:10:42.846: INFO: Pod "pod-projected-configmaps-72dcf3bd-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 40.524707ms | |
Jul 9 19:10:44.883: INFO: Pod "pod-projected-configmaps-72dcf3bd-83e6-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.077334435s | |
STEP: Saw pod success | |
Jul 9 19:10:44.883: INFO: Pod "pod-projected-configmaps-72dcf3bd-83e6-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:10:44.933: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-configmaps-72dcf3bd-83e6-11e8-992b-28d244b00276 container projected-configmap-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:10:45.039: INFO: Waiting for pod pod-projected-configmaps-72dcf3bd-83e6-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:10:45.077: INFO: Pod pod-projected-configmaps-72dcf3bd-83e6-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:45.077: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-dw7b2" for this suite. | |
Jul 9 19:10:51.236: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:55.279: INFO: namespace: e2e-tests-projected-dw7b2, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:55.603: INFO: namespace e2e-tests-projected-dw7b2 deletion completed in 10.483606757s | |
• [SLOW TEST:15.713 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable in multiple volumes in the same pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Downward API volume | |
should set mode on item file [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:41.733: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:10:43.292: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-t27sb | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should set mode on item file [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:10:43.923: INFO: Waiting up to 5m0s for pod "downwardapi-volume-73875f2c-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-downward-api-t27sb" to be "success or failure" | |
Jul 9 19:10:43.955: INFO: Pod "downwardapi-volume-73875f2c-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 31.758628ms | |
Jul 9 19:10:45.984: INFO: Pod "downwardapi-volume-73875f2c-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.060640477s | |
STEP: Saw pod success | |
Jul 9 19:10:45.984: INFO: Pod "downwardapi-volume-73875f2c-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:10:46.012: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-73875f2c-83e6-11e8-8401-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:10:46.078: INFO: Waiting for pod downwardapi-volume-73875f2c-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:10:46.106: INFO: Pod downwardapi-volume-73875f2c-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:46.106: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-t27sb" for this suite. | |
Jul 9 19:10:52.245: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:10:55.554: INFO: namespace: e2e-tests-downward-api-t27sb, resource: bindings, ignored listing per whitelist | |
Jul 9 19:10:55.783: INFO: namespace e2e-tests-downward-api-t27sb deletion completed in 9.633243258s | |
• [SLOW TEST:14.050 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should set mode on item file [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:419 | |
Jul 9 19:10:55.784: INFO: This plugin does not isolate namespaces by default. | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:55.784: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:55.785: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] | |
[Area:Networking] services | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/services.go:10 | |
when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:418 | |
should allow connections to services in the default namespace from a pod in another namespace on a different node [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/services.go:52 | |
Jul 9 19:10:55.784: This plugin does not isolate namespaces by default. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
SS | |
------------------------------ | |
[Feature:Builds][pruning] prune builds based on settings in the buildconfig | |
[Conformance] buildconfigs should have a default history limit set when created via the group api [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:294 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][pruning] prune builds based on settings in the buildconfig | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:55.604: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][pruning] prune builds based on settings in the buildconfig | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:10:57.827: INFO: configPath is now "/tmp/e2e-test-build-pruning-hptxt-user.kubeconfig" | |
Jul 9 19:10:57.827: INFO: The user is now "e2e-test-build-pruning-hptxt-user" | |
Jul 9 19:10:57.827: INFO: Creating project "e2e-test-build-pruning-hptxt" | |
Jul 9 19:10:57.977: INFO: Waiting on permissions in project "e2e-test-build-pruning-hptxt" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:37 | |
Jul 9 19:10:58.038: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:41 | |
STEP: waiting for builder service account | |
STEP: waiting for openshift namespace imagestreams | |
Jul 9 19:10:58.171: INFO: Running scan #0 | |
Jul 9 19:10:58.171: INFO: Checking language ruby | |
Jul 9 19:10:58.202: INFO: Checking tag 2.0 | |
Jul 9 19:10:58.202: INFO: Checking tag 2.2 | |
Jul 9 19:10:58.202: INFO: Checking tag 2.3 | |
Jul 9 19:10:58.202: INFO: Checking tag 2.4 | |
Jul 9 19:10:58.202: INFO: Checking tag 2.5 | |
Jul 9 19:10:58.202: INFO: Checking tag latest | |
Jul 9 19:10:58.202: INFO: Checking language nodejs | |
Jul 9 19:10:58.241: INFO: Checking tag 0.10 | |
Jul 9 19:10:58.241: INFO: Checking tag 4 | |
Jul 9 19:10:58.241: INFO: Checking tag 6 | |
Jul 9 19:10:58.241: INFO: Checking tag 8 | |
Jul 9 19:10:58.241: INFO: Checking tag latest | |
Jul 9 19:10:58.241: INFO: Checking language perl | |
Jul 9 19:10:58.280: INFO: Checking tag 5.16 | |
Jul 9 19:10:58.280: INFO: Checking tag 5.20 | |
Jul 9 19:10:58.280: INFO: Checking tag 5.24 | |
Jul 9 19:10:58.280: INFO: Checking tag latest | |
Jul 9 19:10:58.280: INFO: Checking language php | |
Jul 9 19:10:58.318: INFO: Checking tag latest | |
Jul 9 19:10:58.318: INFO: Checking tag 5.5 | |
Jul 9 19:10:58.318: INFO: Checking tag 5.6 | |
Jul 9 19:10:58.318: INFO: Checking tag 7.0 | |
Jul 9 19:10:58.318: INFO: Checking tag 7.1 | |
Jul 9 19:10:58.318: INFO: Checking language python | |
Jul 9 19:10:58.375: INFO: Checking tag 2.7 | |
Jul 9 19:10:58.375: INFO: Checking tag 3.3 | |
Jul 9 19:10:58.375: INFO: Checking tag 3.4 | |
Jul 9 19:10:58.375: INFO: Checking tag 3.5 | |
Jul 9 19:10:58.375: INFO: Checking tag 3.6 | |
Jul 9 19:10:58.375: INFO: Checking tag latest | |
Jul 9 19:10:58.375: INFO: Checking language wildfly | |
Jul 9 19:10:58.405: INFO: Checking tag 11.0 | |
Jul 9 19:10:58.405: INFO: Checking tag 12.0 | |
Jul 9 19:10:58.405: INFO: Checking tag 8.1 | |
Jul 9 19:10:58.405: INFO: Checking tag 9.0 | |
Jul 9 19:10:58.405: INFO: Checking tag latest | |
Jul 9 19:10:58.405: INFO: Checking tag 10.0 | |
Jul 9 19:10:58.405: INFO: Checking tag 10.1 | |
Jul 9 19:10:58.405: INFO: Checking language mysql | |
Jul 9 19:10:58.444: INFO: Checking tag 5.5 | |
Jul 9 19:10:58.444: INFO: Checking tag 5.6 | |
Jul 9 19:10:58.444: INFO: Checking tag 5.7 | |
Jul 9 19:10:58.444: INFO: Checking tag latest | |
Jul 9 19:10:58.444: INFO: Checking language postgresql | |
Jul 9 19:10:58.476: INFO: Checking tag 9.5 | |
Jul 9 19:10:58.476: INFO: Checking tag 9.6 | |
Jul 9 19:10:58.476: INFO: Checking tag latest | |
Jul 9 19:10:58.476: INFO: Checking tag 9.2 | |
Jul 9 19:10:58.476: INFO: Checking tag 9.4 | |
Jul 9 19:10:58.476: INFO: Checking language mongodb | |
Jul 9 19:10:58.508: INFO: Checking tag 2.4 | |
Jul 9 19:10:58.508: INFO: Checking tag 2.6 | |
Jul 9 19:10:58.508: INFO: Checking tag 3.2 | |
Jul 9 19:10:58.508: INFO: Checking tag 3.4 | |
Jul 9 19:10:58.508: INFO: Checking tag latest | |
Jul 9 19:10:58.508: INFO: Checking language jenkins | |
Jul 9 19:10:58.547: INFO: Checking tag 1 | |
Jul 9 19:10:58.547: INFO: Checking tag 2 | |
Jul 9 19:10:58.547: INFO: Checking tag latest | |
Jul 9 19:10:58.547: INFO: Success! | |
STEP: creating test image stream | |
Jul 9 19:10:58.547: INFO: Running 'oc create --config=/tmp/e2e-test-build-pruning-hptxt-user.kubeconfig --namespace=e2e-test-build-pruning-hptxt -f /tmp/fixture-testdata-dir877664294/test/extended/testdata/builds/build-pruning/imagestream.yaml' | |
imagestream.image.openshift.io "myphp" created | |
[It] [Conformance] buildconfigs should have a default history limit set when created via the group api [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:294 | |
STEP: creating a build config with the group api | |
Jul 9 19:10:58.824: INFO: Running 'oc create --config=/tmp/e2e-test-build-pruning-hptxt-user.kubeconfig --namespace=e2e-test-build-pruning-hptxt -f /tmp/fixture-testdata-dir877664294/test/extended/testdata/builds/build-pruning/default-group-build-config.yaml' | |
buildconfig.build.openshift.io "myphp" created | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:56 | |
[AfterEach] [Feature:Builds][pruning] prune builds based on settings in the buildconfig | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:10:59.264: INFO: namespace : e2e-test-build-pruning-hptxt api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][pruning] prune builds based on settings in the buildconfig | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:05.305: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:9.748 seconds] | |
[Feature:Builds][pruning] prune builds based on settings in the buildconfig | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:21 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:35 | |
[Conformance] buildconfigs should have a default history limit set when created via the group api [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/build_pruning.go:294 | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs | |
should adhere to Three Laws of Controllers [Conformance] [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1137 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:09.487: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:10:11.378: INFO: configPath is now "/tmp/e2e-test-cli-deployment-kj6d8-user.kubeconfig" | |
Jul 9 19:10:11.378: INFO: The user is now "e2e-test-cli-deployment-kj6d8-user" | |
Jul 9 19:10:11.378: INFO: Creating project "e2e-test-cli-deployment-kj6d8" | |
Jul 9 19:10:11.495: INFO: Waiting on permissions in project "e2e-test-cli-deployment-kj6d8" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should adhere to Three Laws of Controllers [Conformance] [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1137 | |
STEP: should create ControllerRef in RCs it creates | |
Jul 9 19:10:24.708: INFO: Latest rollout of dc/deployment-simple (rc/deployment-simple-1) is complete. | |
STEP: releasing RCs that no longer match its selector | |
STEP: adopting RCs that match its selector and have no ControllerRef | |
STEP: deleting owned RCs when deleted | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1132 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:11:00.476: INFO: namespace : e2e-test-cli-deployment-kj6d8 api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:06.509: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:57.069 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1130 | |
should adhere to Three Laws of Controllers [Conformance] [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1137 | |
------------------------------ | |
S | |
------------------------------ | |
[k8s.io] Probing container | |
with readiness probe that fails should never be ready and never restart [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Probing container | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:09:45.979: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:09:47.706: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-container-probe-s5hdx | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] Probing container | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/container_probe.go:48 | |
[It] with readiness probe that fails should never be ready and never restart [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[AfterEach] [k8s.io] Probing container | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:10:48.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-container-probe-s5hdx" for this suite. | |
Jul 9 19:11:10.531: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:11:13.589: INFO: namespace: e2e-tests-container-probe-s5hdx, resource: bindings, ignored listing per whitelist | |
Jul 9 19:11:13.871: INFO: namespace e2e-tests-container-probe-s5hdx deletion completed in 25.444058596s | |
• [SLOW TEST:87.892 seconds] | |
[k8s.io] Probing container | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
with readiness probe that fails should never be ready and never restart [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] Downward API volume | |
should provide container's memory limit [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:06.559: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:08.279: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-snx9g | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should provide container's memory limit [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:11:08.898: INFO: Waiting up to 5m0s for pod "downwardapi-volume-826a7dd9-83e6-11e8-8fe2-28d244b00276" in namespace "e2e-tests-downward-api-snx9g" to be "success or failure" | |
Jul 9 19:11:08.932: INFO: Pod "downwardapi-volume-826a7dd9-83e6-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 33.604579ms | |
Jul 9 19:11:10.971: INFO: Pod "downwardapi-volume-826a7dd9-83e6-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.073107651s | |
STEP: Saw pod success | |
Jul 9 19:11:10.971: INFO: Pod "downwardapi-volume-826a7dd9-83e6-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:11:11.002: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-826a7dd9-83e6-11e8-8fe2-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:11:11.087: INFO: Waiting for pod downwardapi-volume-826a7dd9-83e6-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:11:11.118: INFO: Pod downwardapi-volume-826a7dd9-83e6-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:11.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-snx9g" for this suite. | |
Jul 9 19:11:17.270: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:11:19.183: INFO: namespace: e2e-tests-downward-api-snx9g, resource: bindings, ignored listing per whitelist | |
Jul 9 19:11:21.143: INFO: namespace e2e-tests-downward-api-snx9g deletion completed in 9.977340991s | |
• [SLOW TEST:14.583 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should provide container's memory limit [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[k8s.io] Variable Expansion | |
should allow substituting values in a container's command [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:05.353: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:07.324: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-7r8ws | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should allow substituting values in a container's command [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test substitution in container's command | |
Jul 9 19:11:08.163: INFO: Waiting up to 5m0s for pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276" in namespace "e2e-tests-var-expansion-7r8ws" to be "success or failure" | |
Jul 9 19:11:08.214: INFO: Pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 50.78496ms | |
Jul 9 19:11:10.317: INFO: Pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.153374508s | |
Jul 9 19:11:12.360: INFO: Pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 4.196448037s | |
Jul 9 19:11:14.398: INFO: Pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 6.234583462s | |
Jul 9 19:11:16.495: INFO: Pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 8.331559739s | |
Jul 9 19:11:18.535: INFO: Pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.371487095s | |
STEP: Saw pod success | |
Jul 9 19:11:18.535: INFO: Pod "var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:11:18.572: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276 container dapi-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:11:18.662: INFO: Waiting for pod var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:11:18.698: INFO: Pod var-expansion-81f9e7f4-83e6-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:18.698: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-var-expansion-7r8ws" for this suite. | |
Jul 9 19:11:24.868: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:11:28.792: INFO: namespace: e2e-tests-var-expansion-7r8ws, resource: bindings, ignored listing per whitelist | |
Jul 9 19:11:29.280: INFO: namespace e2e-tests-var-expansion-7r8ws deletion completed in 10.540215906s | |
• [SLOW TEST:23.927 seconds] | |
[k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should allow substituting values in a container's command [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SSSS | |
------------------------------ | |
[Conformance][Area:Networking][Feature:Router] The HAProxy router | |
should override the route host with a custom value [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:109 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:55.787: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:10:57.402: INFO: configPath is now "/tmp/e2e-test-router-scoped-smdsm-user.kubeconfig" | |
Jul 9 19:10:57.402: INFO: The user is now "e2e-test-router-scoped-smdsm-user" | |
Jul 9 19:10:57.402: INFO: Creating project "e2e-test-router-scoped-smdsm" | |
Jul 9 19:10:57.610: INFO: Waiting on permissions in project "e2e-test-router-scoped-smdsm" ... | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:48 | |
Jul 9 19:10:57.705: INFO: Running 'oc new-app --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-router-scoped-smdsm -f /tmp/fixture-testdata-dir180677416/test/extended/testdata/scoped-router.yaml -p IMAGE=openshift/origin-haproxy-router' | |
--> Deploying template "e2e-test-router-scoped-smdsm/" for "/tmp/fixture-testdata-dir180677416/test/extended/testdata/scoped-router.yaml" to project e2e-test-router-scoped-smdsm | |
* With parameters: | |
* IMAGE=openshift/origin-haproxy-router | |
* SCOPE=["--name=test-scoped", "--namespace=$(POD_NAMESPACE)", "--loglevel=4", "--labels=select=first"] | |
--> Creating resources ... | |
pod "router-scoped" created | |
pod "router-override" created | |
pod "router-override-domains" created | |
rolebinding "system-router" created | |
route "route-1" created | |
route "route-2" created | |
route "route-override-domain-1" created | |
route "route-override-domain-2" created | |
service "endpoints" created | |
pod "endpoint-1" created | |
--> Success | |
Access your application via route 'first.example.com' | |
Access your application via route 'second.example.com' | |
Access your application via route 'y.a.null.ptr' | |
Access your application via route 'main.void.str' | |
Run 'oc status' to view your app. | |
[It] should override the route host with a custom value [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:109 | |
Jul 9 19:10:58.736: INFO: Creating new exec pod | |
STEP: creating a scoped router from a config file "/tmp/fixture-testdata-dir180677416/test/extended/testdata/scoped-router.yaml" | |
STEP: waiting for the healthz endpoint to respond | |
Jul 9 19:11:07.875: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-smdsm execpod -- /bin/sh -c | |
set -e | |
for i in $(seq 1 180); do | |
code=$( curl -k -s -o /dev/null -w '%{http_code}\n' --header 'Host: 10.2.2.224' "http://10.2.2.224:1936/healthz" ) || rc=$? | |
if [[ "${rc:-0}" -eq 0 ]]; then | |
echo $code | |
if [[ $code -eq 200 ]]; then | |
exit 0 | |
fi | |
if [[ $code -ne 503 ]]; then | |
exit 1 | |
fi | |
else | |
echo "error ${rc}" 1>&2 | |
fi | |
sleep 1 | |
done | |
' | |
Jul 9 19:11:08.612: INFO: stderr: "" | |
STEP: waiting for the valid route to respond | |
Jul 9 19:11:08.613: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-smdsm execpod -- /bin/sh -c | |
set -e | |
for i in $(seq 1 180); do | |
code=$( curl -k -s -o /dev/null -w '%{http_code}\n' --header 'Host: route-1-e2e-test-router-scoped-smdsm.myapps.mycompany.com' "http://10.2.2.224/Letter" ) || rc=$? | |
if [[ "${rc:-0}" -eq 0 ]]; then | |
echo $code | |
if [[ $code -eq 200 ]]; then | |
exit 0 | |
fi | |
if [[ $code -ne 503 ]]; then | |
exit 1 | |
fi | |
else | |
echo "error ${rc}" 1>&2 | |
fi | |
sleep 1 | |
done | |
' | |
Jul 9 19:11:15.466: INFO: stderr: "" | |
STEP: checking that the stored domain name does not match a route | |
Jul 9 19:11:15.466: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-smdsm execpod -- /bin/sh -c curl -s -o /dev/null -w '%{http_code}' --header 'Host: first.example.com' "http://10.2.2.224/Letter"' | |
Jul 9 19:11:16.104: INFO: stderr: "" | |
STEP: checking that route-1-e2e-test-router-scoped-smdsm.myapps.mycompany.com matches a route | |
Jul 9 19:11:16.104: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-smdsm execpod -- /bin/sh -c curl -s -o /dev/null -w '%{http_code}' --header 'Host: route-1-e2e-test-router-scoped-smdsm.myapps.mycompany.com' "http://10.2.2.224/Letter"' | |
Jul 9 19:11:16.822: INFO: stderr: "" | |
STEP: checking that route-2-e2e-test-router-scoped-smdsm.myapps.mycompany.com matches a route | |
Jul 9 19:11:16.822: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-smdsm execpod -- /bin/sh -c curl -s -o /dev/null -w '%{http_code}' --header 'Host: route-2-e2e-test-router-scoped-smdsm.myapps.mycompany.com' "http://10.2.2.224/Letter"' | |
Jul 9 19:11:17.497: INFO: stderr: "" | |
STEP: checking that the router reported the correct ingress and override | |
Jul 9 19:11:17.550: INFO: Selected: &route.RouteIngress{Host:"route-1-e2e-test-router-scoped-smdsm.myapps.mycompany.com", RouterName:"test-override", Conditions:[]route.RouteIngressCondition{route.RouteIngressCondition{Type:"Admitted", Status:"True", Reason:"", Message:"", LastTransitionTime:(*v1.Time)(0xc420ac0020)}}, WildcardPolicy:"None", RouterCanonicalHostname:""}, All: []route.RouteIngress{route.RouteIngress{Host:"first.example.com", RouterName:"router", Conditions:[]route.RouteIngressCondition{route.RouteIngressCondition{Type:"Admitted", Status:"True", Reason:"", Message:"", LastTransitionTime:(*v1.Time)(0xc421923c00)}}, WildcardPolicy:"None", RouterCanonicalHostname:""}, route.RouteIngress{Host:"first.example.com", RouterName:"test-override-domains", Conditions:[]route.RouteIngressCondition{route.RouteIngressCondition{Type:"Admitted", Status:"True", Reason:"", Message:"", LastTransitionTime:(*v1.Time)(0xc421923d60)}}, WildcardPolicy:"None", RouterCanonicalHostname:""}, route.RouteIngress{Host:"first.example.com", RouterName:"test-scoped", Conditions:[]route.RouteIngressCondition{route.RouteIngressCondition{Type:"Admitted", Status:"True", Reason:"", Message:"", LastTransitionTime:(*v1.Time)(0xc421923ec0)}}, WildcardPolicy:"None", RouterCanonicalHostname:""}, route.RouteIngress{Host:"route-1-e2e-test-router-scoped-smdsm.myapps.mycompany.com", RouterName:"test-override", Conditions:[]route.RouteIngressCondition{route.RouteIngressCondition{Type:"Admitted", Status:"True", Reason:"", Message:"", LastTransitionTime:(*v1.Time)(0xc420ac0020)}}, WildcardPolicy:"None", RouterCanonicalHostname:""}} | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:36 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:11:17.645: INFO: namespace : e2e-test-router-scoped-smdsm api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:29.678: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:33.943 seconds] | |
[Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:26 | |
The HAProxy router | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:67 | |
should override the route host with a custom value [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:109 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] Projected | |
should provide node allocatable (memory) as default memory limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:21.144: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:22.818: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-8qd2x | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should provide node allocatable (memory) as default memory limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:11:23.493: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276" in namespace "e2e-tests-projected-8qd2x" to be "success or failure" | |
Jul 9 19:11:23.539: INFO: Pod "downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 46.274665ms | |
Jul 9 19:11:25.582: INFO: Pod "downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.08953228s | |
Jul 9 19:11:27.618: INFO: Pod "downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.125776966s | |
STEP: Saw pod success | |
Jul 9 19:11:27.618: INFO: Pod "downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:11:27.651: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:11:27.733: INFO: Waiting for pod downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:11:27.763: INFO: Pod downwardapi-volume-8b1ddb7d-83e6-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:27.763: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-8qd2x" for this suite. | |
Jul 9 19:11:33.976: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:11:37.118: INFO: namespace: e2e-tests-projected-8qd2x, resource: bindings, ignored listing per whitelist | |
Jul 9 19:11:37.781: INFO: namespace e2e-tests-projected-8qd2x deletion completed in 9.97551978s | |
• [SLOW TEST:16.637 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should provide node allocatable (memory) as default memory limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:29.284: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:11:31.450: INFO: configPath is now "/tmp/e2e-test-router-reencrypt-rvndf-user.kubeconfig" | |
Jul 9 19:11:31.450: INFO: The user is now "e2e-test-router-reencrypt-rvndf-user" | |
Jul 9 19:11:31.450: INFO: Creating project "e2e-test-router-reencrypt-rvndf" | |
Jul 9 19:11:31.579: INFO: Waiting on permissions in project "e2e-test-router-reencrypt-rvndf" ... | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/reencrypt.go:41 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/reencrypt.go:29 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:11:31.726: INFO: namespace : e2e-test-router-reencrypt-rvndf api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:37.767: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [8.543 seconds] | |
[Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/reencrypt.go:18 | |
The HAProxy router [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/reencrypt.go:52 | |
should support reencrypt to services backed by a serving certificate automatically [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/reencrypt.go:53 | |
no router installed on the cluster | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/reencrypt.go:44 | |
------------------------------ | |
SSS | |
------------------------------ | |
[Conformance][templates] templateinstance cross-namespace test | |
should create and delete objects across namespaces [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_cross_namespace.go:30 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:29.732: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:11:31.293: INFO: configPath is now "/tmp/e2e-test-templates-b585v-user.kubeconfig" | |
Jul 9 19:11:31.293: INFO: The user is now "e2e-test-templates-b585v-user" | |
Jul 9 19:11:31.293: INFO: Creating project "e2e-test-templates-b585v" | |
Jul 9 19:11:31.440: INFO: Waiting on permissions in project "e2e-test-templates-b585v" ... | |
[BeforeEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:31.487: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:11:33.023: INFO: configPath is now "/tmp/e2e-test-templates2-nghzw-user.kubeconfig" | |
Jul 9 19:11:33.023: INFO: The user is now "e2e-test-templates2-nghzw-user" | |
Jul 9 19:11:33.023: INFO: Creating project "e2e-test-templates2-nghzw" | |
Jul 9 19:11:33.263: INFO: Waiting on permissions in project "e2e-test-templates2-nghzw" ... | |
[It] should create and delete objects across namespaces [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_cross_namespace.go:30 | |
Jul 9 19:11:33.304: INFO: Running 'oc adm --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-templates2-nghzw policy add-role-to-user admin e2e-test-templates-b585v-user' | |
role "admin" added: "e2e-test-templates-b585v-user" | |
STEP: creating the templateinstance | |
STEP: deleting the templateinstance | |
[AfterEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:11:35.637: INFO: namespace : e2e-test-templates-b585v api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:41.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:11:41.769: INFO: namespace : e2e-test-templates2-nghzw api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:47.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:18.111 seconds] | |
[Conformance][templates] templateinstance cross-namespace test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_cross_namespace.go:22 | |
should create and delete objects across namespaces [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_cross_namespace.go:30 | |
------------------------------ | |
[sig-storage] EmptyDir volumes when FSGroup is specified | |
nonexistent volume subPath should have the correct mode and owner using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:53 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:37.783: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:39.394: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-t499c | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] nonexistent volume subPath should have the correct mode and owner using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:53 | |
STEP: Creating a pod to test emptydir subpath on tmpfs | |
Jul 9 19:11:40.061: INFO: Waiting up to 5m0s for pod "pod-94fd2bd8-83e6-11e8-8fe2-28d244b00276" in namespace "e2e-tests-emptydir-t499c" to be "success or failure" | |
Jul 9 19:11:40.092: INFO: Pod "pod-94fd2bd8-83e6-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 31.167765ms | |
Jul 9 19:11:42.123: INFO: Pod "pod-94fd2bd8-83e6-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.062039439s | |
STEP: Saw pod success | |
Jul 9 19:11:42.123: INFO: Pod "pod-94fd2bd8-83e6-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:11:42.162: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-94fd2bd8-83e6-11e8-8fe2-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:11:42.262: INFO: Waiting for pod pod-94fd2bd8-83e6-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:11:42.292: INFO: Pod pod-94fd2bd8-83e6-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:42.292: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-t499c" for this suite. | |
Jul 9 19:11:48.465: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:11:50.328: INFO: namespace: e2e-tests-emptydir-t499c, resource: bindings, ignored listing per whitelist | |
Jul 9 19:11:52.155: INFO: namespace e2e-tests-emptydir-t499c deletion completed in 9.826706369s | |
• [SLOW TEST:14.372 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
when FSGroup is specified | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:44 | |
nonexistent volume subPath should have the correct mode and owner using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:53 | |
------------------------------ | |
[sig-storage] EmptyDir volumes when FSGroup is specified | |
volume on default medium should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:61 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:37.830: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:39.688: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-kmg6b | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] volume on default medium should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:61 | |
STEP: Creating a pod to test emptydir volume type on node default medium | |
Jul 9 19:11:40.434: INFO: Waiting up to 5m0s for pod "pod-95357f3c-83e6-11e8-992b-28d244b00276" in namespace "e2e-tests-emptydir-kmg6b" to be "success or failure" | |
Jul 9 19:11:40.471: INFO: Pod "pod-95357f3c-83e6-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 36.789536ms | |
Jul 9 19:11:42.542: INFO: Pod "pod-95357f3c-83e6-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.108162582s | |
STEP: Saw pod success | |
Jul 9 19:11:42.542: INFO: Pod "pod-95357f3c-83e6-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:11:42.581: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-95357f3c-83e6-11e8-992b-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:11:42.668: INFO: Waiting for pod pod-95357f3c-83e6-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:11:42.704: INFO: Pod pod-95357f3c-83e6-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:42.704: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-kmg6b" for this suite. | |
Jul 9 19:11:48.868: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:11:52.324: INFO: namespace: e2e-tests-emptydir-kmg6b, resource: bindings, ignored listing per whitelist | |
Jul 9 19:11:53.097: INFO: namespace e2e-tests-emptydir-kmg6b deletion completed in 10.351115117s | |
• [SLOW TEST:15.267 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
when FSGroup is specified | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:44 | |
volume on default medium should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:61 | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:47.844: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:49.287: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-gr2vb | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating projection with secret that has name projected-secret-test-9ada9ad5-83e6-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:11:49.927: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-9adfcf80-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-projected-gr2vb" to be "success or failure" | |
Jul 9 19:11:49.955: INFO: Pod "pod-projected-secrets-9adfcf80-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 27.730422ms | |
Jul 9 19:11:51.982: INFO: Pod "pod-projected-secrets-9adfcf80-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.055067008s | |
STEP: Saw pod success | |
Jul 9 19:11:51.982: INFO: Pod "pod-projected-secrets-9adfcf80-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:11:52.009: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-secrets-9adfcf80-83e6-11e8-8401-28d244b00276 container projected-secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:11:52.115: INFO: Waiting for pod pod-projected-secrets-9adfcf80-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:11:52.148: INFO: Pod pod-projected-secrets-9adfcf80-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:52.148: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-gr2vb" for this suite. | |
Jul 9 19:11:58.289: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:12:00.976: INFO: namespace: e2e-tests-projected-gr2vb, resource: bindings, ignored listing per whitelist | |
Jul 9 19:12:02.037: INFO: namespace e2e-tests-projected-gr2vb deletion completed in 9.85639901s | |
• [SLOW TEST:14.194 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
should fail resolving unresolvable valueFrom in docker build environment variable references [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:122 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:53.098: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:11:55.298: INFO: configPath is now "/tmp/e2e-test-build-valuefrom-q75fk-user.kubeconfig" | |
Jul 9 19:11:55.298: INFO: The user is now "e2e-test-build-valuefrom-q75fk-user" | |
Jul 9 19:11:55.298: INFO: Creating project "e2e-test-build-valuefrom-q75fk" | |
Jul 9 19:11:55.416: INFO: Waiting on permissions in project "e2e-test-build-valuefrom-q75fk" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:27 | |
Jul 9 19:11:55.477: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:38 | |
STEP: waiting for builder service account | |
STEP: waiting for openshift namespace imagestreams | |
Jul 9 19:11:55.628: INFO: Running scan #0 | |
Jul 9 19:11:55.628: INFO: Checking language ruby | |
Jul 9 19:11:55.683: INFO: Checking tag 2.0 | |
Jul 9 19:11:55.683: INFO: Checking tag 2.2 | |
Jul 9 19:11:55.683: INFO: Checking tag 2.3 | |
Jul 9 19:11:55.683: INFO: Checking tag 2.4 | |
Jul 9 19:11:55.683: INFO: Checking tag 2.5 | |
Jul 9 19:11:55.683: INFO: Checking tag latest | |
Jul 9 19:11:55.683: INFO: Checking language nodejs | |
Jul 9 19:11:55.725: INFO: Checking tag 0.10 | |
Jul 9 19:11:55.725: INFO: Checking tag 4 | |
Jul 9 19:11:55.725: INFO: Checking tag 6 | |
Jul 9 19:11:55.725: INFO: Checking tag 8 | |
Jul 9 19:11:55.725: INFO: Checking tag latest | |
Jul 9 19:11:55.725: INFO: Checking language perl | |
Jul 9 19:11:55.757: INFO: Checking tag 5.16 | |
Jul 9 19:11:55.757: INFO: Checking tag 5.20 | |
Jul 9 19:11:55.757: INFO: Checking tag 5.24 | |
Jul 9 19:11:55.757: INFO: Checking tag latest | |
Jul 9 19:11:55.757: INFO: Checking language php | |
Jul 9 19:11:55.789: INFO: Checking tag 7.1 | |
Jul 9 19:11:55.789: INFO: Checking tag latest | |
Jul 9 19:11:55.789: INFO: Checking tag 5.5 | |
Jul 9 19:11:55.789: INFO: Checking tag 5.6 | |
Jul 9 19:11:55.789: INFO: Checking tag 7.0 | |
Jul 9 19:11:55.789: INFO: Checking language python | |
Jul 9 19:11:55.825: INFO: Checking tag 3.4 | |
Jul 9 19:11:55.825: INFO: Checking tag 3.5 | |
Jul 9 19:11:55.825: INFO: Checking tag 3.6 | |
Jul 9 19:11:55.825: INFO: Checking tag latest | |
Jul 9 19:11:55.825: INFO: Checking tag 2.7 | |
Jul 9 19:11:55.825: INFO: Checking tag 3.3 | |
Jul 9 19:11:55.825: INFO: Checking language wildfly | |
Jul 9 19:11:55.860: INFO: Checking tag 9.0 | |
Jul 9 19:11:55.860: INFO: Checking tag latest | |
Jul 9 19:11:55.860: INFO: Checking tag 10.0 | |
Jul 9 19:11:55.860: INFO: Checking tag 10.1 | |
Jul 9 19:11:55.860: INFO: Checking tag 11.0 | |
Jul 9 19:11:55.860: INFO: Checking tag 12.0 | |
Jul 9 19:11:55.860: INFO: Checking tag 8.1 | |
Jul 9 19:11:55.860: INFO: Checking language mysql | |
Jul 9 19:11:55.890: INFO: Checking tag latest | |
Jul 9 19:11:55.890: INFO: Checking tag 5.5 | |
Jul 9 19:11:55.890: INFO: Checking tag 5.6 | |
Jul 9 19:11:55.890: INFO: Checking tag 5.7 | |
Jul 9 19:11:55.890: INFO: Checking language postgresql | |
Jul 9 19:11:55.924: INFO: Checking tag latest | |
Jul 9 19:11:55.924: INFO: Checking tag 9.2 | |
Jul 9 19:11:55.924: INFO: Checking tag 9.4 | |
Jul 9 19:11:55.924: INFO: Checking tag 9.5 | |
Jul 9 19:11:55.924: INFO: Checking tag 9.6 | |
Jul 9 19:11:55.924: INFO: Checking language mongodb | |
Jul 9 19:11:55.963: INFO: Checking tag 3.4 | |
Jul 9 19:11:55.963: INFO: Checking tag latest | |
Jul 9 19:11:55.963: INFO: Checking tag 2.4 | |
Jul 9 19:11:55.963: INFO: Checking tag 2.6 | |
Jul 9 19:11:55.963: INFO: Checking tag 3.2 | |
Jul 9 19:11:55.963: INFO: Checking language jenkins | |
Jul 9 19:11:55.996: INFO: Checking tag latest | |
Jul 9 19:11:55.996: INFO: Checking tag 1 | |
Jul 9 19:11:55.996: INFO: Checking tag 2 | |
Jul 9 19:11:55.996: INFO: Success! | |
STEP: creating test image stream | |
Jul 9 19:11:55.996: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-q75fk-user.kubeconfig --namespace=e2e-test-build-valuefrom-q75fk -f /tmp/fixture-testdata-dir877664294/test/extended/testdata/builds/valuefrom/test-is.json' | |
imagestream.image.openshift.io "test" created | |
STEP: creating test secret | |
Jul 9 19:11:56.351: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-q75fk-user.kubeconfig --namespace=e2e-test-build-valuefrom-q75fk -f /tmp/fixture-testdata-dir877664294/test/extended/testdata/builds/valuefrom/test-secret.yaml' | |
secret "mysecret" created | |
STEP: creating test configmap | |
Jul 9 19:11:56.911: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-q75fk-user.kubeconfig --namespace=e2e-test-build-valuefrom-q75fk -f /tmp/fixture-testdata-dir877664294/test/extended/testdata/builds/valuefrom/test-configmap.yaml' | |
configmap "myconfigmap" created | |
[It] should fail resolving unresolvable valueFrom in docker build environment variable references [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:122 | |
STEP: creating test build config | |
Jul 9 19:11:57.312: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-q75fk-user.kubeconfig --namespace=e2e-test-build-valuefrom-q75fk -f /tmp/fixture-testdata-dir877664294/test/extended/testdata/builds/valuefrom/failed-docker-build-value-from-config.yaml' | |
buildconfig.build.openshift.io "mydockertest" created | |
STEP: starting test build | |
Jul 9 19:11:57.636: INFO: Running 'oc start-build --config=/tmp/e2e-test-build-valuefrom-q75fk-user.kubeconfig --namespace=e2e-test-build-valuefrom-q75fk mydockertest -o=name' | |
Jul 9 19:11:57.927: INFO: | |
start-build output with args [mydockertest -o=name]: | |
Error><nil> | |
StdOut> | |
build/mydockertest-1 | |
StdErr> | |
Jul 9 19:11:57.928: INFO: Waiting for mydockertest-1 to complete | |
Jul 9 19:12:04.011: INFO: WaitForABuild returning with error: The build "mydockertest-1" status is "Error" | |
Jul 9 19:12:04.011: INFO: Done waiting for mydockertest-1: util.BuildResult{BuildPath:"build/mydockertest-1", BuildName:"mydockertest-1", StartBuildStdErr:"", StartBuildStdOut:"build/mydockertest-1", StartBuildErr:error(nil), BuildConfigName:"", Build:(*build.Build)(0xc421470300), BuildAttempt:true, BuildSuccess:false, BuildFailure:true, BuildCancelled:false, BuildTimeout:false, LogDumper:(util.LogDumperFunc)(nil), Oc:(*util.CLI)(0xc420dca1e0)} | |
with error: The build "mydockertest-1" status is "Error" | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:31 | |
[AfterEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:12:04.100: INFO: namespace : e2e-test-build-valuefrom-q75fk api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:10.145: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:17.089 seconds] | |
[Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:13 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:26 | |
should fail resolving unresolvable valueFrom in docker build environment variable references [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:122 | |
------------------------------ | |
SS | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:431 | |
Jul 9 19:12:10.506: INFO: Could not check network plugin name: exit status 1. Assuming a non-OpenShift plugin | |
Jul 9 19:12:10.506: INFO: This plugin does not implement NetworkPolicy. | |
[AfterEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:10.507: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.316 seconds] | |
NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:48 | |
when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:430 | |
should enforce policy based on PodSelector [Feature:OSNetworkPolicy] [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:86 | |
Jul 9 19:12:10.506: This plugin does not implement NetworkPolicy. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
[k8s.io] Pods | |
should allow activeDeadlineSeconds to be updated [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:12:02.041: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:12:03.665: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-wvqk5 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/pods.go:127 | |
[It] should allow activeDeadlineSeconds to be updated [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: creating the pod | |
STEP: submitting the pod to kubernetes | |
STEP: verifying the pod is in kubernetes | |
STEP: updating the pod | |
Jul 9 19:12:07.146: INFO: Successfully updated pod "pod-update-activedeadlineseconds-a379c7c5-83e6-11e8-8401-28d244b00276" | |
Jul 9 19:12:07.146: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-a379c7c5-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-pods-wvqk5" to be "terminated due to deadline exceeded" | |
Jul 9 19:12:07.174: INFO: Pod "pod-update-activedeadlineseconds-a379c7c5-83e6-11e8-8401-28d244b00276": Phase="Running", Reason="", readiness=true. Elapsed: 27.773489ms | |
Jul 9 19:12:09.205: INFO: Pod "pod-update-activedeadlineseconds-a379c7c5-83e6-11e8-8401-28d244b00276": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 2.058771469s | |
Jul 9 19:12:09.205: INFO: Pod "pod-update-activedeadlineseconds-a379c7c5-83e6-11e8-8401-28d244b00276" satisfied condition "terminated due to deadline exceeded" | |
[AfterEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:09.205: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pods-wvqk5" for this suite. | |
Jul 9 19:12:15.356: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:12:17.208: INFO: namespace: e2e-tests-pods-wvqk5, resource: bindings, ignored listing per whitelist | |
Jul 9 19:12:18.839: INFO: namespace e2e-tests-pods-wvqk5 deletion completed in 9.598590478s | |
• [SLOW TEST:16.798 seconds] | |
[k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should allow activeDeadlineSeconds to be updated [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[k8s.io] InitContainer | |
should invoke init containers on a RestartAlways pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:103 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:13.874: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:15.455: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-9dmbk | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:40 | |
[It] should invoke init containers on a RestartAlways pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:103 | |
STEP: creating the pod | |
Jul 9 19:11:16.159: INFO: PodSpec: initContainers in spec.initContainers | |
[AfterEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:11:56.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-init-container-9dmbk" for this suite. | |
Jul 9 19:12:18.314: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:12:20.767: INFO: namespace: e2e-tests-init-container-9dmbk, resource: bindings, ignored listing per whitelist | |
Jul 9 19:12:22.186: INFO: namespace e2e-tests-init-container-9dmbk deletion completed in 25.976091167s | |
• [SLOW TEST:68.313 seconds] | |
[k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should invoke init containers on a RestartAlways pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:103 | |
------------------------------ | |
SS | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:12:22.189: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:12:24.143: INFO: configPath is now "/tmp/e2e-test-resolve-local-names-x2lpn-user.kubeconfig" | |
Jul 9 19:12:24.143: INFO: The user is now "e2e-test-resolve-local-names-x2lpn-user" | |
Jul 9 19:12:24.143: INFO: Creating project "e2e-test-resolve-local-names-x2lpn" | |
Jul 9 19:12:24.351: INFO: Waiting on permissions in project "e2e-test-resolve-local-names-x2lpn" ... | |
[It] should perform lookup when the pod has the resolve-names annotation [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:73 | |
Jul 9 19:12:24.400: INFO: Running 'oc import-image --config=/tmp/e2e-test-resolve-local-names-x2lpn-user.kubeconfig --namespace=e2e-test-resolve-local-names-x2lpn busybox:latest --confirm' | |
The import completed successfully. | |
Name: busybox | |
Namespace: e2e-test-resolve-local-names-x2lpn | |
Created: Less than a second ago | |
Labels: <none> | |
Annotations: openshift.io/image.dockerRepositoryCheck=2018-07-10T02:12:26Z | |
Docker Pull Spec: docker-registry.default.svc:5000/e2e-test-resolve-local-names-x2lpn/busybox | |
Image Lookup: local=false | |
Unique Images: 1 | |
Tags: 1 | |
latest | |
tagged from busybox:latest | |
* busybox@sha256:74f634b1bc1bd74535d5209589734efbd44a25f4e2dc96d78784576a3eb5b335 | |
Less than a second ago | |
Image Name: busybox:latest | |
Docker Image: busybox@sha256:74f634b1bc1bd74535d5209589734efbd44a25f4e2dc96d78784576a3eb5b335 | |
Name: sha256:74f634b1bc1bd74535d5209589734efbd44a25f4e2dc96d78784576a3eb5b335 | |
Created: Less than a second ago | |
Annotations: image.openshift.io/dockerLayersOrder=ascending | |
Image Size: 724.6kB | |
Image Created: 6 weeks ago | |
Author: <none> | |
Arch: amd64 | |
Command: sh | |
Working Dir: <none> | |
User: <none> | |
Exposes Ports: <none> | |
Docker Labels: <none> | |
Environment: PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin | |
[AfterEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:12:26.312: INFO: namespace : e2e-test-resolve-local-names-x2lpn api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:32.345: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] [10.191 seconds] | |
[Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:14 | |
should perform lookup when the pod has the resolve-names annotation [Suite:openshift/conformance/parallel] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:73 | |
default image resolution is not configured, can't verify pod resolution | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:99 | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs with multiple image change triggers [Conformance] | |
should run a successful deployment with multiple triggers [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:513 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:10:48.830: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:10:51.252: INFO: configPath is now "/tmp/e2e-test-cli-deployment-92vwf-user.kubeconfig" | |
Jul 9 19:10:51.252: INFO: The user is now "e2e-test-cli-deployment-92vwf-user" | |
Jul 9 19:10:51.252: INFO: Creating project "e2e-test-cli-deployment-92vwf" | |
Jul 9 19:10:51.410: INFO: Waiting on permissions in project "e2e-test-cli-deployment-92vwf" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should run a successful deployment with multiple triggers [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:513 | |
STEP: creating DC | |
STEP: verifying the deployment is marked complete | |
Jul 9 19:11:52.159: INFO: Latest rollout of dc/example (rc/example-1) is complete. | |
[AfterEach] with multiple image change triggers [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:509 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:11:54.263: INFO: namespace : e2e-test-cli-deployment-92vwf api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:34.311: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:105.552 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
with multiple image change triggers [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:507 | |
should run a successful deployment with multiple triggers [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:513 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:12:34.383: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:12:36.680: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-hostpath-tscb6 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:37 | |
[It] should support existing single file subPath [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:167 | |
Jul 9 19:12:37.452: INFO: No SSH Key for provider : 'GetSigner(...) not implemented for ' | |
[AfterEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:37.453: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-hostpath-tscb6" for this suite. | |
Jul 9 19:12:43.654: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:12:45.782: INFO: namespace: e2e-tests-hostpath-tscb6, resource: bindings, ignored listing per whitelist | |
Jul 9 19:12:48.312: INFO: namespace e2e-tests-hostpath-tscb6 deletion completed in 10.796412144s | |
S [SKIPPING] [13.929 seconds] | |
[sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:34 | |
should support existing single file subPath [Suite:openshift/conformance/parallel] [Suite:k8s] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:167 | |
Jul 9 19:12:37.452: No SSH Key for provider : 'GetSigner(...) not implemented for ' | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
S | |
------------------------------ | |
[k8s.io] Pods | |
should support remote command execution over websockets [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/pods.go:470 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:12:10.508: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:12:12.627: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-9775r | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/pods.go:127 | |
[It] should support remote command execution over websockets [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/pods.go:470 | |
Jul 9 19:12:13.409: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: creating the pod | |
STEP: submitting the pod to kubernetes | |
[AfterEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:17.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pods-9775r" for this suite. | |
Jul 9 19:12:56.059: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:12:59.715: INFO: namespace: e2e-tests-pods-9775r, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:00.520: INFO: namespace e2e-tests-pods-9775r deletion completed in 42.593555306s | |
• [SLOW TEST:50.012 seconds] | |
[k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should support remote command execution over websockets [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/pods.go:470 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:419 | |
Jul 9 19:13:00.521: INFO: This plugin does not isolate namespaces by default. | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:00.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:00.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] | |
[Area:Networking] services | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/services.go:10 | |
when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:418 | |
should prevent connections to pods in different namespaces on the same node via service IPs [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/services.go:40 | |
Jul 9 19:13:00.521: This plugin does not isolate namespaces by default. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
[Conformance][templates] templateinstance impersonation tests | |
should pass impersonation update tests [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:252 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:12:48.316: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:12:50.508: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-user.kubeconfig" | |
Jul 9 19:12:50.508: INFO: The user is now "e2e-test-templates-b5fkm-user" | |
Jul 9 19:12:50.508: INFO: Creating project "e2e-test-templates-b5fkm" | |
Jul 9 19:12:50.659: INFO: Waiting on permissions in project "e2e-test-templates-b5fkm" ... | |
[BeforeEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:57 | |
Jul 9 19:12:51.908: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-adminuser.kubeconfig" | |
Jul 9 19:12:52.180: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-impersonateuser.kubeconfig" | |
Jul 9 19:12:52.429: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-impersonatebygroupuser.kubeconfig" | |
Jul 9 19:12:52.677: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-edituser1.kubeconfig" | |
Jul 9 19:12:52.922: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-edituser2.kubeconfig" | |
Jul 9 19:12:53.178: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-viewuser.kubeconfig" | |
Jul 9 19:12:53.434: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-impersonatebygroupuser.kubeconfig" | |
[It] should pass impersonation update tests [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:252 | |
STEP: testing as system:admin user | |
STEP: testing as e2e-test-templates-b5fkm-adminuser user | |
Jul 9 19:12:54.343: INFO: configPath is now "/tmp/e2e-test-templates-b5fkm-adminuser.kubeconfig" | |
[AfterEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:12:54.783: INFO: namespace : e2e-test-templates-b5fkm api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:13:00.877: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:221 | |
• Failure [12.949 seconds] | |
[Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:27 | |
should pass impersonation update tests [Suite:openshift/conformance/parallel] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:252 | |
Expected an error to have occurred. Got: | |
<nil>: nil | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:322 | |
------------------------------ | |
[sig-network] Networking Granular Checks: Pods | |
should function for intra-pod communication: http [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:12:18.841: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:12:20.604: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-7j6f8 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should function for intra-pod communication: http [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-7j6f8 | |
STEP: creating a selector | |
STEP: Creating the service pods in kubernetes | |
Jul 9 19:12:21.211: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable | |
STEP: Creating test pods | |
Jul 9 19:12:43.788: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.2.2.242:8080/dial?request=hostName&protocol=http&host=10.2.2.238&port=8080&tries=1'] Namespace:e2e-tests-pod-network-test-7j6f8 PodName:host-test-container-pod ContainerName:hostexec Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} | |
Jul 9 19:12:43.789: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:12:44.203: INFO: Waiting for endpoints: map[] | |
[AfterEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:12:44.203: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pod-network-test-7j6f8" for this suite. | |
Jul 9 19:13:06.338: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:08.366: INFO: namespace: e2e-tests-pod-network-test-7j6f8, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:09.624: INFO: namespace e2e-tests-pod-network-test-7j6f8 deletion completed in 25.380544553s | |
• [SLOW TEST:50.783 seconds] | |
[sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:25 | |
Granular Checks: Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:28 | |
should function for intra-pod communication: http [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] EmptyDir volumes | |
should support (root,0644,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:01.265: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:03.565: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-7mdjl | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should support (root,0644,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test emptydir 0644 on node default medium | |
Jul 9 19:13:04.438: INFO: Waiting up to 5m0s for pod "pod-c740ba1e-83e6-11e8-881a-28d244b00276" in namespace "e2e-tests-emptydir-7mdjl" to be "success or failure" | |
Jul 9 19:13:04.481: INFO: Pod "pod-c740ba1e-83e6-11e8-881a-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 43.665856ms | |
Jul 9 19:13:06.523: INFO: Pod "pod-c740ba1e-83e6-11e8-881a-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.085696002s | |
STEP: Saw pod success | |
Jul 9 19:13:06.523: INFO: Pod "pod-c740ba1e-83e6-11e8-881a-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:13:06.566: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-c740ba1e-83e6-11e8-881a-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:13:06.673: INFO: Waiting for pod pod-c740ba1e-83e6-11e8-881a-28d244b00276 to disappear | |
Jul 9 19:13:06.717: INFO: Pod pod-c740ba1e-83e6-11e8-881a-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:06.717: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-7mdjl" for this suite. | |
Jul 9 19:13:12.909: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:15.123: INFO: namespace: e2e-tests-emptydir-7mdjl, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:17.537: INFO: namespace e2e-tests-emptydir-7mdjl deletion completed in 10.770969057s | |
• [SLOW TEST:16.271 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
should support (root,0644,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs paused [Conformance] | |
should disable actions on deployments [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:742 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:12:32.382: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:12:34.249: INFO: configPath is now "/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig" | |
Jul 9 19:12:34.249: INFO: The user is now "e2e-test-cli-deployment-lzhmc-user" | |
Jul 9 19:12:34.249: INFO: Creating project "e2e-test-cli-deployment-lzhmc" | |
Jul 9 19:12:34.415: INFO: Waiting on permissions in project "e2e-test-cli-deployment-lzhmc" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should disable actions on deployments [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:742 | |
STEP: verifying that we cannot start a new deployment via oc deploy | |
Jul 9 19:12:34.793: INFO: Running 'oc deploy --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc dc/paused --latest' | |
Jul 9 19:12:35.082: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc deploy --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc dc/paused --latest] [] Command "deploy" is deprecated, Use the `rollout latest` and `rollout cancel` commands instead. | |
Flag --latest has been deprecated, use 'oc rollout latest' instead | |
error: cannot deploy a paused deployment config | |
Command "deploy" is deprecated, Use the `rollout latest` and `rollout cancel` commands instead. | |
Flag --latest has been deprecated, use 'oc rollout latest' instead | |
error: cannot deploy a paused deployment config | |
[] <nil> 0xc421067200 exit status 1 <nil> <nil> true [0xc420efe310 0xc420efe390 0xc420efe390] [0xc420efe310 0xc420efe390] [0xc420efe318 0xc420efe370] [0x916090 0x916190] 0xc420ea0600 <nil>}: | |
Command "deploy" is deprecated, Use the `rollout latest` and `rollout cancel` commands instead. | |
Flag --latest has been deprecated, use 'oc rollout latest' instead | |
error: cannot deploy a paused deployment config | |
STEP: verifying that we cannot start a new deployment via oc rollout | |
Jul 9 19:12:35.082: INFO: Running 'oc rollout --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc latest dc/paused' | |
Jul 9 19:12:35.319: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc rollout --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc latest dc/paused] [] error: cannot deploy a paused deployment config | |
error: cannot deploy a paused deployment config | |
[] <nil> 0xc42090aed0 exit status 1 <nil> <nil> true [0xc4219600c0 0xc4219600e8 0xc4219600e8] [0xc4219600c0 0xc4219600e8] [0xc4219600c8 0xc4219600e0] [0x916090 0x916190] 0xc42199b9e0 <nil>}: | |
error: cannot deploy a paused deployment config | |
STEP: verifying that we cannot cancel a deployment | |
Jul 9 19:12:35.319: INFO: Running 'oc rollout --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc cancel dc/paused' | |
Jul 9 19:12:35.670: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc rollout --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc cancel dc/paused] [] unable to cancel paused deployment e2e-test-cli-deployment-lzhmc/paused | |
there have been no replication controllers for e2e-test-cli-deployment-lzhmc/paused | |
unable to cancel paused deployment e2e-test-cli-deployment-lzhmc/paused | |
there have been no replication controllers for e2e-test-cli-deployment-lzhmc/paused | |
[] <nil> 0xc42090b3b0 exit status 1 <nil> <nil> true [0xc4219600f8 0xc421960128 0xc421960128] [0xc4219600f8 0xc421960128] [0xc421960100 0xc421960118] [0x916090 0x916190] 0xc42199baa0 <nil>}: | |
unable to cancel paused deployment e2e-test-cli-deployment-lzhmc/paused | |
there have been no replication controllers for e2e-test-cli-deployment-lzhmc/paused | |
STEP: verifying that we cannot retry a deployment | |
Jul 9 19:12:35.670: INFO: Running 'oc deploy --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc dc/paused --retry' | |
Jul 9 19:12:35.890: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc deploy --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc dc/paused --retry] [] Command "deploy" is deprecated, Use the `rollout latest` and `rollout cancel` commands instead. | |
error: cannot retry a paused deployment config | |
Command "deploy" is deprecated, Use the `rollout latest` and `rollout cancel` commands instead. | |
error: cannot retry a paused deployment config | |
[] <nil> 0xc42090b860 exit status 1 <nil> <nil> true [0xc421960130 0xc421960200 0xc421960200] [0xc421960130 0xc421960200] [0xc421960138 0xc4219601f0] [0x916090 0x916190] 0xc42199bb60 <nil>}: | |
Command "deploy" is deprecated, Use the `rollout latest` and `rollout cancel` commands instead. | |
error: cannot retry a paused deployment config | |
STEP: verifying that we cannot rollout retry a deployment | |
Jul 9 19:12:35.890: INFO: Running 'oc rollout --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc retry dc/paused' | |
Jul 9 19:12:36.152: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc rollout --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc retry dc/paused] [] error: unable to retry paused deployment config "paused" | |
error: unable to retry paused deployment config "paused" | |
[] <nil> 0xc42090bce0 exit status 1 <nil> <nil> true [0xc421960210 0xc421960280 0xc421960280] [0xc421960210 0xc421960280] [0xc421960220 0xc421960270] [0x916090 0x916190] 0xc42199bc20 <nil>}: | |
error: unable to retry paused deployment config "paused" | |
STEP: verifying that we cannot rollback a deployment | |
Jul 9 19:12:36.152: INFO: Running 'oc rollback --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc dc/paused --to-version 1' | |
Jul 9 19:12:36.396: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc rollback --config=/tmp/e2e-test-cli-deployment-lzhmc-user.kubeconfig --namespace=e2e-test-cli-deployment-lzhmc dc/paused --to-version 1] [] error: cannot rollback a paused deployment config | |
error: cannot rollback a paused deployment config | |
[] <nil> 0xc4210ecb70 exit status 1 <nil> <nil> true [0xc421af61e8 0xc421af6220 0xc421af6220] [0xc421af61e8 0xc421af6220] [0xc421af61f8 0xc421af6210] [0x916090 0x916190] 0xc4215e0060 <nil>}: | |
error: cannot rollback a paused deployment config | |
Jul 9 19:12:41.132: INFO: Latest rollout of dc/paused (rc/paused-1) is complete. | |
STEP: making sure it updates observedGeneration after being paused | |
[AfterEach] paused [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:738 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:12:43.581: INFO: namespace : e2e-test-cli-deployment-lzhmc api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:23.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:51.262 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
paused [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:736 | |
should disable actions on deployments [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:742 | |
------------------------------ | |
SS | |
------------------------------ | |
[k8s.io] Docker Containers | |
should use the image defaults if command and args are blank [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Docker Containers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:09.625: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:11.246: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-zdwzn | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should use the image defaults if command and args are blank [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test use defaults | |
Jul 9 19:13:11.887: INFO: Waiting up to 5m0s for pod "client-containers-cbb9f2ce-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-containers-zdwzn" to be "success or failure" | |
Jul 9 19:13:11.916: INFO: Pod "client-containers-cbb9f2ce-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 29.099543ms | |
Jul 9 19:13:13.990: INFO: Pod "client-containers-cbb9f2ce-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.103415267s | |
STEP: Saw pod success | |
Jul 9 19:13:13.990: INFO: Pod "client-containers-cbb9f2ce-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:13:14.019: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod client-containers-cbb9f2ce-83e6-11e8-8401-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:13:14.135: INFO: Waiting for pod client-containers-cbb9f2ce-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:13:14.172: INFO: Pod client-containers-cbb9f2ce-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [k8s.io] Docker Containers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:14.172: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-containers-zdwzn" for this suite. | |
Jul 9 19:13:20.314: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:23.147: INFO: namespace: e2e-tests-containers-zdwzn, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:23.659: INFO: namespace e2e-tests-containers-zdwzn deletion completed in 9.455658364s | |
• [SLOW TEST:14.034 seconds] | |
[k8s.io] Docker Containers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should use the image defaults if command and args are blank [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Secrets | |
optional updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:11:52.157: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:11:53.896: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-kqpsc | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] optional updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
Jul 9 19:11:54.615: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node | |
STEP: Creating secret with name s-test-opt-del-9db12ddf-83e6-11e8-8fe2-28d244b00276 | |
STEP: Creating secret with name s-test-opt-upd-9db12e14-83e6-11e8-8fe2-28d244b00276 | |
STEP: Creating the pod | |
STEP: Deleting secret s-test-opt-del-9db12ddf-83e6-11e8-8fe2-28d244b00276 | |
STEP: Updating secret s-test-opt-upd-9db12e14-83e6-11e8-8fe2-28d244b00276 | |
STEP: Creating secret with name s-test-opt-create-9db12e26-83e6-11e8-8fe2-28d244b00276 | |
STEP: waiting to observe update in volume | |
[AfterEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:05.191: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-kqpsc" for this suite. | |
Jul 9 19:13:27.371: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:29.000: INFO: namespace: e2e-tests-secrets-kqpsc, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:31.240: INFO: namespace e2e-tests-secrets-kqpsc deletion completed in 25.968448725s | |
• [SLOW TEST:99.083 seconds] | |
[sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 | |
optional updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[k8s.io] Variable Expansion | |
should allow composing env vars into new env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:17.540: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:20.288: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-bzdhq | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should allow composing env vars into new env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test env composition | |
Jul 9 19:13:21.123: INFO: Waiting up to 5m0s for pod "var-expansion-d1387e25-83e6-11e8-881a-28d244b00276" in namespace "e2e-tests-var-expansion-bzdhq" to be "success or failure" | |
Jul 9 19:13:21.168: INFO: Pod "var-expansion-d1387e25-83e6-11e8-881a-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 45.213081ms | |
Jul 9 19:13:23.210: INFO: Pod "var-expansion-d1387e25-83e6-11e8-881a-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.086529997s | |
Jul 9 19:13:25.269: INFO: Pod "var-expansion-d1387e25-83e6-11e8-881a-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.146297789s | |
STEP: Saw pod success | |
Jul 9 19:13:25.270: INFO: Pod "var-expansion-d1387e25-83e6-11e8-881a-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:13:25.313: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod var-expansion-d1387e25-83e6-11e8-881a-28d244b00276 container dapi-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:13:25.417: INFO: Waiting for pod var-expansion-d1387e25-83e6-11e8-881a-28d244b00276 to disappear | |
Jul 9 19:13:25.458: INFO: Pod var-expansion-d1387e25-83e6-11e8-881a-28d244b00276 no longer exists | |
[AfterEach] [k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:25.458: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-var-expansion-bzdhq" for this suite. | |
Jul 9 19:13:31.661: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:35.196: INFO: namespace: e2e-tests-var-expansion-bzdhq, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:36.455: INFO: namespace e2e-tests-var-expansion-bzdhq deletion completed in 10.949058678s | |
• [SLOW TEST:18.915 seconds] | |
[k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should allow composing env vars into new env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Secrets | |
should be consumable from pods in volume with mappings [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:23.660: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:25.211: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-d6xjb | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in volume with mappings [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating secret with name secret-test-map-d40d907f-83e6-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:13:25.896: INFO: Waiting up to 5m0s for pod "pod-secrets-d41333e3-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-secrets-d6xjb" to be "success or failure" | |
Jul 9 19:13:25.928: INFO: Pod "pod-secrets-d41333e3-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 31.358634ms | |
Jul 9 19:13:27.974: INFO: Pod "pod-secrets-d41333e3-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.077946634s | |
STEP: Saw pod success | |
Jul 9 19:13:27.974: INFO: Pod "pod-secrets-d41333e3-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:13:28.002: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-secrets-d41333e3-83e6-11e8-8401-28d244b00276 container secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:13:28.077: INFO: Waiting for pod pod-secrets-d41333e3-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:13:28.104: INFO: Pod pod-secrets-d41333e3-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:28.104: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-d6xjb" for this suite. | |
Jul 9 19:13:34.235: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:37.397: INFO: namespace: e2e-tests-secrets-d6xjb, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:37.621: INFO: namespace e2e-tests-secrets-d6xjb deletion completed in 9.480839271s | |
• [SLOW TEST:13.961 seconds] | |
[sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 | |
should be consumable from pods in volume with mappings [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-api-machinery] Downward API | |
should provide container's limits.cpu/memory and requests.cpu/memory as env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:23.646: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:25.194: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-jvv2t | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward api env vars | |
Jul 9 19:13:25.945: INFO: Waiting up to 5m0s for pod "downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276" in namespace "e2e-tests-downward-api-jvv2t" to be "success or failure" | |
Jul 9 19:13:25.974: INFO: Pod "downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 29.397911ms | |
Jul 9 19:13:28.002: INFO: Pod "downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.057000308s | |
Jul 9 19:13:30.036: INFO: Pod "downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.091581003s | |
STEP: Saw pod success | |
Jul 9 19:13:30.036: INFO: Pod "downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:13:30.068: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276 container dapi-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:13:30.149: INFO: Waiting for pod downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:13:30.177: INFO: Pod downward-api-d41a6e73-83e6-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:30.177: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-jvv2t" for this suite. | |
Jul 9 19:13:36.311: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:39.008: INFO: namespace: e2e-tests-downward-api-jvv2t, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:39.763: INFO: namespace e2e-tests-downward-api-jvv2t deletion completed in 9.552506513s | |
• [SLOW TEST:16.117 seconds] | |
[sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downward_api.go:37 | |
should provide container's limits.cpu/memory and requests.cpu/memory as env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Projected | |
should project all components that make up the projection API [Projection] [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:31.244: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:32.967: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-c2mgg | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should project all components that make up the projection API [Projection] [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating configMap with name configmap-projected-all-test-volume-d8b8fea4-83e6-11e8-8fe2-28d244b00276 | |
STEP: Creating secret with name secret-projected-all-test-volume-d8b8fe90-83e6-11e8-8fe2-28d244b00276 | |
STEP: Creating a pod to test Check all projections for projected volume plugin | |
Jul 9 19:13:33.762: INFO: Waiting up to 5m0s for pod "projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276" in namespace "e2e-tests-projected-c2mgg" to be "success or failure" | |
Jul 9 19:13:33.794: INFO: Pod "projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 31.255872ms | |
Jul 9 19:13:35.830: INFO: Pod "projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.068069492s | |
Jul 9 19:13:37.861: INFO: Pod "projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.099079179s | |
STEP: Saw pod success | |
Jul 9 19:13:37.862: INFO: Pod "projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:13:37.894: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276 container projected-all-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:13:37.981: INFO: Waiting for pod projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:13:38.012: INFO: Pod projected-volume-d8b8fe5a-83e6-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:38.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-c2mgg" for this suite. | |
Jul 9 19:13:44.154: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:47.657: INFO: namespace: e2e-tests-projected-c2mgg, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:47.785: INFO: namespace e2e-tests-projected-c2mgg deletion completed in 9.733539364s | |
• [SLOW TEST:16.541 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should project all components that make up the projection API [Projection] [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] EmptyDir volumes | |
should support (non-root,0777,tmpfs) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:36.458: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:38.586: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-8rb45 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should support (non-root,0777,tmpfs) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test emptydir 0777 on tmpfs | |
Jul 9 19:13:39.444: INFO: Waiting up to 5m0s for pod "pod-dc233d27-83e6-11e8-881a-28d244b00276" in namespace "e2e-tests-emptydir-8rb45" to be "success or failure" | |
Jul 9 19:13:39.505: INFO: Pod "pod-dc233d27-83e6-11e8-881a-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 60.863579ms | |
Jul 9 19:13:41.548: INFO: Pod "pod-dc233d27-83e6-11e8-881a-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.103172176s | |
STEP: Saw pod success | |
Jul 9 19:13:41.548: INFO: Pod "pod-dc233d27-83e6-11e8-881a-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:13:41.593: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-dc233d27-83e6-11e8-881a-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:13:41.685: INFO: Waiting for pod pod-dc233d27-83e6-11e8-881a-28d244b00276 to disappear | |
Jul 9 19:13:41.727: INFO: Pod pod-dc233d27-83e6-11e8-881a-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:41.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-8rb45" for this suite. | |
Jul 9 19:13:47.906: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:52.792: INFO: namespace: e2e-tests-emptydir-8rb45, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:52.924: INFO: namespace e2e-tests-emptydir-8rb45 deletion completed in 11.151615087s | |
• [SLOW TEST:16.467 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
should support (non-root,0777,tmpfs) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[sig-network] Networking Granular Checks: Pods | |
should function for intra-pod communication: udp [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:00.523: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:02.562: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-gwdcd | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should function for intra-pod communication: udp [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-gwdcd | |
STEP: creating a selector | |
STEP: Creating the service pods in kubernetes | |
Jul 9 19:13:03.346: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable | |
STEP: Creating test pods | |
Jul 9 19:13:25.983: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.2.2.248:8080/dial?request=hostName&protocol=udp&host=10.2.2.243&port=8081&tries=1'] Namespace:e2e-tests-pod-network-test-gwdcd PodName:host-test-container-pod ContainerName:hostexec Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} | |
Jul 9 19:13:25.983: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:13:26.338: INFO: Waiting for endpoints: map[] | |
[AfterEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:26.338: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pod-network-test-gwdcd" for this suite. | |
Jul 9 19:13:48.527: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:13:52.820: INFO: namespace: e2e-tests-pod-network-test-gwdcd, resource: bindings, ignored listing per whitelist | |
Jul 9 19:13:53.066: INFO: namespace e2e-tests-pod-network-test-gwdcd deletion completed in 26.668327647s | |
• [SLOW TEST:52.543 seconds] | |
[sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:25 | |
Granular Checks: Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:28 | |
should function for intra-pod communication: udp [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:431 | |
Jul 9 19:13:53.079: INFO: This plugin does not implement NetworkPolicy. | |
[AfterEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:53.080: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] | |
NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:48 | |
when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:430 | |
should support allow-all policy [Feature:OSNetworkPolicy] [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:245 | |
Jul 9 19:13:53.079: This plugin does not implement NetworkPolicy. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs should respect image stream tag reference policy [Conformance] | |
resolve the image pull spec [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:272 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:47.786: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:13:49.617: INFO: configPath is now "/tmp/e2e-test-cli-deployment-vz7zx-user.kubeconfig" | |
Jul 9 19:13:49.618: INFO: The user is now "e2e-test-cli-deployment-vz7zx-user" | |
Jul 9 19:13:49.618: INFO: Creating project "e2e-test-cli-deployment-vz7zx" | |
Jul 9 19:13:49.769: INFO: Waiting on permissions in project "e2e-test-cli-deployment-vz7zx" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] resolve the image pull spec [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:272 | |
Jul 9 19:13:49.858: INFO: Running 'oc create --config=/tmp/e2e-test-cli-deployment-vz7zx-user.kubeconfig --namespace=e2e-test-cli-deployment-vz7zx -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/deployments/deployment-image-resolution-is.yaml' | |
imagestream.image.openshift.io "deployment-image-resolution" created | |
[AfterEach] should respect image stream tag reference policy [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:268 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:13:54.625: INFO: namespace : e2e-test-cli-deployment-vz7zx api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:14:00.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:12.916 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
should respect image stream tag reference policy [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:266 | |
resolve the image pull spec [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:272 | |
------------------------------ | |
[k8s.io] Pods | |
should be submitted and removed [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:39.764: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:41.217: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pods-jq7dk | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/pods.go:127 | |
[It] should be submitted and removed [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: creating the pod | |
STEP: setting up watch | |
STEP: submitting the pod to kubernetes | |
STEP: verifying the pod is in kubernetes | |
STEP: verifying pod creation was observed | |
Jul 9 19:13:44.065: INFO: running pod: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-submit-remove-dd98635e-83e6-11e8-bd2e-28d244b00276", GenerateName:"", Namespace:"e2e-tests-pods-jq7dk", SelfLink:"/api/v1/namespaces/e2e-tests-pods-jq7dk/pods/pod-submit-remove-dd98635e-83e6-11e8-bd2e-28d244b00276", UID:"ddae555c-83e6-11e8-84c6-0af96768d57e", ResourceVersion:"73676", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63666785621, loc:(*time.Location)(0x6b11480)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"time":"826854293", "name":"foo"}, Annotations:map[string]string{"openshift.io/scc":"anyuid"}, OwnerReferences:[]v1.OwnerReference(nil), Initializers:(*v1.Initializers)(nil), Finalizers:[]string(nil), ClusterName:""}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"default-token-ttjmn", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(0xc421245a00), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"nginx", Image:"k8s.gcr.io/nginx-slim-amd64:0.20", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"default-token-ttjmn", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil)}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc421245a80), Stdin:false, StdinOnce:false, TTY:false}}, RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc421034538), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"ip-10-0-130-54.us-west-2.compute.internal", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc421245d40), ImagePullSecrets:[]v1.LocalObjectReference{v1.LocalObjectReference{Name:"default-dockercfg-nrch4"}}, Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration(nil), HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(nil), DNSConfig:(*v1.PodDNSConfig)(nil)}, Status:v1.PodStatus{Phase:"Running", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666785621, loc:(*time.Location)(0x6b11480)}}, Reason:"", Message:""}, v1.PodCondition{Type:"Ready", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666785623, loc:(*time.Location)(0x6b11480)}}, Reason:"", Message:""}, v1.PodCondition{Type:"ContainersReady", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, Reason:"", Message:""}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, LastTransitionTime:v1.Time{Time:time.Time{wall:0x0, ext:63666785621, loc:(*time.Location)(0x6b11480)}}, Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"10.0.130.54", PodIP:"10.2.2.9", StartTime:(*v1.Time)(0xc421962340), InitContainerStatuses:[]v1.ContainerStatus(nil), ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"nginx", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(0xc421962360), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:true, RestartCount:0, Image:"k8s.gcr.io/nginx-slim-amd64:0.20", ImageID:"docker-pullable://k8s.gcr.io/nginx-slim-amd64@sha256:6654db6d4028756062edac466454ee5c9cf9b20ef79e35a81e3c840031eb1e2b", ContainerID:"docker://a721f4ab308b2de0f53691e6f5d4ef5382b19d4044df015d2b6a98b0ecb64ed2"}}, QOSClass:"BestEffort"}} | |
STEP: deleting the pod gracefully | |
STEP: verifying the kubelet observed the termination notice | |
STEP: verifying pod deletion was observed | |
[AfterEach] [k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:13:53.213: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pods-jq7dk" for this suite. | |
Jul 9 19:13:59.356: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:14:02.828: INFO: namespace: e2e-tests-pods-jq7dk, resource: bindings, ignored listing per whitelist | |
Jul 9 19:14:03.321: INFO: namespace e2e-tests-pods-jq7dk deletion completed in 10.074916195s | |
• [SLOW TEST:23.557 seconds] | |
[k8s.io] Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should be submitted and removed [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] EmptyDir volumes when FSGroup is specified | |
new files should be created with FSGroup ownership when container is non-root [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:49 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:14:00.704: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:14:02.510: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-dt8nh | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] new files should be created with FSGroup ownership when container is non-root [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:49 | |
STEP: Creating a pod to test emptydir 0644 on tmpfs | |
Jul 9 19:14:03.362: INFO: Waiting up to 5m0s for pod "pod-ea682ef8-83e6-11e8-8fe2-28d244b00276" in namespace "e2e-tests-emptydir-dt8nh" to be "success or failure" | |
Jul 9 19:14:03.404: INFO: Pod "pod-ea682ef8-83e6-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 41.886007ms | |
Jul 9 19:14:05.443: INFO: Pod "pod-ea682ef8-83e6-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.081038301s | |
STEP: Saw pod success | |
Jul 9 19:14:05.443: INFO: Pod "pod-ea682ef8-83e6-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:14:05.488: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-ea682ef8-83e6-11e8-8fe2-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:14:05.567: INFO: Waiting for pod pod-ea682ef8-83e6-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:14:05.599: INFO: Pod pod-ea682ef8-83e6-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:14:05.599: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-dt8nh" for this suite. | |
Jul 9 19:14:11.750: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:14:14.948: INFO: namespace: e2e-tests-emptydir-dt8nh, resource: bindings, ignored listing per whitelist | |
Jul 9 19:14:15.767: INFO: namespace e2e-tests-emptydir-dt8nh deletion completed in 10.11986951s | |
• [SLOW TEST:15.063 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
when FSGroup is specified | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:44 | |
new files should be created with FSGroup ownership when container is non-root [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:49 | |
------------------------------ | |
[Feature:AnnotationTrigger] Annotation trigger | |
reconciles after the image is overwritten [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/trigger/annotation.go:29 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:AnnotationTrigger] Annotation trigger | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:37.623: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:AnnotationTrigger] Annotation trigger | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:13:39.183: INFO: configPath is now "/tmp/e2e-test-cli-deployment-2hv6q-user.kubeconfig" | |
Jul 9 19:13:39.183: INFO: The user is now "e2e-test-cli-deployment-2hv6q-user" | |
Jul 9 19:13:39.183: INFO: Creating project "e2e-test-cli-deployment-2hv6q" | |
Jul 9 19:13:39.325: INFO: Waiting on permissions in project "e2e-test-cli-deployment-2hv6q" ... | |
[It] reconciles after the image is overwritten [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/trigger/annotation.go:29 | |
STEP: creating a Deployment | |
STEP: tagging the docker.io/library/centos:latest as test:v1 image to create ImageStream | |
Jul 9 19:13:39.422: INFO: Running 'oc tag --config=/tmp/e2e-test-cli-deployment-2hv6q-user.kubeconfig --namespace=e2e-test-cli-deployment-2hv6q docker.io/library/centos:latest test:v1' | |
Jul 9 19:13:39.670: INFO: Tag test:v1 set to docker.io/library/centos:latest. | |
STEP: waiting for the initial image to be replaced from ImageStream | |
STEP: setting Deployment image repeatedly to ' ' to fight with annotation trigger | |
STEP: waiting for the image to be injected by annotation trigger | |
[AfterEach] [Feature:AnnotationTrigger] Annotation trigger | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:13:43.532: INFO: namespace : e2e-test-cli-deployment-2hv6q api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:AnnotationTrigger] Annotation trigger | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:14:25.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:47.973 seconds] | |
[Feature:AnnotationTrigger] Annotation trigger | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/trigger/annotation.go:20 | |
reconciles after the image is overwritten [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/trigger/annotation.go:29 | |
------------------------------ | |
SS | |
------------------------------ | |
[sig-storage] Downward API volume | |
should provide podname as non-root with fsgroup and defaultMode [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:100 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:14:25.600: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:14:27.153: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-hhn2q | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should provide podname as non-root with fsgroup and defaultMode [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:100 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:14:27.731: INFO: Waiting up to 5m0s for pod "metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276" in namespace "e2e-tests-downward-api-hhn2q" to be "success or failure" | |
Jul 9 19:14:27.776: INFO: Pod "metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 45.181634ms | |
Jul 9 19:14:29.816: INFO: Pod "metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.085654956s | |
Jul 9 19:14:31.847: INFO: Pod "metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.11590508s | |
STEP: Saw pod success | |
Jul 9 19:14:31.847: INFO: Pod "metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:14:31.881: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:14:31.949: INFO: Waiting for pod metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:14:31.981: INFO: Pod metadata-volume-f8eebb12-83e6-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:14:31.981: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-hhn2q" for this suite. | |
Jul 9 19:14:38.114: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:14:39.634: INFO: namespace: e2e-tests-downward-api-hhn2q, resource: bindings, ignored listing per whitelist | |
Jul 9 19:14:41.402: INFO: namespace e2e-tests-downward-api-hhn2q deletion completed in 9.382069444s | |
• [SLOW TEST:15.803 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should provide podname as non-root with fsgroup and defaultMode [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:100 | |
------------------------------ | |
[sig-network] Networking Granular Checks: Pods | |
should function for node-pod communication: http [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:53.082: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:13:55.129: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-pod-network-test-psv7q | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should function for node-pod communication: http [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Performing setup for networking test in namespace e2e-tests-pod-network-test-psv7q | |
STEP: creating a selector | |
STEP: Creating the service pods in kubernetes | |
Jul 9 19:13:55.838: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable | |
STEP: Creating test pods | |
Jul 9 19:14:16.629: INFO: ExecWithOptions {Command:[/bin/sh -c timeout -t 15 curl -g -q -s --connect-timeout 1 http://10.2.2.11:8080/hostName | grep -v '^\s*$'] Namespace:e2e-tests-pod-network-test-psv7q PodName:host-test-container-pod ContainerName:hostexec Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} | |
Jul 9 19:14:16.629: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:14:16.952: INFO: Found all expected endpoints: [netserver-0] | |
[AfterEach] [sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:14:16.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-pod-network-test-psv7q" for this suite. | |
Jul 9 19:14:39.104: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:14:41.446: INFO: namespace: e2e-tests-pod-network-test-psv7q, resource: bindings, ignored listing per whitelist | |
Jul 9 19:14:43.230: INFO: namespace e2e-tests-pod-network-test-psv7q deletion completed in 26.238510831s | |
• [SLOW TEST:50.148 seconds] | |
[sig-network] Networking | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:25 | |
Granular Checks: Pods | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/networking.go:28 | |
should function for node-pod communication: http [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs with multiple image change triggers [Conformance] | |
should run a successful deployment with a trigger used by different containers [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:522 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:14:15.770: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:14:17.604: INFO: configPath is now "/tmp/e2e-test-cli-deployment-42zb4-user.kubeconfig" | |
Jul 9 19:14:17.604: INFO: The user is now "e2e-test-cli-deployment-42zb4-user" | |
Jul 9 19:14:17.604: INFO: Creating project "e2e-test-cli-deployment-42zb4" | |
Jul 9 19:14:17.731: INFO: Waiting on permissions in project "e2e-test-cli-deployment-42zb4" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should run a successful deployment with a trigger used by different containers [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:522 | |
STEP: verifying the deployment is marked complete | |
Jul 9 19:14:25.461: INFO: Latest rollout of dc/example (rc/example-1) is complete. | |
[AfterEach] with multiple image change triggers [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:509 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:14:27.527: INFO: namespace : e2e-test-cli-deployment-42zb4 api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:07.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:51.836 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
with multiple image change triggers [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:507 | |
should run a successful deployment with a trigger used by different containers [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:522 | |
------------------------------ | |
[Feature:Builds][Conformance] oc new-app | |
should succeed with a --name of 58 characters [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:49 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:14:03.322: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:14:05.052: INFO: configPath is now "/tmp/e2e-test-new-app-xn8nh-user.kubeconfig" | |
Jul 9 19:14:05.052: INFO: The user is now "e2e-test-new-app-xn8nh-user" | |
Jul 9 19:14:05.052: INFO: Creating project "e2e-test-new-app-xn8nh" | |
Jul 9 19:14:05.274: INFO: Waiting on permissions in project "e2e-test-new-app-xn8nh" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:26 | |
Jul 9 19:14:05.347: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:30 | |
STEP: waiting for builder service account | |
STEP: waiting for openshift namespace imagestreams | |
Jul 9 19:14:05.505: INFO: Running scan #0 | |
Jul 9 19:14:05.505: INFO: Checking language ruby | |
Jul 9 19:14:05.552: INFO: Checking tag 2.0 | |
Jul 9 19:14:05.552: INFO: Checking tag 2.2 | |
Jul 9 19:14:05.552: INFO: Checking tag 2.3 | |
Jul 9 19:14:05.552: INFO: Checking tag 2.4 | |
Jul 9 19:14:05.552: INFO: Checking tag 2.5 | |
Jul 9 19:14:05.552: INFO: Checking tag latest | |
Jul 9 19:14:05.552: INFO: Checking language nodejs | |
Jul 9 19:14:05.596: INFO: Checking tag 0.10 | |
Jul 9 19:14:05.596: INFO: Checking tag 4 | |
Jul 9 19:14:05.596: INFO: Checking tag 6 | |
Jul 9 19:14:05.596: INFO: Checking tag 8 | |
Jul 9 19:14:05.596: INFO: Checking tag latest | |
Jul 9 19:14:05.596: INFO: Checking language perl | |
Jul 9 19:14:05.647: INFO: Checking tag 5.20 | |
Jul 9 19:14:05.647: INFO: Checking tag 5.24 | |
Jul 9 19:14:05.647: INFO: Checking tag latest | |
Jul 9 19:14:05.647: INFO: Checking tag 5.16 | |
Jul 9 19:14:05.647: INFO: Checking language php | |
Jul 9 19:14:05.689: INFO: Checking tag 7.1 | |
Jul 9 19:14:05.689: INFO: Checking tag latest | |
Jul 9 19:14:05.689: INFO: Checking tag 5.5 | |
Jul 9 19:14:05.689: INFO: Checking tag 5.6 | |
Jul 9 19:14:05.689: INFO: Checking tag 7.0 | |
Jul 9 19:14:05.689: INFO: Checking language python | |
Jul 9 19:14:05.740: INFO: Checking tag latest | |
Jul 9 19:14:05.740: INFO: Checking tag 2.7 | |
Jul 9 19:14:05.740: INFO: Checking tag 3.3 | |
Jul 9 19:14:05.740: INFO: Checking tag 3.4 | |
Jul 9 19:14:05.740: INFO: Checking tag 3.5 | |
Jul 9 19:14:05.740: INFO: Checking tag 3.6 | |
Jul 9 19:14:05.740: INFO: Checking language wildfly | |
Jul 9 19:14:05.783: INFO: Checking tag latest | |
Jul 9 19:14:05.783: INFO: Checking tag 10.0 | |
Jul 9 19:14:05.783: INFO: Checking tag 10.1 | |
Jul 9 19:14:05.783: INFO: Checking tag 11.0 | |
Jul 9 19:14:05.783: INFO: Checking tag 12.0 | |
Jul 9 19:14:05.783: INFO: Checking tag 8.1 | |
Jul 9 19:14:05.783: INFO: Checking tag 9.0 | |
Jul 9 19:14:05.783: INFO: Checking language mysql | |
Jul 9 19:14:05.830: INFO: Checking tag 5.5 | |
Jul 9 19:14:05.830: INFO: Checking tag 5.6 | |
Jul 9 19:14:05.830: INFO: Checking tag 5.7 | |
Jul 9 19:14:05.830: INFO: Checking tag latest | |
Jul 9 19:14:05.830: INFO: Checking language postgresql | |
Jul 9 19:14:05.879: INFO: Checking tag 9.4 | |
Jul 9 19:14:05.879: INFO: Checking tag 9.5 | |
Jul 9 19:14:05.879: INFO: Checking tag 9.6 | |
Jul 9 19:14:05.879: INFO: Checking tag latest | |
Jul 9 19:14:05.879: INFO: Checking tag 9.2 | |
Jul 9 19:14:05.879: INFO: Checking language mongodb | |
Jul 9 19:14:05.930: INFO: Checking tag 2.6 | |
Jul 9 19:14:05.931: INFO: Checking tag 3.2 | |
Jul 9 19:14:05.931: INFO: Checking tag 3.4 | |
Jul 9 19:14:05.931: INFO: Checking tag latest | |
Jul 9 19:14:05.931: INFO: Checking tag 2.4 | |
Jul 9 19:14:05.931: INFO: Checking language jenkins | |
Jul 9 19:14:05.973: INFO: Checking tag 1 | |
Jul 9 19:14:05.973: INFO: Checking tag 2 | |
Jul 9 19:14:05.973: INFO: Checking tag latest | |
Jul 9 19:14:05.973: INFO: Success! | |
[It] should succeed with a --name of 58 characters [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:49 | |
STEP: calling oc new-app | |
Jul 9 19:14:05.973: INFO: Running 'oc new-app --config=/tmp/e2e-test-new-app-xn8nh-user.kubeconfig --namespace=e2e-test-new-app-xn8nh https://github.com/openshift/nodejs-ex --name a234567890123456789012345678901234567890123456789012345678' | |
--> Found image 5c36a77 (2 weeks old) in image stream "openshift/nodejs" under tag "8" for "nodejs" | |
Node.js 8 | |
--------- | |
Node.js 8 available as container is a base platform for building and running various Node.js 8 applications and frameworks. Node.js is a platform built on Chrome's JavaScript runtime for easily building fast, scalable network applications. Node.js uses an event-driven, non-blocking I/O model that makes it lightweight and efficient, perfect for data-intensive real-time applications that run across distributed devices. | |
Tags: builder, nodejs, nodejs8 | |
* The source repository appears to match: nodejs | |
* A source build using source code from https://github.com/openshift/nodejs-ex will be created | |
* The resulting image will be pushed to image stream "a234567890123456789012345678901234567890123456789012345678:latest" | |
* Use 'start-build' to trigger a new build | |
* This image will be deployed in deployment config "a234567890123456789012345678901234567890123456789012345678" | |
* Port 8080/tcp will be load balanced by service "a234567890123456789012345678901234567890123456789012345678" | |
* Other containers can access this service through the hostname "a234567890123456789012345678901234567890123456789012345678" | |
--> Creating resources ... | |
imagestream "a234567890123456789012345678901234567890123456789012345678" created | |
buildconfig "a234567890123456789012345678901234567890123456789012345678" created | |
deploymentconfig "a234567890123456789012345678901234567890123456789012345678" created | |
service "a234567890123456789012345678901234567890123456789012345678" created | |
--> Success | |
Build scheduled, use 'oc logs -f bc/a234567890123456789012345678901234567890123456789012345678' to track its progress. | |
Application is not exposed. You can expose services to the outside world by executing one or more of the commands below: | |
'oc expose svc/a234567890123456789012345678901234567890123456789012345678' | |
Run 'oc status' to view your app. | |
STEP: waiting for the build to complete | |
STEP: waiting for the deployment to complete | |
Jul 9 19:14:45.592: INFO: waiting for deploymentconfig e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678 to be available with version 1 | |
Jul 9 19:14:49.680: INFO: deploymentconfig e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678 available after 4.08804044s | |
pods: a23456789012345678901234567890123456789012345678901234567895p48 | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:40 | |
[AfterEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:14:49.742: INFO: namespace : e2e-test-new-app-xn8nh api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:11.772: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:68.485 seconds] | |
[Feature:Builds][Conformance] oc new-app | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:16 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:24 | |
should succeed with a --name of 58 characters [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/new_app.go:49 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs when tagging images [Conformance] | |
should successfully tag the deployed image [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:441 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:14:43.232: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:14:45.156: INFO: configPath is now "/tmp/e2e-test-cli-deployment-j6sbn-user.kubeconfig" | |
Jul 9 19:14:45.156: INFO: The user is now "e2e-test-cli-deployment-j6sbn-user" | |
Jul 9 19:14:45.156: INFO: Creating project "e2e-test-cli-deployment-j6sbn" | |
Jul 9 19:14:45.297: INFO: Waiting on permissions in project "e2e-test-cli-deployment-j6sbn" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should successfully tag the deployed image [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:441 | |
STEP: creating the deployment config fixture | |
STEP: verifying the deployment is marked complete | |
Jul 9 19:14:54.877: INFO: Latest rollout of dc/tag-images (rc/tag-images-1) is complete. | |
STEP: verifying the deployer service account can update imagestreamtags and user can get them | |
STEP: verifying the post deployment action happened: tag is set | |
[AfterEach] when tagging images [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:437 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:14:57.102: INFO: namespace : e2e-test-cli-deployment-j6sbn api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:19.155: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:35.977 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
when tagging images [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:435 | |
should successfully tag the deployed image [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:441 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
should fail resolving unresolvable valueFrom in sti build environment variable references [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:105 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:15:07.607: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:15:09.399: INFO: configPath is now "/tmp/e2e-test-build-valuefrom-6mbp7-user.kubeconfig" | |
Jul 9 19:15:09.399: INFO: The user is now "e2e-test-build-valuefrom-6mbp7-user" | |
Jul 9 19:15:09.399: INFO: Creating project "e2e-test-build-valuefrom-6mbp7" | |
Jul 9 19:15:09.541: INFO: Waiting on permissions in project "e2e-test-build-valuefrom-6mbp7" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:27 | |
Jul 9 19:15:09.595: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:38 | |
STEP: waiting for builder service account | |
STEP: waiting for openshift namespace imagestreams | |
Jul 9 19:15:09.731: INFO: Running scan #0 | |
Jul 9 19:15:09.731: INFO: Checking language ruby | |
Jul 9 19:15:09.768: INFO: Checking tag 2.0 | |
Jul 9 19:15:09.768: INFO: Checking tag 2.2 | |
Jul 9 19:15:09.768: INFO: Checking tag 2.3 | |
Jul 9 19:15:09.768: INFO: Checking tag 2.4 | |
Jul 9 19:15:09.768: INFO: Checking tag 2.5 | |
Jul 9 19:15:09.768: INFO: Checking tag latest | |
Jul 9 19:15:09.768: INFO: Checking language nodejs | |
Jul 9 19:15:09.812: INFO: Checking tag 0.10 | |
Jul 9 19:15:09.812: INFO: Checking tag 4 | |
Jul 9 19:15:09.812: INFO: Checking tag 6 | |
Jul 9 19:15:09.812: INFO: Checking tag 8 | |
Jul 9 19:15:09.812: INFO: Checking tag latest | |
Jul 9 19:15:09.812: INFO: Checking language perl | |
Jul 9 19:15:09.850: INFO: Checking tag 5.16 | |
Jul 9 19:15:09.850: INFO: Checking tag 5.20 | |
Jul 9 19:15:09.850: INFO: Checking tag 5.24 | |
Jul 9 19:15:09.850: INFO: Checking tag latest | |
Jul 9 19:15:09.850: INFO: Checking language php | |
Jul 9 19:15:09.892: INFO: Checking tag latest | |
Jul 9 19:15:09.892: INFO: Checking tag 5.5 | |
Jul 9 19:15:09.892: INFO: Checking tag 5.6 | |
Jul 9 19:15:09.892: INFO: Checking tag 7.0 | |
Jul 9 19:15:09.892: INFO: Checking tag 7.1 | |
Jul 9 19:15:09.892: INFO: Checking language python | |
Jul 9 19:15:09.928: INFO: Checking tag 2.7 | |
Jul 9 19:15:09.928: INFO: Checking tag 3.3 | |
Jul 9 19:15:09.928: INFO: Checking tag 3.4 | |
Jul 9 19:15:09.928: INFO: Checking tag 3.5 | |
Jul 9 19:15:09.928: INFO: Checking tag 3.6 | |
Jul 9 19:15:09.928: INFO: Checking tag latest | |
Jul 9 19:15:09.928: INFO: Checking language wildfly | |
Jul 9 19:15:09.966: INFO: Checking tag latest | |
Jul 9 19:15:09.966: INFO: Checking tag 10.0 | |
Jul 9 19:15:09.966: INFO: Checking tag 10.1 | |
Jul 9 19:15:09.966: INFO: Checking tag 11.0 | |
Jul 9 19:15:09.966: INFO: Checking tag 12.0 | |
Jul 9 19:15:09.966: INFO: Checking tag 8.1 | |
Jul 9 19:15:09.966: INFO: Checking tag 9.0 | |
Jul 9 19:15:09.966: INFO: Checking language mysql | |
Jul 9 19:15:10.002: INFO: Checking tag 5.5 | |
Jul 9 19:15:10.002: INFO: Checking tag 5.6 | |
Jul 9 19:15:10.002: INFO: Checking tag 5.7 | |
Jul 9 19:15:10.002: INFO: Checking tag latest | |
Jul 9 19:15:10.002: INFO: Checking language postgresql | |
Jul 9 19:15:10.041: INFO: Checking tag 9.5 | |
Jul 9 19:15:10.041: INFO: Checking tag 9.6 | |
Jul 9 19:15:10.041: INFO: Checking tag latest | |
Jul 9 19:15:10.041: INFO: Checking tag 9.2 | |
Jul 9 19:15:10.041: INFO: Checking tag 9.4 | |
Jul 9 19:15:10.041: INFO: Checking language mongodb | |
Jul 9 19:15:10.081: INFO: Checking tag 3.2 | |
Jul 9 19:15:10.081: INFO: Checking tag 3.4 | |
Jul 9 19:15:10.081: INFO: Checking tag latest | |
Jul 9 19:15:10.081: INFO: Checking tag 2.4 | |
Jul 9 19:15:10.081: INFO: Checking tag 2.6 | |
Jul 9 19:15:10.081: INFO: Checking language jenkins | |
Jul 9 19:15:10.115: INFO: Checking tag 1 | |
Jul 9 19:15:10.115: INFO: Checking tag 2 | |
Jul 9 19:15:10.115: INFO: Checking tag latest | |
Jul 9 19:15:10.115: INFO: Success! | |
STEP: creating test image stream | |
Jul 9 19:15:10.115: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-6mbp7-user.kubeconfig --namespace=e2e-test-build-valuefrom-6mbp7 -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/valuefrom/test-is.json' | |
imagestream.image.openshift.io "test" created | |
STEP: creating test secret | |
Jul 9 19:15:10.486: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-6mbp7-user.kubeconfig --namespace=e2e-test-build-valuefrom-6mbp7 -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/valuefrom/test-secret.yaml' | |
secret "mysecret" created | |
STEP: creating test configmap | |
Jul 9 19:15:10.765: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-6mbp7-user.kubeconfig --namespace=e2e-test-build-valuefrom-6mbp7 -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/valuefrom/test-configmap.yaml' | |
configmap "myconfigmap" created | |
[It] should fail resolving unresolvable valueFrom in sti build environment variable references [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:105 | |
STEP: creating test build config | |
Jul 9 19:15:11.083: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-6mbp7-user.kubeconfig --namespace=e2e-test-build-valuefrom-6mbp7 -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/valuefrom/failed-sti-build-value-from-config.yaml' | |
buildconfig.build.openshift.io "mys2itest" created | |
STEP: starting test build | |
Jul 9 19:15:11.382: INFO: Running 'oc start-build --config=/tmp/e2e-test-build-valuefrom-6mbp7-user.kubeconfig --namespace=e2e-test-build-valuefrom-6mbp7 mys2itest -o=name' | |
Jul 9 19:15:11.722: INFO: | |
start-build output with args [mys2itest -o=name]: | |
Error><nil> | |
StdOut> | |
build/mys2itest-1 | |
StdErr> | |
Jul 9 19:15:11.723: INFO: Waiting for mys2itest-1 to complete | |
Jul 9 19:15:17.798: INFO: WaitForABuild returning with error: The build "mys2itest-1" status is "Error" | |
Jul 9 19:15:17.798: INFO: Done waiting for mys2itest-1: util.BuildResult{BuildPath:"build/mys2itest-1", BuildName:"mys2itest-1", StartBuildStdErr:"", StartBuildStdOut:"build/mys2itest-1", StartBuildErr:error(nil), BuildConfigName:"", Build:(*build.Build)(0xc420eb8f00), BuildAttempt:true, BuildSuccess:false, BuildFailure:true, BuildCancelled:false, BuildTimeout:false, LogDumper:(util.LogDumperFunc)(nil), Oc:(*util.CLI)(0xc42078c1e0)} | |
with error: The build "mys2itest-1" status is "Error" | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:31 | |
[AfterEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:15:17.867: INFO: namespace : e2e-test-build-valuefrom-6mbp7 api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:23.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:16.351 seconds] | |
[Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:13 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:26 | |
should fail resolving unresolvable valueFrom in sti build environment variable references [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:105 | |
------------------------------ | |
[Conformance][Area:Networking][Feature:Router] The HAProxy router | |
should override the route host for overridden domains with a custom value [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:169 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:15:23.960: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:15:25.857: INFO: configPath is now "/tmp/e2e-test-router-scoped-65jwn-user.kubeconfig" | |
Jul 9 19:15:25.857: INFO: The user is now "e2e-test-router-scoped-65jwn-user" | |
Jul 9 19:15:25.857: INFO: Creating project "e2e-test-router-scoped-65jwn" | |
Jul 9 19:15:25.994: INFO: Waiting on permissions in project "e2e-test-router-scoped-65jwn" ... | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:48 | |
Jul 9 19:15:26.062: INFO: Running 'oc new-app --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-router-scoped-65jwn -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/scoped-router.yaml -p IMAGE=openshift/origin-haproxy-router' | |
--> Deploying template "e2e-test-router-scoped-65jwn/" for "/tmp/fixture-testdata-dir333495585/test/extended/testdata/scoped-router.yaml" to project e2e-test-router-scoped-65jwn | |
* With parameters: | |
* IMAGE=openshift/origin-haproxy-router | |
* SCOPE=["--name=test-scoped", "--namespace=$(POD_NAMESPACE)", "--loglevel=4", "--labels=select=first"] | |
--> Creating resources ... | |
pod "router-scoped" created | |
pod "router-override" created | |
pod "router-override-domains" created | |
rolebinding "system-router" created | |
route "route-1" created | |
route "route-2" created | |
route "route-override-domain-1" created | |
route "route-override-domain-2" created | |
service "endpoints" created | |
pod "endpoint-1" created | |
--> Success | |
Access your application via route 'first.example.com' | |
Access your application via route 'second.example.com' | |
Access your application via route 'y.a.null.ptr' | |
Access your application via route 'main.void.str' | |
Run 'oc status' to view your app. | |
[It] should override the route host for overridden domains with a custom value [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:169 | |
Jul 9 19:15:27.169: INFO: Creating new exec pod | |
STEP: creating a scoped router with overridden domains from a config file "/tmp/fixture-testdata-dir333495585/test/extended/testdata/scoped-router.yaml" | |
STEP: waiting for the healthz endpoint to respond | |
Jul 9 19:15:34.315: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-65jwn execpod -- /bin/sh -c | |
set -e | |
for i in $(seq 1 180); do | |
code=$( curl -k -s -o /dev/null -w '%{http_code}\n' --header 'Host: 10.2.2.35' "http://10.2.2.35:1936/healthz" ) || rc=$? | |
if [[ "${rc:-0}" -eq 0 ]]; then | |
echo $code | |
if [[ $code -eq 200 ]]; then | |
exit 0 | |
fi | |
if [[ $code -ne 503 ]]; then | |
exit 1 | |
fi | |
else | |
echo "error ${rc}" 1>&2 | |
fi | |
sleep 1 | |
done | |
' | |
Jul 9 19:15:34.940: INFO: stderr: "" | |
STEP: waiting for the valid route to respond | |
Jul 9 19:15:34.940: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-65jwn execpod -- /bin/sh -c | |
set -e | |
for i in $(seq 1 180); do | |
code=$( curl -k -s -o /dev/null -w '%{http_code}\n' --header 'Host: route-override-domain-1-e2e-test-router-scoped-65jwn.apps.veto.test' "http://10.2.2.35/Letter" ) || rc=$? | |
if [[ "${rc:-0}" -eq 0 ]]; then | |
echo $code | |
if [[ $code -eq 200 ]]; then | |
exit 0 | |
fi | |
if [[ $code -ne 503 ]]; then | |
exit 1 | |
fi | |
else | |
echo "error ${rc}" 1>&2 | |
fi | |
sleep 1 | |
done | |
' | |
Jul 9 19:15:35.607: INFO: stderr: "" | |
STEP: checking that the stored domain name does not match a route | |
Jul 9 19:15:35.607: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-65jwn execpod -- /bin/sh -c curl -s -o /dev/null -w '%{http_code}' --header 'Host: y.a.null.ptr' "http://10.2.2.35/Letter"' | |
Jul 9 19:15:36.246: INFO: stderr: "" | |
STEP: checking that route-override-domain-1-e2e-test-router-scoped-65jwn.apps.veto.test matches a route | |
Jul 9 19:15:36.246: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-65jwn execpod -- /bin/sh -c curl -s -o /dev/null -w '%{http_code}' --header 'Host: route-override-domain-1-e2e-test-router-scoped-65jwn.apps.veto.test' "http://10.2.2.35/Letter"' | |
Jul 9 19:15:36.960: INFO: stderr: "" | |
STEP: checking that route-override-domain-2-e2e-test-router-scoped-65jwn.apps.veto.test matches a route | |
Jul 9 19:15:36.960: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-router-scoped-65jwn execpod -- /bin/sh -c curl -s -o /dev/null -w '%{http_code}' --header 'Host: route-override-domain-2-e2e-test-router-scoped-65jwn.apps.veto.test' "http://10.2.2.35/Letter"' | |
Jul 9 19:15:37.613: INFO: stderr: "" | |
STEP: checking that the router reported the correct ingress and override | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:36 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:15:37.766: INFO: namespace : e2e-test-router-scoped-65jwn api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:51.819: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:27.904 seconds] | |
[Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:26 | |
The HAProxy router | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:67 | |
should override the route host for overridden domains with a custom value [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/scoped.go:169 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs viewing rollout history [Conformance] | |
should print the rollout history [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:602 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:14:41.404: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:14:43.030: INFO: configPath is now "/tmp/e2e-test-cli-deployment-bvczx-user.kubeconfig" | |
Jul 9 19:14:43.030: INFO: The user is now "e2e-test-cli-deployment-bvczx-user" | |
Jul 9 19:14:43.030: INFO: Creating project "e2e-test-cli-deployment-bvczx" | |
Jul 9 19:14:43.210: INFO: Waiting on permissions in project "e2e-test-cli-deployment-bvczx" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should print the rollout history [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:602 | |
STEP: waiting for the first rollout to complete | |
Jul 9 19:14:57.434: INFO: Latest rollout of dc/deployment-simple (rc/deployment-simple-1) is complete. | |
STEP: updating the deployment config in order to trigger a new rollout | |
STEP: waiting for the second rollout to complete | |
Jul 9 19:15:12.069: INFO: Latest rollout of dc/deployment-simple (rc/deployment-simple-2) is complete. | |
Jul 9 19:15:12.069: INFO: Running 'oc rollout --config=/tmp/e2e-test-cli-deployment-bvczx-user.kubeconfig --namespace=e2e-test-cli-deployment-bvczx history dc/deployment-simple' | |
STEP: checking the history for substrings | |
deploymentconfigs "deployment-simple" | |
REVISION STATUS CAUSE | |
1 Complete config change | |
2 Complete config change | |
[AfterEach] viewing rollout history [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:598 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:15:14.466: INFO: namespace : e2e-test-cli-deployment-bvczx api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:52.494: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:71.121 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
viewing rollout history [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:596 | |
should print the rollout history [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:602 | |
------------------------------ | |
[sig-storage] EmptyDir volumes | |
should support (non-root,0666,tmpfs) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:15:52.526: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:15:54.037: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-vj5sj | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should support (non-root,0666,tmpfs) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test emptydir 0666 on tmpfs | |
Jul 9 19:15:54.718: INFO: Waiting up to 5m0s for pod "pod-2cc7bf6f-83e7-11e8-8401-28d244b00276" in namespace "e2e-tests-emptydir-vj5sj" to be "success or failure" | |
Jul 9 19:15:54.748: INFO: Pod "pod-2cc7bf6f-83e7-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 29.638ms | |
Jul 9 19:15:56.780: INFO: Pod "pod-2cc7bf6f-83e7-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.062550676s | |
STEP: Saw pod success | |
Jul 9 19:15:56.780: INFO: Pod "pod-2cc7bf6f-83e7-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:15:56.929: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-2cc7bf6f-83e7-11e8-8401-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:15:57.118: INFO: Waiting for pod pod-2cc7bf6f-83e7-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:15:57.146: INFO: Pod pod-2cc7bf6f-83e7-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:57.146: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-vj5sj" for this suite. | |
Jul 9 19:16:03.264: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:16:04.671: INFO: namespace: e2e-tests-emptydir-vj5sj, resource: bindings, ignored listing per whitelist | |
Jul 9 19:16:06.533: INFO: namespace e2e-tests-emptydir-vj5sj deletion completed in 9.355363525s | |
• [SLOW TEST:14.008 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
should support (non-root,0666,tmpfs) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:431 | |
Jul 9 19:16:06.536: INFO: This plugin does not implement NetworkPolicy. | |
[AfterEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:06.537: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] | |
NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:48 | |
when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:430 | |
should enforce policy based on NamespaceSelector [Feature:OSNetworkPolicy] [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:282 | |
Jul 9 19:16:06.536: This plugin does not implement NetworkPolicy. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
[sig-storage] Downward API volume | |
should provide container's cpu limit [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:15:51.866: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:15:53.579: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-rj487 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should provide container's cpu limit [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:15:54.303: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2c87287f-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-downward-api-rj487" to be "success or failure" | |
Jul 9 19:15:54.377: INFO: Pod "downwardapi-volume-2c87287f-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 74.260454ms | |
Jul 9 19:15:56.621: INFO: Pod "downwardapi-volume-2c87287f-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.317905061s | |
STEP: Saw pod success | |
Jul 9 19:15:56.621: INFO: Pod "downwardapi-volume-2c87287f-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:15:56.683: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-2c87287f-83e7-11e8-8fe2-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:15:56.753: INFO: Waiting for pod downwardapi-volume-2c87287f-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:15:56.784: INFO: Pod downwardapi-volume-2c87287f-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:15:56.784: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-rj487" for this suite. | |
Jul 9 19:16:03.142: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:16:05.110: INFO: namespace: e2e-tests-downward-api-rj487, resource: bindings, ignored listing per whitelist | |
Jul 9 19:16:06.679: INFO: namespace e2e-tests-downward-api-rj487 deletion completed in 9.741913411s | |
• [SLOW TEST:14.813 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should provide container's cpu limit [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:06.538: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:08.064: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-pmhns | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating configMap with name projected-configmap-test-volume-35243cd3-83e7-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:16:08.783: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-352a5988-83e7-11e8-8401-28d244b00276" in namespace "e2e-tests-projected-pmhns" to be "success or failure" | |
Jul 9 19:16:08.837: INFO: Pod "pod-projected-configmaps-352a5988-83e7-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 53.581543ms | |
Jul 9 19:16:10.865: INFO: Pod "pod-projected-configmaps-352a5988-83e7-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.082002351s | |
STEP: Saw pod success | |
Jul 9 19:16:10.865: INFO: Pod "pod-projected-configmaps-352a5988-83e7-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:16:10.900: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-configmaps-352a5988-83e7-11e8-8401-28d244b00276 container projected-configmap-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:16:10.967: INFO: Waiting for pod pod-projected-configmaps-352a5988-83e7-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:16:10.996: INFO: Pod pod-projected-configmaps-352a5988-83e7-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:10.996: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-pmhns" for this suite. | |
Jul 9 19:16:17.132: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:16:19.587: INFO: namespace: e2e-tests-projected-pmhns, resource: bindings, ignored listing per whitelist | |
Jul 9 19:16:20.420: INFO: namespace e2e-tests-projected-pmhns deletion completed in 9.389700495s | |
• [SLOW TEST:13.882 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:419 | |
Jul 9 19:16:20.423: INFO: This plugin does not isolate namespaces by default. | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:20.423: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:20.423: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] | |
[Area:Networking] services | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/services.go:10 | |
when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:418 | |
should allow connections from pods in the default namespace to a service in another namespace on a different node [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/services.go:60 | |
Jul 9 19:16:20.423: This plugin does not isolate namespaces by default. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] EmptyDir volumes | |
should support (root,0777,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:06.680: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:08.428: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-pvdqq | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should support (root,0777,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test emptydir 0777 on node default medium | |
Jul 9 19:16:09.175: INFO: Waiting up to 5m0s for pod "pod-3564eb7c-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-emptydir-pvdqq" to be "success or failure" | |
Jul 9 19:16:09.206: INFO: Pod "pod-3564eb7c-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.751906ms | |
Jul 9 19:16:11.255: INFO: Pod "pod-3564eb7c-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.079832714s | |
Jul 9 19:16:13.287: INFO: Pod "pod-3564eb7c-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.111368217s | |
STEP: Saw pod success | |
Jul 9 19:16:13.287: INFO: Pod "pod-3564eb7c-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:16:13.322: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-3564eb7c-83e7-11e8-8fe2-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:16:13.405: INFO: Waiting for pod pod-3564eb7c-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:16:13.436: INFO: Pod pod-3564eb7c-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:13.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-pvdqq" for this suite. | |
Jul 9 19:16:19.572: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:16:21.562: INFO: namespace: e2e-tests-emptydir-pvdqq, resource: bindings, ignored listing per whitelist | |
Jul 9 19:16:23.503: INFO: namespace e2e-tests-emptydir-pvdqq deletion completed in 10.027196829s | |
• [SLOW TEST:16.823 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
should support (root,0777,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs with enhanced status [Conformance] | |
should include various info in status [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:539 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:15:19.211: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:15:21.149: INFO: configPath is now "/tmp/e2e-test-cli-deployment-25qpw-user.kubeconfig" | |
Jul 9 19:15:21.149: INFO: The user is now "e2e-test-cli-deployment-25qpw-user" | |
Jul 9 19:15:21.149: INFO: Creating project "e2e-test-cli-deployment-25qpw" | |
Jul 9 19:15:21.313: INFO: Waiting on permissions in project "e2e-test-cli-deployment-25qpw" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should include various info in status [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:539 | |
STEP: verifying the deployment is marked complete | |
Jul 9 19:15:34.015: INFO: Latest rollout of dc/deployment-simple (rc/deployment-simple-1) is complete. | |
STEP: verifying that status.replicas is set | |
Jul 9 19:15:34.015: INFO: Running 'oc get --config=/tmp/e2e-test-cli-deployment-25qpw-user.kubeconfig --namespace=e2e-test-cli-deployment-25qpw dc/deployment-simple --output=jsonpath="{.status.replicas}"' | |
STEP: verifying that status.updatedReplicas is set | |
Jul 9 19:15:34.240: INFO: Running 'oc get --config=/tmp/e2e-test-cli-deployment-25qpw-user.kubeconfig --namespace=e2e-test-cli-deployment-25qpw dc/deployment-simple --output=jsonpath="{.status.updatedReplicas}"' | |
STEP: verifying that status.availableReplicas is set | |
Jul 9 19:15:34.503: INFO: Running 'oc get --config=/tmp/e2e-test-cli-deployment-25qpw-user.kubeconfig --namespace=e2e-test-cli-deployment-25qpw dc/deployment-simple --output=jsonpath="{.status.availableReplicas}"' | |
STEP: verifying that status.unavailableReplicas is set | |
Jul 9 19:15:34.763: INFO: Running 'oc get --config=/tmp/e2e-test-cli-deployment-25qpw-user.kubeconfig --namespace=e2e-test-cli-deployment-25qpw dc/deployment-simple --output=jsonpath="{.status.unavailableReplicas}"' | |
[AfterEach] with enhanced status [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:535 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:15:37.115: INFO: namespace : e2e-test-cli-deployment-25qpw api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:29.193: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:70.063 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
with enhanced status [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:532 | |
should include various info in status [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:539 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:23.505: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:16:25.180: INFO: configPath is now "/tmp/e2e-test-router-stress-vn5rl-user.kubeconfig" | |
Jul 9 19:16:25.180: INFO: The user is now "e2e-test-router-stress-vn5rl-user" | |
Jul 9 19:16:25.180: INFO: Creating project "e2e-test-router-stress-vn5rl" | |
Jul 9 19:16:25.318: INFO: Waiting on permissions in project "e2e-test-router-stress-vn5rl" ... | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/stress.go:52 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/stress.go:40 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:16:25.456: INFO: namespace : e2e-test-router-stress-vn5rl api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:31.521: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [8.058 seconds] | |
[Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/stress.go:30 | |
The HAProxy router [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/stress.go:86 | |
converges when multiple routers are writing conflicting status [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/stress.go:168 | |
no router installed on the cluster | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/stress.go:57 | |
------------------------------ | |
[sig-storage] Secrets | |
should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:20.425: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:21.998: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-trb4p | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating secret with name secret-test-3d680430-83e7-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:16:22.645: INFO: Waiting up to 5m0s for pod "pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276" in namespace "e2e-tests-secrets-trb4p" to be "success or failure" | |
Jul 9 19:16:22.676: INFO: Pod "pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.889356ms | |
Jul 9 19:16:24.706: INFO: Pod "pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061005967s | |
Jul 9 19:16:26.735: INFO: Pod "pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.089959761s | |
STEP: Saw pod success | |
Jul 9 19:16:26.735: INFO: Pod "pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:16:26.775: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276 container secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:16:26.841: INFO: Waiting for pod pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:16:26.872: INFO: Pod pod-secrets-3d6cff39-83e7-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:26.872: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-trb4p" for this suite. | |
Jul 9 19:16:33.062: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:16:35.481: INFO: namespace: e2e-tests-secrets-trb4p, resource: bindings, ignored listing per whitelist | |
Jul 9 19:16:36.427: INFO: namespace e2e-tests-secrets-trb4p deletion completed in 9.462154922s | |
• [SLOW TEST:16.002 seconds] | |
[sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 | |
should be consumable from pods in volume with defaultMode set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs rolled back [Conformance] | |
should rollback to an older deployment [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:842 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:15:11.809: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:15:13.424: INFO: configPath is now "/tmp/e2e-test-cli-deployment-fb2n9-user.kubeconfig" | |
Jul 9 19:15:13.424: INFO: The user is now "e2e-test-cli-deployment-fb2n9-user" | |
Jul 9 19:15:13.424: INFO: Creating project "e2e-test-cli-deployment-fb2n9" | |
Jul 9 19:15:13.552: INFO: Waiting on permissions in project "e2e-test-cli-deployment-fb2n9" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should rollback to an older deployment [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:842 | |
Jul 9 19:15:27.148: INFO: Latest rollout of dc/deployment-simple (rc/deployment-simple-1) is complete. | |
Jul 9 19:15:27.148: INFO: Running 'oc rollout --config=/tmp/e2e-test-cli-deployment-fb2n9-user.kubeconfig --namespace=e2e-test-cli-deployment-fb2n9 latest deployment-simple' | |
STEP: verifying that we are on the second version | |
Jul 9 19:15:27.463: INFO: Running 'oc get --config=/tmp/e2e-test-cli-deployment-fb2n9-user.kubeconfig --namespace=e2e-test-cli-deployment-fb2n9 dc/deployment-simple --output=jsonpath="{.status.latestVersion}"' | |
Jul 9 19:15:42.267: INFO: Latest rollout of dc/deployment-simple (rc/deployment-simple-2) is complete. | |
STEP: verifying that we can rollback | |
Jul 9 19:15:42.267: INFO: Running 'oc rollout --config=/tmp/e2e-test-cli-deployment-fb2n9-user.kubeconfig --namespace=e2e-test-cli-deployment-fb2n9 undo dc/deployment-simple' | |
Jul 9 19:15:56.225: INFO: Latest rollout of dc/deployment-simple (rc/deployment-simple-3) is complete. | |
STEP: verifying that we are on the third version | |
Jul 9 19:15:56.225: INFO: Running 'oc get --config=/tmp/e2e-test-cli-deployment-fb2n9-user.kubeconfig --namespace=e2e-test-cli-deployment-fb2n9 dc/deployment-simple --output=jsonpath="{.status.latestVersion}"' | |
[AfterEach] rolled back [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:838 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:15:58.788: INFO: namespace : e2e-test-cli-deployment-fb2n9 api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:44.823: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:93.049 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
rolled back [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:836 | |
should rollback to an older deployment [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:842 | |
------------------------------ | |
[sig-storage] Projected | |
should provide podname as non-root with fsgroup [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:907 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:36.429: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:37.910: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-78tdl | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should provide podname as non-root with fsgroup [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:907 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:16:38.533: INFO: Waiting up to 5m0s for pod "metadata-volume-46e4ef9c-83e7-11e8-8401-28d244b00276" in namespace "e2e-tests-projected-78tdl" to be "success or failure" | |
Jul 9 19:16:38.566: INFO: Pod "metadata-volume-46e4ef9c-83e7-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 33.032762ms | |
Jul 9 19:16:40.599: INFO: Pod "metadata-volume-46e4ef9c-83e7-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.066337047s | |
STEP: Saw pod success | |
Jul 9 19:16:40.599: INFO: Pod "metadata-volume-46e4ef9c-83e7-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:16:40.628: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod metadata-volume-46e4ef9c-83e7-11e8-8401-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:16:40.698: INFO: Waiting for pod metadata-volume-46e4ef9c-83e7-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:16:40.726: INFO: Pod metadata-volume-46e4ef9c-83e7-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:40.726: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-78tdl" for this suite. | |
Jul 9 19:16:46.857: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:16:48.296: INFO: namespace: e2e-tests-projected-78tdl, resource: bindings, ignored listing per whitelist | |
Jul 9 19:16:50.296: INFO: namespace e2e-tests-projected-78tdl deletion completed in 9.537525694s | |
• [SLOW TEST:13.866 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should provide podname as non-root with fsgroup [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:907 | |
------------------------------ | |
S | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:44.859: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:46.729: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-hostpath-67sp9 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:37 | |
[It] should support existing directory subPath [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:121 | |
Jul 9 19:16:47.360: INFO: No SSH Key for provider : 'GetSigner(...) not implemented for ' | |
[AfterEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:47.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-hostpath-67sp9" for this suite. | |
Jul 9 19:16:53.507: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:16:55.401: INFO: namespace: e2e-tests-hostpath-67sp9, resource: bindings, ignored listing per whitelist | |
Jul 9 19:16:57.014: INFO: namespace e2e-tests-hostpath-67sp9 deletion completed in 9.602960881s | |
S [SKIPPING] [12.155 seconds] | |
[sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:34 | |
should support existing directory subPath [Suite:openshift/conformance/parallel] [Suite:k8s] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:121 | |
Jul 9 19:16:47.360: No SSH Key for provider : 'GetSigner(...) not implemented for ' | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:50.298: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:51.762: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-sz99l | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating projection with secret that has name projected-secret-test-4f2e0a0f-83e7-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:16:52.463: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4f330743-83e7-11e8-8401-28d244b00276" in namespace "e2e-tests-projected-sz99l" to be "success or failure" | |
Jul 9 19:16:52.494: INFO: Pod "pod-projected-secrets-4f330743-83e7-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.28511ms | |
Jul 9 19:16:54.554: INFO: Pod "pod-projected-secrets-4f330743-83e7-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.090533248s | |
STEP: Saw pod success | |
Jul 9 19:16:54.554: INFO: Pod "pod-projected-secrets-4f330743-83e7-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:16:54.607: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-secrets-4f330743-83e7-11e8-8401-28d244b00276 container projected-secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:16:54.693: INFO: Waiting for pod pod-projected-secrets-4f330743-83e7-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:16:54.721: INFO: Pod pod-projected-secrets-4f330743-83e7-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:16:54.721: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-sz99l" for this suite. | |
Jul 9 19:17:00.850: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:17:03.092: INFO: namespace: e2e-tests-projected-sz99l, resource: bindings, ignored listing per whitelist | |
Jul 9 19:17:05.032: INFO: namespace e2e-tests-projected-sz99l deletion completed in 10.271422297s | |
• [SLOW TEST:14.735 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] ConfigMap | |
should be consumable from pods in volume as non-root [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:57.017: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:58.550: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-h5zmm | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in volume as non-root [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating configMap with name configmap-test-volume-5343002a-83e7-11e8-bd2e-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:16:59.308: INFO: Waiting up to 5m0s for pod "pod-configmaps-53479f06-83e7-11e8-bd2e-28d244b00276" in namespace "e2e-tests-configmap-h5zmm" to be "success or failure" | |
Jul 9 19:16:59.342: INFO: Pod "pod-configmaps-53479f06-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 33.851956ms | |
Jul 9 19:17:01.372: INFO: Pod "pod-configmaps-53479f06-83e7-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.063354557s | |
STEP: Saw pod success | |
Jul 9 19:17:01.372: INFO: Pod "pod-configmaps-53479f06-83e7-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:17:01.401: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-configmaps-53479f06-83e7-11e8-bd2e-28d244b00276 container configmap-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:17:01.579: INFO: Waiting for pod pod-configmaps-53479f06-83e7-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:17:01.636: INFO: Pod pod-configmaps-53479f06-83e7-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:17:01.636: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-configmap-h5zmm" for this suite. | |
Jul 9 19:17:07.805: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:17:11.117: INFO: namespace: e2e-tests-configmap-h5zmm, resource: bindings, ignored listing per whitelist | |
Jul 9 19:17:11.562: INFO: namespace e2e-tests-configmap-h5zmm deletion completed in 9.885698637s | |
• [SLOW TEST:14.545 seconds] | |
[sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 | |
should be consumable from pods in volume as non-root [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[k8s.io] Sysctls | |
should reject invalid sysctls [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:142 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:29.275: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:16:31.583: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sysctl-vp8nj | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:56 | |
[It] should reject invalid sysctls [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:142 | |
STEP: Creating a pod with one valid and two invalid sysctls | |
[AfterEach] [k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Collecting events from namespace "e2e-tests-sysctl-vp8nj". | |
STEP: Found 1 events. | |
Jul 9 19:16:32.407: INFO: At 2018-07-09 19:16:32 -0700 PDT - event for sysctl-433830bf-83e7-11e8-992b-28d244b00276: {default-scheduler } Scheduled: Successfully assigned e2e-tests-sysctl-vp8nj/sysctl-433830bf-83e7-11e8-992b-28d244b00276 to ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:16:32.747: INFO: POD NODE PHASE GRACE CONDITIONS | |
Jul 9 19:16:32.747: INFO: registry-6559c8c4db-45526 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: deployment-simple-2-x8xwq ip-10-0-130-54.us-west-2.compute.internal Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:15:31 -0700 PDT } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:23 -0700 PDT ContainersNotReady containers with unready status: [myapp]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [myapp]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:15:31 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: deployment-simple-3-htj8x ip-10-0-130-54.us-west-2.compute.internal Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:15:45 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:15:52 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:15:45 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: execpod98j4h ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:16 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: frontend-1-build ip-10-0-130-54.us-west-2.compute.internal Succeeded [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:27 -0700 PDT PodCompleted } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:50 -0700 PDT PodCompleted } {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC PodCompleted } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:21 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: sysctl-433830bf-83e7-11e8-992b-28d244b00276 ip-10-0-130-54.us-west-2.compute.internal Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:32 -0700 PDT } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:32 -0700 PDT ContainersNotReady containers with unready status: [test-container]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [test-container]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:32 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-apiserver-cn2ps ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:45 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-controller-manager-558dc6fb98-q6vr5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:34 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-core-operator-75d546fbbb-c7ctx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:20 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-dns-787c975867-txmxv ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:22 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-flannel-bgv4g ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:59 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-flannel-m5wph ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:58 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-flannel-xcck7 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:17 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-proxy-5td7p ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:54 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-proxy-l2cnn ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-proxy-zsgcb ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-scheduler-68f8875b5c-s5tdr ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: metrics-server-5767bfc576-gfbwb ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: openshift-apiserver-rkms5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:19 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: openshift-controller-manager-99d6586b-qq685 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:55 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: pod-checkpointer-4882g ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:03 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: prometheus-0 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:40 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-network-operator-jwwmp ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:13 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-node-controller-2ctqd ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:08 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: webconsole-6698d4fbbc-rgsw2 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: default-http-backend-6985d557bb-8h44n ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: router-6796c95fdf-2k4wk ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:37 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:46 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: directory-sync-d84d84d9f-j7pr6 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: kube-addon-operator-675f99d7f8-c6pdt ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:29 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-alm-operator-79b6996f74-prs9h ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:35 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-channel-operator-5d878cd785-l66n4 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-clu-6b8d87785f-fswbx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-node-agent-r77mj ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:37:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-node-agent-rrwlg ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:12:57 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-stats-emitter-d87f669fd-988nl ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:29 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:36 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:23 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: tectonic-utility-operator-786b69fc8b-4xffz ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:41 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:16:32.747: INFO: | |
Jul 9 19:16:32.800: INFO: | |
Logging node info for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:16:32.837: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-130-54.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-130-54.us-west-2.compute.internal,UID:2f71bed0-83b7-11e8-84c6-0af96768d57e,ResourceVersion:76056,Generation:0,CreationTimestamp:2018-07-09 13:32:23 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-130-54,node-role.kubernetes.io/worker: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"3e:08:91:8f:b9:a5"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.130.54,node-configuration.v1.coreos.com/currentConfig: worker-2650561509,node-configuration.v1.coreos.com/desiredConfig: worker-2650561509,node-configuration.v1.coreos.com/targetConfig: worker-2650561509,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.2.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-0cb9cec2620663d39,Unschedulable:false,Taints:[],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365150208 0} {<nil>} 8169092Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260292608 0} {<nil>} 8066692Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:16:26 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:16:26 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:16:26 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:16:26 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:16:26 -0700 PDT 2018-07-09 13:33:23 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.130.54} {InternalDNS ip-10-0-130-54.us-west-2.compute.internal} {Hostname ip-10-0-130-54}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC283016-6CE7-ACE7-0F9A-02CE10505945,BootID:cfad64a2-03d7-403a-bd51-76866880a650,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[openshift/origin-haproxy-router@sha256:f0a71ada9e9ee48529540c2d4938b9caa55f9a0ac8a3be598e269ca5cebf70c0 openshift/origin-haproxy-router:v3.10.0-alpha.0] 1284960820} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test@sha256:6daa01a6f7f0784905bf9dcbce49826d73d7c3c1d62a802f875ee7c10db02960 docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test:latest] 613134454} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test@sha256:92c5e723d97318711a71afb9ee5c12c3c48b98d0f2aaa5e954095fabbcb505ee docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test:latest] 613133841} {[docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example@sha256:02d80c750d1e71afc7792f55f935c3dd6cde1788bee2b53ab554d29c903ca064 docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example:latest] 603384691} {[docker-registry.default.svc:5000/openshift/php@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7:latest] 589408618} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass@sha256:880359284c1e0933fe5f2db29b8c4d948b70da3dfb26a0462f68b23397740b0a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass:latest] 568094192} {[docker-registry.default.svc:5000/openshift/php@sha256:59c3d53372cd7097494187f5a58bab58a1d956a340b70a23c84a0d000a565cbe] 567254500} {[docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test@sha256:539e80a4de02794f6126cffce75562bcb721041c6d443c5ced15ba286d70e229 docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test:latest] 566117187} {[docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test@sha256:e0eeef684e9de55219871fa9e360d73a1163cfc407c626eade862cbee5a9bbc5 docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test:latest] 566117040} {[centos/ruby-22-centos7@sha256:a18c8706118a5c4c9f1adf045024d2abf06ba632b5674b23421019ee4d3edcae centos/ruby-22-centos7:latest] 566117040} {[centos/nodejs-6-centos7@sha256:b2867b5008d9e975b3d4710ec0f31cdc96b079b83334b17e03a60602a7a590fc] 560696751} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot@sha256:0397f7e12d87d62c539356a4936348d0a8deb40e1b5e970cdd1744d3e6ffa05a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot:latest] 560696751} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot@sha256:4084131a9910c10780186608faf5a9643de0f18d09c27fe828499a8d180abfba docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot:latest] 560696751} {[docker-registry.default.svc:5000/openshift/ruby@sha256:2e83b9e07e85960060096b6aff7ee202a5f52e0e18447641b080b1f3879e0901] 536571487} {[docker-registry.default.svc:5000/openshift/ruby@sha256:8f00b7a5789887b72db0415355830c87e18804b774a922a424736f5237a44933] 518934530} {[docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678@sha256:a9ecb5931f283c598dcaf3aca9025599eb71115bd0f2cd0f1989a9f37394efad docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678:latest] 511744495} {[docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample@sha256:95a78c60dc1709c2212cd8cc48cd3fffe6cdcdd847674497d9aa5d7891551699 docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample:latest] 511744370} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:3bb2aed7578ab5b6ba2bf22993df3c73ef91bdb02e273cc0ce8e529de7ee5660] 506453985} {[docker-registry.default.svc:5000/openshift/ruby@sha256:0eaaed9fae1b0d9bc8ed73b93d581c6ab019a92277484c9acf52fa60b3269a7c] 504578679} {[docker-registry.default.svc:5000/openshift/nodejs@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653 centos/nodejs-8-centos7@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653] 504452018} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:896482969cd659b419bc444c153a74d11820655c7ed19b5094b8eb041f0065d6] 487132847} {[openshift/origin-docker-builder@sha256:4fe8032f87d2f8485a711ec60a9ffb330e42a6cd8d232ad3cf63c42471cfab29 openshift/origin-docker-builder:latest] 447580928} {[docker-registry.default.svc:5000/openshift/mysql@sha256:d03537ef57d51b13e6ad4a73a382ca180a0e02d975c8237790410f45865aae3c] 429435940} {[openshift/origin-haproxy-router@sha256:485fa86ac97b0d289411b3216fb8970989cd580817ebb5fcbb0f83a6dc2466f5 openshift/origin-haproxy-router:latest] 394965919} {[openshift/origin-deployer@sha256:1295e5be56fc03d4c482194378a882f2e96a8d23eadaf6dd32d603d3e877df99 openshift/origin-deployer:latest] 371674595} {[openshift/origin-web-console@sha256:d2cbbb533d26996226add8cb327cb2060e7a03c6aa96ad94cd236d4064c094ce openshift/origin-web-console:latest] 336636057} {[openshift/prometheus@sha256:35e2e0efc874c055be60a025874256816c98b9cebc10f259d7fb806bbe68badf openshift/prometheus:v2.2.1] 317896379} {[openshift/origin-docker-registry@sha256:c40ebb707721327c3b9c79f0e8e7f02483f034355d4149479333cc134b72967c openshift/origin-docker-registry:latest] 302637209} {[openshift/origin-pod@sha256:8fbd41f21824f5981716568790c5f78a4710bb0709ce9c473eb21ad2fbc5e877 openshift/origin-pod:latest] 251747200} {[openshift/origin-base@sha256:43dd97db435025eee02606658cfcccbc0a8ac4135e0d8870e91930d6cab8d1fd openshift/origin-base:latest] 228695137} {[openshift/oauth-proxy@sha256:4b73830ee6f7447d0921eedc3946de50016eb8f048d66ea3969abc4116f1e42a openshift/oauth-proxy:v1.0.0] 228241928} {[openshift/prometheus-alertmanager@sha256:35443abf6c5cf99b080307fe0f98098334f299780537a3e61ac5604cbfe48f7e openshift/prometheus-alertmanager:v0.14.0] 221857684} {[openshift/prometheus-alert-buffer@sha256:076f8dd576806f5c2dde7e536d020c31aa7d2ec7dcea52da6cbb944895def7ba openshift/prometheus-alert-buffer:v0.0.2] 200521084} {[docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage@sha256:df3e69e3fe1bc86897717b020b6caa000f1f97c14dc0b3853ca0d7149412da54 docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage:v1] 199835207} {[centos@sha256:b67d21dfe609ddacf404589e04631d90a342921e81c40aeaf3391f6717fa5322 centos@sha256:eed5b251b615d1e70b10bcec578d64e8aa839d2785c2ffd5424e472818c42755 centos:7 centos:centos7] 199678471} {[docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1@sha256:3967cd8851952bbba0b3a4d9c038f36dc5001463c8521d6955ab0f3f4598d779 docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1:latest] 199678471} {[k8s.gcr.io/nginx-slim-amd64@sha256:6654db6d4028756062edac466454ee5c9cf9b20ef79e35a81e3c840031eb1e2b k8s.gcr.io/nginx-slim-amd64:0.20] 103591055} {[gcr.io/google_containers/metrics-server-amd64@sha256:54d2cf293e01f72d9be0e7c4f2c98e31f599088a9426a6415fe62426d446f5b2 gcr.io/google_containers/metrics-server-amd64:v0.2.0] 96501893} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/directory-sync@sha256:e5e7fe901868853d89c2c0697cc88f0686c6ba1178ca045ec57bfd18e7000048 quay.io/coreos/directory-sync:v0.0.2] 38433928} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[k8s.gcr.io/addon-resizer@sha256:d00afd42fc267fa3275a541083cfe67d160f966c788174b44597434760e1e1eb k8s.gcr.io/addon-resizer:2.1] 26450138} {[quay.io/coreos/tectonic-error-server@sha256:aefa0a012e103bee299c17e798e5830128588b6ef5d4d1f6bc8ae5804bc4d8cd quay.io/coreos/tectonic-error-server:1.1] 12714516} {[gcr.io/google_containers/dnsutils@sha256:cd9182f6d74e616942db1cef6f25e1e54b49ba0330c2e19d3ec061f027666cc0 gcr.io/google_containers/dnsutils:e2e] 8897789} {[gcr.io/kubernetes-e2e-test-images/hostexec-amd64@sha256:bdaecec5adfa7c79e9525c0992fdab36c2d68066f5e91eff0d1d9e8d73c654ea gcr.io/kubernetes-e2e-test-images/hostexec-amd64:1.1] 8407119} {[gcr.io/kubernetes-e2e-test-images/netexec-amd64@sha256:2edfad424a541b9e024f26368d3a5b7dcc1d7cd27a4ee8c1d8c3f81d9209ab2e gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0] 6227659} {[openshift/hello-openshift@sha256:aaea76ff622d2f8bcb32e538e7b3cd0ef6d291953f3e7c9f556c1ba5baf47e2e openshift/hello-openshift:latest] 6089990}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:16:32.837: INFO: | |
Logging kubelet events for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:16:32.875: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:16:32.991: INFO: kube-flannel-xcck7 started at 2018-07-09 13:32:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: metrics-server-5767bfc576-gfbwb started at 2018-07-09 13:33:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container metrics-server ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: Container metrics-server-nanny ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: execpod98j4h started at 2018-07-09 19:14:15 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container exec ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: tectonic-node-agent-rrwlg started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container node-agent ready: true, restart count 3 | |
Jul 9 19:16:32.991: INFO: directory-sync-d84d84d9f-j7pr6 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container directory-sync ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: webconsole-6698d4fbbc-rgsw2 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container webconsole ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: default-http-backend-6985d557bb-8h44n started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container default-http-backend ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: router-6796c95fdf-2k4wk started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container router ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: deployment-simple-3-htj8x started at 2018-07-09 19:15:45 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container myapp ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: frontend-1-build started at 2018-07-09 19:14:21 -0700 PDT (2+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Init container git-clone ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: Init container manage-dockerfile ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: Container sti-build ready: false, restart count 0 | |
Jul 9 19:16:32.991: INFO: kube-proxy-5td7p started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: registry-6559c8c4db-45526 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container registry ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: deployment-simple-2-x8xwq started at 2018-07-09 19:15:31 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container myapp ready: false, restart count 0 | |
Jul 9 19:16:32.991: INFO: prometheus-0 started at 2018-07-09 13:50:04 -0700 PDT (0+6 container statuses recorded) | |
Jul 9 19:16:32.991: INFO: Container alert-buffer ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: Container alertmanager ready: true, restart count 0 | |
Jul 9 19:16:32.991: INFO: Container alertmanager-proxy ready: true, restart count 0 | |
Jul 9 19:16:32.992: INFO: Container alerts-proxy ready: true, restart count 0 | |
Jul 9 19:16:32.992: INFO: Container prom-proxy ready: true, restart count 0 | |
Jul 9 19:16:32.992: INFO: Container prometheus ready: true, restart count 0 | |
Jul 9 19:16:32.992: INFO: sysctl-433830bf-83e7-11e8-992b-28d244b00276 started at 2018-07-09 19:16:32 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:16:32.992: INFO: Container test-container ready: false, restart count 0 | |
W0709 19:16:33.039551 11713 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:16:33.169: INFO: | |
Latency metrics for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:16:33.169: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:39.952369s} | |
Jul 9 19:16:33.169: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.138495s} | |
Jul 9 19:16:33.169: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:18.561353s} | |
Jul 9 19:16:33.169: INFO: | |
Logging node info for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:16:33.236: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-141-201.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-141-201.us-west-2.compute.internal,UID:ab76db34-83b4-11e8-8888-0af96768d57e,ResourceVersion:76069,Generation:0,CreationTimestamp:2018-07-09 13:14:22 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-141-201,node-role.kubernetes.io/etcd: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"b6:11:a8:d0:6d:85"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.141.201,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.1.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-03457d640f9c71dd1,Unschedulable:false,Taints:[{node-role.kubernetes.io/etcd NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:16:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:16:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:16:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:16:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:16:30 -0700 PDT 2018-07-09 13:16:04 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.141.201} {InternalDNS ip-10-0-141-201.us-west-2.compute.internal} {Hostname ip-10-0-141-201}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2F6BCA-4D59-F6AA-8C7B-027F94D52D78,BootID:92773d40-1311-4ad5-b294-38db65faf16c,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/kube-client-agent@sha256:8564ab65bcb1064006d2fc9c6e32a5ca3f4326cdd2da9a2efc4fb7cc0e0b6041 quay.io/coreos/kube-client-agent:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 33236131} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:16:33.236: INFO: | |
Logging kubelet events for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:16:33.286: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:17:03.379: INFO: Unable to retrieve kubelet pods for node ip-10-0-141-201.us-west-2.compute.internal: the server is currently unable to handle the request (get nodes ip-10-0-141-201.us-west-2.compute.internal:10250) | |
Jul 9 19:17:03.379: INFO: | |
Logging node info for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:17:03.416: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-35-213.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-35-213.us-west-2.compute.internal,UID:a83cf873-83b4-11e8-8888-0af96768d57e,ResourceVersion:76531,Generation:0,CreationTimestamp:2018-07-09 13:14:17 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2c,kubernetes.io/hostname: ip-10-0-35-213,node-role.kubernetes.io/master: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"5e:08:be:54:0d:9f"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.35.213,node-configuration.v1.coreos.com/currentConfig: master-2063737633,node-configuration.v1.coreos.com/desiredConfig: master-2063737633,node-configuration.v1.coreos.com/targetConfig: master-2063737633,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.0.0/24,ExternalID:,ProviderID:aws:///us-west-2c/i-0e1d36783c9705b28,Unschedulable:false,Taints:[{node-role.kubernetes.io/master NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:16:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:16:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:16:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:16:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:16:59 -0700 PDT 2018-07-09 13:16:08 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.35.213} {ExternalIP 34.220.249.237} {InternalDNS ip-10-0-35-213.us-west-2.compute.internal} {ExternalDNS ec2-34-220-249-237.us-west-2.compute.amazonaws.com} {Hostname ip-10-0-35-213}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2ED297-E036-AA0D-C4ED-9057B3EA9001,BootID:7f784e0b-09a6-495a-b787-3d8619214f8a,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[openshift/origin-hypershift@sha256:3b26011ae771a6036a7533d970052be5c04bc1f6e6812314ffefd902f40910fd openshift/origin-hypershift:latest] 518022163} {[openshift/origin-hyperkube@sha256:11a08060b48d226d64d4bb5234f2386bf22472a0835c5b91f0fb0db25b0a7e19 openshift/origin-hyperkube:latest] 498702039} {[quay.io/coreos/awscli@sha256:1d6ea2f37c248a4f4f2a70126f0b8555fd0804d4e65af3b30c3a949247ea13a6 quay.io/coreos/awscli:025a357f05242fdad6a81e8a6b520098aa65a600] 97521631} {[quay.io/coreos/bootkube@sha256:63afddd30deedff273d65607f4fcf0b331f4418838a00c69b6ab7a5754a24f5a quay.io/coreos/bootkube:v0.10.0] 84921995} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[gcr.io/google_containers/k8s-dns-kube-dns-amd64@sha256:6d8e0da4fb46e9ea2034a3f4cab0e095618a2ead78720c12e791342738e5f85d gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.8] 50456751} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/tectonic-stats@sha256:e800fe60dd1a0f89f8ae85caae9209201254e17d889d664d633ed08e274e2a39 quay.io/coreos/tectonic-stats:6e882361357fe4b773adbf279cddf48cb50164c1] 48779830} {[quay.io/coreos/pod-checkpointer@sha256:1e1e48228f872d56c8a57a5e12adb5239ae9e6206536baf2904e4bf03314c8e8 quay.io/coreos/pod-checkpointer:9dc83e1ab3bc36ca25c9f7c18ddef1b91d4a0558] 47992230} {[quay.io/coreos/tectonic-network-operator-dev@sha256:e29d797f5740cf6f5c0ccc0de2b3e606d187acbdc0bb79a4397c058d8840c8fe quay.io/coreos/tectonic-network-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44068170} {[quay.io/coreos/tectonic-node-controller-operator-dev@sha256:7a31568c6c2e398cffa7e8387cf51543e3bf1f01b4a050a5d00a9b593c3dace0 quay.io/coreos/tectonic-node-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44053165} {[quay.io/coreos/kube-addon-operator-dev@sha256:e327727a93813c31f6d65f76f2998722754b8ccb5110949153e55f2adbc2374e quay.io/coreos/kube-addon-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44052211} {[quay.io/coreos/tectonic-utility-operator-dev@sha256:4fb4de52c7aa64ce124e1bf73fb27989356c414101ecc19ca4ec9ab80e00a88d quay.io/coreos/tectonic-utility-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43818409} {[quay.io/coreos/tectonic-ingress-controller-operator-dev@sha256:5e96253c8fe8357473d4806b116fcf03fe18dcad466a88083f9b9310045821f1 quay.io/coreos/tectonic-ingress-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43808038} {[quay.io/coreos/tectonic-alm-operator@sha256:ce32e6d4745040be8807d09eb925b2b076b60fb0a93e33302b74a5cc8f294ca5 quay.io/coreos/tectonic-alm-operator:v0.3.1] 43202998} {[gcr.io/google_containers/k8s-dns-sidecar-amd64@sha256:23df717980b4aa08d2da6c4cfa327f1b730d92ec9cf740959d2d5911830d82fb gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.8] 42210862} {[gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64@sha256:93c827f018cf3322f1ff2aa80324a0306048b0a69bc274e423071fb0d2d29d8b gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.8] 40951779} {[quay.io/coreos/kube-core-operator-dev@sha256:6cc0dd2405f19014b41a0eed57c39160aeb92c2380ac8f8a067ce7dee476cba2 quay.io/coreos/kube-core-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40849618} {[quay.io/coreos/tectonic-channel-operator-dev@sha256:6eeb84c385333755a2189c199587bc26db6c5d897e1962d7e1047dec2531e85e quay.io/coreos/tectonic-channel-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40523592} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[quay.io/coreos/kube-core-renderer-dev@sha256:a595dfe57b7992971563fcea8ac1858c306529a465f9b690911f4220d93d3c5c quay.io/coreos/kube-core-renderer-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 36535818} {[quay.io/coreos/kube-etcd-signer-server@sha256:c4c0becf6779523af5b644b53375d61bed9c4688d496cb2f88d4f08024ac5390 quay.io/coreos/kube-etcd-signer-server:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 34655544} {[quay.io/coreos/tectonic-node-controller-dev@sha256:c9c17f7c4c738e519e36224ae8c71d3a881b92ffb86fdb75f358efebafa27d84 quay.io/coreos/tectonic-node-controller-dev:a437848532713f2fa4137e9a0f4f6a689cf554a8] 25570332} {[quay.io/coreos/tectonic-clu@sha256:4e6a907a433e741632c8f9a7d9d9009bc08ac494dce05e0a19f8fa0a440a3926 quay.io/coreos/tectonic-clu:v0.0.1] 5081911} {[quay.io/coreos/tectonic-stats-extender@sha256:6e7fe41ca2d63791c08d2cc4b4311d9e01b37fa3dc116d3e77e7306cbe29a0f1 quay.io/coreos/tectonic-stats-extender:487b3da4e175da96dabfb44fba65cdb8b823db2e] 2818916} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:17:03.417: INFO: | |
Logging kubelet events for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:17:03.452: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:17:03.641: INFO: kube-apiserver-cn2ps started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.641: INFO: Container kube-apiserver ready: true, restart count 4 | |
Jul 9 19:17:03.641: INFO: tectonic-node-controller-2ctqd started at 2018-07-09 13:18:05 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.641: INFO: Container tectonic-node-controller ready: true, restart count 0 | |
Jul 9 19:17:03.641: INFO: tectonic-alm-operator-79b6996f74-prs9h started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.641: INFO: Container tectonic-alm-operator ready: true, restart count 0 | |
Jul 9 19:17:03.641: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.641: INFO: Container tectonic-ingress-controller-operator ready: true, restart count 0 | |
Jul 9 19:17:03.641: INFO: tectonic-node-agent-r77mj started at 2018-07-09 13:19:20 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.641: INFO: Container node-agent ready: true, restart count 4 | |
Jul 9 19:17:03.642: INFO: pod-checkpointer-4882g started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container pod-checkpointer ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: kube-flannel-m5wph started at 2018-07-09 13:15:39 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: openshift-controller-manager-99d6586b-qq685 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container openshift-controller-manager ready: true, restart count 3 | |
Jul 9 19:17:03.642: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container tectonic-node-controller-operator ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: kube-core-operator-75d546fbbb-c7ctx started at 2018-07-09 13:18:11 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container kube-core-operator ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: tectonic-utility-operator-786b69fc8b-4xffz started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container tectonic-utility-operator ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: kube-addon-operator-675f99d7f8-c6pdt started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container kube-addon-operator ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal started at <nil> (0+0 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: kube-controller-manager-558dc6fb98-q6vr5 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container kube-controller-manager ready: true, restart count 1 | |
Jul 9 19:17:03.642: INFO: tectonic-stats-emitter-d87f669fd-988nl started at 2018-07-09 13:19:23 -0700 PDT (1+2 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Init container tectonic-stats-extender-init ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: Container tectonic-stats-emitter ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: Container tectonic-stats-extender ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: tectonic-channel-operator-5d878cd785-l66n4 started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container tectonic-channel-operator ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: kube-proxy-l2cnn started at 2018-07-09 13:14:22 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: openshift-apiserver-rkms5 started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container openshift-apiserver ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: tectonic-network-operator-jwwmp started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container tectonic-network-operator ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: kube-dns-787c975867-txmxv started at 2018-07-09 13:16:08 -0700 PDT (0+3 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container dnsmasq ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: Container kubedns ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: Container sidecar ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: kube-scheduler-68f8875b5c-s5tdr started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container kube-scheduler ready: true, restart count 0 | |
Jul 9 19:17:03.642: INFO: tectonic-clu-6b8d87785f-fswbx started at 2018-07-09 13:19:06 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:03.642: INFO: Container tectonic-clu ready: true, restart count 0 | |
W0709 19:17:03.683471 11713 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:17:03.841: INFO: | |
Latency metrics for node ip-10-0-35-213.us-west-2.compute.internal | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:17:03.884: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-sysctl-vp8nj" for this suite. | |
Jul 9 19:17:10.036: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:17:13.714: INFO: namespace: e2e-tests-sysctl-vp8nj, resource: bindings, ignored listing per whitelist | |
Jul 9 19:17:14.686: INFO: namespace e2e-tests-sysctl-vp8nj deletion completed in 10.762470175s | |
• Failure [45.411 seconds] | |
[k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should reject invalid sysctls [Suite:openshift/conformance/parallel] [Suite:k8s] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:142 | |
Expected | |
<nil>: nil | |
not to be nil | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:177 | |
------------------------------ | |
S | |
------------------------------ | |
[Conformance][templates] templateinstance impersonation tests | |
should pass impersonation creation tests [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:231 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:05.033: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:17:06.734: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-user.kubeconfig" | |
Jul 9 19:17:06.734: INFO: The user is now "e2e-test-templates-6rmnp-user" | |
Jul 9 19:17:06.734: INFO: Creating project "e2e-test-templates-6rmnp" | |
Jul 9 19:17:06.939: INFO: Waiting on permissions in project "e2e-test-templates-6rmnp" ... | |
[BeforeEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:57 | |
Jul 9 19:17:08.098: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-adminuser.kubeconfig" | |
Jul 9 19:17:08.432: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-impersonateuser.kubeconfig" | |
Jul 9 19:17:08.778: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-impersonatebygroupuser.kubeconfig" | |
Jul 9 19:17:09.017: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-edituser1.kubeconfig" | |
Jul 9 19:17:09.331: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-edituser2.kubeconfig" | |
Jul 9 19:17:09.592: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-viewuser.kubeconfig" | |
Jul 9 19:17:09.870: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-impersonatebygroupuser.kubeconfig" | |
[It] should pass impersonation creation tests [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:231 | |
STEP: testing as system:admin user | |
STEP: testing as e2e-test-templates-6rmnp-adminuser user | |
Jul 9 19:17:10.199: INFO: configPath is now "/tmp/e2e-test-templates-6rmnp-adminuser.kubeconfig" | |
[AfterEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:17:10.333: INFO: namespace : e2e-test-templates-6rmnp api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:17:16.455: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] [Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:221 | |
• Failure [11.727 seconds] | |
[Conformance][templates] templateinstance impersonation tests | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:27 | |
should pass impersonation creation tests [Suite:openshift/conformance/parallel] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:231 | |
Expected an error to have occurred. Got: | |
<nil>: nil | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateinstance_impersonation.go:241 | |
------------------------------ | |
[Conformance][templates] templateservicebroker security test | |
should pass security tests [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:164 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][templates] templateservicebroker security test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:16.761: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][templates] templateservicebroker security test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:17:18.405: INFO: configPath is now "/tmp/e2e-test-templates-m7s7v-user.kubeconfig" | |
Jul 9 19:17:18.405: INFO: The user is now "e2e-test-templates-m7s7v-user" | |
Jul 9 19:17:18.405: INFO: Creating project "e2e-test-templates-m7s7v" | |
Jul 9 19:17:18.548: INFO: Waiting on permissions in project "e2e-test-templates-m7s7v" ... | |
[BeforeEach] [Conformance][templates] templateservicebroker security test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:45 | |
[AfterEach] [Conformance][templates] templateservicebroker security test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:17:18.846: INFO: namespace : e2e-test-templates-m7s7v api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][templates] templateservicebroker security test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:17:24.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] [Conformance][templates] templateservicebroker security test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:78 | |
• Failure in Spec Setup (BeforeEach) [8.202 seconds] | |
[Conformance][templates] templateservicebroker security test | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:28 | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:150 | |
should pass security tests [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:164 | |
Expected error: | |
<*errors.StatusError | 0xc42182d170>: { | |
ErrStatus: { | |
TypeMeta: {Kind: "", APIVersion: ""}, | |
ListMeta: {SelfLink: "", ResourceVersion: "", Continue: ""}, | |
Status: "Failure", | |
Message: "services \"apiserver\" not found", | |
Reason: "NotFound", | |
Details: {Name: "apiserver", Group: "", Kind: "services", UID: "", Causes: nil, RetryAfterSeconds: 0}, | |
Code: 404, | |
}, | |
} | |
services "apiserver" not found | |
not to have occurred | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:52 | |
------------------------------ | |
SS | |
------------------------------ | |
[k8s.io] Variable Expansion | |
should allow substituting values in a container's args [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:14.689: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:17:16.748: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-var-expansion-cw78h | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should allow substituting values in a container's args [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test substitution in container's args | |
Jul 9 19:17:17.550: INFO: Waiting up to 5m0s for pod "var-expansion-5e255d23-83e7-11e8-992b-28d244b00276" in namespace "e2e-tests-var-expansion-cw78h" to be "success or failure" | |
Jul 9 19:17:17.594: INFO: Pod "var-expansion-5e255d23-83e7-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 43.983466ms | |
Jul 9 19:17:19.634: INFO: Pod "var-expansion-5e255d23-83e7-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.083132661s | |
Jul 9 19:17:21.681: INFO: Pod "var-expansion-5e255d23-83e7-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.131014186s | |
STEP: Saw pod success | |
Jul 9 19:17:21.681: INFO: Pod "var-expansion-5e255d23-83e7-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:17:21.721: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod var-expansion-5e255d23-83e7-11e8-992b-28d244b00276 container dapi-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:17:21.869: INFO: Waiting for pod var-expansion-5e255d23-83e7-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:17:21.908: INFO: Pod var-expansion-5e255d23-83e7-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:17:21.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-var-expansion-cw78h" for this suite. | |
Jul 9 19:17:28.073: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:17:31.972: INFO: namespace: e2e-tests-var-expansion-cw78h, resource: bindings, ignored listing per whitelist | |
Jul 9 19:17:32.438: INFO: namespace e2e-tests-var-expansion-cw78h deletion completed in 10.478758253s | |
• [SLOW TEST:17.749 seconds] | |
[k8s.io] Variable Expansion | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should allow substituting values in a container's args [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs with minimum ready seconds set [Conformance] | |
should not transition the deployment to Complete before satisfied [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1008 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:16:31.565: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:16:33.458: INFO: configPath is now "/tmp/e2e-test-cli-deployment-gtldl-user.kubeconfig" | |
Jul 9 19:16:33.458: INFO: The user is now "e2e-test-cli-deployment-gtldl-user" | |
Jul 9 19:16:33.458: INFO: Creating project "e2e-test-cli-deployment-gtldl" | |
Jul 9 19:16:33.574: INFO: Waiting on permissions in project "e2e-test-cli-deployment-gtldl" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should not transition the deployment to Complete before satisfied [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1008 | |
STEP: verifying the deployment is created | |
STEP: verifying that all pods are ready | |
Jul 9 19:16:37.910: INFO: All replicas are ready. | |
STEP: verifying that the deployment is still running | |
STEP: waiting for the deployment to finish | |
Jul 9 19:17:38.001: INFO: Finished waiting for deployment. | |
[AfterEach] with minimum ready seconds set [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1004 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:17:40.444: INFO: namespace : e2e-test-cli-deployment-gtldl api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:17:46.476: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:74.949 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
with minimum ready seconds set [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1002 | |
should not transition the deployment to Complete before satisfied [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:1008 | |
------------------------------ | |
[sig-storage] EmptyDir volumes | |
should support (non-root,0666,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:32.446: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:17:34.454: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-d7w4d | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should support (non-root,0666,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test emptydir 0666 on node default medium | |
Jul 9 19:17:35.304: INFO: Waiting up to 5m0s for pod "pod-68bac026-83e7-11e8-992b-28d244b00276" in namespace "e2e-tests-emptydir-d7w4d" to be "success or failure" | |
Jul 9 19:17:35.340: INFO: Pod "pod-68bac026-83e7-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 36.039848ms | |
Jul 9 19:17:37.380: INFO: Pod "pod-68bac026-83e7-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.076062691s | |
STEP: Saw pod success | |
Jul 9 19:17:37.380: INFO: Pod "pod-68bac026-83e7-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:17:37.422: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-68bac026-83e7-11e8-992b-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:17:37.505: INFO: Waiting for pod pod-68bac026-83e7-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:17:37.550: INFO: Pod pod-68bac026-83e7-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:17:37.550: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-d7w4d" for this suite. | |
Jul 9 19:17:43.723: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:17:46.551: INFO: namespace: e2e-tests-emptydir-d7w4d, resource: bindings, ignored listing per whitelist | |
Jul 9 19:17:47.894: INFO: namespace e2e-tests-emptydir-d7w4d deletion completed in 10.300527545s | |
• [SLOW TEST:15.449 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
should support (non-root,0666,default) [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SSSSSS | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:46.516: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:17:48.338: INFO: configPath is now "/tmp/e2e-test-router-stress-lzm6c-user.kubeconfig" | |
Jul 9 19:17:48.338: INFO: The user is now "e2e-test-router-stress-lzm6c-user" | |
Jul 9 19:17:48.338: INFO: Creating project "e2e-test-router-stress-lzm6c" | |
Jul 9 19:17:48.461: INFO: Waiting on permissions in project "e2e-test-router-stress-lzm6c" ... | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:45 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:32 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:17:48.594: INFO: namespace : e2e-test-router-stress-lzm6c api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:17:54.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [8.153 seconds] | |
[Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:21 | |
The HAProxy router [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:68 | |
should serve routes that were created from an ingress [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:79 | |
no router installed on the cluster | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/router.go:48 | |
------------------------------ | |
S | |
------------------------------ | |
[k8s.io] Sysctls | |
should support sysctls [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:60 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:11.563: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:17:13.354: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-sysctl-xg4rw | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:56 | |
[It] should support sysctls [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:60 | |
STEP: Creating a pod with the kernel.shm_rmid_forced sysctl | |
STEP: Watching for error events or started pod | |
STEP: Waiting for pod completion | |
STEP: Checking that the pod succeeded | |
STEP: Getting logs from the pod | |
STEP: Checking that the sysctl is actually updated | |
[AfterEach] [k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Collecting events from namespace "e2e-tests-sysctl-xg4rw". | |
STEP: Found 5 events. | |
Jul 9 19:17:18.215: INFO: At 2018-07-09 19:17:14 -0700 PDT - event for sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276: {default-scheduler } Scheduled: Successfully assigned e2e-tests-sysctl-xg4rw/sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276 to ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:17:18.215: INFO: At 2018-07-09 19:17:14 -0700 PDT - event for sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulling: pulling image "busybox" | |
Jul 9 19:17:18.215: INFO: At 2018-07-09 19:17:16 -0700 PDT - event for sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Successfully pulled image "busybox" | |
Jul 9 19:17:18.215: INFO: At 2018-07-09 19:17:16 -0700 PDT - event for sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:17:18.215: INFO: At 2018-07-09 19:17:16 -0700 PDT - event for sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Started: Started container | |
Jul 9 19:17:18.370: INFO: POD NODE PHASE GRACE CONDITIONS | |
Jul 9 19:17:18.370: INFO: registry-6559c8c4db-45526 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: minreadytest-1-chctk ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:35 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:37 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:35 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: minreadytest-1-deploy ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:34 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:35 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:34 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: minreadytest-1-fddc7 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:35 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:37 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:16:35 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: execpod98j4h ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:16 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: frontend-1-build ip-10-0-130-54.us-west-2.compute.internal Succeeded [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:27 -0700 PDT PodCompleted } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:50 -0700 PDT PodCompleted } {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC PodCompleted } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:21 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276 ip-10-0-130-54.us-west-2.compute.internal Succeeded [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:14 -0700 PDT PodCompleted } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:14 -0700 PDT PodCompleted } {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC PodCompleted } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:14 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: var-expansion-5e255d23-83e7-11e8-992b-28d244b00276 ip-10-0-130-54.us-west-2.compute.internal Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:17 -0700 PDT } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:17 -0700 PDT ContainersNotReady containers with unready status: [dapi-container]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [dapi-container]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:17 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-apiserver-cn2ps ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:45 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-controller-manager-558dc6fb98-q6vr5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:34 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-core-operator-75d546fbbb-c7ctx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:20 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-dns-787c975867-txmxv ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:22 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-flannel-bgv4g ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:59 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-flannel-m5wph ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:58 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-flannel-xcck7 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:17 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-proxy-5td7p ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:54 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-proxy-l2cnn ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-proxy-zsgcb ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: kube-scheduler-68f8875b5c-s5tdr ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: metrics-server-5767bfc576-gfbwb ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: openshift-apiserver-rkms5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:19 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: openshift-controller-manager-99d6586b-qq685 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:55 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: pod-checkpointer-4882g ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:03 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:17:18.371: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: prometheus-0 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:40 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-network-operator-jwwmp ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:13 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-node-controller-2ctqd ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:08 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: webconsole-6698d4fbbc-rgsw2 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: default-http-backend-6985d557bb-8h44n ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: router-6796c95fdf-2k4wk ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:37 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:46 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: directory-sync-d84d84d9f-j7pr6 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: kube-addon-operator-675f99d7f8-c6pdt ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:29 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-alm-operator-79b6996f74-prs9h ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:35 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-channel-operator-5d878cd785-l66n4 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-clu-6b8d87785f-fswbx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-node-agent-r77mj ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:37:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-node-agent-rrwlg ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:12:57 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-stats-emitter-d87f669fd-988nl ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:29 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:36 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:23 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: tectonic-utility-operator-786b69fc8b-4xffz ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:41 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:17:18.372: INFO: | |
Jul 9 19:17:18.405: INFO: | |
Logging node info for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:17:18.439: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-130-54.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-130-54.us-west-2.compute.internal,UID:2f71bed0-83b7-11e8-84c6-0af96768d57e,ResourceVersion:76832,Generation:0,CreationTimestamp:2018-07-09 13:32:23 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-130-54,node-role.kubernetes.io/worker: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"3e:08:91:8f:b9:a5"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.130.54,node-configuration.v1.coreos.com/currentConfig: worker-2650561509,node-configuration.v1.coreos.com/desiredConfig: worker-2650561509,node-configuration.v1.coreos.com/targetConfig: worker-2650561509,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.2.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-0cb9cec2620663d39,Unschedulable:false,Taints:[],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365150208 0} {<nil>} 8169092Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260292608 0} {<nil>} 8066692Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:17:16 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:17:16 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:17:16 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:17:16 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:17:16 -0700 PDT 2018-07-09 13:33:23 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.130.54} {InternalDNS ip-10-0-130-54.us-west-2.compute.internal} {Hostname ip-10-0-130-54}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC283016-6CE7-ACE7-0F9A-02CE10505945,BootID:cfad64a2-03d7-403a-bd51-76866880a650,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[openshift/origin-haproxy-router@sha256:f0a71ada9e9ee48529540c2d4938b9caa55f9a0ac8a3be598e269ca5cebf70c0 openshift/origin-haproxy-router:v3.10.0-alpha.0] 1284960820} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test@sha256:6daa01a6f7f0784905bf9dcbce49826d73d7c3c1d62a802f875ee7c10db02960 docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test:latest] 613134454} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test@sha256:92c5e723d97318711a71afb9ee5c12c3c48b98d0f2aaa5e954095fabbcb505ee docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test:latest] 613133841} {[docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example@sha256:02d80c750d1e71afc7792f55f935c3dd6cde1788bee2b53ab554d29c903ca064 docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example:latest] 603384691} {[docker-registry.default.svc:5000/openshift/php@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7:latest] 589408618} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass@sha256:880359284c1e0933fe5f2db29b8c4d948b70da3dfb26a0462f68b23397740b0a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass:latest] 568094192} {[docker-registry.default.svc:5000/openshift/php@sha256:59c3d53372cd7097494187f5a58bab58a1d956a340b70a23c84a0d000a565cbe] 567254500} {[docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test@sha256:539e80a4de02794f6126cffce75562bcb721041c6d443c5ced15ba286d70e229 docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test:latest] 566117187} {[docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test@sha256:e0eeef684e9de55219871fa9e360d73a1163cfc407c626eade862cbee5a9bbc5 docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test:latest] 566117040} {[centos/ruby-22-centos7@sha256:a18c8706118a5c4c9f1adf045024d2abf06ba632b5674b23421019ee4d3edcae centos/ruby-22-centos7:latest] 566117040} {[centos/nodejs-6-centos7@sha256:b2867b5008d9e975b3d4710ec0f31cdc96b079b83334b17e03a60602a7a590fc] 560696751} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot@sha256:0397f7e12d87d62c539356a4936348d0a8deb40e1b5e970cdd1744d3e6ffa05a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot:latest] 560696751} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot@sha256:4084131a9910c10780186608faf5a9643de0f18d09c27fe828499a8d180abfba docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot:latest] 560696751} {[docker-registry.default.svc:5000/openshift/ruby@sha256:2e83b9e07e85960060096b6aff7ee202a5f52e0e18447641b080b1f3879e0901] 536571487} {[docker-registry.default.svc:5000/openshift/ruby@sha256:8f00b7a5789887b72db0415355830c87e18804b774a922a424736f5237a44933] 518934530} {[docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678@sha256:a9ecb5931f283c598dcaf3aca9025599eb71115bd0f2cd0f1989a9f37394efad docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678:latest] 511744495} {[docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample@sha256:95a78c60dc1709c2212cd8cc48cd3fffe6cdcdd847674497d9aa5d7891551699 docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample:latest] 511744370} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:3bb2aed7578ab5b6ba2bf22993df3c73ef91bdb02e273cc0ce8e529de7ee5660] 506453985} {[docker-registry.default.svc:5000/openshift/ruby@sha256:0eaaed9fae1b0d9bc8ed73b93d581c6ab019a92277484c9acf52fa60b3269a7c] 504578679} {[docker-registry.default.svc:5000/openshift/nodejs@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653 centos/nodejs-8-centos7@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653] 504452018} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:896482969cd659b419bc444c153a74d11820655c7ed19b5094b8eb041f0065d6] 487132847} {[openshift/origin-docker-builder@sha256:4fe8032f87d2f8485a711ec60a9ffb330e42a6cd8d232ad3cf63c42471cfab29 openshift/origin-docker-builder:latest] 447580928} {[docker-registry.default.svc:5000/openshift/mysql@sha256:d03537ef57d51b13e6ad4a73a382ca180a0e02d975c8237790410f45865aae3c] 429435940} {[openshift/origin-haproxy-router@sha256:485fa86ac97b0d289411b3216fb8970989cd580817ebb5fcbb0f83a6dc2466f5 openshift/origin-haproxy-router:latest] 394965919} {[openshift/origin-deployer@sha256:1295e5be56fc03d4c482194378a882f2e96a8d23eadaf6dd32d603d3e877df99 openshift/origin-deployer:latest] 371674595} {[openshift/origin-web-console@sha256:d2cbbb533d26996226add8cb327cb2060e7a03c6aa96ad94cd236d4064c094ce openshift/origin-web-console:latest] 336636057} {[openshift/prometheus@sha256:35e2e0efc874c055be60a025874256816c98b9cebc10f259d7fb806bbe68badf openshift/prometheus:v2.2.1] 317896379} {[openshift/origin-docker-registry@sha256:c40ebb707721327c3b9c79f0e8e7f02483f034355d4149479333cc134b72967c openshift/origin-docker-registry:latest] 302637209} {[openshift/origin-pod@sha256:8fbd41f21824f5981716568790c5f78a4710bb0709ce9c473eb21ad2fbc5e877 openshift/origin-pod:latest] 251747200} {[openshift/origin-base@sha256:43dd97db435025eee02606658cfcccbc0a8ac4135e0d8870e91930d6cab8d1fd openshift/origin-base:latest] 228695137} {[openshift/oauth-proxy@sha256:4b73830ee6f7447d0921eedc3946de50016eb8f048d66ea3969abc4116f1e42a openshift/oauth-proxy:v1.0.0] 228241928} {[openshift/prometheus-alertmanager@sha256:35443abf6c5cf99b080307fe0f98098334f299780537a3e61ac5604cbfe48f7e openshift/prometheus-alertmanager:v0.14.0] 221857684} {[openshift/prometheus-alert-buffer@sha256:076f8dd576806f5c2dde7e536d020c31aa7d2ec7dcea52da6cbb944895def7ba openshift/prometheus-alert-buffer:v0.0.2] 200521084} {[docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage@sha256:df3e69e3fe1bc86897717b020b6caa000f1f97c14dc0b3853ca0d7149412da54 docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage:v1] 199835207} {[centos@sha256:b67d21dfe609ddacf404589e04631d90a342921e81c40aeaf3391f6717fa5322 centos@sha256:eed5b251b615d1e70b10bcec578d64e8aa839d2785c2ffd5424e472818c42755 centos:7 centos:centos7] 199678471} {[docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1@sha256:3967cd8851952bbba0b3a4d9c038f36dc5001463c8521d6955ab0f3f4598d779 docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1:latest] 199678471} {[k8s.gcr.io/nginx-slim-amd64@sha256:6654db6d4028756062edac466454ee5c9cf9b20ef79e35a81e3c840031eb1e2b k8s.gcr.io/nginx-slim-amd64:0.20] 103591055} {[gcr.io/google_containers/metrics-server-amd64@sha256:54d2cf293e01f72d9be0e7c4f2c98e31f599088a9426a6415fe62426d446f5b2 gcr.io/google_containers/metrics-server-amd64:v0.2.0] 96501893} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/directory-sync@sha256:e5e7fe901868853d89c2c0697cc88f0686c6ba1178ca045ec57bfd18e7000048 quay.io/coreos/directory-sync:v0.0.2] 38433928} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[k8s.gcr.io/addon-resizer@sha256:d00afd42fc267fa3275a541083cfe67d160f966c788174b44597434760e1e1eb k8s.gcr.io/addon-resizer:2.1] 26450138} {[quay.io/coreos/tectonic-error-server@sha256:aefa0a012e103bee299c17e798e5830128588b6ef5d4d1f6bc8ae5804bc4d8cd quay.io/coreos/tectonic-error-server:1.1] 12714516} {[gcr.io/google_containers/dnsutils@sha256:cd9182f6d74e616942db1cef6f25e1e54b49ba0330c2e19d3ec061f027666cc0 gcr.io/google_containers/dnsutils:e2e] 8897789} {[gcr.io/kubernetes-e2e-test-images/hostexec-amd64@sha256:bdaecec5adfa7c79e9525c0992fdab36c2d68066f5e91eff0d1d9e8d73c654ea gcr.io/kubernetes-e2e-test-images/hostexec-amd64:1.1] 8407119} {[gcr.io/kubernetes-e2e-test-images/netexec-amd64@sha256:2edfad424a541b9e024f26368d3a5b7dcc1d7cd27a4ee8c1d8c3f81d9209ab2e gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0] 6227659} {[openshift/hello-openshift@sha256:aaea76ff622d2f8bcb32e538e7b3cd0ef6d291953f3e7c9f556c1ba5baf47e2e openshift/hello-openshift:latest] 6089990}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:17:18.440: INFO: | |
Logging kubelet events for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:17:18.472: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:17:18.584: INFO: default-http-backend-6985d557bb-8h44n started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container default-http-backend ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: router-6796c95fdf-2k4wk started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container router ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: minreadytest-1-chctk started at 2018-07-09 19:16:35 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container myapp ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: frontend-1-build started at 2018-07-09 19:14:21 -0700 PDT (2+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Init container git-clone ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Init container manage-dockerfile ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container sti-build ready: false, restart count 0 | |
Jul 9 19:17:18.584: INFO: kube-proxy-5td7p started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: registry-6559c8c4db-45526 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container registry ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: minreadytest-1-deploy started at 2018-07-09 19:16:34 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container deployment ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: prometheus-0 started at 2018-07-09 13:50:04 -0700 PDT (0+6 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container alert-buffer ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container alertmanager ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container alertmanager-proxy ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container alerts-proxy ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container prom-proxy ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container prometheus ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: kube-flannel-xcck7 started at 2018-07-09 13:32:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: metrics-server-5767bfc576-gfbwb started at 2018-07-09 13:33:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container metrics-server ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: Container metrics-server-nanny ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: sysctl-5c0d9804-83e7-11e8-bd2e-28d244b00276 started at 2018-07-09 19:17:14 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container test-container ready: false, restart count 0 | |
Jul 9 19:17:18.584: INFO: execpod98j4h started at 2018-07-09 19:14:15 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container exec ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: tectonic-node-agent-rrwlg started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container node-agent ready: true, restart count 3 | |
Jul 9 19:17:18.584: INFO: directory-sync-d84d84d9f-j7pr6 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container directory-sync ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: webconsole-6698d4fbbc-rgsw2 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container webconsole ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: minreadytest-1-fddc7 started at 2018-07-09 19:16:35 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container myapp ready: true, restart count 0 | |
Jul 9 19:17:18.584: INFO: var-expansion-5e255d23-83e7-11e8-992b-28d244b00276 started at 2018-07-09 19:17:17 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:18.584: INFO: Container dapi-container ready: false, restart count 0 | |
W0709 19:17:18.623181 11748 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:17:18.703: INFO: | |
Latency metrics for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:17:18.703: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:39.952369s} | |
Jul 9 19:17:18.703: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.138495s} | |
Jul 9 19:17:18.703: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:18.561353s} | |
Jul 9 19:17:18.703: INFO: | |
Logging node info for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:17:18.737: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-141-201.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-141-201.us-west-2.compute.internal,UID:ab76db34-83b4-11e8-8888-0af96768d57e,ResourceVersion:76717,Generation:0,CreationTimestamp:2018-07-09 13:14:22 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-141-201,node-role.kubernetes.io/etcd: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"b6:11:a8:d0:6d:85"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.141.201,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.1.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-03457d640f9c71dd1,Unschedulable:false,Taints:[{node-role.kubernetes.io/etcd NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:17:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:17:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:17:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:17:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:17:10 -0700 PDT 2018-07-09 13:16:04 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.141.201} {InternalDNS ip-10-0-141-201.us-west-2.compute.internal} {Hostname ip-10-0-141-201}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2F6BCA-4D59-F6AA-8C7B-027F94D52D78,BootID:92773d40-1311-4ad5-b294-38db65faf16c,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/kube-client-agent@sha256:8564ab65bcb1064006d2fc9c6e32a5ca3f4326cdd2da9a2efc4fb7cc0e0b6041 quay.io/coreos/kube-client-agent:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 33236131} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:17:18.737: INFO: | |
Logging kubelet events for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:17:18.767: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:17:48.799: INFO: Unable to retrieve kubelet pods for node ip-10-0-141-201.us-west-2.compute.internal: the server is currently unable to handle the request (get nodes ip-10-0-141-201.us-west-2.compute.internal:10250) | |
Jul 9 19:17:48.799: INFO: | |
Logging node info for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:17:48.838: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-35-213.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-35-213.us-west-2.compute.internal,UID:a83cf873-83b4-11e8-8888-0af96768d57e,ResourceVersion:77099,Generation:0,CreationTimestamp:2018-07-09 13:14:17 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2c,kubernetes.io/hostname: ip-10-0-35-213,node-role.kubernetes.io/master: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"5e:08:be:54:0d:9f"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.35.213,node-configuration.v1.coreos.com/currentConfig: master-2063737633,node-configuration.v1.coreos.com/desiredConfig: master-2063737633,node-configuration.v1.coreos.com/targetConfig: master-2063737633,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.0.0/24,ExternalID:,ProviderID:aws:///us-west-2c/i-0e1d36783c9705b28,Unschedulable:false,Taints:[{node-role.kubernetes.io/master NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:17:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:17:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:17:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:17:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:17:39 -0700 PDT 2018-07-09 13:16:08 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.35.213} {ExternalIP 34.220.249.237} {InternalDNS ip-10-0-35-213.us-west-2.compute.internal} {ExternalDNS ec2-34-220-249-237.us-west-2.compute.amazonaws.com} {Hostname ip-10-0-35-213}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2ED297-E036-AA0D-C4ED-9057B3EA9001,BootID:7f784e0b-09a6-495a-b787-3d8619214f8a,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[openshift/origin-hypershift@sha256:3b26011ae771a6036a7533d970052be5c04bc1f6e6812314ffefd902f40910fd openshift/origin-hypershift:latest] 518022163} {[openshift/origin-hyperkube@sha256:11a08060b48d226d64d4bb5234f2386bf22472a0835c5b91f0fb0db25b0a7e19 openshift/origin-hyperkube:latest] 498702039} {[quay.io/coreos/awscli@sha256:1d6ea2f37c248a4f4f2a70126f0b8555fd0804d4e65af3b30c3a949247ea13a6 quay.io/coreos/awscli:025a357f05242fdad6a81e8a6b520098aa65a600] 97521631} {[quay.io/coreos/bootkube@sha256:63afddd30deedff273d65607f4fcf0b331f4418838a00c69b6ab7a5754a24f5a quay.io/coreos/bootkube:v0.10.0] 84921995} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[gcr.io/google_containers/k8s-dns-kube-dns-amd64@sha256:6d8e0da4fb46e9ea2034a3f4cab0e095618a2ead78720c12e791342738e5f85d gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.8] 50456751} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/tectonic-stats@sha256:e800fe60dd1a0f89f8ae85caae9209201254e17d889d664d633ed08e274e2a39 quay.io/coreos/tectonic-stats:6e882361357fe4b773adbf279cddf48cb50164c1] 48779830} {[quay.io/coreos/pod-checkpointer@sha256:1e1e48228f872d56c8a57a5e12adb5239ae9e6206536baf2904e4bf03314c8e8 quay.io/coreos/pod-checkpointer:9dc83e1ab3bc36ca25c9f7c18ddef1b91d4a0558] 47992230} {[quay.io/coreos/tectonic-network-operator-dev@sha256:e29d797f5740cf6f5c0ccc0de2b3e606d187acbdc0bb79a4397c058d8840c8fe quay.io/coreos/tectonic-network-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44068170} {[quay.io/coreos/tectonic-node-controller-operator-dev@sha256:7a31568c6c2e398cffa7e8387cf51543e3bf1f01b4a050a5d00a9b593c3dace0 quay.io/coreos/tectonic-node-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44053165} {[quay.io/coreos/kube-addon-operator-dev@sha256:e327727a93813c31f6d65f76f2998722754b8ccb5110949153e55f2adbc2374e quay.io/coreos/kube-addon-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44052211} {[quay.io/coreos/tectonic-utility-operator-dev@sha256:4fb4de52c7aa64ce124e1bf73fb27989356c414101ecc19ca4ec9ab80e00a88d quay.io/coreos/tectonic-utility-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43818409} {[quay.io/coreos/tectonic-ingress-controller-operator-dev@sha256:5e96253c8fe8357473d4806b116fcf03fe18dcad466a88083f9b9310045821f1 quay.io/coreos/tectonic-ingress-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43808038} {[quay.io/coreos/tectonic-alm-operator@sha256:ce32e6d4745040be8807d09eb925b2b076b60fb0a93e33302b74a5cc8f294ca5 quay.io/coreos/tectonic-alm-operator:v0.3.1] 43202998} {[gcr.io/google_containers/k8s-dns-sidecar-amd64@sha256:23df717980b4aa08d2da6c4cfa327f1b730d92ec9cf740959d2d5911830d82fb gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.8] 42210862} {[gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64@sha256:93c827f018cf3322f1ff2aa80324a0306048b0a69bc274e423071fb0d2d29d8b gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.8] 40951779} {[quay.io/coreos/kube-core-operator-dev@sha256:6cc0dd2405f19014b41a0eed57c39160aeb92c2380ac8f8a067ce7dee476cba2 quay.io/coreos/kube-core-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40849618} {[quay.io/coreos/tectonic-channel-operator-dev@sha256:6eeb84c385333755a2189c199587bc26db6c5d897e1962d7e1047dec2531e85e quay.io/coreos/tectonic-channel-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40523592} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[quay.io/coreos/kube-core-renderer-dev@sha256:a595dfe57b7992971563fcea8ac1858c306529a465f9b690911f4220d93d3c5c quay.io/coreos/kube-core-renderer-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 36535818} {[quay.io/coreos/kube-etcd-signer-server@sha256:c4c0becf6779523af5b644b53375d61bed9c4688d496cb2f88d4f08024ac5390 quay.io/coreos/kube-etcd-signer-server:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 34655544} {[quay.io/coreos/tectonic-node-controller-dev@sha256:c9c17f7c4c738e519e36224ae8c71d3a881b92ffb86fdb75f358efebafa27d84 quay.io/coreos/tectonic-node-controller-dev:a437848532713f2fa4137e9a0f4f6a689cf554a8] 25570332} {[quay.io/coreos/tectonic-clu@sha256:4e6a907a433e741632c8f9a7d9d9009bc08ac494dce05e0a19f8fa0a440a3926 quay.io/coreos/tectonic-clu:v0.0.1] 5081911} {[quay.io/coreos/tectonic-stats-extender@sha256:6e7fe41ca2d63791c08d2cc4b4311d9e01b37fa3dc116d3e77e7306cbe29a0f1 quay.io/coreos/tectonic-stats-extender:487b3da4e175da96dabfb44fba65cdb8b823db2e] 2818916} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:17:48.838: INFO: | |
Logging kubelet events for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:17:48.875: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:17:49.001: INFO: tectonic-clu-6b8d87785f-fswbx started at 2018-07-09 13:19:06 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container tectonic-clu ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: tectonic-stats-emitter-d87f669fd-988nl started at 2018-07-09 13:19:23 -0700 PDT (1+2 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Init container tectonic-stats-extender-init ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: Container tectonic-stats-emitter ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: Container tectonic-stats-extender ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: tectonic-channel-operator-5d878cd785-l66n4 started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container tectonic-channel-operator ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: kube-proxy-l2cnn started at 2018-07-09 13:14:22 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: openshift-apiserver-rkms5 started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container openshift-apiserver ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: tectonic-network-operator-jwwmp started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container tectonic-network-operator ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: kube-dns-787c975867-txmxv started at 2018-07-09 13:16:08 -0700 PDT (0+3 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container dnsmasq ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: Container kubedns ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: Container sidecar ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: kube-scheduler-68f8875b5c-s5tdr started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container kube-scheduler ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: kube-apiserver-cn2ps started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container kube-apiserver ready: true, restart count 4 | |
Jul 9 19:17:49.001: INFO: tectonic-node-controller-2ctqd started at 2018-07-09 13:18:05 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container tectonic-node-controller ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: tectonic-alm-operator-79b6996f74-prs9h started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container tectonic-alm-operator ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.001: INFO: Container tectonic-ingress-controller-operator ready: true, restart count 0 | |
Jul 9 19:17:49.001: INFO: tectonic-node-agent-r77mj started at 2018-07-09 13:19:20 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container node-agent ready: true, restart count 4 | |
Jul 9 19:17:49.002: INFO: pod-checkpointer-4882g started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container pod-checkpointer ready: true, restart count 0 | |
Jul 9 19:17:49.002: INFO: kube-flannel-m5wph started at 2018-07-09 13:15:39 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:17:49.002: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:17:49.002: INFO: openshift-controller-manager-99d6586b-qq685 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container openshift-controller-manager ready: true, restart count 3 | |
Jul 9 19:17:49.002: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container tectonic-node-controller-operator ready: true, restart count 0 | |
Jul 9 19:17:49.002: INFO: kube-core-operator-75d546fbbb-c7ctx started at 2018-07-09 13:18:11 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container kube-core-operator ready: true, restart count 0 | |
Jul 9 19:17:49.002: INFO: tectonic-utility-operator-786b69fc8b-4xffz started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container tectonic-utility-operator ready: true, restart count 0 | |
Jul 9 19:17:49.002: INFO: kube-addon-operator-675f99d7f8-c6pdt started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container kube-addon-operator ready: true, restart count 0 | |
Jul 9 19:17:49.002: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal started at <nil> (0+0 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: kube-controller-manager-558dc6fb98-q6vr5 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:17:49.002: INFO: Container kube-controller-manager ready: true, restart count 1 | |
W0709 19:17:49.037552 11748 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:17:49.151: INFO: | |
Latency metrics for node ip-10-0-35-213.us-west-2.compute.internal | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:17:49.188: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-sysctl-xg4rw" for this suite. | |
Jul 9 19:17:55.321: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:17:58.581: INFO: namespace: e2e-tests-sysctl-xg4rw, resource: bindings, ignored listing per whitelist | |
Jul 9 19:17:59.396: INFO: namespace e2e-tests-sysctl-xg4rw deletion completed in 10.163694124s | |
• Failure [47.833 seconds] | |
[k8s.io] Sysctls | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should support sysctls [Suite:openshift/conformance/parallel] [Suite:k8s] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:60 | |
Expected | |
<string>: kernel.shm_rmid_forced = 0 | |
to contain substring | |
<string>: kernel.shm_rmid_forced = 1 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/sysctl.go:98 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] Downward API volume | |
should provide node allocatable (memory) as default memory limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:47.900: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:17:49.996: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-gd54g | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should provide node allocatable (memory) as default memory limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:17:50.746: INFO: Waiting up to 5m0s for pod "downwardapi-volume-71eeb0f9-83e7-11e8-992b-28d244b00276" in namespace "e2e-tests-downward-api-gd54g" to be "success or failure" | |
Jul 9 19:17:50.785: INFO: Pod "downwardapi-volume-71eeb0f9-83e7-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 39.183911ms | |
Jul 9 19:17:52.822: INFO: Pod "downwardapi-volume-71eeb0f9-83e7-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.07606692s | |
STEP: Saw pod success | |
Jul 9 19:17:52.822: INFO: Pod "downwardapi-volume-71eeb0f9-83e7-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:17:52.862: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-71eeb0f9-83e7-11e8-992b-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:17:52.960: INFO: Waiting for pod downwardapi-volume-71eeb0f9-83e7-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:17:52.996: INFO: Pod downwardapi-volume-71eeb0f9-83e7-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:17:52.996: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-gd54g" for this suite. | |
Jul 9 19:17:59.179: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:18:01.251: INFO: namespace: e2e-tests-downward-api-gd54g, resource: bindings, ignored listing per whitelist | |
Jul 9 19:18:03.494: INFO: namespace e2e-tests-downward-api-gd54g deletion completed in 10.449564971s | |
• [SLOW TEST:15.594 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should provide node allocatable (memory) as default memory limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:431 | |
Jul 9 19:18:03.496: INFO: This plugin does not implement NetworkPolicy. | |
[AfterEach] when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:18:03.498: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.002 seconds] | |
NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:48 | |
when using a plugin that implements NetworkPolicy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:430 | |
should enforce multiple, stacked policies with overlapping podSelectors [Feature:OSNetworkPolicy] [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/networkpolicy.go:177 | |
Jul 9 19:18:03.496: This plugin does not implement NetworkPolicy. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
[sig-storage] Secrets | |
should be consumable from pods in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:59.398: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:18:01.024: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-qlgsc | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating secret with name secret-test-7871b510-83e7-11e8-bd2e-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:18:01.696: INFO: Waiting up to 5m0s for pod "pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276" in namespace "e2e-tests-secrets-qlgsc" to be "success or failure" | |
Jul 9 19:18:01.727: INFO: Pod "pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.880909ms | |
Jul 9 19:18:03.756: INFO: Pod "pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.060054186s | |
Jul 9 19:18:05.790: INFO: Pod "pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.09359757s | |
STEP: Saw pod success | |
Jul 9 19:18:05.790: INFO: Pod "pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:18:05.821: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276 container secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:18:05.887: INFO: Waiting for pod pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:18:05.925: INFO: Pod pod-secrets-78768152-83e7-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:18:05.925: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-qlgsc" for this suite. | |
Jul 9 19:18:12.054: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:18:15.286: INFO: namespace: e2e-tests-secrets-qlgsc, resource: bindings, ignored listing per whitelist | |
Jul 9 19:18:15.316: INFO: namespace e2e-tests-secrets-qlgsc deletion completed in 9.35852938s | |
• [SLOW TEST:15.917 seconds] | |
[sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 | |
should be consumable from pods in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] Downward API volume | |
should provide podname only [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:18:15.317: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:18:16.827: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-g87bq | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should provide podname only [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:18:17.558: INFO: Waiting up to 5m0s for pod "downwardapi-volume-81ebe892-83e7-11e8-bd2e-28d244b00276" in namespace "e2e-tests-downward-api-g87bq" to be "success or failure" | |
Jul 9 19:18:17.591: INFO: Pod "downwardapi-volume-81ebe892-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 32.518953ms | |
Jul 9 19:18:19.623: INFO: Pod "downwardapi-volume-81ebe892-83e7-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.064857681s | |
STEP: Saw pod success | |
Jul 9 19:18:19.623: INFO: Pod "downwardapi-volume-81ebe892-83e7-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:18:19.654: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-81ebe892-83e7-11e8-bd2e-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:18:19.721: INFO: Waiting for pod downwardapi-volume-81ebe892-83e7-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:18:19.755: INFO: Pod downwardapi-volume-81ebe892-83e7-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:18:19.756: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-g87bq" for this suite. | |
Jul 9 19:18:25.880: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:18:28.874: INFO: namespace: e2e-tests-downward-api-g87bq, resource: bindings, ignored listing per whitelist | |
Jul 9 19:18:29.199: INFO: namespace e2e-tests-downward-api-g87bq deletion completed in 9.409267552s | |
• [SLOW TEST:13.882 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should provide podname only [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[image_ecosystem][mongodb] openshift mongodb image creating from a template | |
should instantiate the template [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:34 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [image_ecosystem][mongodb] openshift mongodb image | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:54.671: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [image_ecosystem][mongodb] openshift mongodb image | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:17:56.971: INFO: configPath is now "/tmp/e2e-test-mongodb-create-mxp59-user.kubeconfig" | |
Jul 9 19:17:56.971: INFO: The user is now "e2e-test-mongodb-create-mxp59-user" | |
Jul 9 19:17:56.971: INFO: Creating project "e2e-test-mongodb-create-mxp59" | |
Jul 9 19:17:57.125: INFO: Waiting on permissions in project "e2e-test-mongodb-create-mxp59" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:22 | |
Jul 9 19:17:57.190: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[It] should instantiate the template [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:34 | |
openshift namespace image streams OK | |
STEP: creating a new app | |
Jul 9 19:17:57.626: INFO: Running 'oc new-app --config=/tmp/e2e-test-mongodb-create-mxp59-user.kubeconfig --namespace=e2e-test-mongodb-create-mxp59 -f /tmp/fixture-testdata-dir333495585/examples/db-templates/mongodb-ephemeral-template.json' | |
--> Deploying template "e2e-test-mongodb-create-mxp59/mongodb-ephemeral" for "/tmp/fixture-testdata-dir333495585/examples/db-templates/mongodb-ephemeral-template.json" to project e2e-test-mongodb-create-mxp59 | |
MongoDB (Ephemeral) | |
--------- | |
MongoDB database service, without persistent storage. For more information about using this template, including OpenShift considerations, see https://github.com/sclorg/mongodb-container/blob/master/3.2/README.md. | |
WARNING: Any data stored will be lost upon pod destruction. Only use this template for testing | |
The following service(s) have been created in your project: mongodb. | |
Username: userSH5 | |
Password: rYXxAfAPgqyge1eS | |
Database Name: sampledb | |
Connection URL: mongodb://userSH5:rYXxAfAPgqyge1eS@mongodb/sampledb | |
For more information about using this template, including OpenShift considerations, see https://github.com/sclorg/mongodb-container/blob/master/3.2/README.md. | |
* With parameters: | |
* Memory Limit=512Mi | |
* Namespace=openshift | |
* Database Service Name=mongodb | |
* MongoDB Connection Username=userSH5 # generated | |
* MongoDB Connection Password=rYXxAfAPgqyge1eS # generated | |
* MongoDB Database Name=sampledb | |
* MongoDB Admin Password=b2glYaLmorpxNjYS # generated | |
* Version of MongoDB Image=3.2 | |
--> Creating resources ... | |
secret "mongodb" created | |
service "mongodb" created | |
deploymentconfig "mongodb" created | |
--> Success | |
Application is not exposed. You can expose services to the outside world by executing one or more of the commands below: | |
'oc expose svc/mongodb' | |
Run 'oc status' to view your app. | |
STEP: waiting for the deployment to complete | |
Jul 9 19:17:59.934: INFO: waiting for deploymentconfig e2e-test-mongodb-create-mxp59/mongodb to be available with version 1 | |
Jul 9 19:18:21.005: INFO: deploymentconfig e2e-test-mongodb-create-mxp59/mongodb available after 21.071302293s | |
pods: mongodb-1-zkx79 | |
STEP: expecting the mongodb pod is running | |
STEP: expecting the mongodb service is answering for ping | |
Jul 9 19:18:22.044: INFO: Running 'oc exec --config=/tmp/e2e-test-mongodb-create-mxp59-user.kubeconfig --namespace=e2e-test-mongodb-create-mxp59 mongodb-1-zkx79 -- bash -c mongo --quiet --eval '{"ping", 1}'' | |
STEP: expecting that we can insert a new record | |
Jul 9 19:18:22.755: INFO: Running 'oc exec --config=/tmp/e2e-test-mongodb-create-mxp59-user.kubeconfig --namespace=e2e-test-mongodb-create-mxp59 mongodb-1-zkx79 -- bash -c mongo --quiet "$MONGODB_DATABASE" --username "$MONGODB_USER" --password "$MONGODB_PASSWORD" --eval 'db.foo.save({ "status": "passed" })'' | |
STEP: expecting that we can read a record | |
Jul 9 19:18:23.426: INFO: Running 'oc exec --config=/tmp/e2e-test-mongodb-create-mxp59-user.kubeconfig --namespace=e2e-test-mongodb-create-mxp59 mongodb-1-zkx79 -- bash -c mongo --quiet "$MONGODB_DATABASE" --username "$MONGODB_USER" --password "$MONGODB_PASSWORD" --eval 'printjson(db.foo.find({}, {_id: 0}).toArray())'' | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:26 | |
[AfterEach] [image_ecosystem][mongodb] openshift mongodb image | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:18:24.226: INFO: namespace : e2e-test-mongodb-create-mxp59 api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [image_ecosystem][mongodb] openshift mongodb image | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:18:46.258: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:51.631 seconds] | |
[image_ecosystem][mongodb] openshift mongodb image | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:15 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:21 | |
creating from a template | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:33 | |
should instantiate the template [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/image_ecosystem/mongodb_ephemeral.go:34 | |
------------------------------ | |
[k8s.io] InitContainer | |
should invoke init containers on a RestartNever pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:44 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:18:29.200: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:18:30.691: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-init-container-ljvgk | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:40 | |
[It] should invoke init containers on a RestartNever pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:44 | |
STEP: creating the pod | |
Jul 9 19:18:31.273: INFO: PodSpec: initContainers in spec.initContainers | |
[AfterEach] [k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:18:40.517: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-init-container-ljvgk" for this suite. | |
Jul 9 19:18:46.688: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:18:50.027: INFO: namespace: e2e-tests-init-container-ljvgk, resource: bindings, ignored listing per whitelist | |
Jul 9 19:18:50.057: INFO: namespace e2e-tests-init-container-ljvgk deletion completed in 9.488643845s | |
• [SLOW TEST:20.857 seconds] | |
[k8s.io] InitContainer | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should invoke init containers on a RestartNever pod [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/init_container.go:44 | |
------------------------------ | |
[sig-storage] Secrets | |
should be consumable in multiple volumes in a pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:18:50.060: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:18:51.647: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-mtlqm | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable in multiple volumes in a pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating secret with name secret-test-96a7f343-83e7-11e8-bd2e-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:18:52.378: INFO: Waiting up to 5m0s for pod "pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276" in namespace "e2e-tests-secrets-mtlqm" to be "success or failure" | |
Jul 9 19:18:52.411: INFO: Pod "pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 32.682627ms | |
Jul 9 19:18:54.449: INFO: Pod "pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.07052138s | |
Jul 9 19:18:56.479: INFO: Pod "pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.10078221s | |
STEP: Saw pod success | |
Jul 9 19:18:56.479: INFO: Pod "pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:18:56.530: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276 container secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:18:56.599: INFO: Waiting for pod pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:18:56.627: INFO: Pod pod-secrets-96ac7b2f-83e7-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:18:56.627: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-mtlqm" for this suite. | |
Jul 9 19:19:02.789: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:19:05.783: INFO: namespace: e2e-tests-secrets-mtlqm, resource: bindings, ignored listing per whitelist | |
Jul 9 19:19:06.308: INFO: namespace e2e-tests-secrets-mtlqm deletion completed in 9.638494379s | |
• [SLOW TEST:16.248 seconds] | |
[sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 | |
should be consumable in multiple volumes in a pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[Area:Networking] network isolation when using a plugin that does not isolate namespaces by default | |
should allow communication between pods in different namespaces on the same node [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/isolation.go:15 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that does not isolate namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:407 | |
Jul 9 19:18:46.575: INFO: Could not check network plugin name: exit status 1. Assuming a non-OpenShift plugin | |
[BeforeEach] when using a plugin that does not isolate namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:18:46.575: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-net-isolation1-f5qp8 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] when using a plugin that does not isolate namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:18:48.506: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-net-isolation2-fvc9k | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should allow communication between pods in different namespaces on the same node [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/isolation.go:15 | |
Jul 9 19:18:50.646: INFO: Using ip-10-0-130-54.us-west-2.compute.internal for test ([ip-10-0-130-54.us-west-2.compute.internal] out of [ip-10-0-130-54.us-west-2.compute.internal]) | |
Jul 9 19:18:52.794: INFO: Target pod IP:port is 10.2.2.72:8080 | |
Jul 9 19:18:52.794: INFO: Creating an exec pod on node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:18:52.794: INFO: Creating new exec pod | |
Jul 9 19:18:56.956: INFO: Waiting up to 10s to wget 10.2.2.72:8080 | |
Jul 9 19:18:56.956: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-tests-net-isolation2-fvc9k execpod-sourceip-ip-10-0-130-54.us-west-2.compute.internaltjbwd -- /bin/sh -c wget -T 30 -qO- 10.2.2.72:8080' | |
Jul 9 19:18:57.616: INFO: stderr: "" | |
Jul 9 19:18:57.616: INFO: Cleaning up the exec pod | |
[AfterEach] when using a plugin that does not isolate namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:18:57.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-net-isolation1-f5qp8" for this suite. | |
Jul 9 19:19:03.874: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:19:05.852: INFO: namespace: e2e-tests-net-isolation1-f5qp8, resource: bindings, ignored listing per whitelist | |
Jul 9 19:19:07.599: INFO: namespace e2e-tests-net-isolation1-f5qp8 deletion completed in 9.836500252s | |
[AfterEach] when using a plugin that does not isolate namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:19:07.599: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-net-isolation2-fvc9k" for this suite. | |
Jul 9 19:19:13.760: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:19:15.952: INFO: namespace: e2e-tests-net-isolation2-fvc9k, resource: bindings, ignored listing per whitelist | |
Jul 9 19:19:17.605: INFO: namespace e2e-tests-net-isolation2-fvc9k deletion completed in 9.968094137s | |
• [SLOW TEST:31.301 seconds] | |
[Area:Networking] network isolation | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/isolation.go:10 | |
when using a plugin that does not isolate namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:406 | |
should allow communication between pods in different namespaces on the same node [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/isolation.go:15 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:19:17.606: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:19:19.438: INFO: configPath is now "/tmp/e2e-test-unprivileged-router-q8gbw-user.kubeconfig" | |
Jul 9 19:19:19.438: INFO: The user is now "e2e-test-unprivileged-router-q8gbw-user" | |
Jul 9 19:19:19.438: INFO: Creating project "e2e-test-unprivileged-router-q8gbw" | |
Jul 9 19:19:19.591: INFO: Waiting on permissions in project "e2e-test-unprivileged-router-q8gbw" ... | |
[BeforeEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/unprivileged.go:41 | |
Jul 9 19:19:19.625: INFO: Running 'oc new-app --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-unprivileged-router-q8gbw -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/scoped-router.yaml -p=IMAGE=openshift/origin-haproxy-router -p=SCOPE=["--name=test-unprivileged", "--namespace=$(POD_NAMESPACE)", "--loglevel=4", "--labels=select=first", "--update-status=false"]' | |
warning: --param no longer accepts comma-separated lists of values. "SCOPE=[\"--name=test-unprivileged\", \"--namespace=$(POD_NAMESPACE)\", \"--loglevel=4\", \"--labels=select=first\", \"--update-status=false\"]" will be treated as a single key-value pair. | |
--> Deploying template "e2e-test-unprivileged-router-q8gbw/" for "/tmp/fixture-testdata-dir333495585/test/extended/testdata/scoped-router.yaml" to project e2e-test-unprivileged-router-q8gbw | |
* With parameters: | |
* IMAGE=openshift/origin-haproxy-router | |
* SCOPE=["--name=test-unprivileged", "--namespace=$(POD_NAMESPACE)", "--loglevel=4", "--labels=select=first", "--update-status=false"] | |
--> Creating resources ... | |
pod "router-scoped" created | |
pod "router-override" created | |
pod "router-override-domains" created | |
rolebinding "system-router" created | |
route "route-1" created | |
route "route-2" created | |
route "route-override-domain-1" created | |
route "route-override-domain-2" created | |
service "endpoints" created | |
pod "endpoint-1" created | |
--> Success | |
Access your application via route 'first.example.com' | |
Access your application via route 'second.example.com' | |
Access your application via route 'y.a.null.ptr' | |
Access your application via route 'main.void.str' | |
Run 'oc status' to view your app. | |
[It] should run even if it has no access to update status [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/unprivileged.go:55 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/unprivileged.go:29 | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:19:20.742: INFO: namespace : e2e-test-unprivileged-router-q8gbw api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:19:40.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] [23.219 seconds] | |
[Conformance][Area:Networking][Feature:Router] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/unprivileged.go:19 | |
The HAProxy router | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/unprivileged.go:54 | |
should run even if it has no access to update status [Suite:openshift/conformance/parallel] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/unprivileged.go:55 | |
test temporarily disabled | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/router/unprivileged.go:56 | |
------------------------------ | |
[Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
should successfully resolve valueFrom in s2i build environment variables [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:61 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:19:06.309: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:19:07.882: INFO: configPath is now "/tmp/e2e-test-build-valuefrom-frlkw-user.kubeconfig" | |
Jul 9 19:19:07.882: INFO: The user is now "e2e-test-build-valuefrom-frlkw-user" | |
Jul 9 19:19:07.882: INFO: Creating project "e2e-test-build-valuefrom-frlkw" | |
Jul 9 19:19:08.044: INFO: Waiting on permissions in project "e2e-test-build-valuefrom-frlkw" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:27 | |
Jul 9 19:19:08.105: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:38 | |
STEP: waiting for builder service account | |
STEP: waiting for openshift namespace imagestreams | |
Jul 9 19:19:08.242: INFO: Running scan #0 | |
Jul 9 19:19:08.242: INFO: Checking language ruby | |
Jul 9 19:19:08.287: INFO: Checking tag latest | |
Jul 9 19:19:08.287: INFO: Checking tag 2.0 | |
Jul 9 19:19:08.287: INFO: Checking tag 2.2 | |
Jul 9 19:19:08.287: INFO: Checking tag 2.3 | |
Jul 9 19:19:08.287: INFO: Checking tag 2.4 | |
Jul 9 19:19:08.287: INFO: Checking tag 2.5 | |
Jul 9 19:19:08.287: INFO: Checking language nodejs | |
Jul 9 19:19:08.327: INFO: Checking tag 4 | |
Jul 9 19:19:08.327: INFO: Checking tag 6 | |
Jul 9 19:19:08.327: INFO: Checking tag 8 | |
Jul 9 19:19:08.327: INFO: Checking tag latest | |
Jul 9 19:19:08.327: INFO: Checking tag 0.10 | |
Jul 9 19:19:08.327: INFO: Checking language perl | |
Jul 9 19:19:08.376: INFO: Checking tag 5.16 | |
Jul 9 19:19:08.376: INFO: Checking tag 5.20 | |
Jul 9 19:19:08.376: INFO: Checking tag 5.24 | |
Jul 9 19:19:08.376: INFO: Checking tag latest | |
Jul 9 19:19:08.376: INFO: Checking language php | |
Jul 9 19:19:08.415: INFO: Checking tag 7.1 | |
Jul 9 19:19:08.415: INFO: Checking tag latest | |
Jul 9 19:19:08.415: INFO: Checking tag 5.5 | |
Jul 9 19:19:08.415: INFO: Checking tag 5.6 | |
Jul 9 19:19:08.415: INFO: Checking tag 7.0 | |
Jul 9 19:19:08.415: INFO: Checking language python | |
Jul 9 19:19:08.455: INFO: Checking tag latest | |
Jul 9 19:19:08.455: INFO: Checking tag 2.7 | |
Jul 9 19:19:08.455: INFO: Checking tag 3.3 | |
Jul 9 19:19:08.455: INFO: Checking tag 3.4 | |
Jul 9 19:19:08.456: INFO: Checking tag 3.5 | |
Jul 9 19:19:08.456: INFO: Checking tag 3.6 | |
Jul 9 19:19:08.456: INFO: Checking language wildfly | |
Jul 9 19:19:08.510: INFO: Checking tag 10.1 | |
Jul 9 19:19:08.510: INFO: Checking tag 11.0 | |
Jul 9 19:19:08.510: INFO: Checking tag 12.0 | |
Jul 9 19:19:08.510: INFO: Checking tag 8.1 | |
Jul 9 19:19:08.510: INFO: Checking tag 9.0 | |
Jul 9 19:19:08.510: INFO: Checking tag latest | |
Jul 9 19:19:08.510: INFO: Checking tag 10.0 | |
Jul 9 19:19:08.510: INFO: Checking language mysql | |
Jul 9 19:19:08.550: INFO: Checking tag 5.6 | |
Jul 9 19:19:08.550: INFO: Checking tag 5.7 | |
Jul 9 19:19:08.550: INFO: Checking tag latest | |
Jul 9 19:19:08.550: INFO: Checking tag 5.5 | |
Jul 9 19:19:08.550: INFO: Checking language postgresql | |
Jul 9 19:19:08.614: INFO: Checking tag 9.2 | |
Jul 9 19:19:08.615: INFO: Checking tag 9.4 | |
Jul 9 19:19:08.615: INFO: Checking tag 9.5 | |
Jul 9 19:19:08.615: INFO: Checking tag 9.6 | |
Jul 9 19:19:08.615: INFO: Checking tag latest | |
Jul 9 19:19:08.615: INFO: Checking language mongodb | |
Jul 9 19:19:08.657: INFO: Checking tag 2.6 | |
Jul 9 19:19:08.657: INFO: Checking tag 3.2 | |
Jul 9 19:19:08.657: INFO: Checking tag 3.4 | |
Jul 9 19:19:08.657: INFO: Checking tag latest | |
Jul 9 19:19:08.657: INFO: Checking tag 2.4 | |
Jul 9 19:19:08.657: INFO: Checking language jenkins | |
Jul 9 19:19:08.711: INFO: Checking tag 2 | |
Jul 9 19:19:08.711: INFO: Checking tag latest | |
Jul 9 19:19:08.711: INFO: Checking tag 1 | |
Jul 9 19:19:08.711: INFO: Success! | |
STEP: creating test image stream | |
Jul 9 19:19:08.711: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-frlkw-user.kubeconfig --namespace=e2e-test-build-valuefrom-frlkw -f /tmp/fixture-testdata-dir225659500/test/extended/testdata/builds/valuefrom/test-is.json' | |
imagestream.image.openshift.io "test" created | |
STEP: creating test secret | |
Jul 9 19:19:08.959: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-frlkw-user.kubeconfig --namespace=e2e-test-build-valuefrom-frlkw -f /tmp/fixture-testdata-dir225659500/test/extended/testdata/builds/valuefrom/test-secret.yaml' | |
secret "mysecret" created | |
STEP: creating test configmap | |
Jul 9 19:19:09.260: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-frlkw-user.kubeconfig --namespace=e2e-test-build-valuefrom-frlkw -f /tmp/fixture-testdata-dir225659500/test/extended/testdata/builds/valuefrom/test-configmap.yaml' | |
configmap "myconfigmap" created | |
[It] should successfully resolve valueFrom in s2i build environment variables [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:61 | |
STEP: creating test successful build config | |
Jul 9 19:19:09.551: INFO: Running 'oc create --config=/tmp/e2e-test-build-valuefrom-frlkw-user.kubeconfig --namespace=e2e-test-build-valuefrom-frlkw -f /tmp/fixture-testdata-dir225659500/test/extended/testdata/builds/valuefrom/successful-sti-build-value-from-config.yaml' | |
buildconfig.build.openshift.io "mys2itest" created | |
STEP: starting test build | |
Jul 9 19:19:09.872: INFO: Running 'oc start-build --config=/tmp/e2e-test-build-valuefrom-frlkw-user.kubeconfig --namespace=e2e-test-build-valuefrom-frlkw mys2itest -o=name' | |
Jul 9 19:19:10.143: INFO: | |
start-build output with args [mys2itest -o=name]: | |
Error><nil> | |
StdOut> | |
build/mys2itest-1 | |
StdErr> | |
Jul 9 19:19:10.144: INFO: Waiting for mys2itest-1 to complete | |
Jul 9 19:19:36.260: INFO: Done waiting for mys2itest-1: util.BuildResult{BuildPath:"build/mys2itest-1", BuildName:"mys2itest-1", StartBuildStdErr:"", StartBuildStdOut:"build/mys2itest-1", StartBuildErr:error(nil), BuildConfigName:"", Build:(*build.Build)(0xc421e63b00), BuildAttempt:true, BuildSuccess:true, BuildFailure:false, BuildCancelled:false, BuildTimeout:false, LogDumper:(util.LogDumperFunc)(nil), Oc:(*util.CLI)(0xc42096a1e0)} | |
with error: <nil> | |
Jul 9 19:19:36.260: INFO: Running 'oc logs --config=/tmp/e2e-test-build-valuefrom-frlkw-user.kubeconfig --namespace=e2e-test-build-valuefrom-frlkw -f build/mys2itest-1 --timestamps' | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:31 | |
[AfterEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:19:36.858: INFO: namespace : e2e-test-build-valuefrom-frlkw api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:19:42.890: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:36.620 seconds] | |
[Feature:Builds][Conformance][valueFrom] process valueFrom in build strategy environment variables | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:13 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:26 | |
should successfully resolve valueFrom in s2i build environment variables [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/valuefrom.go:61 | |
------------------------------ | |
SS | |
------------------------------ | |
[k8s.io] KubeletManagedEtcHosts | |
should test kubelet managed /etc/hosts file [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] KubeletManagedEtcHosts | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:18:03.500: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:18:05.498: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-e2e-kubelet-etc-hosts-mt788 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should test kubelet managed /etc/hosts file [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Setting up the test | |
STEP: Creating hostNetwork=false pod | |
STEP: Creating hostNetwork=true pod | |
STEP: Running the test | |
STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false | |
Jul 9 19:18:14.519: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-mt788 PodName:test-pod ContainerName:busybox-1 Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} | |
Jul 9 19:18:14.520: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:18:14.844: INFO: Exec stderr: "" | |
Jul 9 19:18:14.844: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-mt788 PodName:test-pod ContainerName:busybox-2 Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} | |
Jul 9 19:18:14.844: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:18:15.158: INFO: Exec stderr: "" | |
STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount | |
Jul 9 19:18:15.159: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-tests-e2e-kubelet-etc-hosts-mt788 PodName:test-pod ContainerName:busybox-3 Stdin:<nil> CaptureStdout:true CaptureStderr:true PreserveWhitespace:false} | |
Jul 9 19:18:15.159: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
Jul 9 19:18:15.361: INFO: Exec stderr: "" | |
[AfterEach] [k8s.io] KubeletManagedEtcHosts | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Collecting events from namespace "e2e-tests-e2e-kubelet-etc-hosts-mt788". | |
STEP: Found 19 events. | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:06 -0700 PDT - event for test-pod: {default-scheduler } Scheduled: Successfully assigned e2e-tests-e2e-kubelet-etc-hosts-mt788/test-pod to ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:07 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Container image "gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0" already present on machine | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:08 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Container image "gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0" already present on machine | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:08 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:08 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Container image "gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0" already present on machine | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:08 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Started: Started container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:08 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:08 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Started: Started container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:09 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Failed: Error: failed to start container "busybox-3": Error response from daemon: OCI runtime create failed: container_linux.go:348: starting container process caused "process_linux.go:402: container init caused \"rootfs_linux.go:58: mounting \\\"/tmp/etc-hosts291332576\\\" to rootfs \\\"/var/lib/docker/overlay2/28b4fd916cf3ee847aa8b641cf8791ffc74146925446e90f4b63b4739457a8a4/merged\\\" at \\\"/var/lib/docker/overlay2/28b4fd916cf3ee847aa8b641cf8791ffc74146925446e90f4b63b4739457a8a4/merged/etc/hosts\\\" caused \\\"not a directory\\\"\"": unknown: Are you trying to mount a directory onto a file (or vice-versa)? Check if the specified host path exists and is the expected type | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:09 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:11 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Failed: Error: failed to start container "busybox-3": Error response from daemon: OCI runtime create failed: container_linux.go:348: starting container process caused "process_linux.go:402: container init caused \"rootfs_linux.go:58: mounting \\\"/tmp/etc-hosts291332576\\\" to rootfs \\\"/var/lib/docker/overlay2/fb1e65fec82cd9b220a6b4c104980f69c2be4b9ca18b8032df78bc0a3c65cac6/merged\\\" at \\\"/var/lib/docker/overlay2/fb1e65fec82cd9b220a6b4c104980f69c2be4b9ca18b8032df78bc0a3c65cac6/merged/etc/hosts\\\" caused \\\"not a directory\\\"\"": unknown: Are you trying to mount a directory onto a file (or vice-versa)? Check if the specified host path exists and is the expected type | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:11 -0700 PDT - event for test-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} BackOff: Back-off restarting failed container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:12 -0700 PDT - event for test-host-network-pod: {default-scheduler } Scheduled: Successfully assigned e2e-tests-e2e-kubelet-etc-hosts-mt788/test-host-network-pod to ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:13 -0700 PDT - event for test-host-network-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:13 -0700 PDT - event for test-host-network-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Container image "gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0" already present on machine | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:13 -0700 PDT - event for test-host-network-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Started: Started container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:13 -0700 PDT - event for test-host-network-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:13 -0700 PDT - event for test-host-network-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Container image "gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0" already present on machine | |
Jul 9 19:18:15.400: INFO: At 2018-07-09 19:18:13 -0700 PDT - event for test-host-network-pod: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Started: Started container | |
Jul 9 19:18:15.580: INFO: POD NODE PHASE GRACE CONDITIONS | |
Jul 9 19:18:15.580: INFO: registry-6559c8c4db-45526 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: mongodb-1-deploy ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:01 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:02 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:01 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: mongodb-1-zkx79 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:02 -0700 PDT } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:02 -0700 PDT ContainersNotReady containers with unready status: [mongodb]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [mongodb]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:02 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: execpod98j4h ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:16 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: frontend-1-build ip-10-0-130-54.us-west-2.compute.internal Succeeded [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:27 -0700 PDT PodCompleted } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:50 -0700 PDT PodCompleted } {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC PodCompleted } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:21 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: test-host-network-pod ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:13 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:12 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: test-pod ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:06 -0700 PDT } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:06 -0700 PDT ContainersNotReady containers with unready status: [busybox-3]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [busybox-3]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:18:06 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: pod-host-path-test ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:27 -0700 PDT } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:27 -0700 PDT ContainersNotReady containers with unready status: [test-container-1]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [test-container-1]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:17:27 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-apiserver-cn2ps ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:45 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-controller-manager-558dc6fb98-q6vr5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:34 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-core-operator-75d546fbbb-c7ctx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:20 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-dns-787c975867-txmxv ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:22 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-flannel-bgv4g ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:59 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-flannel-m5wph ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:58 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-flannel-xcck7 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:17 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-proxy-5td7p ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:54 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-proxy-l2cnn ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-proxy-zsgcb ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-scheduler-68f8875b5c-s5tdr ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: metrics-server-5767bfc576-gfbwb ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: openshift-apiserver-rkms5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:19 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: openshift-controller-manager-99d6586b-qq685 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:55 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: pod-checkpointer-4882g ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:03 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: prometheus-0 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:40 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-network-operator-jwwmp ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:13 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-node-controller-2ctqd ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:08 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: webconsole-6698d4fbbc-rgsw2 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: default-http-backend-6985d557bb-8h44n ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: router-6796c95fdf-2k4wk ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:37 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:46 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: directory-sync-d84d84d9f-j7pr6 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: kube-addon-operator-675f99d7f8-c6pdt ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:29 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-alm-operator-79b6996f74-prs9h ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:35 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-channel-operator-5d878cd785-l66n4 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-clu-6b8d87785f-fswbx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-node-agent-r77mj ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:37:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-node-agent-rrwlg ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:12:57 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-stats-emitter-d87f669fd-988nl ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:29 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:36 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:23 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: tectonic-utility-operator-786b69fc8b-4xffz ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:41 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:18:15.580: INFO: | |
Jul 9 19:18:15.618: INFO: | |
Logging node info for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:18:15.658: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-130-54.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-130-54.us-west-2.compute.internal,UID:2f71bed0-83b7-11e8-84c6-0af96768d57e,ResourceVersion:77546,Generation:0,CreationTimestamp:2018-07-09 13:32:23 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-130-54,node-role.kubernetes.io/worker: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"3e:08:91:8f:b9:a5"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.130.54,node-configuration.v1.coreos.com/currentConfig: worker-2650561509,node-configuration.v1.coreos.com/desiredConfig: worker-2650561509,node-configuration.v1.coreos.com/targetConfig: worker-2650561509,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.2.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-0cb9cec2620663d39,Unschedulable:false,Taints:[],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365150208 0} {<nil>} 8169092Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260292608 0} {<nil>} 8066692Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:18:06 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:18:06 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:18:06 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:18:06 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:18:06 -0700 PDT 2018-07-09 13:33:23 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.130.54} {InternalDNS ip-10-0-130-54.us-west-2.compute.internal} {Hostname ip-10-0-130-54}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC283016-6CE7-ACE7-0F9A-02CE10505945,BootID:cfad64a2-03d7-403a-bd51-76866880a650,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[openshift/origin-haproxy-router@sha256:f0a71ada9e9ee48529540c2d4938b9caa55f9a0ac8a3be598e269ca5cebf70c0 openshift/origin-haproxy-router:v3.10.0-alpha.0] 1284960820} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test@sha256:6daa01a6f7f0784905bf9dcbce49826d73d7c3c1d62a802f875ee7c10db02960 docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test:latest] 613134454} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test@sha256:92c5e723d97318711a71afb9ee5c12c3c48b98d0f2aaa5e954095fabbcb505ee docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test:latest] 613133841} {[docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example@sha256:02d80c750d1e71afc7792f55f935c3dd6cde1788bee2b53ab554d29c903ca064 docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example:latest] 603384691} {[docker-registry.default.svc:5000/openshift/php@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7:latest] 589408618} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass@sha256:880359284c1e0933fe5f2db29b8c4d948b70da3dfb26a0462f68b23397740b0a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass:latest] 568094192} {[docker-registry.default.svc:5000/openshift/php@sha256:59c3d53372cd7097494187f5a58bab58a1d956a340b70a23c84a0d000a565cbe] 567254500} {[docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test@sha256:539e80a4de02794f6126cffce75562bcb721041c6d443c5ced15ba286d70e229 docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test:latest] 566117187} {[docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test@sha256:e0eeef684e9de55219871fa9e360d73a1163cfc407c626eade862cbee5a9bbc5 docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test:latest] 566117040} {[centos/ruby-22-centos7@sha256:a18c8706118a5c4c9f1adf045024d2abf06ba632b5674b23421019ee4d3edcae centos/ruby-22-centos7:latest] 566117040} {[centos/nodejs-6-centos7@sha256:b2867b5008d9e975b3d4710ec0f31cdc96b079b83334b17e03a60602a7a590fc] 560696751} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot@sha256:0397f7e12d87d62c539356a4936348d0a8deb40e1b5e970cdd1744d3e6ffa05a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot:latest] 560696751} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot@sha256:4084131a9910c10780186608faf5a9643de0f18d09c27fe828499a8d180abfba docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot:latest] 560696751} {[docker-registry.default.svc:5000/openshift/ruby@sha256:2e83b9e07e85960060096b6aff7ee202a5f52e0e18447641b080b1f3879e0901] 536571487} {[docker-registry.default.svc:5000/openshift/ruby@sha256:8f00b7a5789887b72db0415355830c87e18804b774a922a424736f5237a44933] 518934530} {[docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678@sha256:a9ecb5931f283c598dcaf3aca9025599eb71115bd0f2cd0f1989a9f37394efad docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678:latest] 511744495} {[docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample@sha256:95a78c60dc1709c2212cd8cc48cd3fffe6cdcdd847674497d9aa5d7891551699 docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample:latest] 511744370} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:3bb2aed7578ab5b6ba2bf22993df3c73ef91bdb02e273cc0ce8e529de7ee5660] 506453985} {[docker-registry.default.svc:5000/openshift/ruby@sha256:0eaaed9fae1b0d9bc8ed73b93d581c6ab019a92277484c9acf52fa60b3269a7c] 504578679} {[docker-registry.default.svc:5000/openshift/nodejs@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653 centos/nodejs-8-centos7@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653] 504452018} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:896482969cd659b419bc444c153a74d11820655c7ed19b5094b8eb041f0065d6] 487132847} {[openshift/origin-docker-builder@sha256:4fe8032f87d2f8485a711ec60a9ffb330e42a6cd8d232ad3cf63c42471cfab29 openshift/origin-docker-builder:latest] 447580928} {[docker-registry.default.svc:5000/openshift/mysql@sha256:d03537ef57d51b13e6ad4a73a382ca180a0e02d975c8237790410f45865aae3c] 429435940} {[openshift/origin-haproxy-router@sha256:485fa86ac97b0d289411b3216fb8970989cd580817ebb5fcbb0f83a6dc2466f5 openshift/origin-haproxy-router:latest] 394965919} {[openshift/origin-deployer@sha256:1295e5be56fc03d4c482194378a882f2e96a8d23eadaf6dd32d603d3e877df99 openshift/origin-deployer:latest] 371674595} {[openshift/origin-web-console@sha256:d2cbbb533d26996226add8cb327cb2060e7a03c6aa96ad94cd236d4064c094ce openshift/origin-web-console:latest] 336636057} {[openshift/prometheus@sha256:35e2e0efc874c055be60a025874256816c98b9cebc10f259d7fb806bbe68badf openshift/prometheus:v2.2.1] 317896379} {[openshift/origin-docker-registry@sha256:c40ebb707721327c3b9c79f0e8e7f02483f034355d4149479333cc134b72967c openshift/origin-docker-registry:latest] 302637209} {[openshift/origin-pod@sha256:8fbd41f21824f5981716568790c5f78a4710bb0709ce9c473eb21ad2fbc5e877 openshift/origin-pod:latest] 251747200} {[openshift/origin-base@sha256:43dd97db435025eee02606658cfcccbc0a8ac4135e0d8870e91930d6cab8d1fd openshift/origin-base:latest] 228695137} {[openshift/oauth-proxy@sha256:4b73830ee6f7447d0921eedc3946de50016eb8f048d66ea3969abc4116f1e42a openshift/oauth-proxy:v1.0.0] 228241928} {[openshift/prometheus-alertmanager@sha256:35443abf6c5cf99b080307fe0f98098334f299780537a3e61ac5604cbfe48f7e openshift/prometheus-alertmanager:v0.14.0] 221857684} {[openshift/prometheus-alert-buffer@sha256:076f8dd576806f5c2dde7e536d020c31aa7d2ec7dcea52da6cbb944895def7ba openshift/prometheus-alert-buffer:v0.0.2] 200521084} {[docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage@sha256:df3e69e3fe1bc86897717b020b6caa000f1f97c14dc0b3853ca0d7149412da54 docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage:v1] 199835207} {[centos@sha256:b67d21dfe609ddacf404589e04631d90a342921e81c40aeaf3391f6717fa5322 centos@sha256:eed5b251b615d1e70b10bcec578d64e8aa839d2785c2ffd5424e472818c42755 centos:7 centos:centos7] 199678471} {[docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1@sha256:3967cd8851952bbba0b3a4d9c038f36dc5001463c8521d6955ab0f3f4598d779 docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1:latest] 199678471} {[k8s.gcr.io/nginx-slim-amd64@sha256:6654db6d4028756062edac466454ee5c9cf9b20ef79e35a81e3c840031eb1e2b k8s.gcr.io/nginx-slim-amd64:0.20] 103591055} {[gcr.io/google_containers/metrics-server-amd64@sha256:54d2cf293e01f72d9be0e7c4f2c98e31f599088a9426a6415fe62426d446f5b2 gcr.io/google_containers/metrics-server-amd64:v0.2.0] 96501893} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/directory-sync@sha256:e5e7fe901868853d89c2c0697cc88f0686c6ba1178ca045ec57bfd18e7000048 quay.io/coreos/directory-sync:v0.0.2] 38433928} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[k8s.gcr.io/addon-resizer@sha256:d00afd42fc267fa3275a541083cfe67d160f966c788174b44597434760e1e1eb k8s.gcr.io/addon-resizer:2.1] 26450138} {[quay.io/coreos/tectonic-error-server@sha256:aefa0a012e103bee299c17e798e5830128588b6ef5d4d1f6bc8ae5804bc4d8cd quay.io/coreos/tectonic-error-server:1.1] 12714516} {[gcr.io/google_containers/dnsutils@sha256:cd9182f6d74e616942db1cef6f25e1e54b49ba0330c2e19d3ec061f027666cc0 gcr.io/google_containers/dnsutils:e2e] 8897789} {[gcr.io/kubernetes-e2e-test-images/hostexec-amd64@sha256:bdaecec5adfa7c79e9525c0992fdab36c2d68066f5e91eff0d1d9e8d73c654ea gcr.io/kubernetes-e2e-test-images/hostexec-amd64:1.1] 8407119} {[gcr.io/kubernetes-e2e-test-images/netexec-amd64@sha256:2edfad424a541b9e024f26368d3a5b7dcc1d7cd27a4ee8c1d8c3f81d9209ab2e gcr.io/kubernetes-e2e-test-images/netexec-amd64:1.0] 6227659} {[openshift/hello-openshift@sha256:aaea76ff622d2f8bcb32e538e7b3cd0ef6d291953f3e7c9f556c1ba5baf47e2e openshift/hello-openshift:latest] 6089990}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:18:15.658: INFO: | |
Logging kubelet events for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:18:15.703: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:18:15.838: INFO: router-6796c95fdf-2k4wk started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container router ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: mongodb-1-deploy started at 2018-07-09 19:18:01 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container deployment ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: pod-host-path-test started at 2018-07-09 19:17:27 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container test-container-1 ready: false, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container test-container-2 ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: mongodb-1-zkx79 started at 2018-07-09 19:18:02 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container mongodb ready: false, restart count 0 | |
Jul 9 19:18:15.838: INFO: frontend-1-build started at 2018-07-09 19:14:21 -0700 PDT (2+1 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Init container git-clone ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Init container manage-dockerfile ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container sti-build ready: false, restart count 0 | |
Jul 9 19:18:15.838: INFO: default-http-backend-6985d557bb-8h44n started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container default-http-backend ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: registry-6559c8c4db-45526 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container registry ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: test-pod started at 2018-07-09 19:18:06 -0700 PDT (0+3 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container busybox-1 ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container busybox-2 ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container busybox-3 ready: false, restart count 1 | |
Jul 9 19:18:15.838: INFO: prometheus-0 started at 2018-07-09 13:50:04 -0700 PDT (0+6 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container alert-buffer ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container alertmanager ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container alertmanager-proxy ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container alerts-proxy ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container prom-proxy ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container prometheus ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: kube-proxy-5td7p started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: metrics-server-5767bfc576-gfbwb started at 2018-07-09 13:33:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container metrics-server ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: Container metrics-server-nanny ready: true, restart count 0 | |
Jul 9 19:18:15.838: INFO: test-host-network-pod started at 2018-07-09 19:18:12 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:18:15.838: INFO: Container busybox-1 ready: true, restart count 0 | |
Jul 9 19:18:15.839: INFO: Container busybox-2 ready: true, restart count 0 | |
Jul 9 19:18:15.839: INFO: execpod98j4h started at 2018-07-09 19:14:15 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.839: INFO: Container exec ready: true, restart count 0 | |
Jul 9 19:18:15.839: INFO: kube-flannel-xcck7 started at 2018-07-09 13:32:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:18:15.839: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:18:15.839: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:18:15.839: INFO: directory-sync-d84d84d9f-j7pr6 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.839: INFO: Container directory-sync ready: true, restart count 0 | |
Jul 9 19:18:15.839: INFO: webconsole-6698d4fbbc-rgsw2 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.839: INFO: Container webconsole ready: true, restart count 0 | |
Jul 9 19:18:15.839: INFO: tectonic-node-agent-rrwlg started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:15.839: INFO: Container node-agent ready: true, restart count 3 | |
W0709 19:18:15.878394 11713 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:18:16.036: INFO: | |
Latency metrics for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:18:16.036: INFO: {Operation: Method:pod_start_latency_microseconds Quantile:0.99 Latency:39.952369s} | |
Jul 9 19:18:16.036: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.138495s} | |
Jul 9 19:18:16.036: INFO: {Operation:pull_image Method:docker_operations_latency_microseconds Quantile:0.99 Latency:18.561353s} | |
Jul 9 19:18:16.036: INFO: | |
Logging node info for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:18:16.083: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-141-201.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-141-201.us-west-2.compute.internal,UID:ab76db34-83b4-11e8-8888-0af96768d57e,ResourceVersion:77561,Generation:0,CreationTimestamp:2018-07-09 13:14:22 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-141-201,node-role.kubernetes.io/etcd: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"b6:11:a8:d0:6d:85"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.141.201,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.1.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-03457d640f9c71dd1,Unschedulable:false,Taints:[{node-role.kubernetes.io/etcd NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:18:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:18:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:18:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:18:10 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:18:10 -0700 PDT 2018-07-09 13:16:04 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.141.201} {InternalDNS ip-10-0-141-201.us-west-2.compute.internal} {Hostname ip-10-0-141-201}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2F6BCA-4D59-F6AA-8C7B-027F94D52D78,BootID:92773d40-1311-4ad5-b294-38db65faf16c,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/kube-client-agent@sha256:8564ab65bcb1064006d2fc9c6e32a5ca3f4326cdd2da9a2efc4fb7cc0e0b6041 quay.io/coreos/kube-client-agent:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 33236131} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:18:16.083: INFO: | |
Logging kubelet events for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:18:16.120: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:18:46.159: INFO: Unable to retrieve kubelet pods for node ip-10-0-141-201.us-west-2.compute.internal: the server is currently unable to handle the request (get nodes ip-10-0-141-201.us-west-2.compute.internal:10250) | |
Jul 9 19:18:46.159: INFO: | |
Logging node info for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:18:46.202: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-35-213.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-35-213.us-west-2.compute.internal,UID:a83cf873-83b4-11e8-8888-0af96768d57e,ResourceVersion:77848,Generation:0,CreationTimestamp:2018-07-09 13:14:17 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2c,kubernetes.io/hostname: ip-10-0-35-213,node-role.kubernetes.io/master: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"5e:08:be:54:0d:9f"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.35.213,node-configuration.v1.coreos.com/currentConfig: master-2063737633,node-configuration.v1.coreos.com/desiredConfig: master-2063737633,node-configuration.v1.coreos.com/targetConfig: master-2063737633,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.0.0/24,ExternalID:,ProviderID:aws:///us-west-2c/i-0e1d36783c9705b28,Unschedulable:false,Taints:[{node-role.kubernetes.io/master NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:18:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:18:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:18:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:18:39 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:18:39 -0700 PDT 2018-07-09 13:16:08 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.35.213} {ExternalIP 34.220.249.237} {InternalDNS ip-10-0-35-213.us-west-2.compute.internal} {ExternalDNS ec2-34-220-249-237.us-west-2.compute.amazonaws.com} {Hostname ip-10-0-35-213}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2ED297-E036-AA0D-C4ED-9057B3EA9001,BootID:7f784e0b-09a6-495a-b787-3d8619214f8a,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[openshift/origin-hypershift@sha256:3b26011ae771a6036a7533d970052be5c04bc1f6e6812314ffefd902f40910fd openshift/origin-hypershift:latest] 518022163} {[openshift/origin-hyperkube@sha256:11a08060b48d226d64d4bb5234f2386bf22472a0835c5b91f0fb0db25b0a7e19 openshift/origin-hyperkube:latest] 498702039} {[quay.io/coreos/awscli@sha256:1d6ea2f37c248a4f4f2a70126f0b8555fd0804d4e65af3b30c3a949247ea13a6 quay.io/coreos/awscli:025a357f05242fdad6a81e8a6b520098aa65a600] 97521631} {[quay.io/coreos/bootkube@sha256:63afddd30deedff273d65607f4fcf0b331f4418838a00c69b6ab7a5754a24f5a quay.io/coreos/bootkube:v0.10.0] 84921995} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[gcr.io/google_containers/k8s-dns-kube-dns-amd64@sha256:6d8e0da4fb46e9ea2034a3f4cab0e095618a2ead78720c12e791342738e5f85d gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.8] 50456751} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/tectonic-stats@sha256:e800fe60dd1a0f89f8ae85caae9209201254e17d889d664d633ed08e274e2a39 quay.io/coreos/tectonic-stats:6e882361357fe4b773adbf279cddf48cb50164c1] 48779830} {[quay.io/coreos/pod-checkpointer@sha256:1e1e48228f872d56c8a57a5e12adb5239ae9e6206536baf2904e4bf03314c8e8 quay.io/coreos/pod-checkpointer:9dc83e1ab3bc36ca25c9f7c18ddef1b91d4a0558] 47992230} {[quay.io/coreos/tectonic-network-operator-dev@sha256:e29d797f5740cf6f5c0ccc0de2b3e606d187acbdc0bb79a4397c058d8840c8fe quay.io/coreos/tectonic-network-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44068170} {[quay.io/coreos/tectonic-node-controller-operator-dev@sha256:7a31568c6c2e398cffa7e8387cf51543e3bf1f01b4a050a5d00a9b593c3dace0 quay.io/coreos/tectonic-node-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44053165} {[quay.io/coreos/kube-addon-operator-dev@sha256:e327727a93813c31f6d65f76f2998722754b8ccb5110949153e55f2adbc2374e quay.io/coreos/kube-addon-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44052211} {[quay.io/coreos/tectonic-utility-operator-dev@sha256:4fb4de52c7aa64ce124e1bf73fb27989356c414101ecc19ca4ec9ab80e00a88d quay.io/coreos/tectonic-utility-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43818409} {[quay.io/coreos/tectonic-ingress-controller-operator-dev@sha256:5e96253c8fe8357473d4806b116fcf03fe18dcad466a88083f9b9310045821f1 quay.io/coreos/tectonic-ingress-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43808038} {[quay.io/coreos/tectonic-alm-operator@sha256:ce32e6d4745040be8807d09eb925b2b076b60fb0a93e33302b74a5cc8f294ca5 quay.io/coreos/tectonic-alm-operator:v0.3.1] 43202998} {[gcr.io/google_containers/k8s-dns-sidecar-amd64@sha256:23df717980b4aa08d2da6c4cfa327f1b730d92ec9cf740959d2d5911830d82fb gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.8] 42210862} {[gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64@sha256:93c827f018cf3322f1ff2aa80324a0306048b0a69bc274e423071fb0d2d29d8b gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.8] 40951779} {[quay.io/coreos/kube-core-operator-dev@sha256:6cc0dd2405f19014b41a0eed57c39160aeb92c2380ac8f8a067ce7dee476cba2 quay.io/coreos/kube-core-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40849618} {[quay.io/coreos/tectonic-channel-operator-dev@sha256:6eeb84c385333755a2189c199587bc26db6c5d897e1962d7e1047dec2531e85e quay.io/coreos/tectonic-channel-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40523592} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[quay.io/coreos/kube-core-renderer-dev@sha256:a595dfe57b7992971563fcea8ac1858c306529a465f9b690911f4220d93d3c5c quay.io/coreos/kube-core-renderer-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 36535818} {[quay.io/coreos/kube-etcd-signer-server@sha256:c4c0becf6779523af5b644b53375d61bed9c4688d496cb2f88d4f08024ac5390 quay.io/coreos/kube-etcd-signer-server:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 34655544} {[quay.io/coreos/tectonic-node-controller-dev@sha256:c9c17f7c4c738e519e36224ae8c71d3a881b92ffb86fdb75f358efebafa27d84 quay.io/coreos/tectonic-node-controller-dev:a437848532713f2fa4137e9a0f4f6a689cf554a8] 25570332} {[quay.io/coreos/tectonic-clu@sha256:4e6a907a433e741632c8f9a7d9d9009bc08ac494dce05e0a19f8fa0a440a3926 quay.io/coreos/tectonic-clu:v0.0.1] 5081911} {[quay.io/coreos/tectonic-stats-extender@sha256:6e7fe41ca2d63791c08d2cc4b4311d9e01b37fa3dc116d3e77e7306cbe29a0f1 quay.io/coreos/tectonic-stats-extender:487b3da4e175da96dabfb44fba65cdb8b823db2e] 2818916} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:18:46.202: INFO: | |
Logging kubelet events for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:18:46.238: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:18:46.385: INFO: kube-addon-operator-675f99d7f8-c6pdt started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container kube-addon-operator ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal started at <nil> (0+0 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: kube-controller-manager-558dc6fb98-q6vr5 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container kube-controller-manager ready: true, restart count 1 | |
Jul 9 19:18:46.385: INFO: kube-scheduler-68f8875b5c-s5tdr started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container kube-scheduler ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: tectonic-clu-6b8d87785f-fswbx started at 2018-07-09 13:19:06 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-clu ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: tectonic-stats-emitter-d87f669fd-988nl started at 2018-07-09 13:19:23 -0700 PDT (1+2 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Init container tectonic-stats-extender-init ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: Container tectonic-stats-emitter ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: Container tectonic-stats-extender ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: tectonic-channel-operator-5d878cd785-l66n4 started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-channel-operator ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: kube-proxy-l2cnn started at 2018-07-09 13:14:22 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: openshift-apiserver-rkms5 started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container openshift-apiserver ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: tectonic-network-operator-jwwmp started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-network-operator ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: kube-dns-787c975867-txmxv started at 2018-07-09 13:16:08 -0700 PDT (0+3 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container dnsmasq ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: Container kubedns ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: Container sidecar ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: kube-apiserver-cn2ps started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container kube-apiserver ready: true, restart count 4 | |
Jul 9 19:18:46.385: INFO: tectonic-node-controller-2ctqd started at 2018-07-09 13:18:05 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-node-controller ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: tectonic-alm-operator-79b6996f74-prs9h started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-alm-operator ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-ingress-controller-operator ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: tectonic-node-agent-r77mj started at 2018-07-09 13:19:20 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container node-agent ready: true, restart count 4 | |
Jul 9 19:18:46.385: INFO: tectonic-utility-operator-786b69fc8b-4xffz started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-utility-operator ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: pod-checkpointer-4882g started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container pod-checkpointer ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: kube-flannel-m5wph started at 2018-07-09 13:15:39 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: openshift-controller-manager-99d6586b-qq685 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container openshift-controller-manager ready: true, restart count 3 | |
Jul 9 19:18:46.385: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container tectonic-node-controller-operator ready: true, restart count 0 | |
Jul 9 19:18:46.385: INFO: kube-core-operator-75d546fbbb-c7ctx started at 2018-07-09 13:18:11 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:18:46.385: INFO: Container kube-core-operator ready: true, restart count 0 | |
W0709 19:18:46.427641 11713 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:18:46.567: INFO: | |
Latency metrics for node ip-10-0-35-213.us-west-2.compute.internal | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:18:46.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-e2e-kubelet-etc-hosts-mt788" for this suite. | |
Jul 9 19:19:44.791: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:19:47.721: INFO: namespace: e2e-tests-e2e-kubelet-etc-hosts-mt788, resource: bindings, ignored listing per whitelist | |
Jul 9 19:19:48.939: INFO: namespace e2e-tests-e2e-kubelet-etc-hosts-mt788 deletion completed in 1m2.271310319s | |
• Failure [105.439 seconds] | |
[k8s.io] KubeletManagedEtcHosts | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should test kubelet managed /etc/hosts file [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
failed to execute command in pod test-pod, container busybox-3: unable to upgrade connection: container not found ("busybox-3") | |
Expected error: | |
<*errors.errorString | 0xc421007490>: { | |
s: "unable to upgrade connection: container not found (\"busybox-3\")", | |
} | |
unable to upgrade connection: container not found ("busybox-3") | |
not to have occurred | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/exec_util.go:104 | |
------------------------------ | |
[sig-storage] ConfigMap | |
should be consumable from pods in volume with mappings as non-root [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:19:40.826: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:19:42.560: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-mq4jz | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in volume with mappings as non-root [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating configMap with name configmap-test-volume-map-b4fc7a45-83e7-11e8-8fe2-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:19:43.269: INFO: Waiting up to 5m0s for pod "pod-configmaps-b50214c9-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-configmap-mq4jz" to be "success or failure" | |
Jul 9 19:19:43.308: INFO: Pod "pod-configmaps-b50214c9-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 39.425503ms | |
Jul 9 19:19:45.341: INFO: Pod "pod-configmaps-b50214c9-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.071906859s | |
STEP: Saw pod success | |
Jul 9 19:19:45.341: INFO: Pod "pod-configmaps-b50214c9-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:19:45.374: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-configmaps-b50214c9-83e7-11e8-8fe2-28d244b00276 container configmap-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:19:45.459: INFO: Waiting for pod pod-configmaps-b50214c9-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:19:45.490: INFO: Pod pod-configmaps-b50214c9-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:19:45.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-configmap-mq4jz" for this suite. | |
Jul 9 19:19:51.668: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:19:55.501: INFO: namespace: e2e-tests-configmap-mq4jz, resource: bindings, ignored listing per whitelist | |
Jul 9 19:19:55.605: INFO: namespace e2e-tests-configmap-mq4jz deletion completed in 10.039097537s | |
• [SLOW TEST:14.779 seconds] | |
[sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 | |
should be consumable from pods in volume with mappings as non-root [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable from pods in volume with mappings as non-root with FSGroup [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:469 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:19:55.608: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:19:57.646: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-zd4xr | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable from pods in volume with mappings as non-root with FSGroup [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:469 | |
STEP: Creating configMap with name projected-configmap-test-volume-map-bdf766a8-83e7-11e8-8fe2-28d244b00276 | |
STEP: Creating a pod to test consume configMaps | |
Jul 9 19:19:58.338: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-bdfcfa30-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-projected-zd4xr" to be "success or failure" | |
Jul 9 19:19:58.369: INFO: Pod "pod-projected-configmaps-bdfcfa30-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 31.096945ms | |
Jul 9 19:20:00.402: INFO: Pod "pod-projected-configmaps-bdfcfa30-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.063701216s | |
STEP: Saw pod success | |
Jul 9 19:20:00.402: INFO: Pod "pod-projected-configmaps-bdfcfa30-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:20:00.441: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-configmaps-bdfcfa30-83e7-11e8-8fe2-28d244b00276 container projected-configmap-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:20:00.515: INFO: Waiting for pod pod-projected-configmaps-bdfcfa30-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:20:00.548: INFO: Pod pod-projected-configmaps-bdfcfa30-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:00.548: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-zd4xr" for this suite. | |
Jul 9 19:20:06.685: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:20:08.966: INFO: namespace: e2e-tests-projected-zd4xr, resource: bindings, ignored listing per whitelist | |
Jul 9 19:20:10.513: INFO: namespace e2e-tests-projected-zd4xr deletion completed in 9.924000537s | |
• [SLOW TEST:14.905 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable from pods in volume with mappings as non-root with FSGroup [Feature:FSGroup] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:469 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:DeploymentConfig] deploymentconfigs with failing hook [Conformance] | |
should get all logs from retried hooks [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:819 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:19:48.942: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:19:50.938: INFO: configPath is now "/tmp/e2e-test-cli-deployment-rhvgs-user.kubeconfig" | |
Jul 9 19:19:50.938: INFO: The user is now "e2e-test-cli-deployment-rhvgs-user" | |
Jul 9 19:19:50.938: INFO: Creating project "e2e-test-cli-deployment-rhvgs" | |
Jul 9 19:19:51.216: INFO: Waiting on permissions in project "e2e-test-cli-deployment-rhvgs" ... | |
[JustBeforeEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:43 | |
[It] should get all logs from retried hooks [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:819 | |
Jul 9 19:19:55.728: INFO: Running 'oc logs --config=/tmp/e2e-test-cli-deployment-rhvgs-user.kubeconfig --namespace=e2e-test-cli-deployment-rhvgs dc/hook' | |
STEP: checking the logs for substrings | |
--> pre: Running hook pod ... | |
pre hook logs | |
--> pre: Retrying hook pod (retry #1) | |
pre hook logs | |
[AfterEach] with failing hook [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:815 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:62 | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:19:58.080: INFO: namespace : e2e-test-cli-deployment-rhvgs api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:16.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:27.219 seconds] | |
[Feature:DeploymentConfig] deploymentconfigs | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:37 | |
with failing hook [Conformance] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:813 | |
should get all logs from retried hooks [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/deployments/deployments.go:819 | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:20:16.165: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:20:18.277: INFO: configPath is now "/tmp/e2e-test-resolve-local-names-m7cbv-user.kubeconfig" | |
Jul 9 19:20:18.277: INFO: The user is now "e2e-test-resolve-local-names-m7cbv-user" | |
Jul 9 19:20:18.277: INFO: Creating project "e2e-test-resolve-local-names-m7cbv" | |
Jul 9 19:20:18.629: INFO: Waiting on permissions in project "e2e-test-resolve-local-names-m7cbv" ... | |
[It] should update standard Kube object image fields when local names are on [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:19 | |
Jul 9 19:20:18.671: INFO: Running 'oc import-image --config=/tmp/e2e-test-resolve-local-names-m7cbv-user.kubeconfig --namespace=e2e-test-resolve-local-names-m7cbv busybox:latest --confirm' | |
The import completed successfully. | |
Name: busybox | |
Namespace: e2e-test-resolve-local-names-m7cbv | |
Created: Less than a second ago | |
Labels: <none> | |
Annotations: openshift.io/image.dockerRepositoryCheck=2018-07-10T02:20:20Z | |
Docker Pull Spec: docker-registry.default.svc:5000/e2e-test-resolve-local-names-m7cbv/busybox | |
Image Lookup: local=false | |
Unique Images: 1 | |
Tags: 1 | |
latest | |
tagged from busybox:latest | |
* busybox@sha256:74f634b1bc1bd74535d5209589734efbd44a25f4e2dc96d78784576a3eb5b335 | |
Less than a second ago | |
Image Name: busybox:latest | |
Docker Image: busybox@sha256:74f634b1bc1bd74535d5209589734efbd44a25f4e2dc96d78784576a3eb5b335 | |
Name: sha256:74f634b1bc1bd74535d5209589734efbd44a25f4e2dc96d78784576a3eb5b335 | |
Created: Less than a second ago | |
Annotations: image.openshift.io/dockerLayersOrder=ascending | |
Image Size: 724.6kB | |
Image Created: 6 weeks ago | |
Author: <none> | |
Arch: amd64 | |
Command: sh | |
Working Dir: <none> | |
User: <none> | |
Exposes Ports: <none> | |
Docker Labels: <none> | |
Environment: PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin | |
Jul 9 19:20:20.358: INFO: Running 'oc set image-lookup --config=/tmp/e2e-test-resolve-local-names-m7cbv-user.kubeconfig --namespace=e2e-test-resolve-local-names-m7cbv busybox' | |
imagestream "busybox" updated | |
[AfterEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:20:20.894: INFO: namespace : e2e-test-resolve-local-names-m7cbv api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:26.939: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] [10.827 seconds] | |
[Feature:ImageLookup][registry] Image policy | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:14 | |
should update standard Kube object image fields when local names are on [Suite:openshift/conformance/parallel] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:19 | |
default image resolution is not configured, can't verify pod resolution | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/images/resolve.go:43 | |
------------------------------ | |
[sig-api-machinery] Secrets | |
should be consumable from pods in env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-api-machinery] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:20:10.518: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:20:12.297: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-gp26s | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating secret with name secret-test-c6b47a70-83e7-11e8-8fe2-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:20:12.999: INFO: Waiting up to 5m0s for pod "pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-secrets-gp26s" to be "success or failure" | |
Jul 9 19:20:13.038: INFO: Pod "pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 39.646084ms | |
Jul 9 19:20:15.071: INFO: Pod "pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.072119091s | |
Jul 9 19:20:17.105: INFO: Pod "pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.10677299s | |
STEP: Saw pod success | |
Jul 9 19:20:17.105: INFO: Pod "pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:20:17.157: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276 container secret-env-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:20:17.233: INFO: Waiting for pod pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:20:17.276: INFO: Pod pod-secrets-c6b9ab6b-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-api-machinery] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:17.276: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-gp26s" for this suite. | |
Jul 9 19:20:23.428: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:20:26.387: INFO: namespace: e2e-tests-secrets-gp26s, resource: bindings, ignored listing per whitelist | |
Jul 9 19:20:27.380: INFO: namespace e2e-tests-secrets-gp26s deletion completed in 10.067981102s | |
• [SLOW TEST:16.862 seconds] | |
[sig-api-machinery] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets.go:30 | |
should be consumable from pods in env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[sig-storage] Projected | |
should be consumable in multiple volumes in a pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:20:26.993: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:20:28.974: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-lknkb | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be consumable in multiple volumes in a pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating secret with name projected-secret-test-d0c46100-83e7-11e8-992b-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:20:29.903: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276" in namespace "e2e-tests-projected-lknkb" to be "success or failure" | |
Jul 9 19:20:29.944: INFO: Pod "pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 40.806839ms | |
Jul 9 19:20:32.008: INFO: Pod "pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.104830832s | |
STEP: Saw pod success | |
Jul 9 19:20:32.008: INFO: Pod "pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:20:32.058: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276 container secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:20:32.142: INFO: Waiting for pod pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276 to disappear | |
Jul 9 19:20:32.178: INFO: Pod pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:32.178: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-lknkb" for this suite. | |
Jul 9 19:20:38.341: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:20:40.743: INFO: namespace: e2e-tests-projected-lknkb, resource: bindings, ignored listing per whitelist | |
Jul 9 19:20:42.537: INFO: namespace e2e-tests-projected-lknkb deletion completed in 10.317801375s | |
• [SLOW TEST:15.544 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be consumable in multiple volumes in a pod [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[Feature:Builds][pullsecret][Conformance] docker build using a pull secret Building from a template | |
should create a docker build that pulls using a secret run it [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:44 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][pullsecret][Conformance] docker build using a pull secret | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:20:27.383: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][pullsecret][Conformance] docker build using a pull secret | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:20:29.233: INFO: configPath is now "/tmp/e2e-test-docker-build-pullsecret-r2mt4-user.kubeconfig" | |
Jul 9 19:20:29.233: INFO: The user is now "e2e-test-docker-build-pullsecret-r2mt4-user" | |
Jul 9 19:20:29.233: INFO: Creating project "e2e-test-docker-build-pullsecret-r2mt4" | |
Jul 9 19:20:29.392: INFO: Waiting on permissions in project "e2e-test-docker-build-pullsecret-r2mt4" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:26 | |
Jul 9 19:20:29.469: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:30 | |
STEP: waiting for builder service account | |
[It] should create a docker build that pulls using a secret run it [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:44 | |
STEP: calling oc create -f "/tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/test-docker-build-pullsecret.json" | |
Jul 9 19:20:29.604: INFO: Running 'oc create --config=/tmp/e2e-test-docker-build-pullsecret-r2mt4-user.kubeconfig --namespace=e2e-test-docker-build-pullsecret-r2mt4 -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/test-docker-build-pullsecret.json' | |
imagestream.image.openshift.io "image1" created | |
buildconfig.build.openshift.io "docker-build" created | |
buildconfig.build.openshift.io "docker-build-pull" created | |
STEP: starting a build | |
Jul 9 19:20:29.987: INFO: Running 'oc start-build --config=/tmp/e2e-test-docker-build-pullsecret-r2mt4-user.kubeconfig --namespace=e2e-test-docker-build-pullsecret-r2mt4 docker-build -o=name' | |
Jul 9 19:20:30.249: INFO: | |
start-build output with args [docker-build -o=name]: | |
Error><nil> | |
StdOut> | |
build/docker-build-1 | |
StdErr> | |
Jul 9 19:20:30.251: INFO: Waiting for docker-build-1 to complete | |
Jul 9 19:20:36.331: INFO: Done waiting for docker-build-1: util.BuildResult{BuildPath:"build/docker-build-1", BuildName:"docker-build-1", StartBuildStdErr:"", StartBuildStdOut:"build/docker-build-1", StartBuildErr:error(nil), BuildConfigName:"", Build:(*build.Build)(0xc421140900), BuildAttempt:true, BuildSuccess:true, BuildFailure:false, BuildCancelled:false, BuildTimeout:false, LogDumper:(util.LogDumperFunc)(nil), Oc:(*util.CLI)(0xc42004c5a0)} | |
with error: <nil> | |
STEP: starting a second build that pulls the image from the first build | |
Jul 9 19:20:36.331: INFO: Running 'oc start-build --config=/tmp/e2e-test-docker-build-pullsecret-r2mt4-user.kubeconfig --namespace=e2e-test-docker-build-pullsecret-r2mt4 docker-build-pull -o=name' | |
Jul 9 19:20:36.645: INFO: | |
start-build output with args [docker-build-pull -o=name]: | |
Error><nil> | |
StdOut> | |
build/docker-build-pull-1 | |
StdErr> | |
Jul 9 19:20:36.646: INFO: Waiting for docker-build-pull-1 to complete | |
Jul 9 19:20:42.742: INFO: Done waiting for docker-build-pull-1: util.BuildResult{BuildPath:"build/docker-build-pull-1", BuildName:"docker-build-pull-1", StartBuildStdErr:"", StartBuildStdOut:"build/docker-build-pull-1", StartBuildErr:error(nil), BuildConfigName:"", Build:(*build.Build)(0xc420bb4f00), BuildAttempt:true, BuildSuccess:true, BuildFailure:false, BuildCancelled:false, BuildTimeout:false, LogDumper:(util.LogDumperFunc)(nil), Oc:(*util.CLI)(0xc42004c5a0)} | |
with error: <nil> | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:36 | |
[AfterEach] [Feature:Builds][pullsecret][Conformance] docker build using a pull secret | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:20:42.812: INFO: namespace : e2e-test-docker-build-pullsecret-r2mt4 api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][pullsecret][Conformance] docker build using a pull secret | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:48.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:21.558 seconds] | |
[Feature:Builds][pullsecret][Conformance] docker build using a pull secret | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:12 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:24 | |
Building from a template | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:43 | |
should create a docker build that pulls using a secret run it [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/docker_pullsecret.go:44 | |
------------------------------ | |
SS | |
------------------------------ | |
[sig-storage] Downward API volume | |
should provide node allocatable (cpu) as default cpu limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:20:48.947: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:20:50.734: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-hfpvx | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:20:51.453: INFO: Waiting up to 5m0s for pod "downwardapi-volume-dda26113-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-downward-api-hfpvx" to be "success or failure" | |
Jul 9 19:20:51.493: INFO: Pod "downwardapi-volume-dda26113-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 39.998508ms | |
Jul 9 19:20:53.529: INFO: Pod "downwardapi-volume-dda26113-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.075876986s | |
STEP: Saw pod success | |
Jul 9 19:20:53.529: INFO: Pod "downwardapi-volume-dda26113-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:20:53.565: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-dda26113-83e7-11e8-8fe2-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:20:53.637: INFO: Waiting for pod downwardapi-volume-dda26113-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:20:53.669: INFO: Pod downwardapi-volume-dda26113-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:53.669: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-hfpvx" for this suite. | |
Jul 9 19:20:59.807: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:03.187: INFO: namespace: e2e-tests-downward-api-hfpvx, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:03.809: INFO: namespace e2e-tests-downward-api-hfpvx deletion completed in 10.105937921s | |
• [SLOW TEST:14.862 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should provide node allocatable (cpu) as default cpu limit if the limit is not set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] HostPath | |
should support subPath [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:89 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:17:24.970: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:17:26.667: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-hostpath-2j5jw | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:37 | |
[It] should support subPath [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:89 | |
STEP: Creating a pod to test hostPath subPath | |
Jul 9 19:17:27.363: INFO: Waiting up to 5m0s for pod "pod-host-path-test" in namespace "e2e-tests-hostpath-2j5jw" to be "success or failure" | |
Jul 9 19:17:27.412: INFO: Pod "pod-host-path-test": Phase="Pending", Reason="", readiness=false. Elapsed: 48.766725ms | |
Jul 9 19:17:29.443: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2.080036384s | |
Jul 9 19:17:31.472: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 4.10831245s | |
Jul 9 19:17:33.500: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 6.137065362s | |
Jul 9 19:17:35.529: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 8.165850876s | |
Jul 9 19:17:37.584: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 10.220718947s | |
Jul 9 19:17:39.613: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 12.250143104s | |
Jul 9 19:17:41.652: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 14.288646069s | |
Jul 9 19:17:43.682: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 16.318657765s | |
Jul 9 19:17:45.716: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 18.353057718s | |
Jul 9 19:17:47.748: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 20.384946151s | |
Jul 9 19:17:49.778: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 22.415270934s | |
Jul 9 19:17:51.875: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 24.511994921s | |
Jul 9 19:17:53.910: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 26.546765122s | |
Jul 9 19:17:55.966: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 28.602672218s | |
Jul 9 19:17:57.994: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 30.63105058s | |
Jul 9 19:18:00.022: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 32.658880066s | |
Jul 9 19:18:02.060: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 34.696653256s | |
Jul 9 19:18:04.090: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 36.726588775s | |
Jul 9 19:18:06.125: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 38.761327977s | |
Jul 9 19:18:08.157: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 40.793418512s | |
Jul 9 19:18:10.187: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 42.823558149s | |
Jul 9 19:18:12.219: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 44.855600786s | |
Jul 9 19:18:14.247: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 46.883459371s | |
Jul 9 19:18:16.279: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 48.916003037s | |
Jul 9 19:18:18.308: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 50.945136203s | |
Jul 9 19:18:20.346: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 52.982891108s | |
Jul 9 19:18:22.375: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 55.012244154s | |
Jul 9 19:18:24.405: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 57.04207056s | |
Jul 9 19:18:26.441: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 59.077526327s | |
Jul 9 19:18:28.481: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m1.118107634s | |
Jul 9 19:18:30.517: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m3.153982095s | |
Jul 9 19:18:32.568: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m5.204918284s | |
Jul 9 19:18:34.606: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m7.242487448s | |
Jul 9 19:18:36.649: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m9.285458494s | |
Jul 9 19:18:38.696: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m11.332911416s | |
Jul 9 19:18:40.743: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m13.380168508s | |
Jul 9 19:18:42.772: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m15.409187535s | |
Jul 9 19:18:44.810: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m17.447200031s | |
Jul 9 19:18:46.872: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m19.508415647s | |
Jul 9 19:18:48.899: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m21.536103758s | |
Jul 9 19:18:50.930: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m23.566458835s | |
Jul 9 19:18:52.958: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m25.594760198s | |
Jul 9 19:18:54.994: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m27.630472752s | |
Jul 9 19:18:57.024: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m29.660624298s | |
Jul 9 19:18:59.056: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m31.692742004s | |
Jul 9 19:19:01.087: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m33.72369698s | |
Jul 9 19:19:03.138: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m35.775192157s | |
Jul 9 19:19:05.168: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m37.805126824s | |
Jul 9 19:19:07.198: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m39.83462179s | |
Jul 9 19:19:09.245: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m41.881375253s | |
Jul 9 19:19:11.274: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m43.910492024s | |
Jul 9 19:19:13.308: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m45.944920471s | |
Jul 9 19:19:15.359: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m47.996208568s | |
Jul 9 19:19:17.413: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m50.049810858s | |
Jul 9 19:19:19.456: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m52.092516178s | |
Jul 9 19:19:21.485: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m54.121965085s | |
Jul 9 19:19:23.516: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m56.153138837s | |
Jul 9 19:19:25.545: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 1m58.181907531s | |
Jul 9 19:19:27.578: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m0.215177602s | |
Jul 9 19:19:29.623: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m2.259343602s | |
Jul 9 19:19:31.653: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m4.289689214s | |
Jul 9 19:19:33.683: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m6.319717386s | |
Jul 9 19:19:35.718: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m8.354924772s | |
Jul 9 19:19:37.754: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m10.390980101s | |
Jul 9 19:19:39.785: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m12.422225392s | |
Jul 9 19:19:41.816: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m14.452619505s | |
Jul 9 19:19:43.846: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m16.482444513s | |
Jul 9 19:19:45.885: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m18.521624607s | |
Jul 9 19:19:47.915: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m20.551836216s | |
Jul 9 19:19:49.946: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m22.583179984s | |
Jul 9 19:19:51.975: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m24.611292454s | |
Jul 9 19:19:54.004: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m26.640859662s | |
Jul 9 19:19:56.033: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m28.66970714s | |
Jul 9 19:19:58.061: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m30.698201213s | |
Jul 9 19:20:00.092: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m32.72898788s | |
Jul 9 19:20:02.125: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m34.761317261s | |
Jul 9 19:20:04.155: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m36.791796274s | |
Jul 9 19:20:06.184: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m38.820527217s | |
Jul 9 19:20:08.223: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m40.85949819s | |
Jul 9 19:20:10.254: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m42.890730071s | |
Jul 9 19:20:12.289: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m44.92582505s | |
Jul 9 19:20:14.323: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m46.959632845s | |
Jul 9 19:20:16.355: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m48.991470927s | |
Jul 9 19:20:18.388: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m51.024328865s | |
Jul 9 19:20:20.423: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m53.059502388s | |
Jul 9 19:20:22.470: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m55.106788934s | |
Jul 9 19:20:24.508: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m57.144505221s | |
Jul 9 19:20:26.537: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 2m59.173553392s | |
Jul 9 19:20:28.572: INFO: Pod "pod-host-path-test": Phase="Running", Reason="", readiness=false. Elapsed: 3m1.209025858s | |
Jul 9 19:20:30.605: INFO: Pod "pod-host-path-test": Phase="Failed", Reason="", readiness=false. Elapsed: 3m3.242171498s | |
Jul 9 19:20:30.665: INFO: Output of node "ip-10-0-130-54.us-west-2.compute.internal" pod "pod-host-path-test" container "test-container-1": content of file "/test-volume/test-file": mount-tester new file | |
mode of file "/test-volume/test-file": -rw-r--r-- | |
Jul 9 19:20:30.734: INFO: Output of node "ip-10-0-130-54.us-west-2.compute.internal" pod "pod-host-path-test" container "test-container-2": Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
Error reading file /test-volume/sub-path/test-file: open /test-volume/sub-path/test-file: no such file or directory, retrying | |
STEP: delete the pod | |
Jul 9 19:20:30.894: INFO: Waiting for pod pod-host-path-test to disappear | |
Jul 9 19:20:30.929: INFO: Pod pod-host-path-test no longer exists | |
Jul 9 19:20:30.929: INFO: Unexpected error occurred: expected pod "pod-host-path-test" success: pod "pod-host-path-test" failed with status: {Phase:Failed Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason:ContainersNotReady Message:containers with unready status: [test-container-1 test-container-2]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:0001-01-01 00:00:00 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [test-container-1 test-container-2]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:10.0.130.54 PodIP:10.2.2.61 StartTime:2018-07-09 19:17:27 -0700 PDT InitContainerStatuses:[] ContainerStatuses:[{Name:test-container-1 State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2018-07-09 19:17:28 -0700 PDT,FinishedAt:2018-07-09 19:17:28 -0700 PDT,ContainerID:docker://2cd77b44fb6fdc32e044424e85163cc9d9a912bcc3ab095019a727af01cab8f8,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0 ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/mounttest-amd64@sha256:dc4e2dcfbde16249c4662de673295d00778577bc2e2ca7013a1b85d4f47398ca ContainerID:docker://2cd77b44fb6fdc32e044424e85163cc9d9a912bcc3ab095019a727af01cab8f8} {Name:test-container-2 State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:1,Signal:0,Reason:Error,Message:,StartedAt:2018-07-09 19:17:28 -0700 PDT,FinishedAt:2018-07-09 19:20:28 -0700 PDT,ContainerID:docker://f08d07a6f13f69f8f0450200ecff44ec62016d3cee4a8dc6778b39ab9588becd,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0 ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/mounttest-amd64@sha256:dc4e2dcfbde16249c4662de673295d00778577bc2e2ca7013a1b85d4f47398ca ContainerID:docker://f08d07a6f13f69f8f0450200ecff44ec62016d3cee4a8dc6778b39ab9588becd}] QOSClass:BestEffort} | |
[AfterEach] [sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Collecting events from namespace "e2e-tests-hostpath-2j5jw". | |
STEP: Found 7 events. | |
Jul 9 19:20:30.965: INFO: At 2018-07-09 19:17:27 -0700 PDT - event for pod-host-path-test: {default-scheduler } Scheduled: Successfully assigned e2e-tests-hostpath-2j5jw/pod-host-path-test to ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:20:30.965: INFO: At 2018-07-09 19:17:28 -0700 PDT - event for pod-host-path-test: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Container image "gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0" already present on machine | |
Jul 9 19:20:30.965: INFO: At 2018-07-09 19:17:28 -0700 PDT - event for pod-host-path-test: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:20:30.965: INFO: At 2018-07-09 19:17:28 -0700 PDT - event for pod-host-path-test: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Started: Started container | |
Jul 9 19:20:30.965: INFO: At 2018-07-09 19:17:28 -0700 PDT - event for pod-host-path-test: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Pulled: Container image "gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0" already present on machine | |
Jul 9 19:20:30.965: INFO: At 2018-07-09 19:17:28 -0700 PDT - event for pod-host-path-test: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Created: Created container | |
Jul 9 19:20:30.965: INFO: At 2018-07-09 19:17:28 -0700 PDT - event for pod-host-path-test: {kubelet ip-10-0-130-54.us-west-2.compute.internal} Started: Started container | |
Jul 9 19:20:31.095: INFO: POD NODE PHASE GRACE CONDITIONS | |
Jul 9 19:20:31.095: INFO: registry-6559c8c4db-45526 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: docker-build-1-build ip-10-0-130-54.us-west-2.compute.internal Pending [{Initialized False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:20:30 -0700 PDT ContainersNotInitialized containers with incomplete status: [manage-dockerfile]} {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:20:30 -0700 PDT ContainersNotReady containers with unready status: [docker-build]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [docker-build]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:20:30 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: execpod98j4h ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:16 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:15 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: frontend-1-build ip-10-0-130-54.us-west-2.compute.internal Succeeded [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:27 -0700 PDT PodCompleted } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:50 -0700 PDT PodCompleted } {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC PodCompleted } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:14:21 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: pod-configmaps-b625b422-83e7-11e8-bd2e-28d244b00276 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:19:45 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:19:46 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:19:45 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276 ip-10-0-130-54.us-west-2.compute.internal Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:20:29 -0700 PDT } {Ready False 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:20:29 -0700 PDT ContainersNotReady containers with unready status: [secret-volume-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC ContainersNotReady containers with unready status: [secret-volume-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 19:20:29 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-apiserver-cn2ps ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:45 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-controller-manager-558dc6fb98-q6vr5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:34 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-core-operator-75d546fbbb-c7ctx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:20 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:11 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-dns-787c975867-txmxv ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:22 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-flannel-bgv4g ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:59 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-flannel-m5wph ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:58 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:39 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-flannel-xcck7 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:17 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-proxy-5td7p ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:54 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-proxy-l2cnn ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:22 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-proxy-zsgcb ip-10-0-141-201.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:53 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-scheduler-68f8875b5c-s5tdr ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: metrics-server-5767bfc576-gfbwb ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: openshift-apiserver-rkms5 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:19 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: openshift-controller-manager-99d6586b-qq685 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:17:55 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: pod-checkpointer-4882g ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:03 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:08 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: prometheus-0 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:40 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:50:04 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-network-operator-jwwmp ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:15:13 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:14:24 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-node-controller-2ctqd ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:08 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:05 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:14 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:16:08 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: webconsole-6698d4fbbc-rgsw2 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:44 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: default-http-backend-6985d557bb-8h44n ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:38 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: router-6796c95fdf-2k4wk ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:37 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:46 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: directory-sync-d84d84d9f-j7pr6 ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:34:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:33:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: kube-addon-operator-675f99d7f8-c6pdt ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:29 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-alm-operator-79b6996f74-prs9h ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:35 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-channel-operator-5d878cd785-l66n4 ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:12 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-clu-6b8d87785f-fswbx ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:11 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:06 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-node-agent-r77mj ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:37:33 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:20 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-node-agent-rrwlg ip-10-0-130-54.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 18:12:57 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:32:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-stats-emitter-d87f669fd-988nl ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:29 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:36 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:19:23 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: tectonic-utility-operator-786b69fc8b-4xffz ip-10-0-35-213.us-west-2.compute.internal Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT } {Ready True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:41 -0700 PDT } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2018-07-09 13:18:13 -0700 PDT }] | |
Jul 9 19:20:31.095: INFO: | |
Jul 9 19:20:31.137: INFO: | |
Logging node info for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:20:31.173: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-130-54.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-130-54.us-west-2.compute.internal,UID:2f71bed0-83b7-11e8-84c6-0af96768d57e,ResourceVersion:79048,Generation:0,CreationTimestamp:2018-07-09 13:32:23 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-130-54,node-role.kubernetes.io/worker: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"3e:08:91:8f:b9:a5"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.130.54,node-configuration.v1.coreos.com/currentConfig: worker-2650561509,node-configuration.v1.coreos.com/desiredConfig: worker-2650561509,node-configuration.v1.coreos.com/targetConfig: worker-2650561509,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.2.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-0cb9cec2620663d39,Unschedulable:false,Taints:[],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365150208 0} {<nil>} 8169092Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260292608 0} {<nil>} 8066692Ki BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:20:27 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:20:27 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:20:27 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:20:27 -0700 PDT 2018-07-09 13:32:23 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:20:27 -0700 PDT 2018-07-09 13:33:23 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.130.54} {InternalDNS ip-10-0-130-54.us-west-2.compute.internal} {Hostname ip-10-0-130-54}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC283016-6CE7-ACE7-0F9A-02CE10505945,BootID:cfad64a2-03d7-403a-bd51-76866880a650,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[openshift/origin-haproxy-router@sha256:f0a71ada9e9ee48529540c2d4938b9caa55f9a0ac8a3be598e269ca5cebf70c0 openshift/origin-haproxy-router:v3.10.0-alpha.0] 1284960820} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-frlkw/test@sha256:ee11e7c7dbb2d609aaa42c8806ef1bf5663df95dd925e6ab424b4439dbaf75fd docker-registry.default.svc:5000/e2e-test-build-valuefrom-frlkw/test:latest] 613134548} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test@sha256:6daa01a6f7f0784905bf9dcbce49826d73d7c3c1d62a802f875ee7c10db02960 docker-registry.default.svc:5000/e2e-test-build-valuefrom-lxhjr/test:latest] 613134454} {[docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test@sha256:92c5e723d97318711a71afb9ee5c12c3c48b98d0f2aaa5e954095fabbcb505ee docker-registry.default.svc:5000/e2e-test-build-valuefrom-vmtzs/test:latest] 613133841} {[docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example@sha256:02d80c750d1e71afc7792f55f935c3dd6cde1788bee2b53ab554d29c903ca064 docker-registry.default.svc:5000/e2e-test-templates-l8j5q/cakephp-mysql-example:latest] 603384691} {[docker-registry.default.svc:5000/openshift/php@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7@sha256:164fa6fc06cdd98c98be09ecff5398feef3d479652899f983cf41e644d604ff0 centos/php-70-centos7:latest] 589408618} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass@sha256:880359284c1e0933fe5f2db29b8c4d948b70da3dfb26a0462f68b23397740b0a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejspass:latest] 568094192} {[docker-registry.default.svc:5000/openshift/php@sha256:59c3d53372cd7097494187f5a58bab58a1d956a340b70a23c84a0d000a565cbe] 567254500} {[docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test@sha256:539e80a4de02794f6126cffce75562bcb721041c6d443c5ced15ba286d70e229 docker-registry.default.svc:5000/e2e-test-build-timing-bl897/test:latest] 566117187} {[docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test@sha256:e0eeef684e9de55219871fa9e360d73a1163cfc407c626eade862cbee5a9bbc5 docker-registry.default.svc:5000/e2e-test-build-timing-wkh4w/test:latest] 566117040} {[centos/ruby-22-centos7@sha256:a18c8706118a5c4c9f1adf045024d2abf06ba632b5674b23421019ee4d3edcae centos/ruby-22-centos7:latest] 566117040} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot@sha256:4084131a9910c10780186608faf5a9643de0f18d09c27fe828499a8d180abfba docker-registry.default.svc:5000/e2e-test-s2i-build-root-phq84/nodejsroot:latest] 560696751} {[docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot@sha256:0397f7e12d87d62c539356a4936348d0a8deb40e1b5e970cdd1744d3e6ffa05a docker-registry.default.svc:5000/e2e-test-s2i-build-root-4l2gq/nodejsroot:latest] 560696751} {[centos/nodejs-6-centos7@sha256:b2867b5008d9e975b3d4710ec0f31cdc96b079b83334b17e03a60602a7a590fc] 560696751} {[docker-registry.default.svc:5000/openshift/ruby@sha256:2e83b9e07e85960060096b6aff7ee202a5f52e0e18447641b080b1f3879e0901] 536571487} {[docker-registry.default.svc:5000/openshift/ruby@sha256:8f00b7a5789887b72db0415355830c87e18804b774a922a424736f5237a44933] 518934530} {[docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678@sha256:a9ecb5931f283c598dcaf3aca9025599eb71115bd0f2cd0f1989a9f37394efad docker-registry.default.svc:5000/e2e-test-new-app-xn8nh/a234567890123456789012345678901234567890123456789012345678:latest] 511744495} {[docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample@sha256:95a78c60dc1709c2212cd8cc48cd3fffe6cdcdd847674497d9aa5d7891551699 docker-registry.default.svc:5000/e2e-test-prometheus-4fqst/origin-nodejs-sample:latest] 511744370} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:3bb2aed7578ab5b6ba2bf22993df3c73ef91bdb02e273cc0ce8e529de7ee5660] 506453985} {[docker-registry.default.svc:5000/openshift/ruby@sha256:0eaaed9fae1b0d9bc8ed73b93d581c6ab019a92277484c9acf52fa60b3269a7c] 504578679} {[docker-registry.default.svc:5000/openshift/nodejs@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653 centos/nodejs-8-centos7@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653] 504452018} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:896482969cd659b419bc444c153a74d11820655c7ed19b5094b8eb041f0065d6] 487132847} {[docker-registry.default.svc:5000/openshift/mongodb@sha256:91955c14f978a0f48918eecc8b3772faf1615e943daccf9bb051a51cba30422f] 465041680} {[openshift/origin-docker-builder@sha256:4fe8032f87d2f8485a711ec60a9ffb330e42a6cd8d232ad3cf63c42471cfab29 openshift/origin-docker-builder:latest] 447580928} {[docker-registry.default.svc:5000/openshift/mysql@sha256:d03537ef57d51b13e6ad4a73a382ca180a0e02d975c8237790410f45865aae3c] 429435940} {[openshift/origin-haproxy-router@sha256:485fa86ac97b0d289411b3216fb8970989cd580817ebb5fcbb0f83a6dc2466f5 openshift/origin-haproxy-router:latest] 394965919} {[openshift/origin-deployer@sha256:1295e5be56fc03d4c482194378a882f2e96a8d23eadaf6dd32d603d3e877df99 openshift/origin-deployer:latest] 371674595} {[openshift/origin-web-console@sha256:d2cbbb533d26996226add8cb327cb2060e7a03c6aa96ad94cd236d4064c094ce openshift/origin-web-console:latest] 336636057} {[openshift/prometheus@sha256:35e2e0efc874c055be60a025874256816c98b9cebc10f259d7fb806bbe68badf openshift/prometheus:v2.2.1] 317896379} {[openshift/origin-docker-registry@sha256:c40ebb707721327c3b9c79f0e8e7f02483f034355d4149479333cc134b72967c openshift/origin-docker-registry:latest] 302637209} {[openshift/origin-pod@sha256:8fbd41f21824f5981716568790c5f78a4710bb0709ce9c473eb21ad2fbc5e877 openshift/origin-pod:latest] 251747200} {[openshift/origin-base@sha256:43dd97db435025eee02606658cfcccbc0a8ac4135e0d8870e91930d6cab8d1fd openshift/origin-base:latest] 228695137} {[openshift/oauth-proxy@sha256:4b73830ee6f7447d0921eedc3946de50016eb8f048d66ea3969abc4116f1e42a openshift/oauth-proxy:v1.0.0] 228241928} {[openshift/prometheus-alertmanager@sha256:35443abf6c5cf99b080307fe0f98098334f299780537a3e61ac5604cbfe48f7e openshift/prometheus-alertmanager:v0.14.0] 221857684} {[openshift/prometheus-alert-buffer@sha256:076f8dd576806f5c2dde7e536d020c31aa7d2ec7dcea52da6cbb944895def7ba openshift/prometheus-alert-buffer:v0.0.2] 200521084} {[docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage@sha256:df3e69e3fe1bc86897717b020b6caa000f1f97c14dc0b3853ca0d7149412da54 docker-registry.default.svc:5000/e2e-test-build-multistage-9cdjf/multi-stage:v1] 199835207} {[centos@sha256:b67d21dfe609ddacf404589e04631d90a342921e81c40aeaf3391f6717fa5322 centos@sha256:eed5b251b615d1e70b10bcec578d64e8aa839d2785c2ffd5424e472818c42755 centos:7 centos:centos7] 199678471} {[docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1@sha256:3967cd8851952bbba0b3a4d9c038f36dc5001463c8521d6955ab0f3f4598d779 docker-registry.default.svc:5000/e2e-test-docker-build-pullsecret-s4b2n/image1:latest] 199678471} {[k8s.gcr.io/nginx-slim-amd64@sha256:6654db6d4028756062edac466454ee5c9cf9b20ef79e35a81e3c840031eb1e2b k8s.gcr.io/nginx-slim-amd64:0.20] 103591055} {[gcr.io/google_containers/metrics-server-amd64@sha256:54d2cf293e01f72d9be0e7c4f2c98e31f599088a9426a6415fe62426d446f5b2 gcr.io/google_containers/metrics-server-amd64:v0.2.0] 96501893} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/directory-sync@sha256:e5e7fe901868853d89c2c0697cc88f0686c6ba1178ca045ec57bfd18e7000048 quay.io/coreos/directory-sync:v0.0.2] 38433928} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[k8s.gcr.io/addon-resizer@sha256:d00afd42fc267fa3275a541083cfe67d160f966c788174b44597434760e1e1eb k8s.gcr.io/addon-resizer:2.1] 26450138} {[quay.io/coreos/tectonic-error-server@sha256:aefa0a012e103bee299c17e798e5830128588b6ef5d4d1f6bc8ae5804bc4d8cd quay.io/coreos/tectonic-error-server:1.1] 12714516} {[gcr.io/google_containers/dnsutils@sha256:cd9182f6d74e616942db1cef6f25e1e54b49ba0330c2e19d3ec061f027666cc0 gcr.io/google_containers/dnsutils:e2e] 8897789} {[gcr.io/kubernetes-e2e-test-images/hostexec-amd64@sha256:bdaecec5adfa7c79e9525c0992fdab36c2d68066f5e91eff0d1d9e8d73c654ea gcr.io/kubernetes-e2e-test-images/hostexec-amd64:1.1] 8407119}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:20:31.173: INFO: | |
Logging kubelet events for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:20:31.202: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:20:31.311: INFO: prometheus-0 started at 2018-07-09 13:50:04 -0700 PDT (0+6 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container alert-buffer ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container alertmanager ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container alertmanager-proxy ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container alerts-proxy ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container prom-proxy ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container prometheus ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: kube-proxy-5td7p started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: registry-6559c8c4db-45526 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container registry ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: docker-build-1-build started at 2018-07-09 19:20:30 -0700 PDT (1+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Init container manage-dockerfile ready: false, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container docker-build ready: false, restart count 0 | |
Jul 9 19:20:31.311: INFO: execpod98j4h started at 2018-07-09 19:14:15 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container exec ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: kube-flannel-xcck7 started at 2018-07-09 13:32:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: metrics-server-5767bfc576-gfbwb started at 2018-07-09 13:33:23 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container metrics-server ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container metrics-server-nanny ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: webconsole-6698d4fbbc-rgsw2 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container webconsole ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: tectonic-node-agent-rrwlg started at 2018-07-09 13:32:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container node-agent ready: true, restart count 3 | |
Jul 9 19:20:31.311: INFO: directory-sync-d84d84d9f-j7pr6 started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container directory-sync ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: pod-projected-secrets-d0caac8f-83e7-11e8-992b-28d244b00276 started at 2018-07-09 19:20:29 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container secret-volume-test ready: false, restart count 0 | |
Jul 9 19:20:31.311: INFO: pod-configmaps-b625b422-83e7-11e8-bd2e-28d244b00276 started at 2018-07-09 19:19:45 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container configmap-volume-test ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: frontend-1-build started at 2018-07-09 19:14:21 -0700 PDT (2+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Init container git-clone ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Init container manage-dockerfile ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: Container sti-build ready: false, restart count 0 | |
Jul 9 19:20:31.311: INFO: default-http-backend-6985d557bb-8h44n started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container default-http-backend ready: true, restart count 0 | |
Jul 9 19:20:31.311: INFO: router-6796c95fdf-2k4wk started at 2018-07-09 13:33:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:20:31.311: INFO: Container router ready: true, restart count 0 | |
W0709 19:20:31.345070 11714 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:20:31.473: INFO: | |
Latency metrics for node ip-10-0-130-54.us-west-2.compute.internal | |
Jul 9 19:20:31.473: INFO: {Operation:stop_container Method:docker_operations_latency_microseconds Quantile:0.99 Latency:30.274794s} | |
Jul 9 19:20:31.473: INFO: | |
Logging node info for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:20:31.505: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-141-201.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-141-201.us-west-2.compute.internal,UID:ab76db34-83b4-11e8-8888-0af96768d57e,ResourceVersion:79139,Generation:0,CreationTimestamp:2018-07-09 13:14:22 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2b,kubernetes.io/hostname: ip-10-0-141-201,node-role.kubernetes.io/etcd: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"b6:11:a8:d0:6d:85"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.141.201,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.1.0/24,ExternalID:,ProviderID:aws:///us-west-2b/i-03457d640f9c71dd1,Unschedulable:false,Taints:[{node-role.kubernetes.io/etcd NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:20:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:20:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:20:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:20:30 -0700 PDT 2018-07-09 13:14:22 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:20:30 -0700 PDT 2018-07-09 13:16:04 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.141.201} {InternalDNS ip-10-0-141-201.us-west-2.compute.internal} {Hostname ip-10-0-141-201}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2F6BCA-4D59-F6AA-8C7B-027F94D52D78,BootID:92773d40-1311-4ad5-b294-38db65faf16c,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/kube-client-agent@sha256:8564ab65bcb1064006d2fc9c6e32a5ca3f4326cdd2da9a2efc4fb7cc0e0b6041 quay.io/coreos/kube-client-agent:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 33236131} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:20:31.505: INFO: | |
Logging kubelet events for node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:20:31.534: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-141-201.us-west-2.compute.internal | |
Jul 9 19:21:01.575: INFO: Unable to retrieve kubelet pods for node ip-10-0-141-201.us-west-2.compute.internal: the server is currently unable to handle the request (get nodes ip-10-0-141-201.us-west-2.compute.internal:10250) | |
Jul 9 19:21:01.575: INFO: | |
Logging node info for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:21:01.605: INFO: Node Info: &Node{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:ip-10-0-35-213.us-west-2.compute.internal,GenerateName:,Namespace:,SelfLink:/api/v1/nodes/ip-10-0-35-213.us-west-2.compute.internal,UID:a83cf873-83b4-11e8-8888-0af96768d57e,ResourceVersion:79463,Generation:0,CreationTimestamp:2018-07-09 13:14:17 -0700 PDT,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{beta.kubernetes.io/arch: amd64,beta.kubernetes.io/instance-type: m4.large,beta.kubernetes.io/os: linux,failure-domain.beta.kubernetes.io/region: us-west-2,failure-domain.beta.kubernetes.io/zone: us-west-2c,kubernetes.io/hostname: ip-10-0-35-213,node-role.kubernetes.io/master: ,},Annotations:map[string]string{flannel.alpha.coreos.com/backend-data: {"VtepMAC":"5e:08:be:54:0d:9f"},flannel.alpha.coreos.com/backend-type: vxlan,flannel.alpha.coreos.com/kube-subnet-manager: true,flannel.alpha.coreos.com/public-ip: 10.0.35.213,node-configuration.v1.coreos.com/currentConfig: master-2063737633,node-configuration.v1.coreos.com/desiredConfig: master-2063737633,node-configuration.v1.coreos.com/targetConfig: master-2063737633,volumes.kubernetes.io/controller-managed-attach-detach: true,},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Spec:NodeSpec{PodCIDR:10.2.0.0/24,ExternalID:,ProviderID:aws:///us-west-2c/i-0e1d36783c9705b28,Unschedulable:false,Taints:[{node-role.kubernetes.io/master NoSchedule <nil>}],ConfigSource:nil,},Status:NodeStatus{Capacity:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8365146112 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Allocatable:ResourceList{cpu: {{2 0} {<nil>} 2 DecimalSI},hugepages-1Gi: {{0 0} {<nil>} 0 DecimalSI},hugepages-2Mi: {{0 0} {<nil>} 0 DecimalSI},memory: {{8260288512 0} {<nil>} BinarySI},pods: {{110 0} {<nil>} 110 DecimalSI},},Phase:,Conditions:[{OutOfDisk False 2018-07-09 19:20:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientDisk kubelet has sufficient disk space available} {MemoryPressure False 2018-07-09 19:20:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientMemory kubelet has sufficient memory available} {DiskPressure False 2018-07-09 19:20:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasNoDiskPressure kubelet has no disk pressure} {PIDPressure False 2018-07-09 19:20:59 -0700 PDT 2018-07-09 13:14:17 -0700 PDT KubeletHasSufficientPID kubelet has sufficient PID available} {Ready True 2018-07-09 19:20:59 -0700 PDT 2018-07-09 13:16:08 -0700 PDT KubeletReady kubelet is posting ready status}],Addresses:[{InternalIP 10.0.35.213} {ExternalIP 34.220.249.237} {InternalDNS ip-10-0-35-213.us-west-2.compute.internal} {ExternalDNS ec2-34-220-249-237.us-west-2.compute.amazonaws.com} {Hostname ip-10-0-35-213}],DaemonEndpoints:NodeDaemonEndpoints{KubeletEndpoint:DaemonEndpoint{Port:10250,},},NodeInfo:NodeSystemInfo{MachineID:,SystemUUID:EC2ED297-E036-AA0D-C4ED-9057B3EA9001,BootID:7f784e0b-09a6-495a-b787-3d8619214f8a,KernelVersion:4.14.48-coreos-r2,OSImage:CentOS Linux 7 (Core),ContainerRuntimeVersion:docker://18.3.1,KubeletVersion:v1.11.0+d4cacc0,KubeProxyVersion:v1.11.0+d4cacc0,OperatingSystem:linux,Architecture:amd64,},Images:[{[openshift/origin-node@sha256:e6d4cf7a40809bd51834316e10248ea33bb523573c25a5740e1ff34388f29f70 openshift/origin-node:latest] 1302551512} {[quay.io/coreos/hyperkube@sha256:b5eb7d69ca52899d0cbc395a68129b6ca35d1f632a189d7869b25968e2065a19 quay.io/coreos/hyperkube:v1.9.3_coreos.0] 652363757} {[openshift/origin-hypershift@sha256:3b26011ae771a6036a7533d970052be5c04bc1f6e6812314ffefd902f40910fd openshift/origin-hypershift:latest] 518022163} {[openshift/origin-hyperkube@sha256:11a08060b48d226d64d4bb5234f2386bf22472a0835c5b91f0fb0db25b0a7e19 openshift/origin-hyperkube:latest] 498702039} {[quay.io/coreos/awscli@sha256:1d6ea2f37c248a4f4f2a70126f0b8555fd0804d4e65af3b30c3a949247ea13a6 quay.io/coreos/awscli:025a357f05242fdad6a81e8a6b520098aa65a600] 97521631} {[quay.io/coreos/bootkube@sha256:63afddd30deedff273d65607f4fcf0b331f4418838a00c69b6ab7a5754a24f5a quay.io/coreos/bootkube:v0.10.0] 84921995} {[quay.io/coreos/flannel@sha256:056cf57fd3bbe7264c0be1a3b34ec2e289b33e51c70f332f4e88aa83970ad891 quay.io/coreos/flannel:v0.9.1-amd64] 51338831} {[gcr.io/google_containers/k8s-dns-kube-dns-amd64@sha256:6d8e0da4fb46e9ea2034a3f4cab0e095618a2ead78720c12e791342738e5f85d gcr.io/google_containers/k8s-dns-kube-dns-amd64:1.14.8] 50456751} {[quay.io/coreos/flannel-cni@sha256:734b4222110980abd0abe74974b6ca36452d26bdd2a20e25f37fdf7fdc2da170 quay.io/coreos/flannel-cni:v0.3.0] 49786179} {[quay.io/coreos/tectonic-stats@sha256:e800fe60dd1a0f89f8ae85caae9209201254e17d889d664d633ed08e274e2a39 quay.io/coreos/tectonic-stats:6e882361357fe4b773adbf279cddf48cb50164c1] 48779830} {[quay.io/coreos/pod-checkpointer@sha256:1e1e48228f872d56c8a57a5e12adb5239ae9e6206536baf2904e4bf03314c8e8 quay.io/coreos/pod-checkpointer:9dc83e1ab3bc36ca25c9f7c18ddef1b91d4a0558] 47992230} {[quay.io/coreos/tectonic-network-operator-dev@sha256:e29d797f5740cf6f5c0ccc0de2b3e606d187acbdc0bb79a4397c058d8840c8fe quay.io/coreos/tectonic-network-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44068170} {[quay.io/coreos/tectonic-node-controller-operator-dev@sha256:7a31568c6c2e398cffa7e8387cf51543e3bf1f01b4a050a5d00a9b593c3dace0 quay.io/coreos/tectonic-node-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44053165} {[quay.io/coreos/kube-addon-operator-dev@sha256:e327727a93813c31f6d65f76f2998722754b8ccb5110949153e55f2adbc2374e quay.io/coreos/kube-addon-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 44052211} {[quay.io/coreos/tectonic-utility-operator-dev@sha256:4fb4de52c7aa64ce124e1bf73fb27989356c414101ecc19ca4ec9ab80e00a88d quay.io/coreos/tectonic-utility-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43818409} {[quay.io/coreos/tectonic-ingress-controller-operator-dev@sha256:5e96253c8fe8357473d4806b116fcf03fe18dcad466a88083f9b9310045821f1 quay.io/coreos/tectonic-ingress-controller-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 43808038} {[quay.io/coreos/tectonic-alm-operator@sha256:ce32e6d4745040be8807d09eb925b2b076b60fb0a93e33302b74a5cc8f294ca5 quay.io/coreos/tectonic-alm-operator:v0.3.1] 43202998} {[gcr.io/google_containers/k8s-dns-sidecar-amd64@sha256:23df717980b4aa08d2da6c4cfa327f1b730d92ec9cf740959d2d5911830d82fb gcr.io/google_containers/k8s-dns-sidecar-amd64:1.14.8] 42210862} {[gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64@sha256:93c827f018cf3322f1ff2aa80324a0306048b0a69bc274e423071fb0d2d29d8b gcr.io/google_containers/k8s-dns-dnsmasq-nanny-amd64:1.14.8] 40951779} {[quay.io/coreos/kube-core-operator-dev@sha256:6cc0dd2405f19014b41a0eed57c39160aeb92c2380ac8f8a067ce7dee476cba2 quay.io/coreos/kube-core-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40849618} {[quay.io/coreos/tectonic-channel-operator-dev@sha256:6eeb84c385333755a2189c199587bc26db6c5d897e1962d7e1047dec2531e85e quay.io/coreos/tectonic-channel-operator-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 40523592} {[quay.io/coreos/etcd@sha256:688e6c102955fe927c34db97e6352d0e0962554735b2db5f2f66f3f94cfe8fd1 quay.io/coreos/etcd:v3.2.14] 37228348} {[quay.io/coreos/tectonic-node-agent@sha256:1eb929223a1ecc12b246cbe4ba366092b139067aadb809fac8189b8ca274fab4 quay.io/coreos/tectonic-node-agent:92fb13f8b2a46702b8c10f278408f37d825ee2cb] 36791814} {[quay.io/coreos/kube-core-renderer-dev@sha256:a595dfe57b7992971563fcea8ac1858c306529a465f9b690911f4220d93d3c5c quay.io/coreos/kube-core-renderer-dev:c3cee2bc5673011e88ac7b0ab1659c2c7243a499] 36535818} {[quay.io/coreos/kube-etcd-signer-server@sha256:c4c0becf6779523af5b644b53375d61bed9c4688d496cb2f88d4f08024ac5390 quay.io/coreos/kube-etcd-signer-server:678cc8e6841e2121ebfdb6e2db568fce290b67d6] 34655544} {[quay.io/coreos/tectonic-node-controller-dev@sha256:c9c17f7c4c738e519e36224ae8c71d3a881b92ffb86fdb75f358efebafa27d84 quay.io/coreos/tectonic-node-controller-dev:a437848532713f2fa4137e9a0f4f6a689cf554a8] 25570332} {[quay.io/coreos/tectonic-clu@sha256:4e6a907a433e741632c8f9a7d9d9009bc08ac494dce05e0a19f8fa0a440a3926 quay.io/coreos/tectonic-clu:v0.0.1] 5081911} {[quay.io/coreos/tectonic-stats-extender@sha256:6e7fe41ca2d63791c08d2cc4b4311d9e01b37fa3dc116d3e77e7306cbe29a0f1 quay.io/coreos/tectonic-stats-extender:487b3da4e175da96dabfb44fba65cdb8b823db2e] 2818916} {[k8s.gcr.io/pause@sha256:f78411e19d84a252e53bff71a4407a5686c46983a2c2eeed83929b888179acea k8s.gcr.io/pause:3.1] 742472}],VolumesInUse:[],VolumesAttached:[],},} | |
Jul 9 19:21:01.605: INFO: | |
Logging kubelet events for node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:21:01.638: INFO: | |
Logging pods the kubelet thinks is on node ip-10-0-35-213.us-west-2.compute.internal | |
Jul 9 19:21:01.741: INFO: tectonic-stats-emitter-d87f669fd-988nl started at 2018-07-09 13:19:23 -0700 PDT (1+2 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Init container tectonic-stats-extender-init ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: Container tectonic-stats-emitter ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: Container tectonic-stats-extender ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: tectonic-channel-operator-5d878cd785-l66n4 started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-channel-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: kube-proxy-l2cnn started at 2018-07-09 13:14:22 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container kube-proxy ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: openshift-apiserver-rkms5 started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container openshift-apiserver ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: tectonic-network-operator-jwwmp started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-network-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: kube-dns-787c975867-txmxv started at 2018-07-09 13:16:08 -0700 PDT (0+3 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container dnsmasq ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: Container kubedns ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: Container sidecar ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: kube-scheduler-68f8875b5c-s5tdr started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container kube-scheduler ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: tectonic-clu-6b8d87785f-fswbx started at 2018-07-09 13:19:06 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-clu ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: kube-apiserver-cn2ps started at 2018-07-09 13:14:23 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container kube-apiserver ready: true, restart count 4 | |
Jul 9 19:21:01.741: INFO: tectonic-node-controller-2ctqd started at 2018-07-09 13:18:05 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-node-controller ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: tectonic-alm-operator-79b6996f74-prs9h started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-alm-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: tectonic-ingress-controller-operator-78c4d6b9fc-bknh9 started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-ingress-controller-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: tectonic-node-agent-r77mj started at 2018-07-09 13:19:20 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container node-agent ready: true, restart count 4 | |
Jul 9 19:21:01.741: INFO: pod-checkpointer-4882g started at 2018-07-09 13:14:24 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container pod-checkpointer ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: kube-flannel-m5wph started at 2018-07-09 13:15:39 -0700 PDT (0+2 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container install-cni ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: Container kube-flannel ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: openshift-controller-manager-99d6586b-qq685 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container openshift-controller-manager ready: true, restart count 3 | |
Jul 9 19:21:01.741: INFO: tectonic-node-controller-operator-648b9f5d6d-nxvlb started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-node-controller-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: kube-core-operator-75d546fbbb-c7ctx started at 2018-07-09 13:18:11 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container kube-core-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: tectonic-utility-operator-786b69fc8b-4xffz started at 2018-07-09 13:18:13 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container tectonic-utility-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: kube-addon-operator-675f99d7f8-c6pdt started at 2018-07-09 13:18:12 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container kube-addon-operator ready: true, restart count 0 | |
Jul 9 19:21:01.741: INFO: pod-checkpointer-4882g-ip-10-0-35-213.us-west-2.compute.internal started at <nil> (0+0 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: kube-controller-manager-558dc6fb98-q6vr5 started at 2018-07-09 13:16:08 -0700 PDT (0+1 container statuses recorded) | |
Jul 9 19:21:01.741: INFO: Container kube-controller-manager ready: true, restart count 1 | |
W0709 19:21:01.775908 11714 metrics_grabber.go:81] Master node is not registered. Grabbing metrics from Scheduler, ControllerManager and ClusterAutoscaler is disabled. | |
Jul 9 19:21:01.867: INFO: | |
Latency metrics for node ip-10-0-35-213.us-west-2.compute.internal | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:21:01.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-hostpath-2j5jw" for this suite. | |
Jul 9 19:21:08.030: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:10.794: INFO: namespace: e2e-tests-hostpath-2j5jw, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:11.468: INFO: namespace e2e-tests-hostpath-2j5jw deletion completed in 9.534952564s | |
• Failure [226.498 seconds] | |
[sig-storage] HostPath | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:34 | |
should support subPath [Suite:openshift/conformance/parallel] [Suite:k8s] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/host_path.go:89 | |
Expected error: | |
<*errors.errorString | 0xc422109f10>: { | |
s: "expected pod \"pod-host-path-test\" success: pod \"pod-host-path-test\" failed with status: {Phase:Failed Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason:ContainersNotReady Message:containers with unready status: [test-container-1 test-container-2]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:0001-01-01 00:00:00 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [test-container-1 test-container-2]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:10.0.130.54 PodIP:10.2.2.61 StartTime:2018-07-09 19:17:27 -0700 PDT InitContainerStatuses:[] ContainerStatuses:[{Name:test-container-1 State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2018-07-09 19:17:28 -0700 PDT,FinishedAt:2018-07-09 19:17:28 -0700 PDT,ContainerID:docker://2cd77b44fb6fdc32e044424e85163cc9d9a912bcc3ab095019a727af01cab8f8,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0 ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/mounttest-amd64@sha256:dc4e2dcfbde16249c4662de673295d00778577bc2e2ca7013a1b85d4f47398ca ContainerID:docker://2cd77b44fb6fdc32e044424e85163cc9d9a912bcc3ab095019a727af01cab8f8} {Name:test-container-2 State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:1,Signal:0,Reason:Error,Message:,StartedAt:2018-07-09 19:17:28 -0700 PDT,FinishedAt:2018-07-09 19:20:28 -0700 PDT,ContainerID:docker://f08d07a6f13f69f8f0450200ecff44ec62016d3cee4a8dc6778b39ab9588becd,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0 ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/mounttest-amd64@sha256:dc4e2dcfbde16249c4662de673295d00778577bc2e2ca7013a1b85d4f47398ca ContainerID:docker://f08d07a6f13f69f8f0450200ecff44ec62016d3cee4a8dc6778b39ab9588becd}] QOSClass:BestEffort}", | |
} | |
expected pod "pod-host-path-test" success: pod "pod-host-path-test" failed with status: {Phase:Failed Conditions:[{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason: Message:} {Type:Ready Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason:ContainersNotReady Message:containers with unready status: [test-container-1 test-container-2]} {Type:ContainersReady Status:False LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:0001-01-01 00:00:00 +0000 UTC Reason:ContainersNotReady Message:containers with unready status: [test-container-1 test-container-2]} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2018-07-09 19:17:27 -0700 PDT Reason: Message:}] Message: Reason: NominatedNodeName: HostIP:10.0.130.54 PodIP:10.2.2.61 StartTime:2018-07-09 19:17:27 -0700 PDT InitContainerStatuses:[] ContainerStatuses:[{Name:test-container-1 State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:0,Signal:0,Reason:Completed,Message:,StartedAt:2018-07-09 19:17:28 -0700 PDT,FinishedAt:2018-07-09 19:17:28 -0700 PDT,ContainerID:docker://2cd77b44fb6fdc32e044424e85163cc9d9a912bcc3ab095019a727af01cab8f8,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0 ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/mounttest-amd64@sha256:dc4e2dcfbde16249c4662de673295d00778577bc2e2ca7013a1b85d4f47398ca ContainerID:docker://2cd77b44fb6fdc32e044424e85163cc9d9a912bcc3ab095019a727af01cab8f8} {Name:test-container-2 State:{Waiting:nil Running:nil Terminated:&ContainerStateTerminated{ExitCode:1,Signal:0,Reason:Error,Message:,StartedAt:2018-07-09 19:17:28 -0700 PDT,FinishedAt:2018-07-09 19:20:28 -0700 PDT,ContainerID:docker://f08d07a6f13f69f8f0450200ecff44ec62016d3cee4a8dc6778b39ab9588becd,}} LastTerminationState:{Waiting:nil Running:nil Terminated:nil} Ready:false RestartCount:0 Image:gcr.io/kubernetes-e2e-test-images/mounttest-amd64:1.0 ImageID:docker-pullable://gcr.io/kubernetes-e2e-test-images/mounttest-amd64@sha256:dc4e2dcfbde16249c4662de673295d00778577bc2e2ca7013a1b85d4f47398ca ContainerID:docker://f08d07a6f13f69f8f0450200ecff44ec62016d3cee4a8dc6778b39ab9588becd}] QOSClass:BestEffort} | |
not to have occurred | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:2290 | |
------------------------------ | |
[sig-storage] Downward API volume | |
should update labels on modification [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:20:42.538: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:20:44.601: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-qpnsn | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:38 | |
[It] should update labels on modification [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating the pod | |
Jul 9 19:20:48.175: INFO: Successfully updated pod "labelsupdateda06d733-83e7-11e8-992b-28d244b00276" | |
[AfterEach] [sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:50.256: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-qpnsn" for this suite. | |
Jul 9 19:21:12.411: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:15.398: INFO: namespace: e2e-tests-downward-api-qpnsn, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:16.709: INFO: namespace e2e-tests-downward-api-qpnsn deletion completed in 26.411067994s | |
• [SLOW TEST:34.170 seconds] | |
[sig-storage] Downward API volume | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downwardapi_volume.go:33 | |
should update labels on modification [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-api-machinery] Downward API | |
should provide pod UID as env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:21:03.810: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:21:05.487: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-downward-api-bcvkn | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should provide pod UID as env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward api env vars | |
Jul 9 19:21:06.209: INFO: Waiting up to 5m0s for pod "downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-downward-api-bcvkn" to be "success or failure" | |
Jul 9 19:21:06.241: INFO: Pod "downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 32.021781ms | |
Jul 9 19:21:08.278: INFO: Pod "downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.068938105s | |
Jul 9 19:21:10.317: INFO: Pod "downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.107594978s | |
STEP: Saw pod success | |
Jul 9 19:21:10.317: INFO: Pod "downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:21:10.352: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276 container dapi-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:21:10.434: INFO: Waiting for pod downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:21:10.471: INFO: Pod downward-api-e67058f8-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:21:10.472: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-downward-api-bcvkn" for this suite. | |
Jul 9 19:21:16.624: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:19.200: INFO: namespace: e2e-tests-downward-api-bcvkn, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:20.501: INFO: namespace e2e-tests-downward-api-bcvkn deletion completed in 9.989426573s | |
• [SLOW TEST:16.691 seconds] | |
[sig-api-machinery] Downward API | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/downward_api.go:37 | |
should provide pod UID as env vars [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
S | |
------------------------------ | |
[sig-storage] ConfigMap | |
updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:19:42.933: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:19:44.473: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-configmap-bc6g9 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
Jul 9 19:19:45.108: INFO: Couldn't get node TTL annotation (using default value of 0): No TTL annotation found on the node | |
STEP: Creating configMap with name configmap-test-upd-b620d07a-83e7-11e8-bd2e-28d244b00276 | |
STEP: Creating the pod | |
STEP: Updating configmap configmap-test-upd-b620d07a-83e7-11e8-bd2e-28d244b00276 | |
STEP: waiting to observe update in volume | |
[AfterEach] [sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:20:58.821: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-configmap-bc6g9" for this suite. | |
Jul 9 19:21:20.943: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:23.203: INFO: namespace: e2e-tests-configmap-bc6g9, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:24.227: INFO: namespace e2e-tests-configmap-bc6g9 deletion completed in 25.372270298s | |
• [SLOW TEST:101.294 seconds] | |
[sig-storage] ConfigMap | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/configmap_volume.go:32 | |
updates should be reflected in volume [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SSS | |
------------------------------ | |
[k8s.io] Docker Containers | |
should be able to override the image's default command and arguments [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [k8s.io] Docker Containers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:21:20.503: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:21:22.197: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-containers-f9hmf | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be able to override the image's default command and arguments [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test override all | |
Jul 9 19:21:22.891: INFO: Waiting up to 5m0s for pod "client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-containers-f9hmf" to be "success or failure" | |
Jul 9 19:21:22.922: INFO: Pod "client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.522108ms | |
Jul 9 19:21:24.953: INFO: Pod "client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.061966293s | |
Jul 9 19:21:26.984: INFO: Pod "client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.093040223s | |
STEP: Saw pod success | |
Jul 9 19:21:26.984: INFO: Pod "client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:21:27.020: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:21:27.092: INFO: Waiting for pod client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:21:27.124: INFO: Pod client-containers-f06198fc-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [k8s.io] Docker Containers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:21:27.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-containers-f9hmf" for this suite. | |
Jul 9 19:21:33.256: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:36.736: INFO: namespace: e2e-tests-containers-f9hmf, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:37.289: INFO: namespace e2e-tests-containers-f9hmf deletion completed in 10.128596748s | |
• [SLOW TEST:16.786 seconds] | |
[k8s.io] Docker Containers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:669 | |
should be able to override the image's default command and arguments [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[sig-storage] Projected | |
should be able to mount in a volume regardless of a different secret existing with same name in different namespace [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:87 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:21:24.230: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:21:25.848: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-pjp6w | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:87 | |
Jul 9 19:21:26.743: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secret-namespace-gr7ll | |
STEP: Creating projection with secret that has name projected-secret-test-f29dc6de-83e7-11e8-bd2e-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:21:27.344: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276" in namespace "e2e-tests-projected-pjp6w" to be "success or failure" | |
Jul 9 19:21:27.373: INFO: Pod "pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 28.949029ms | |
Jul 9 19:21:29.402: INFO: Pod "pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.058308202s | |
Jul 9 19:21:31.439: INFO: Pod "pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.095428007s | |
STEP: Saw pod success | |
Jul 9 19:21:31.439: INFO: Pod "pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:21:31.470: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276 container projected-secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:21:31.543: INFO: Waiting for pod pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276 to disappear | |
Jul 9 19:21:31.572: INFO: Pod pod-projected-secrets-f30ac01d-83e7-11e8-bd2e-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:21:31.572: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-pjp6w" for this suite. | |
Jul 9 19:21:37.698: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:39.688: INFO: namespace: e2e-tests-projected-pjp6w, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:41.290: INFO: namespace e2e-tests-projected-pjp6w deletion completed in 9.679346612s | |
STEP: Destroying namespace "e2e-tests-secret-namespace-gr7ll" for this suite. | |
Jul 9 19:21:47.385: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:49.741: INFO: namespace: e2e-tests-secret-namespace-gr7ll, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:50.910: INFO: namespace e2e-tests-secret-namespace-gr7ll deletion completed in 9.619386792s | |
• [SLOW TEST:26.680 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should be able to mount in a volume regardless of a different secret existing with same name in different namespace [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:87 | |
------------------------------ | |
[sig-storage] Projected | |
should provide podname only [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:21:37.290: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:21:39.041: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-projected-k98n9 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[BeforeEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:858 | |
[It] should provide podname only [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating a pod to test downward API volume plugin | |
Jul 9 19:21:39.759: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276" in namespace "e2e-tests-projected-k98n9" to be "success or failure" | |
Jul 9 19:21:39.790: INFO: Pod "downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 30.60253ms | |
Jul 9 19:21:41.824: INFO: Pod "downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 2.064754136s | |
Jul 9 19:21:43.862: INFO: Pod "downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.102430482s | |
STEP: Saw pod success | |
Jul 9 19:21:43.862: INFO: Pod "downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:21:43.893: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276 container client-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:21:43.966: INFO: Waiting for pod downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:21:44.000: INFO: Pod downwardapi-volume-fa70d37c-83e7-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:21:44.000: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-projected-k98n9" for this suite. | |
Jul 9 19:21:50.150: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:21:53.641: INFO: namespace: e2e-tests-projected-k98n9, resource: bindings, ignored listing per whitelist | |
Jul 9 19:21:54.031: INFO: namespace e2e-tests-projected-k98n9 deletion completed in 9.979733746s | |
• [SLOW TEST:16.741 seconds] | |
[sig-storage] Projected | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/projected.go:34 | |
should provide podname only [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
[Feature:Builds] build have source revision metadata started build | |
should contain source revision information [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:41 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds] build have source revision metadata | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:21:11.475: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds] build have source revision metadata | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:21:13.068: INFO: configPath is now "/tmp/e2e-test-cli-build-revision-fnrv9-user.kubeconfig" | |
Jul 9 19:21:13.068: INFO: The user is now "e2e-test-cli-build-revision-fnrv9-user" | |
Jul 9 19:21:13.068: INFO: Creating project "e2e-test-cli-build-revision-fnrv9" | |
Jul 9 19:21:13.247: INFO: Waiting on permissions in project "e2e-test-cli-build-revision-fnrv9" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:22 | |
Jul 9 19:21:13.309: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:26 | |
STEP: waiting for builder service account | |
Jul 9 19:21:13.452: INFO: Running 'oc create --config=/tmp/e2e-test-cli-build-revision-fnrv9-user.kubeconfig --namespace=e2e-test-cli-build-revision-fnrv9 -f /tmp/fixture-testdata-dir180677416/test/extended/testdata/builds/test-build-revision.json' | |
buildconfig.build.openshift.io "sample-build" created | |
[It] should contain source revision information [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:41 | |
STEP: starting the build | |
Jul 9 19:21:13.718: INFO: Running 'oc start-build --config=/tmp/e2e-test-cli-build-revision-fnrv9-user.kubeconfig --namespace=e2e-test-cli-build-revision-fnrv9 sample-build -o=name' | |
Jul 9 19:21:13.999: INFO: | |
start-build output with args [sample-build -o=name]: | |
Error><nil> | |
StdOut> | |
build/sample-build-1 | |
StdErr> | |
Jul 9 19:21:14.000: INFO: Waiting for sample-build-1 to complete | |
Jul 9 19:21:50.092: INFO: Done waiting for sample-build-1: util.BuildResult{BuildPath:"build/sample-build-1", BuildName:"sample-build-1", StartBuildStdErr:"", StartBuildStdOut:"build/sample-build-1", StartBuildErr:error(nil), BuildConfigName:"", Build:(*build.Build)(0xc4214f4c00), BuildAttempt:true, BuildSuccess:true, BuildFailure:false, BuildCancelled:false, BuildTimeout:false, LogDumper:(util.LogDumperFunc)(nil), Oc:(*util.CLI)(0xc4200e3860)} | |
with error: <nil> | |
STEP: verifying the status of "build/sample-build-1" | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:33 | |
[AfterEach] [Feature:Builds] build have source revision metadata | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:21:50.244: INFO: namespace : e2e-test-cli-build-revision-fnrv9 api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds] build have source revision metadata | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:21:56.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• [SLOW TEST:44.870 seconds] | |
[Feature:Builds] build have source revision metadata | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:14 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:21 | |
started build | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:40 | |
should contain source revision information [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/revision.go:41 | |
------------------------------ | |
S | |
------------------------------ | |
[Feature:Builds][Conformance] imagechangetriggers | |
imagechangetriggers should trigger builds of all types [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:42 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Builds][Conformance] imagechangetriggers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:21:54.032: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Builds][Conformance] imagechangetriggers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:21:55.831: INFO: configPath is now "/tmp/e2e-test-imagechangetriggers-bdpwn-user.kubeconfig" | |
Jul 9 19:21:55.831: INFO: The user is now "e2e-test-imagechangetriggers-bdpwn-user" | |
Jul 9 19:21:55.831: INFO: Creating project "e2e-test-imagechangetriggers-bdpwn" | |
Jul 9 19:21:56.015: INFO: Waiting on permissions in project "e2e-test-imagechangetriggers-bdpwn" ... | |
[BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:25 | |
Jul 9 19:21:56.062: INFO: | |
docker info output: | |
Containers: 4 | |
Running: 0 | |
Paused: 0 | |
Stopped: 4 | |
Images: 4 | |
Server Version: 1.13.1 | |
Storage Driver: aufs | |
Root Dir: /var/lib/docker/aufs | |
Backing Filesystem: extfs | |
Dirs: 20 | |
Dirperm1 Supported: true | |
Logging Driver: json-file | |
Cgroup Driver: cgroupfs | |
Plugins: | |
Volume: local | |
Network: bridge host macvlan null overlay | |
Swarm: inactive | |
Runtimes: runc | |
Default Runtime: runc | |
Init Binary: docker-init | |
containerd version: (expected: aa8187dbd3b7ad67d8e5e3a15115d3eef43a7ed1) | |
runc version: N/A (expected: 9df8b306d01f59d3a8029be411de015b7304dd8f) | |
init version: N/A (expected: 949e6facb77383876aeff8a6944dde66b3089574) | |
Security Options: | |
apparmor | |
seccomp | |
Profile: default | |
Kernel Version: 4.4.0-128-generic | |
Operating System: Ubuntu 16.04.4 LTS | |
OSType: linux | |
Architecture: x86_64 | |
CPUs: 4 | |
Total Memory: 7.495 GiB | |
Name: yifan-coreos | |
ID: ORNN:EABZ:BH7E:KUJC:6IIM:25VE:YORH:CRWG:6TBV:UF4G:AZXO:BS6Z | |
Docker Root Dir: /var/lib/docker | |
Debug Mode (client): false | |
Debug Mode (server): false | |
Username: yifan | |
Registry: https://index.docker.io/v1/ | |
Experimental: false | |
Insecure Registries: | |
127.0.0.0/8 | |
Live Restore Enabled: false | |
[JustBeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:29 | |
STEP: waiting for builder service account | |
[It] imagechangetriggers should trigger builds of all types [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:42 | |
Jul 9 19:21:56.207: INFO: Running 'oc create --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-imagechangetriggers-bdpwn -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/test-imagechangetriggers.yaml' | |
Jul 9 19:21:57.056: INFO: Error running &{/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/oc [oc create --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-imagechangetriggers-bdpwn -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/test-imagechangetriggers.yaml] [] imagestream.image.openshift.io "nodejs-ex" created | |
buildconfig.build.openshift.io "bc-source" created | |
buildconfig.build.openshift.io "bc-docker" created | |
buildconfig.build.openshift.io "bc-custom" created | |
Error from server: Jenkins pipeline template openshift/jenkins-ephemeral not found | |
imagestream.image.openshift.io "nodejs-ex" created | |
buildconfig.build.openshift.io "bc-source" created | |
buildconfig.build.openshift.io "bc-docker" created | |
buildconfig.build.openshift.io "bc-custom" created | |
Error from server: Jenkins pipeline template openshift/jenkins-ephemeral not found | |
[] <nil> 0xc42157e6c0 exit status 1 <nil> <nil> true [0xc42113e058 0xc42113e080 0xc42113e080] [0xc42113e058 0xc42113e080] [0xc42113e060 0xc42113e078] [0x916090 0x916190] 0xc4210f51a0 <nil>}: | |
imagestream.image.openshift.io "nodejs-ex" created | |
buildconfig.build.openshift.io "bc-source" created | |
buildconfig.build.openshift.io "bc-docker" created | |
buildconfig.build.openshift.io "bc-custom" created | |
Error from server: Jenkins pipeline template openshift/jenkins-ephemeral not found | |
imagestream.image.openshift.io "nodejs-ex" created | |
buildconfig.build.openshift.io "bc-source" created | |
buildconfig.build.openshift.io "bc-docker" created | |
buildconfig.build.openshift.io "bc-custom" created | |
Error from server: Jenkins pipeline template openshift/jenkins-ephemeral not found | |
[AfterEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:35 | |
Jul 9 19:21:57.091: INFO: Dumping pod state for namespace e2e-test-imagechangetriggers-bdpwn | |
Jul 9 19:21:57.091: INFO: Running 'oc get --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-imagechangetriggers-bdpwn pods -o yaml' | |
Jul 9 19:21:57.369: INFO: apiVersion: v1 | |
items: [] | |
kind: List | |
metadata: | |
resourceVersion: "" | |
selfLink: "" | |
[AfterEach] [Feature:Builds][Conformance] imagechangetriggers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:72 | |
STEP: Deleting namespaces | |
Jul 9 19:21:57.477: INFO: namespace : e2e-test-imagechangetriggers-bdpwn api call to delete is complete | |
STEP: Waiting for namespaces to vanish | |
[AfterEach] [Feature:Builds][Conformance] imagechangetriggers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
STEP: Dumping a list of prepulled images on each node... | |
Jul 9 19:22:03.546: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
• Failure [9.575 seconds] | |
[Feature:Builds][Conformance] imagechangetriggers | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:16 | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:24 | |
imagechangetriggers should trigger builds of all types [Suite:openshift/conformance/parallel] [It] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:42 | |
Expected error: | |
<*util.ExitError | 0xc4220cff20>: { | |
Cmd: "oc create --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-imagechangetriggers-bdpwn -f /tmp/fixture-testdata-dir333495585/test/extended/testdata/builds/test-imagechangetriggers.yaml", | |
StdErr: "imagestream.image.openshift.io \"nodejs-ex\" created\nbuildconfig.build.openshift.io \"bc-source\" created\nbuildconfig.build.openshift.io \"bc-docker\" created\nbuildconfig.build.openshift.io \"bc-custom\" created\nError from server: Jenkins pipeline template openshift/jenkins-ephemeral not found", | |
ExitError: { | |
ProcessState: { | |
pid: 16527, | |
status: 256, | |
rusage: { | |
Utime: {Sec: 0, Usec: 136000}, | |
Stime: {Sec: 0, Usec: 4000}, | |
Maxrss: 97516, | |
Ixrss: 0, | |
Idrss: 0, | |
Isrss: 0, | |
Minflt: 6969, | |
Majflt: 0, | |
Nswap: 0, | |
Inblock: 0, | |
Oublock: 0, | |
Msgsnd: 0, | |
Msgrcv: 0, | |
Nsignals: 0, | |
Nvcsw: 736, | |
Nivcsw: 58, | |
}, | |
}, | |
Stderr: nil, | |
}, | |
} | |
exit status 1 | |
not to have occurred | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/builds/imagechangetriggers.go:44 | |
------------------------------ | |
SS | |
------------------------------ | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:419 | |
Jul 9 19:22:03.610: INFO: This plugin does not isolate namespaces by default. | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:22:03.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
[AfterEach] when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:22:03.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] | |
[Area:Networking] network isolation | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/isolation.go:10 | |
when using a plugin that isolates namespaces by default | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/util.go:418 | |
should allow communication from non-default to default namespace on a different node [Suite:openshift/conformance/parallel] [BeforeEach] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/networking/isolation.go:53 | |
Jul 9 19:22:03.610: This plugin does not isolate namespaces by default. | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/util.go:296 | |
------------------------------ | |
SS | |
------------------------------ | |
[sig-storage] Secrets | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:21:56.347: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:21:58.562: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-secrets-tmdcm | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
STEP: Creating secret with name secret-test-0618e1eb-83e8-11e8-8401-28d244b00276 | |
STEP: Creating a pod to test consume secrets | |
Jul 9 19:21:59.352: INFO: Waiting up to 5m0s for pod "pod-secrets-061de3b2-83e8-11e8-8401-28d244b00276" in namespace "e2e-tests-secrets-tmdcm" to be "success or failure" | |
Jul 9 19:21:59.388: INFO: Pod "pod-secrets-061de3b2-83e8-11e8-8401-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 35.321902ms | |
Jul 9 19:22:01.417: INFO: Pod "pod-secrets-061de3b2-83e8-11e8-8401-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.065057597s | |
STEP: Saw pod success | |
Jul 9 19:22:01.417: INFO: Pod "pod-secrets-061de3b2-83e8-11e8-8401-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:22:01.455: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-secrets-061de3b2-83e8-11e8-8401-28d244b00276 container secret-volume-test: <nil> | |
STEP: delete the pod | |
Jul 9 19:22:01.536: INFO: Waiting for pod pod-secrets-061de3b2-83e8-11e8-8401-28d244b00276 to disappear | |
Jul 9 19:22:01.575: INFO: Pod pod-secrets-061de3b2-83e8-11e8-8401-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:22:01.575: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-secrets-tmdcm" for this suite. | |
Jul 9 19:22:07.704: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:22:10.011: INFO: namespace: e2e-tests-secrets-tmdcm, resource: bindings, ignored listing per whitelist | |
Jul 9 19:22:11.419: INFO: namespace e2e-tests-secrets-tmdcm deletion completed in 9.808353935s | |
• [SLOW TEST:15.072 seconds] | |
[sig-storage] Secrets | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/secrets_volume.go:33 | |
should be consumable from pods in volume as non-root with defaultMode and fsGroup set [Conformance] [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:674 | |
------------------------------ | |
SS | |
------------------------------ | |
[sig-storage] EmptyDir volumes when FSGroup is specified | |
volume on tmpfs should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:65 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:22:03.613: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
STEP: Building a namespace api object | |
Jul 9 19:22:05.347: INFO: About to run a Kube e2e test, ensuring namespace is privileged | |
STEP: Binding the e2e-test-privileged-psp PodSecurityPolicy to the default service account in e2e-tests-emptydir-qbpr9 | |
STEP: Waiting for a default service account to be provisioned in namespace | |
[It] volume on tmpfs should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:65 | |
STEP: Creating a pod to test emptydir volume type on tmpfs | |
Jul 9 19:22:06.225: INFO: Waiting up to 5m0s for pod "pod-0a34e8ac-83e8-11e8-8fe2-28d244b00276" in namespace "e2e-tests-emptydir-qbpr9" to be "success or failure" | |
Jul 9 19:22:06.258: INFO: Pod "pod-0a34e8ac-83e8-11e8-8fe2-28d244b00276": Phase="Pending", Reason="", readiness=false. Elapsed: 33.066381ms | |
Jul 9 19:22:08.296: INFO: Pod "pod-0a34e8ac-83e8-11e8-8fe2-28d244b00276": Phase="Succeeded", Reason="", readiness=false. Elapsed: 2.071268095s | |
STEP: Saw pod success | |
Jul 9 19:22:08.296: INFO: Pod "pod-0a34e8ac-83e8-11e8-8fe2-28d244b00276" satisfied condition "success or failure" | |
Jul 9 19:22:08.333: INFO: Trying to get logs from node ip-10-0-130-54.us-west-2.compute.internal pod pod-0a34e8ac-83e8-11e8-8fe2-28d244b00276 container test-container: <nil> | |
STEP: delete the pod | |
Jul 9 19:22:08.409: INFO: Waiting for pod pod-0a34e8ac-83e8-11e8-8fe2-28d244b00276 to disappear | |
Jul 9 19:22:08.439: INFO: Pod pod-0a34e8ac-83e8-11e8-8fe2-28d244b00276 no longer exists | |
[AfterEach] [sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:142 | |
Jul 9 19:22:08.439: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready | |
STEP: Destroying namespace "e2e-tests-emptydir-qbpr9" for this suite. | |
Jul 9 19:22:14.654: INFO: Waiting up to 30s for server preferred namespaced resources to be successfully discovered | |
Jul 9 19:22:16.785: INFO: namespace: e2e-tests-emptydir-qbpr9, resource: bindings, ignored listing per whitelist | |
Jul 9 19:22:18.670: INFO: namespace e2e-tests-emptydir-qbpr9 deletion completed in 10.197573968s | |
• [SLOW TEST:15.057 seconds] | |
[sig-storage] EmptyDir volumes | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:41 | |
when FSGroup is specified | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:44 | |
volume on tmpfs should have the correct mode using FSGroup [Suite:openshift/conformance/parallel] [Suite:k8s] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/common/empty_dir.go:65 | |
------------------------------ | |
[Feature:Prometheus][Feature:Builds] Prometheus when installed to the cluster | |
should start and expose a secured proxy and verify build metrics [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/prometheus/prometheus_builds.go:36 | |
[BeforeEach] [Top Level] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/test.go:51 | |
[BeforeEach] [Feature:Prometheus][Feature:Builds] Prometheus | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/vendor/k8s.io/kubernetes/test/e2e/framework/framework.go:141 | |
STEP: Creating a kubernetes client | |
Jul 9 19:13:52.928: INFO: >>> kubeConfig: /home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig | |
[BeforeEach] [Feature:Prometheus][Feature:Builds] Prometheus | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/util/cli.go:83 | |
Jul 9 19:13:55.279: INFO: configPath is now "/tmp/e2e-test-prometheus-4fqst-user.kubeconfig" | |
Jul 9 19:13:55.279: INFO: The user is now "e2e-test-prometheus-4fqst-user" | |
Jul 9 19:13:55.279: INFO: Creating project "e2e-test-prometheus-4fqst" | |
Jul 9 19:13:55.406: INFO: Waiting on permissions in project "e2e-test-prometheus-4fqst" ... | |
[BeforeEach] [Feature:Prometheus][Feature:Builds] Prometheus | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/prometheus/prometheus_builds.go:31 | |
[It] should start and expose a secured proxy and verify build metrics [Suite:openshift/conformance/parallel] | |
/home/yifan/gopher/src/github.com/openshift/origin/_output/local/go/src/github.com/openshift/origin/test/extended/prometheus/prometheus_builds.go:36 | |
Jul 9 19:14:15.291: INFO: Creating new exec pod | |
STEP: verifying the oauth-proxy reports a 403 on the root URL | |
Jul 9 19:14:17.440: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -k -s -o /dev/null -w '%{http_code}' "https://prometheus.kube-system.svc:443"' | |
Jul 9 19:14:18.221: INFO: stderr: "" | |
STEP: verifying a service account token is able to authenticate | |
Jul 9 19:14:18.221: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -k -s -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' -o /dev/null -w '%{http_code}' "https://prometheus.kube-system.svc:443/graph"' | |
Jul 9 19:14:18.987: INFO: stderr: "" | |
STEP: waiting for builder service account | |
STEP: calling oc new-app /tmp/fixture-testdata-dir574852015/examples/jenkins/application-template.json | |
Jul 9 19:14:19.193: INFO: Running 'oc new-app --config=/tmp/e2e-test-prometheus-4fqst-user.kubeconfig --namespace=e2e-test-prometheus-4fqst /tmp/fixture-testdata-dir574852015/examples/jenkins/application-template.json' | |
--> Deploying template "e2e-test-prometheus-4fqst/nodejs-helloworld-sample" for "/tmp/fixture-testdata-dir574852015/examples/jenkins/application-template.json" to project e2e-test-prometheus-4fqst | |
nodejs-helloworld-sample | |
--------- | |
This example shows how to create a simple nodejs application in openshift origin v3 | |
* With parameters: | |
* Memory Limit=512Mi | |
* Namespace=openshift | |
* Administrator Username=adminFGD # generated | |
* Administrator Password=D1026mG7 # generated | |
--> Creating resources ... | |
service "frontend-prod" created | |
route "frontend" created | |
deploymentconfig "frontend-prod" created | |
service "frontend" created | |
imagestream "origin-nodejs-sample" created | |
imagestream "origin-nodejs-sample2" created | |
imagestream "origin-nodejs-sample3" created | |
imagestream "nodejs-010-centos7" created | |
buildconfig "frontend" created | |
deploymentconfig "frontend" created | |
--> Success | |
Access your application via route 'frontend-e2e-test-prometheus-4fqst.yifan-test-cluster.coreservices.team.coreos.systems' | |
Use 'oc start-build frontend' to start a build. | |
Run 'oc status' to view your app. | |
STEP: wait on imagestreams used by build | |
Jul 9 19:14:20.197: INFO: Running scan #0 | |
Jul 9 19:14:20.197: INFO: Checking language ruby | |
Jul 9 19:14:20.231: INFO: Checking tag 2.0 | |
Jul 9 19:14:20.231: INFO: Checking tag 2.2 | |
Jul 9 19:14:20.231: INFO: Checking tag 2.3 | |
Jul 9 19:14:20.231: INFO: Checking tag 2.4 | |
Jul 9 19:14:20.231: INFO: Checking tag 2.5 | |
Jul 9 19:14:20.231: INFO: Checking tag latest | |
Jul 9 19:14:20.231: INFO: Checking language nodejs | |
Jul 9 19:14:20.263: INFO: Checking tag 6 | |
Jul 9 19:14:20.263: INFO: Checking tag 8 | |
Jul 9 19:14:20.263: INFO: Checking tag latest | |
Jul 9 19:14:20.263: INFO: Checking tag 0.10 | |
Jul 9 19:14:20.263: INFO: Checking tag 4 | |
Jul 9 19:14:20.263: INFO: Checking language perl | |
Jul 9 19:14:20.298: INFO: Checking tag 5.24 | |
Jul 9 19:14:20.299: INFO: Checking tag latest | |
Jul 9 19:14:20.299: INFO: Checking tag 5.16 | |
Jul 9 19:14:20.299: INFO: Checking tag 5.20 | |
Jul 9 19:14:20.299: INFO: Checking language php | |
Jul 9 19:14:20.345: INFO: Checking tag 5.5 | |
Jul 9 19:14:20.345: INFO: Checking tag 5.6 | |
Jul 9 19:14:20.345: INFO: Checking tag 7.0 | |
Jul 9 19:14:20.345: INFO: Checking tag 7.1 | |
Jul 9 19:14:20.345: INFO: Checking tag latest | |
Jul 9 19:14:20.345: INFO: Checking language python | |
Jul 9 19:14:20.384: INFO: Checking tag 3.6 | |
Jul 9 19:14:20.384: INFO: Checking tag latest | |
Jul 9 19:14:20.384: INFO: Checking tag 2.7 | |
Jul 9 19:14:20.384: INFO: Checking tag 3.3 | |
Jul 9 19:14:20.384: INFO: Checking tag 3.4 | |
Jul 9 19:14:20.384: INFO: Checking tag 3.5 | |
Jul 9 19:14:20.384: INFO: Checking language wildfly | |
Jul 9 19:14:20.421: INFO: Checking tag 12.0 | |
Jul 9 19:14:20.421: INFO: Checking tag 8.1 | |
Jul 9 19:14:20.421: INFO: Checking tag 9.0 | |
Jul 9 19:14:20.421: INFO: Checking tag latest | |
Jul 9 19:14:20.421: INFO: Checking tag 10.0 | |
Jul 9 19:14:20.421: INFO: Checking tag 10.1 | |
Jul 9 19:14:20.421: INFO: Checking tag 11.0 | |
Jul 9 19:14:20.421: INFO: Checking language mysql | |
Jul 9 19:14:20.457: INFO: Checking tag 5.5 | |
Jul 9 19:14:20.457: INFO: Checking tag 5.6 | |
Jul 9 19:14:20.457: INFO: Checking tag 5.7 | |
Jul 9 19:14:20.457: INFO: Checking tag latest | |
Jul 9 19:14:20.457: INFO: Checking language postgresql | |
Jul 9 19:14:20.499: INFO: Checking tag 9.2 | |
Jul 9 19:14:20.499: INFO: Checking tag 9.4 | |
Jul 9 19:14:20.499: INFO: Checking tag 9.5 | |
Jul 9 19:14:20.499: INFO: Checking tag 9.6 | |
Jul 9 19:14:20.499: INFO: Checking tag latest | |
Jul 9 19:14:20.499: INFO: Checking language mongodb | |
Jul 9 19:14:20.536: INFO: Checking tag 2.4 | |
Jul 9 19:14:20.536: INFO: Checking tag 2.6 | |
Jul 9 19:14:20.536: INFO: Checking tag 3.2 | |
Jul 9 19:14:20.536: INFO: Checking tag 3.4 | |
Jul 9 19:14:20.536: INFO: Checking tag latest | |
Jul 9 19:14:20.536: INFO: Checking language jenkins | |
Jul 9 19:14:20.572: INFO: Checking tag 1 | |
Jul 9 19:14:20.572: INFO: Checking tag 2 | |
Jul 9 19:14:20.572: INFO: Checking tag latest | |
Jul 9 19:14:20.572: INFO: Success! | |
STEP: explicitly set up image stream tag, avoid timing window | |
Jul 9 19:14:20.572: INFO: Running 'oc tag --config=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig --namespace=e2e-test-prometheus-4fqst openshift/nodejs:latest e2e-test-prometheus-4fqst/nodejs-010-centos7:latest' | |
Tag nodejs-010-centos7:latest set to openshift/nodejs@sha256:3f262ed6ed7ec7497576e1117ddfafacaa23fecf34693e05a0abcb652b93e653. | |
STEP: start build | |
Jul 9 19:14:20.995: INFO: Running 'oc start-build --config=/tmp/e2e-test-prometheus-4fqst-user.kubeconfig --namespace=e2e-test-prometheus-4fqst frontend -o=name' | |
Jul 9 19:14:21.289: INFO: | |
start-build output with args [frontend -o=name]: | |
Error><nil> | |
StdOut> | |
build/frontend-1 | |
StdErr> | |
STEP: verifying build completed successfully | |
Jul 9 19:14:21.289: INFO: Waiting for frontend-1 to complete | |
Jul 9 19:14:52.358: INFO: Done waiting for frontend-1: util.BuildResult{BuildPath:"build/frontend-1", BuildName:"frontend-1", StartBuildStdErr:"", StartBuildStdOut:"build/frontend-1", StartBuildErr:error(nil), BuildConfigName:"", Build:(*build.Build)(0xc421143b00), BuildAttempt:true, BuildSuccess:true, BuildFailure:false, BuildCancelled:false, BuildTimeout:false, LogDumper:(util.LogDumperFunc)(nil), Oc:(*util.CLI)(0xc4213a0f00)} | |
with error: <nil> | |
STEP: verifying a service account token is able to query terminal build metrics from the Prometheus API | |
STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:14:52.358: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:14:53.120: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:14:54.120: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:14:54.934: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:14:55.934: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:14:56.718: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:14:57.719: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:14:58.476: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:14:59.476: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:00.218: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:01.218: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:02.077: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:03.078: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:03.776: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:04.777: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:05.641: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:06.642: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:07.364: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:08.364: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:09.095: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:10.095: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:10.990: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:11.991: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:12.773: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:13.773: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:14.625: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:15.625: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:16.414: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:17.418: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:18.229: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:19.229: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:20.069: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:21.069: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:21.826: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:22.826: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:23.812: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:24.812: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:25.761: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:26.762: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:28.055: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:29.055: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:30.219: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:31.220: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:32.062: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:33.062: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:33.842: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:34.842: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:35.569: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:36.569: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:37.322: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:38.323: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:39.067: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:40.067: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:40.882: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:41.883: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:42.809: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:43.809: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:44.699: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:45.699: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:46.455: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:47.456: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:48.256: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:49.256: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:50.301: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:51.302: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:52.189: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:53.190: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:53.968: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:54.968: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:55.783: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:56.783: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:57.758: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:15:58.759: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:15:59.653: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:00.653: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:01.398: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:02.398: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:03.192: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:04.192: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:04.957: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:05.958: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:06.765: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:07.766: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:08.582: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:09.583: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:10.604: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:11.604: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:12.413: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:13.414: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:14.499: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:15.499: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:16.253: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:17.253: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:18.003: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:19.003: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:19.824: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:20.824: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:21.662: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:22.662: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:23.635: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:24.635: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:25.388: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:26.388: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:27.262: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:28.262: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:29.065: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:30.065: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:30.786: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:31.786: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:32.580: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:33.580: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:34.487: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:35.488: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:36.475: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:37.476: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:38.330: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:39.331: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:40.115: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:41.116: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:41.920: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:42.921: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:43.671: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:44.671: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:45.433: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:46.433: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:47.207: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:48.208: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:48.930: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:49.930: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:50.801: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:51.801: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:52.660: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:53.660: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:54.407: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:55.408: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:56.151: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:57.152: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:57.934: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:16:58.934: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:16:59.832: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:17:00.833: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJrdWJlLXN5c3RlbSIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VjcmV0Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlci10b2tlbi1nbW16NCIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50Lm5hbWUiOiJwcm9tZXRoZXVzLXJlYWRlciIsImt1YmVybmV0ZXMuaW8vc2VydmljZWFjY291bnQvc2VydmljZS1hY2NvdW50LnVpZCI6ImE3YTRlZGIzLTgzYjktMTFlOC04NGM2LTBhZjk2NzY4ZDU3ZSIsInN1YiI6InN5c3RlbTpzZXJ2aWNlYWNjb3VudDprdWJlLXN5c3RlbTpwcm9tZXRoZXVzLXJlYWRlciJ9.I7Tfj6MfIbWy1XD5yyNl9TA5mIhFW_lAwzupgAqZ_dCqHWzNSoiGQuM1oB521nap-7AxRwuMvYlHPHTKm07z3ntJQB6GFFThIjx965ExnOy9B1fHleads4rm1mtwP1QFmHEmkhMGK98QitmRf0rXJsEmAgk5H47uPh0gh2MZSms8y7ERvjM9OwQKUx5CYFjSwh5uTbGad6IrvVjvDrBcfNweo8cMgT1EHI2bU3fUedjPcj7Y_94rpAOKOQP8QFMZRAl0sX6CcDNe1VVvDuwx9yi7JhPMSbQDNU0HHgNB3hqJO8ov2HFx6ws2e52eCWxjs-oVzFj5Z0Gki2JJaDIZ_w' "https://prometheus.kube-system.svc:443/api/v1/query?query=openshift_build_total"' | |
Jul 9 19:17:01.884: INFO: stderr: "" | |
query openshift_build_total for tests []prometheus.metricTest{prometheus.metricTest{labels:map[string]string{"phase":"Complete"}, greaterThanEqual:true, value:0, success:false}} had results {"status":"success","data":{"resultType":"vector","result":[]}}STEP: perform prometheus metric query openshift_build_total | |
Jul 9 19:17:02.884: INFO: Running '/home/yifan/gopher/src/github.com/openshift/origin/_output/local/bin/linux/amd64/kubectl --server=https://yifan-test-cluster-api.coreservices.team.coreos.systems:6443 --kubeconfig=/home/yifan/tectonic-installer/yifan-test-cluster/generated/auth/kubeconfig exec --namespace=e2e-test-prometheus-4fqst execpod98j4h -- /bin/sh -c curl -s -k -H 'Authorization: Bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IiJ9.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment