+ export WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release + WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release + [[ k8s-1.11.0-release =~ openshift-.* ]] + [[ k8s-1.11.0-release =~ .*-1.10.4-.* ]] + export KUBEVIRT_PROVIDER=k8s-1.11.0 + KUBEVIRT_PROVIDER=k8s-1.11.0 + export KUBEVIRT_NUM_NODES=2 + KUBEVIRT_NUM_NODES=2 + export NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + export NAMESPACE=kube-system + NAMESPACE=kube-system + trap '{ make cluster-down; }' EXIT SIGINT SIGTERM SIGSTOP + make cluster-down ./cluster/down.sh + make cluster-up ./cluster/up.sh Downloading ......................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................... Downloading ....... 2018/07/27 08:58:25 Waiting for host: 192.168.66.101:22 2018/07/27 08:58:28 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/07/27 08:58:36 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/07/27 08:58:41 Connected to tcp://192.168.66.101:22 ++ wc -l ++ systemctl status docker ++ grep active + [[ 0 -eq 0 ]] + sleep 2 ++ systemctl status docker ++ grep active ++ wc -l + [[ 1 -eq 0 ]] + kubeadm init --config /etc/kubernetes/kubeadm.conf [init] using Kubernetes version: v1.11.0 [preflight] running pre-flight checks I0727 08:58:44.622121 1263 feature_gate.go:230] feature gates: &{map[]} I0727 08:58:44.695159 1263 kernel_validator.go:81] Validating kernel version I0727 08:58:44.695492 1263 kernel_validator.go:96] Validating kernel config [preflight/images] Pulling images required for setting up a Kubernetes cluster [preflight/images] This might take a minute or two, depending on the speed of your internet connection [preflight/images] You can also perform this action in beforehand using 'kubeadm config images pull' [kubelet] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env" [kubelet] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml" [preflight] Activating the kubelet service [certificates] Generated ca certificate and key. [certificates] Generated apiserver certificate and key. [certificates] apiserver serving cert is signed for DNS names [node01 kubernetes kubernetes.default kubernetes.default.svc kubernetes.default.svc.cluster.local] and IPs [10.96.0.1 192.168.66.101] [certificates] Generated apiserver-kubelet-client certificate and key. [certificates] Generated sa key and public key. [certificates] Generated front-proxy-ca certificate and key. [certificates] Generated front-proxy-client certificate and key. [certificates] Generated etcd/ca certificate and key. [certificates] Generated etcd/server certificate and key. [certificates] etcd/server serving cert is signed for DNS names [node01 localhost] and IPs [127.0.0.1 ::1] [certificates] Generated etcd/peer certificate and key. [certificates] etcd/peer serving cert is signed for DNS names [node01 localhost] and IPs [192.168.66.101 127.0.0.1 ::1] [certificates] Generated etcd/healthcheck-client certificate and key. [certificates] Generated apiserver-etcd-client certificate and key. [certificates] valid certificates and keys now exist in "/etc/kubernetes/pki" [kubeconfig] Wrote KubeConfig file to disk: "/etc/kubernetes/admin.conf" [kubeconfig] Wrote KubeConfig file to disk: "/etc/kubernetes/kubelet.conf" [kubeconfig] Wrote KubeConfig file to disk: "/etc/kubernetes/controller-manager.conf" [kubeconfig] Wrote KubeConfig file to disk: "/etc/kubernetes/scheduler.conf" [controlplane] wrote Static Pod manifest for component kube-apiserver to "/etc/kubernetes/manifests/kube-apiserver.yaml" [controlplane] wrote Static Pod manifest for component kube-controller-manager to "/etc/kubernetes/manifests/kube-controller-manager.yaml" [controlplane] wrote Static Pod manifest for component kube-scheduler to "/etc/kubernetes/manifests/kube-scheduler.yaml" [etcd] Wrote Static Pod manifest for a local etcd instance to "/etc/kubernetes/manifests/etcd.yaml" [init] waiting for the kubelet to boot up the control plane as Static Pods from directory "/etc/kubernetes/manifests" [init] this might take a minute or longer if the control plane images have to be pulled [apiclient] All control plane components are healthy after 45.004984 seconds [uploadconfig] storing the configuration used in ConfigMap "kubeadm-config" in the "kube-system" Namespace [kubelet] Creating a ConfigMap "kubelet-config-1.11" in namespace kube-system with the configuration for the kubelets in the cluster [markmaster] Marking the node node01 as master by adding the label "node-role.kubernetes.io/master=''" [markmaster] Marking the node node01 as master by adding the taints [node-role.kubernetes.io/master:NoSchedule] [patchnode] Uploading the CRI Socket information "/var/run/dockershim.sock" to the Node API object "node01" as an annotation [bootstraptoken] using token: abcdef.1234567890123456 [bootstraptoken] configured RBAC rules to allow Node Bootstrap tokens to post CSRs in order for nodes to get long term certificate credentials [bootstraptoken] configured RBAC rules to allow the csrapprover controller automatically approve CSRs from a Node Bootstrap Token [bootstraptoken] configured RBAC rules to allow certificate rotation for all node client certificates in the cluster [bootstraptoken] creating the "cluster-info" ConfigMap in the "kube-public" namespace [addons] Applied essential addon: CoreDNS [addons] Applied essential addon: kube-proxy Your Kubernetes master has initialized successfully! To start using your cluster, you need to run the following as a regular user: mkdir -p $HOME/.kube sudo cp -i /etc/kubernetes/admin.conf $HOME/.kube/config sudo chown $(id -u):$(id -g) $HOME/.kube/config You should now deploy a pod network to the cluster. Run "kubectl apply -f [podnetwork].yaml" with one of the options listed at: https://kubernetes.io/docs/concepts/cluster-administration/addons/ You can now join any number of machines by running the following on each node as root: kubeadm join 192.168.66.101:6443 --token abcdef.1234567890123456 --discovery-token-ca-cert-hash sha256:b5818c36d651b7e0215f5953881b19fe9f45d216649496043b67d06e151df050 + kubectl --kubeconfig=/etc/kubernetes/admin.conf apply -f https://raw.githubusercontent.com/coreos/flannel/v0.9.1/Documentation/kube-flannel.yml clusterrole.rbac.authorization.k8s.io/flannel created clusterrolebinding.rbac.authorization.k8s.io/flannel created serviceaccount/flannel created configmap/kube-flannel-cfg created daemonset.extensions/kube-flannel-ds created + kubectl --kubeconfig=/etc/kubernetes/admin.conf taint nodes node01 node-role.kubernetes.io/master:NoSchedule- node/node01 untainted + kubectl --kubeconfig=/etc/kubernetes/admin.conf create -f /tmp/local-volume.yaml storageclass.storage.k8s.io/local created configmap/local-storage-config created clusterrolebinding.rbac.authorization.k8s.io/local-storage-provisioner-pv-binding created clusterrole.rbac.authorization.k8s.io/local-storage-provisioner-node-clusterrole created clusterrolebinding.rbac.authorization.k8s.io/local-storage-provisioner-node-binding created role.rbac.authorization.k8s.io/local-storage-provisioner-jobs-role created rolebinding.rbac.authorization.k8s.io/local-storage-provisioner-jobs-rolebinding created serviceaccount/local-storage-admin created daemonset.extensions/local-volume-provisioner created 2018/07/27 08:59:45 Waiting for host: 192.168.66.102:22 2018/07/27 08:59:48 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/07/27 09:00:00 Connected to tcp://192.168.66.102:22 ++ systemctl status docker ++ wc -l ++ grep active + [[ 1 -eq 0 ]] + kubeadm join --token abcdef.1234567890123456 192.168.66.101:6443 --ignore-preflight-errors=all --discovery-token-unsafe-skip-ca-verification=true [preflight] running pre-flight checks [WARNING RequiredIPVSKernelModulesAvailable]: the IPVS proxier will not be used, because the following required kernel modules are not loaded: [ip_vs ip_vs_rr ip_vs_wrr ip_vs_sh] or no builtin kernel ipvs support: map[ip_vs_sh:{} nf_conntrack_ipv4:{} ip_vs:{} ip_vs_rr:{} ip_vs_wrr:{}] you can solve this problem with following methods: 1. Run 'modprobe -- ' to load missing kernel modules; 2. Provide the missing builtin kernel ipvs support I0727 09:00:00.926243 1268 kernel_validator.go:81] Validating kernel version I0727 09:00:00.926502 1268 kernel_validator.go:96] Validating kernel config [discovery] Trying to connect to API Server "192.168.66.101:6443" [discovery] Created cluster-info discovery client, requesting info from "https://192.168.66.101:6443" [discovery] Cluster info signature and contents are valid and no TLS pinning was specified, will use API Server "192.168.66.101:6443" [discovery] Successfully established connection with API Server "192.168.66.101:6443" [kubelet] Downloading configuration for the kubelet from the "kubelet-config-1.11" ConfigMap in the kube-system namespace [kubelet] Writing kubelet configuration to file "/var/lib/kubelet/config.yaml" [kubelet] Writing kubelet environment file with flags to file "/var/lib/kubelet/kubeadm-flags.env" [preflight] Activating the kubelet service [tlsbootstrap] Waiting for the kubelet to perform the TLS Bootstrap... [patchnode] Uploading the CRI Socket information "/var/run/dockershim.sock" to the Node API object "node02" as an annotation This node has joined the cluster: * Certificate signing request was sent to master and a response was received. * The Kubelet was informed of the new secure connection details. Run 'kubectl get nodes' on the master to see this node join the cluster. Sending file modes: C0755 38739968 kubectl Sending file modes: C0600 5454 admin.conf Cluster "kubernetes" set. Cluster "kubernetes" set. + set +e + kubectl get nodes --no-headers + cluster/kubectl.sh get nodes --no-headers node01 Ready master 37s v1.11.0 node02 Ready 15s v1.11.0 + kubectl_rc=0 + '[' 0 -ne 0 ']' ++ kubectl get nodes --no-headers ++ grep NotReady ++ cluster/kubectl.sh get nodes --no-headers + '[' -n '' ']' + set -e + echo 'Nodes are ready:' Nodes are ready: + kubectl get nodes + cluster/kubectl.sh get nodes NAME STATUS ROLES AGE VERSION node01 Ready master 38s v1.11.0 node02 Ready 16s v1.11.0 + make cluster-sync ./cluster/build.sh Building ... sha256:ceba12cbc33e4e37a707840478a630db561e2427b78c8c9f9cd6d0b73276ab32 go version go1.10 linux/amd64 Waiting for rsyncd to be ready go version go1.10 linux/amd64 make[1]: Entering directory `/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt' hack/dockerized "./hack/check.sh && KUBEVIRT_VERSION= ./hack/build-go.sh install " && ./hack/build-copy-artifacts.sh sha256:ceba12cbc33e4e37a707840478a630db561e2427b78c8c9f9cd6d0b73276ab32 go version go1.10 linux/amd64 go version go1.10 linux/amd64 find: '/root/go/src/kubevirt.io/kubevirt/_out/cmd': No such file or directory Compiling tests... compiled tests.test hack/build-docker.sh build Sending build context to Docker daemon 40.37 MB Step 1/8 : FROM fedora:28 ---> cc510acfcd70 Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 84920e004a40 Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-controller ---> Using cache ---> b4f3251c6468 Step 4/8 : WORKDIR /home/virt-controller ---> Using cache ---> 813752072d9d Step 5/8 : USER 1001 ---> Using cache ---> 88b3556f36b4 Step 6/8 : COPY virt-controller /usr/bin/virt-controller ---> b22ffba78162 Removing intermediate container 54372dd6da64 Step 7/8 : ENTRYPOINT /usr/bin/virt-controller ---> Running in c3776fe383b8 ---> 5bf41312168d Removing intermediate container c3776fe383b8 Step 8/8 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "virt-controller" '' ---> Running in 9b3a2a0c0d30 ---> 8ff4e4cf9f3c Removing intermediate container 9b3a2a0c0d30 Successfully built 8ff4e4cf9f3c Sending build context to Docker daemon 43.3 MB Step 1/10 : FROM kubevirt/libvirt:4.2.0 ---> 5f0bfe81a3e0 Step 2/10 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 09010a005182 Step 3/10 : RUN dnf -y install socat genisoimage util-linux libcgroup-tools ethtool net-tools sudo && dnf -y clean all && test $(id -u qemu) = 107 # make sure that the qemu user really is 107 ---> Using cache ---> fc9481693838 Step 4/10 : COPY virt-launcher /usr/bin/virt-launcher ---> b68fa91430ca Removing intermediate container 1f41f7f20e07 Step 5/10 : COPY kubevirt-sudo /etc/sudoers.d/kubevirt ---> fbbe5a00f8c8 Removing intermediate container 80ea98ce8a1c Step 6/10 : RUN setcap CAP_NET_BIND_SERVICE=+eip /usr/bin/qemu-system-x86_64 ---> Running in 958167960dcd  ---> 9b53fda8af0d Removing intermediate container 958167960dcd Step 7/10 : RUN mkdir -p /usr/share/kubevirt/virt-launcher ---> Running in c36ce9305bff  ---> 230ada0c0cb4 Removing intermediate container c36ce9305bff Step 8/10 : COPY entrypoint.sh libvirtd.sh sock-connector /usr/share/kubevirt/virt-launcher/ ---> 48d4268277d9 Removing intermediate container 01f8262227e7 Step 9/10 : ENTRYPOINT /usr/share/kubevirt/virt-launcher/entrypoint.sh ---> Running in 83aeaa6b1920 ---> c740fd9aa442 Removing intermediate container 83aeaa6b1920 Step 10/10 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "virt-launcher" '' ---> Running in 33a9b413d03c ---> 778d18ea0b55 Removing intermediate container 33a9b413d03c Successfully built 778d18ea0b55 Sending build context to Docker daemon 41.67 MB Step 1/5 : FROM fedora:28 ---> cc510acfcd70 Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 84920e004a40 Step 3/5 : COPY virt-handler /usr/bin/virt-handler ---> 22295a1ae764 Removing intermediate container eb4b0fcc20c8 Step 4/5 : ENTRYPOINT /usr/bin/virt-handler ---> Running in 8078aeb8211f ---> 16aedacbf860 Removing intermediate container 8078aeb8211f Step 5/5 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "virt-handler" '' ---> Running in b234c18f24e6 ---> 615e97ef9d7b Removing intermediate container b234c18f24e6 Successfully built 615e97ef9d7b Sending build context to Docker daemon 38.81 MB Step 1/8 : FROM fedora:28 ---> cc510acfcd70 Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 84920e004a40 Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-api ---> Using cache ---> 3cff23506e80 Step 4/8 : WORKDIR /home/virt-api ---> Using cache ---> e94c5606b96b Step 5/8 : USER 1001 ---> Using cache ---> af16317199f5 Step 6/8 : COPY virt-api /usr/bin/virt-api ---> 9ba6f4bdf9f8 Removing intermediate container 14a81c5c3863 Step 7/8 : ENTRYPOINT /usr/bin/virt-api ---> Running in 834e1ed09630 ---> 96b30bc5b70a Removing intermediate container 834e1ed09630 Step 8/8 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "virt-api" '' ---> Running in 761bceff7024 ---> 29729f367063 Removing intermediate container 761bceff7024 Successfully built 29729f367063 Sending build context to Docker daemon 4.096 kB Step 1/7 : FROM fedora:28 ---> cc510acfcd70 Step 2/7 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 84920e004a40 Step 3/7 : ENV container docker ---> Using cache ---> aed3ca4ac3a3 Step 4/7 : RUN mkdir -p /images/custom /images/alpine && truncate -s 64M /images/custom/disk.img && curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /images/alpine/disk.img ---> Using cache ---> c7d0cf8fc982 Step 5/7 : ADD entrypoint.sh / ---> Using cache ---> 0393e5ee0c37 Step 6/7 : CMD /entrypoint.sh ---> Using cache ---> 23798f49dea3 Step 7/7 : LABEL "disks-images-provider" '' "kubevirt-functional-tests-k8s-1.11.0-release1" '' ---> Using cache ---> c2a355bd725c Successfully built c2a355bd725c Sending build context to Docker daemon 2.56 kB Step 1/5 : FROM fedora:28 ---> cc510acfcd70 Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 84920e004a40 Step 3/5 : ENV container docker ---> Using cache ---> aed3ca4ac3a3 Step 4/5 : RUN dnf -y install procps-ng nmap-ncat && dnf -y clean all ---> Using cache ---> d8c990eaf575 Step 5/5 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "vm-killer" '' ---> Using cache ---> a35b53eec8f8 Successfully built a35b53eec8f8 Sending build context to Docker daemon 5.12 kB Step 1/7 : FROM debian:sid ---> 68f33cf86aab Step 2/7 : MAINTAINER "David Vossel" \ ---> Using cache ---> 50fc79ebe51c Step 3/7 : ENV container docker ---> Using cache ---> b8e063496923 Step 4/7 : RUN apt-get update && apt-get install -y bash curl bzip2 qemu-utils && mkdir -p /disk && rm -rf /var/lib/apt/lists/* ---> Using cache ---> 8adb1572b35c Step 5/7 : ADD entry-point.sh / ---> Using cache ---> 8c0c5a52e4df Step 6/7 : CMD /entry-point.sh ---> Using cache ---> 1a4b838e5dee Step 7/7 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "registry-disk-v1alpha" '' ---> Using cache ---> d2597d227e63 Successfully built d2597d227e63 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33000/kubevirt/registry-disk-v1alpha:devel ---> d2597d227e63 Step 2/4 : MAINTAINER "David Vossel" \ ---> Using cache ---> 31d87a3e7551 Step 3/4 : RUN curl https://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img > /disk/cirros.img ---> Using cache ---> 8b729e3d9cc8 Step 4/4 : LABEL "cirros-registry-disk-demo" '' "kubevirt-functional-tests-k8s-1.11.0-release1" '' ---> Using cache ---> e164bf7eb7ce Successfully built e164bf7eb7ce Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33000/kubevirt/registry-disk-v1alpha:devel ---> d2597d227e63 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 8b026f4503d0 Step 3/4 : RUN curl -g -L https://download.fedoraproject.org/pub/fedora/linux/releases/27/CloudImages/x86_64/images/Fedora-Cloud-Base-27-1.6.x86_64.qcow2 > /disk/fedora.qcow2 ---> Using cache ---> 8108a03f9df8 Step 4/4 : LABEL "fedora-cloud-registry-disk-demo" '' "kubevirt-functional-tests-k8s-1.11.0-release1" '' ---> Using cache ---> b3c420487c6f Successfully built b3c420487c6f Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33000/kubevirt/registry-disk-v1alpha:devel ---> d2597d227e63 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 8b026f4503d0 Step 3/4 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /disk/alpine.iso ---> Using cache ---> 8e751825aa41 Step 4/4 : LABEL "alpine-registry-disk-demo" '' "kubevirt-functional-tests-k8s-1.11.0-release1" '' ---> Using cache ---> c3fe0249b081 Successfully built c3fe0249b081 Sending build context to Docker daemon 35.59 MB Step 1/8 : FROM fedora:28 ---> cc510acfcd70 Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 84920e004a40 Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virtctl ---> Using cache ---> d74088d7a4fc Step 4/8 : WORKDIR /home/virtctl ---> Using cache ---> c8c857bf8d96 Step 5/8 : USER 1001 ---> Using cache ---> 36730a67b946 Step 6/8 : COPY subresource-access-test /subresource-access-test ---> 07da40a5e323 Removing intermediate container d9050afcda5a Step 7/8 : ENTRYPOINT /subresource-access-test ---> Running in 53896c27ecf7 ---> 13c48eff7685 Removing intermediate container 53896c27ecf7 Step 8/8 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "subresource-access-test" '' ---> Running in 033928882c97 ---> c3ab82b0d5c2 Removing intermediate container 033928882c97 Successfully built c3ab82b0d5c2 Sending build context to Docker daemon 3.072 kB Step 1/9 : FROM fedora:28 ---> cc510acfcd70 Step 2/9 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 84920e004a40 Step 3/9 : ENV container docker ---> Using cache ---> aed3ca4ac3a3 Step 4/9 : RUN dnf -y install make git gcc && dnf -y clean all ---> Using cache ---> 6050b24a5d85 Step 5/9 : ENV GIMME_GO_VERSION 1.9.2 ---> Using cache ---> 0447d2178073 Step 6/9 : RUN mkdir -p /gimme && curl -sL https://raw.githubusercontent.com/travis-ci/gimme/master/gimme | HOME=/gimme bash >> /etc/profile.d/gimme.sh ---> Using cache ---> 291db82d955f Step 7/9 : ENV GOPATH "/go" GOBIN "/usr/bin" ---> Using cache ---> 793556477837 Step 8/9 : RUN mkdir -p /go && source /etc/profile.d/gimme.sh && go get github.com/masterzen/winrm-cli ---> Using cache ---> fd5c6e1f9461 Step 9/9 : LABEL "kubevirt-functional-tests-k8s-1.11.0-release1" '' "winrmcli" '' ---> Using cache ---> 17a9aaffc747 Successfully built 17a9aaffc747 Sending build context to Docker daemon 36.79 MB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71a8c548e503 Step 3/5 : COPY example-hook-sidecar /example-hook-sidecar ---> 2a66d92d1a70 Removing intermediate container 4d84d4f4bdff Step 4/5 : ENTRYPOINT /example-hook-sidecar ---> Running in f32c918d27ca ---> 5c13b822d2d1 Removing intermediate container f32c918d27ca Step 5/5 : LABEL "example-hook-sidecar" '' "kubevirt-functional-tests-k8s-1.11.0-release1" '' ---> Running in 94959b663914 ---> dada39a17c19 Removing intermediate container 94959b663914 Successfully built dada39a17c19 hack/build-docker.sh push The push refers to a repository [localhost:33000/kubevirt/virt-controller] a2ba080747c5: Preparing b2f5abdac324: Preparing 891e1e4ef82a: Preparing b2f5abdac324: Pushed a2ba080747c5: Pushed 891e1e4ef82a: Pushed devel: digest: sha256:34e25a358a2c1b1a01eb8c9097ea34259878dbd0d81d4dea7d213265aae4244d size: 949 The push refers to a repository [localhost:33000/kubevirt/virt-launcher] a78f2226bbde: Preparing f788c0a95d6e: Preparing a0e268093341: Preparing fb1a5f061646: Preparing 512ecbe44938: Preparing 0b99c4111657: Preparing da38cf808aa5: Preparing b83399358a92: Preparing 186d8b3e4fd8: Preparing 0b99c4111657: Waiting da38cf808aa5: Waiting 186d8b3e4fd8: Waiting b83399358a92: Waiting fa6154170bf5: Preparing 5eefb9960a36: Preparing 891e1e4ef82a: Preparing fa6154170bf5: Waiting 5eefb9960a36: Waiting 891e1e4ef82a: Waiting fb1a5f061646: Pushed a78f2226bbde: Pushed f788c0a95d6e: Pushed da38cf808aa5: Pushed b83399358a92: Pushed 186d8b3e4fd8: Pushed fa6154170bf5: Pushed a0e268093341: Pushed 891e1e4ef82a: Mounted from kubevirt/virt-controller 512ecbe44938: Pushed 0b99c4111657: Pushed 5eefb9960a36: Pushed devel: digest: sha256:46e4357b4df5a8912f2200afe4c94e50e6edeee31b48062a2f7fad62412e96b0 size: 2828 The push refers to a repository [localhost:33000/kubevirt/virt-handler] 9adefec9aa5d: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/virt-launcher 9adefec9aa5d: Pushed devel: digest: sha256:6ee477fe03bf01b6769eba33944ef268397a3fc5518e03f0c45bcfa67df9b937 size: 741 The push refers to a repository [localhost:33000/kubevirt/virt-api] 3878d29f2ed5: Preparing afd1d781e4d1: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/virt-handler afd1d781e4d1: Pushed 3878d29f2ed5: Pushed devel: digest: sha256:0bb36fdd7467f2e0cf20c04a5c623be2af7a2294dd5f8a1e3210c5c4fbd583d1 size: 948 The push refers to a repository [localhost:33000/kubevirt/disks-images-provider] dc0875c44573: Preparing 8fc77a44094f: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/virt-api dc0875c44573: Pushed 8fc77a44094f: Pushed devel: digest: sha256:cb6348df94e8afd7658fdc2dddc7cf282373350661587253e1644793a1b1a8de size: 948 The push refers to a repository [localhost:33000/kubevirt/vm-killer] d1b69e768421: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/disks-images-provider d1b69e768421: Pushed devel: digest: sha256:018237675286c2660cb5fc9687f728d8390fbb946c8a9f3173cf1e0332f2b48a size: 740 The push refers to a repository [localhost:33000/kubevirt/registry-disk-v1alpha] 2a15632f54d4: Preparing 91a924e03d7c: Preparing 25edbec0eaea: Preparing 2a15632f54d4: Pushed 91a924e03d7c: Pushed 25edbec0eaea: Pushed devel: digest: sha256:9ee50b797053306bd5f3e02582611d3af223831ab5cd25047070b588b9e8c7c9 size: 948 The push refers to a repository [localhost:33000/kubevirt/cirros-registry-disk-demo] a4d34e46c815: Preparing 2a15632f54d4: Preparing 91a924e03d7c: Preparing 25edbec0eaea: Preparing 91a924e03d7c: Mounted from kubevirt/registry-disk-v1alpha 25edbec0eaea: Mounted from kubevirt/registry-disk-v1alpha 2a15632f54d4: Mounted from kubevirt/registry-disk-v1alpha a4d34e46c815: Pushed devel: digest: sha256:39d7e91e514c3c4edb49dd2167f75ccf2ec5d81389804655bf95ff577efaad38 size: 1160 The push refers to a repository [localhost:33000/kubevirt/fedora-cloud-registry-disk-demo] d811f437a779: Preparing 2a15632f54d4: Preparing 91a924e03d7c: Preparing 25edbec0eaea: Preparing 91a924e03d7c: Mounted from kubevirt/cirros-registry-disk-demo 25edbec0eaea: Mounted from kubevirt/cirros-registry-disk-demo 2a15632f54d4: Mounted from kubevirt/cirros-registry-disk-demo d811f437a779: Pushed devel: digest: sha256:bc40df962eee7e2654678bd77ec073c5257dfad1ebbaacf9c29872e035a91a55 size: 1161 The push refers to a repository [localhost:33000/kubevirt/alpine-registry-disk-demo] ccc249ff5804: Preparing 2a15632f54d4: Preparing 91a924e03d7c: Preparing 25edbec0eaea: Preparing 25edbec0eaea: Mounted from kubevirt/fedora-cloud-registry-disk-demo 91a924e03d7c: Mounted from kubevirt/fedora-cloud-registry-disk-demo 2a15632f54d4: Mounted from kubevirt/fedora-cloud-registry-disk-demo ccc249ff5804: Pushed devel: digest: sha256:2e01f3baf16a9bdba2dbd8e07b2acbb03ffbd493e29d56baef08b214b6ca3c36 size: 1160 The push refers to a repository [localhost:33000/kubevirt/subresource-access-test] bb87d5682829: Preparing 4052ce9d0aff: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/vm-killer 4052ce9d0aff: Pushed bb87d5682829: Pushed devel: digest: sha256:125cb3263951cdab536ed0cbef1da33a7996f217a1d267bc3c1d1d22954ac9a2 size: 948 The push refers to a repository [localhost:33000/kubevirt/winrmcli] 64ccc7ac4271: Preparing 4242962b50c3: Preparing 0e374d8c733e: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/subresource-access-test 64ccc7ac4271: Pushed 0e374d8c733e: Pushed 4242962b50c3: Pushed devel: digest: sha256:36f030f5f71fd7b4686092ec1e8f115463056fcd62248338b1db2f82587b5a28 size: 1165 The push refers to a repository [localhost:33000/kubevirt/example-hook-sidecar] 463ab3c605cc: Preparing 39bae602f753: Preparing 463ab3c605cc: Pushed 39bae602f753: Pushed devel: digest: sha256:6a78ea2235d1ecfbe893209eb600cb2ea589170df96519d05dea111231d36422 size: 740 make[1]: Leaving directory `/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt' Done ./cluster/clean.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ MANIFEST_TEMPLATES_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/templates/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ KUBEVIRT_PROVIDER=k8s-1.11.0 ++ KUBEVIRT_NUM_NODES=2 ++ '[' -z kubevirt-functional-tests-k8s-1.11.0-release ']' ++ provider_prefix=kubevirt-functional-tests-k8s-1.11.0-release1 ++ job_prefix=kubevirt-functional-tests-k8s-1.11.0-release1 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.7.0-137-g8f207d1 ++ KUBEVIRT_VERSION=v0.7.0-137-g8f207d1 + source cluster/k8s-1.11.0/provider.sh ++ set -e ++ image=k8s-1.11.0@sha256:6c1caf5559eb02a144bf606de37eb0194c06ace4d77ad4561459f3bde876151c ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --net=host --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/gocli@sha256:aa7f295a7908fa333ab5e98ef3af0bfafbabfd3cee2b83f9af47f722e3000f6a' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ KUBEVIRT_PROVIDER=k8s-1.11.0 ++ KUBEVIRT_PROVIDER=k8s-1.11.0 ++ source hack/config-default.sh source hack/config-k8s-1.11.0.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test cmd/example-hook-sidecar' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/disks-images-provider images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli cmd/example-hook-sidecar' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ namespace=kube-system ++ test -f hack/config-provider-k8s-1.11.0.sh ++ source hack/config-provider-k8s-1.11.0.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/cluster/k8s-1.11.0/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/cluster/k8s-1.11.0/.kubectl +++ docker_prefix=localhost:33000/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Cleaning up ...' Cleaning up ... + cluster/kubectl.sh get vmis --all-namespaces -o=custom-columns=NAME:.metadata.name,NAMESPACE:.metadata.namespace,FINALIZERS:.metadata.finalizers --no-headers + grep foregroundDeleteVirtualMachine + read p error: the server doesn't have a resource type "vmis" + _kubectl delete ds -l kubevirt.io -n kube-system --cascade=false --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=libvirt --force --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=virt-handler --force --grace-period 0 No resources found + namespaces=(default ${namespace}) + for i in '${namespaces[@]}' + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete deployment -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete deployment -l kubevirt.io No resources found + _kubectl -n default delete rs -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete rs -l kubevirt.io No resources found + _kubectl -n default delete services -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete services -l kubevirt.io No resources found + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n default delete secrets -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete secrets -l kubevirt.io No resources found + _kubectl -n default delete pv -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete pv -l kubevirt.io No resources found + _kubectl -n default delete pvc -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete pvc -l kubevirt.io No resources found + _kubectl -n default delete ds -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete ds -l kubevirt.io No resources found + _kubectl -n default delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n default delete pods -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete pods -l kubevirt.io No resources found + _kubectl -n default delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n default delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete rolebinding -l kubevirt.io No resources found + _kubectl -n default delete roles -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete roles -l kubevirt.io No resources found + _kubectl -n default delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete clusterroles -l kubevirt.io No resources found + _kubectl -n default delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n default delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n default get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig ++ KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig ++ cluster/k8s-1.11.0/.kubectl -n default get crd offlinevirtualmachines.kubevirt.io No resources found. Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + for i in '${namespaces[@]}' + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete deployment -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete deployment -l kubevirt.io No resources found + _kubectl -n kube-system delete rs -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete rs -l kubevirt.io No resources found + _kubectl -n kube-system delete services -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete services -l kubevirt.io No resources found + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n kube-system delete secrets -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete secrets -l kubevirt.io No resources found + _kubectl -n kube-system delete pv -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete pv -l kubevirt.io No resources found + _kubectl -n kube-system delete pvc -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete pvc -l kubevirt.io No resources found + _kubectl -n kube-system delete ds -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete ds -l kubevirt.io No resources found + _kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n kube-system delete pods -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete pods -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete rolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete roles -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete roles -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete clusterroles -l kubevirt.io No resources found + _kubectl -n kube-system delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl -n kube-system delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig ++ KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig ++ cluster/k8s-1.11.0/.kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io No resources found. Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + sleep 2 + echo Done Done ./cluster/deploy.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ MANIFEST_TEMPLATES_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/templates/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ KUBEVIRT_PROVIDER=k8s-1.11.0 ++ KUBEVIRT_NUM_NODES=2 ++ '[' -z kubevirt-functional-tests-k8s-1.11.0-release ']' ++ provider_prefix=kubevirt-functional-tests-k8s-1.11.0-release1 ++ job_prefix=kubevirt-functional-tests-k8s-1.11.0-release1 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.7.0-137-g8f207d1 ++ KUBEVIRT_VERSION=v0.7.0-137-g8f207d1 + source cluster/k8s-1.11.0/provider.sh ++ set -e ++ image=k8s-1.11.0@sha256:6c1caf5559eb02a144bf606de37eb0194c06ace4d77ad4561459f3bde876151c ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --net=host --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/gocli@sha256:aa7f295a7908fa333ab5e98ef3af0bfafbabfd3cee2b83f9af47f722e3000f6a' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ KUBEVIRT_PROVIDER=k8s-1.11.0 ++ KUBEVIRT_PROVIDER=k8s-1.11.0 ++ source hack/config-default.sh source hack/config-k8s-1.11.0.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test cmd/example-hook-sidecar' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/disks-images-provider images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli cmd/example-hook-sidecar' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ namespace=kube-system ++ test -f hack/config-provider-k8s-1.11.0.sh ++ source hack/config-provider-k8s-1.11.0.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/cluster/k8s-1.11.0/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/cluster/k8s-1.11.0/.kubectl +++ docker_prefix=localhost:33000/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Deploying ...' Deploying ... + [[ -z k8s-1.11.0-release ]] + [[ k8s-1.11.0-release =~ .*-dev ]] + [[ k8s-1.11.0-release =~ .*-release ]] + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/demo-content.yaml =~ .*demo.* ]] + continue + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml =~ .*demo.* ]] + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml clusterrole.rbac.authorization.k8s.io/kubevirt.io:admin created clusterrole.rbac.authorization.k8s.io/kubevirt.io:edit created clusterrole.rbac.authorization.k8s.io/kubevirt.io:view created serviceaccount/kubevirt-apiserver created clusterrolebinding.rbac.authorization.k8s.io/kubevirt-apiserver created clusterrolebinding.rbac.authorization.k8s.io/kubevirt-apiserver-auth-delegator created rolebinding.rbac.authorization.k8s.io/kubevirt-apiserver created role.rbac.authorization.k8s.io/kubevirt-apiserver created clusterrole.rbac.authorization.k8s.io/kubevirt-apiserver created clusterrole.rbac.authorization.k8s.io/kubevirt-controller created serviceaccount/kubevirt-controller created serviceaccount/kubevirt-privileged created clusterrolebinding.rbac.authorization.k8s.io/kubevirt-controller created clusterrolebinding.rbac.authorization.k8s.io/kubevirt-controller-cluster-admin created clusterrolebinding.rbac.authorization.k8s.io/kubevirt-privileged-cluster-admin created clusterrole.rbac.authorization.k8s.io/kubevirt.io:default created clusterrolebinding.rbac.authorization.k8s.io/kubevirt.io:default created service/virt-api created deployment.extensions/virt-api created deployment.extensions/virt-controller created daemonset.extensions/virt-handler created customresourcedefinition.apiextensions.k8s.io/virtualmachineinstances.kubevirt.io created customresourcedefinition.apiextensions.k8s.io/virtualmachineinstancereplicasets.kubevirt.io created customresourcedefinition.apiextensions.k8s.io/virtualmachineinstancepresets.kubevirt.io created customresourcedefinition.apiextensions.k8s.io/virtualmachines.kubevirt.io created + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R + export KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + KUBECONFIG=cluster/k8s-1.11.0/.kubeconfig + cluster/k8s-1.11.0/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R persistentvolumeclaim/disk-alpine created persistentvolume/host-path-disk-alpine created persistentvolumeclaim/disk-custom created persistentvolume/host-path-disk-custom created daemonset.extensions/disks-images-provider created serviceaccount/kubevirt-testing created clusterrolebinding.rbac.authorization.k8s.io/kubevirt-testing-cluster-admin created + [[ k8s-1.11.0 =~ os-* ]] + echo Done Done + namespaces=(kube-system default) + [[ kube-system != \k\u\b\e\-\s\y\s\t\e\m ]] + timeout=300 + sample=30 + for i in '${namespaces[@]}' + current_time=0 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'virt-api-bcc6b587d-d5h4w 0/1 ContainerCreating 0 4s virt-api-bcc6b587d-k269w 0/1 ContainerCreating 0 4s virt-controller-67dcdd8464-8pvbc 0/1 ContainerCreating 0 4s virt-controller-67dcdd8464-z6bgg 0/1 ContainerCreating 0 4s virt-handler-jxj94 0/1 ContainerCreating 0 3s virt-handler-tsdfg 0/1 ContainerCreating 0 3s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers virt-api-bcc6b587d-d5h4w 0/1 ContainerCreating 0 4s virt-api-bcc6b587d-k269w 0/1 ContainerCreating 0 4s virt-controller-67dcdd8464-8pvbc 0/1 ContainerCreating 0 4s virt-controller-67dcdd8464-z6bgg 0/1 ContainerCreating 0 4s virt-handler-jxj94 0/1 ContainerCreating 0 3s virt-handler-tsdfg 0/1 ContainerCreating 0 3s + sleep 30 + current_time=30 + '[' 30 -gt 300 ']' ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n '' ']' + current_time=0 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers ++ grep false ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers + '[' -n '' ']' + kubectl get pods -n kube-system + cluster/kubectl.sh get pods -n kube-system NAME READY STATUS RESTARTS AGE coredns-78fcdf6894-7f6vg 1/1 Running 0 7m coredns-78fcdf6894-mqcbk 1/1 Running 0 7m disks-images-provider-h2skr 1/1 Running 0 38s disks-images-provider-wv5xv 1/1 Running 0 38s etcd-node01 1/1 Running 0 6m kube-apiserver-node01 1/1 Running 0 6m kube-controller-manager-node01 1/1 Running 0 7m kube-flannel-ds-48vtc 1/1 Running 0 7m kube-flannel-ds-dqklx 1/1 Running 0 7m kube-proxy-7mrdp 1/1 Running 0 7m kube-proxy-z9gts 1/1 Running 0 7m kube-scheduler-node01 1/1 Running 0 6m virt-api-bcc6b587d-d5h4w 1/1 Running 0 43s virt-api-bcc6b587d-k269w 1/1 Running 0 43s virt-controller-67dcdd8464-8pvbc 1/1 Running 0 43s virt-controller-67dcdd8464-z6bgg 1/1 Running 0 43s virt-handler-jxj94 1/1 Running 0 42s virt-handler-tsdfg 1/1 Running 0 42s + for i in '${namespaces[@]}' + current_time=0 ++ kubectl get pods -n default --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n default --no-headers + '[' -n '' ']' + current_time=0 ++ kubectl get pods -n default '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers ++ grep false ++ cluster/kubectl.sh get pods -n default '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers + '[' -n '' ']' + kubectl get pods -n default + cluster/kubectl.sh get pods -n default NAME READY STATUS RESTARTS AGE local-volume-provisioner-d6jhb 1/1 Running 0 7m local-volume-provisioner-pb72b 1/1 Running 0 7m + kubectl version + cluster/kubectl.sh version Client Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.0", GitCommit:"91e7b4fd31fcd3d5f436da26c980becec37ceefe", GitTreeState:"clean", BuildDate:"2018-06-27T20:17:28Z", GoVersion:"go1.10.2", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"11", GitVersion:"v1.11.0", GitCommit:"91e7b4fd31fcd3d5f436da26c980becec37ceefe", GitTreeState:"clean", BuildDate:"2018-06-27T20:08:34Z", GoVersion:"go1.10.2", Compiler:"gc", Platform:"linux/amd64"} + ginko_params='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/junit.xml' + [[ k8s-1.11.0-release =~ windows.* ]] + FUNC_TEST_ARGS='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-k8s-1.11.0-release/junit.xml' + make functest hack/dockerized "hack/build-func-tests.sh" sha256:ceba12cbc33e4e37a707840478a630db561e2427b78c8c9f9cd6d0b73276ab32 go version go1.10 linux/amd64 go version go1.10 linux/amd64 Compiling tests... compiled tests.test hack/functests.sh Running Suite: Tests Suite ========================== Random Seed: 1532682494 Will run 148 of 148 specs • [SLOW TEST:35.280 seconds] VNC /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:46 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:54 with VNC connection /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:62 should allow accessing the VNC device /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:64 ------------------------------ •• ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.016 seconds] Windows VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 should succeed to start a vmi [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:133 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1352 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.006 seconds] Windows VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 should succeed to stop a running vmi [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:139 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1352 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.005 seconds] Windows VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with winrm connection [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:150 should have correct UUID /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:192 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1352 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.004 seconds] Windows VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with winrm connection [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:150 should have pod IP /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:208 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1352 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.003 seconds] Windows VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with kubectl command [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:226 should succeed to start a vmi /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:242 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1352 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.006 seconds] Windows VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with kubectl command [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:226 should succeed to stop a vmi /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:250 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1352 ------------------------------ ••••••••••• ------------------------------ • [SLOW TEST:51.823 seconds] Console 2018/07/27 05:09:47 read closing down: EOF /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:37 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 with a cirros image /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:67 should return that we are running cirros /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:68 ------------------------------ • [SLOW TEST:56.211 seconds] 2018/07/27 05:10:44 read closing down: EOF Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:37 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 with a fedora image /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:77 should return that we are running fedora /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:78 ------------------------------ 2018/07/27 05:11:30 read closing down: EOF 2018/07/27 05:11:32 read closing down: EOF 2018/07/27 05:11:33 read closing down: EOF 2018/07/27 05:11:33 read closing down: EOF • [SLOW TEST:49.737 seconds] 2018/07/27 05:11:33 read closing down: EOF Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:37 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 should be able to reconnect to console multiple times /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:87 ------------------------------ • [SLOW TEST:16.108 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:37 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 should wait until the virtual machine is in running state and return a stream interface /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:103 ------------------------------ • [SLOW TEST:30.219 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:37 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 should fail waiting for the virtual machine instance to be running /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:111 ------------------------------ • [SLOW TEST:30.219 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:37 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 should fail waiting for the expecter /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:134 ------------------------------ • ------------------------------ • [SLOW TEST:5.641 seconds] User Access /root/go/src/kubevirt.io/kubevirt/tests/access_test.go:33 With default kubevirt service accounts /root/go/src/kubevirt.io/kubevirt/tests/access_test.go:41 should verify permissions are correct for view, edit, and admin /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 given a vmi /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • [SLOW TEST:5.125 seconds] User Access /root/go/src/kubevirt.io/kubevirt/tests/access_test.go:33 With default kubevirt service accounts /root/go/src/kubevirt.io/kubevirt/tests/access_test.go:41 should verify permissions are correct for view, edit, and admin /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 given an vm /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • ------------------------------ • [SLOW TEST:5.017 seconds] User Access /root/go/src/kubevirt.io/kubevirt/tests/access_test.go:33 With default kubevirt service accounts /root/go/src/kubevirt.io/kubevirt/tests/access_test.go:41 should verify permissions are correct for view, edit, and admin /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 given a vmi replica set /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ 2018/07/27 05:14:00 read closing down: EOF Service cluster-ip-vmi successfully exposed for virtualmachineinstance testvmi6z4bw • [SLOW TEST:52.163 seconds] Expose /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:53 Expose service on a VM /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:61 Expose ClusterIP service /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:68 Should expose a Cluster IP service on a VMI and connect to it /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:71 ------------------------------ Service cluster-ip-target-vmi successfully exposed for virtualmachineinstance testvmi6z4bw •Service node-port-vmi successfully exposed for virtualmachineinstance testvmi6z4bw ------------------------------ • [SLOW TEST:9.160 seconds] Expose /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:53 Expose service on a VM /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:61 Expose NodePort service /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:124 Should expose a NodePort service on a VMI and connect to it /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:129 ------------------------------ 2018/07/27 05:15:01 read closing down: EOF Service cluster-ip-udp-vmi successfully exposed for virtualmachineinstance testvmilxvnw • [SLOW TEST:52.930 seconds] Expose /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:53 Expose UDP service on a VMI /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:166 Expose ClusterIP UDP service /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:173 Should expose a ClusterIP service on a VMI and connect to it /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:177 ------------------------------ Service node-port-udp-vmi successfully exposed for virtualmachineinstance testvmilxvnw • [SLOW TEST:8.283 seconds] Expose /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:53 Expose UDP service on a VMI /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:166 Expose NodePort UDP service /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:205 Should expose a NodePort service on a VMI and connect to it /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:210 ------------------------------ 2018/07/27 05:16:17 read closing down: EOF 2018/07/27 05:16:27 read closing down: EOF Service cluster-ip-vmirs successfully exposed for vmirs replicasetjv84w • [SLOW TEST:78.061 seconds] Expose /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:53 Expose service on a VMI replica set /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:253 Expose ClusterIP service /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:286 Should create a ClusterIP service on VMRS and connect to it /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:290 ------------------------------ Service cluster-ip-vm successfully exposed for virtualmachine testvmiwwcpp VM testvmiwwcpp was scheduled to start 2018/07/27 05:17:26 read closing down: EOF • [SLOW TEST:58.889 seconds] Expose /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:53 Expose service on an VM /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:318 Expose ClusterIP service /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:362 Connect to ClusterIP services that was set when VM was offline /root/go/src/kubevirt.io/kubevirt/tests/expose_test.go:363 ------------------------------ 2018/07/27 05:18:25 read closing down: EOF 2018/07/27 05:19:24 read closing down: EOF • [SLOW TEST:116.009 seconds] 2018/07/27 05:19:26 read closing down: EOF Slirp /root/go/src/kubevirt.io/kubevirt/tests/vmi_slirp_interface_test.go:39 should be able to /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 VirtualMachineInstance with slirp interface /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ •2018/07/27 05:19:28 read closing down: EOF ------------------------------ • [SLOW TEST:74.083 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:41 Starting and stopping the same VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:90 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:91 should success multiple times /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:92 ------------------------------ • [SLOW TEST:14.649 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:41 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:111 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:112 should not modify the spec on status update /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:113 ------------------------------ • [SLOW TEST:32.180 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:41 Starting multiple VMIs /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:129 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:130 should success /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:131 ------------------------------ volumedisk0 compute • [SLOW TEST:52.229 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 VirtualMachineInstance definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:55 with 3 CPU cores /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:56 should report 3 cpu cores under guest OS 2018/07/27 05:22:21 read closing down: EOF /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:62 ------------------------------ • [SLOW TEST:17.589 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 VirtualMachineInstance definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:55 with hugepages /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:108 should consume hugepages /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 hugepages-2Mi /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ S [SKIPPING] [0.234 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 VirtualMachineInstance definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:55 with hugepages /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:108 should consume hugepages /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 hugepages-1Gi [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 No node with hugepages hugepages-1Gi capacity /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:160 ------------------------------ • 2018/07/27 05:24:33 read closing down: EOF ------------------------------ • [SLOW TEST:113.115 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 with CPU spec /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:238 when CPU model defined /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:284 should report defined CPU model /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:285 ------------------------------ 2018/07/27 05:26:39 read closing down: EOF • [SLOW TEST:125.376 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 with CPU spec /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:238 when CPU model equals to passthrough /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:312 should report exactly the same model as node CPU /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:313 ------------------------------ 2018/07/27 05:28:32 read closing down: EOF • [SLOW TEST:113.201 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 with CPU spec /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:238 when CPU model not defined /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:336 should report CPU model from libvirt capabilities /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:337 ------------------------------ • [SLOW TEST:50.231 seconds] 2018/07/27 05:29:22 read closing down: EOF Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 New VirtualMachineInstance with all supported drives /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:357 should have all the device nodes /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:380 ------------------------------ 2018/07/27 05:30:54 read closing down: EOF 2018/07/27 05:31:04 read closing down: EOF 2018/07/27 05:31:14 read closing down: EOF 2018/07/27 05:31:24 read closing down: EOF 2018/07/27 05:31:25 read closing down: EOF 2018/07/27 05:31:26 read closing down: EOF • [SLOW TEST:125.449 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be able to reach /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 the Inbound VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ 2018/07/27 05:31:27 read closing down: EOF 2018/07/27 05:31:27 read closing down: EOF 2018/07/27 05:31:29 read closing down: EOF •2018/07/27 05:31:29 read closing down: EOF 2018/07/27 05:31:29 read closing down: EOF 2018/07/27 05:31:30 read closing down: EOF •2018/07/27 05:31:31 read closing down: EOF 2018/07/27 05:31:31 read closing down: EOF 2018/07/27 05:31:32 read closing down: EOF 2018/07/27 05:31:32 read closing down: EOF •2018/07/27 05:31:32 read closing down: EOF • ------------------------------ • [SLOW TEST:5.075 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be reachable via the propagated IP from a Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 on a different node from Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ •••••2018/07/27 05:32:51 read closing down: EOF 2018/07/27 05:32:52 read closing down: EOF ------------------------------ • [SLOW TEST:53.588 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 VirtualMachineInstance with custom interface model /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:379 should expose the right device type to the guest /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:380 ------------------------------ 2018/07/27 05:32:52 read closing down: EOF 2018/07/27 05:32:53 read closing down: EOF •2018/07/27 05:33:45 read closing down: EOF ------------------------------ 2018/07/27 05:33:45 read closing down: EOF • [SLOW TEST:52.395 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 VirtualMachineInstance with custom MAC address /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:413 should configure custom MAC address /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:414 ------------------------------ 2018/07/27 05:34:40 read closing down: EOF • [SLOW TEST:55.431 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 VirtualMachineInstance with custom MAC address in non-conventional format /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:425 should configure custom MAC address /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:426 ------------------------------ 2018/07/27 05:34:41 read closing down: EOF 2018/07/27 05:35:37 read closing down: EOF • [SLOW TEST:57.060 seconds] 2018/07/27 05:35:38 read closing down: EOF Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 VirtualMachineInstance with custom MAC address and slirp interface /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:438 should configure custom MAC address /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:439 ------------------------------ 2018/07/27 05:36:33 read closing down: EOF • [SLOW TEST:56.451 seconds] 2018/07/27 05:36:34 read closing down: EOF Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 VirtualMachineInstance with disabled automatic attachment of interfaces /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:451 should not configure any external interfaces /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:452 ------------------------------ •• ------------------------------ • [SLOW TEST:16.334 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 should update VirtualMachine once VMIs are up /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:195 ------------------------------ •• ------------------------------ • [SLOW TEST:39.515 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 should recreate VirtualMachineInstance if it gets deleted /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:245 ------------------------------ • [SLOW TEST:35.429 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 should recreate VirtualMachineInstance if the VirtualMachineInstance's pod gets deleted /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:265 ------------------------------ • [SLOW TEST:44.474 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 should stop VirtualMachineInstance if running set to false /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:325 ------------------------------ • [SLOW TEST:178.407 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 should start and stop VirtualMachineInstance multiple times /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:333 ------------------------------ • [SLOW TEST:56.399 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 should not update the VirtualMachineInstance spec if Running /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:346 ------------------------------ 2018/07/27 05:45:49 read closing down: EOF • [SLOW TEST:179.586 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 should survive guest shutdown, multiple times /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:387 ------------------------------ 2018/07/27 05:45:49 read closing down: EOF 2018/07/27 05:45:49 read closing down: EOF VM testvmi422h4 was scheduled to start • [SLOW TEST:16.310 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 Using virtctl interface /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:435 should start a VirtualMachineInstance once /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:436 ------------------------------ VM testvmits8qh was scheduled to stop • [SLOW TEST:28.627 seconds] VirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:47 A valid VirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:115 Using virtctl interface /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:435 should stop a VirtualMachineInstance once /root/go/src/kubevirt.io/kubevirt/tests/vm_test.go:467 ------------------------------ ••••• ------------------------------ • [SLOW TEST:8.196 seconds] VirtualMachineInstanceReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should scale /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 to five, to six and then to zero replicas /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ •• ------------------------------ • [SLOW TEST:17.364 seconds] VirtualMachineInstanceReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should update readyReplicas once VMIs are up /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:157 ------------------------------ ••• ------------------------------ • [SLOW TEST:5.528 seconds] VirtualMachineInstanceReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should remove the finished VM /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:279 ------------------------------ 2018/07/27 05:48:39 read closing down: EOF • [SLOW TEST:61.681 seconds] Health Monitoring /root/go/src/kubevirt.io/kubevirt/tests/vmi_monitoring_test.go:37 A VirtualMachineInstance with a watchdog device /root/go/src/kubevirt.io/kubevirt/tests/vmi_monitoring_test.go:56 should be shut down when the watchdog expires /root/go/src/kubevirt.io/kubevirt/tests/vmi_monitoring_test.go:57 ------------------------------ • Pod name: disks-images-provider-h2skr Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-wv5xv Pod phase: Running copy all images to host mount directory Pod name: virt-api-bcc6b587d-d5h4w Pod phase: Running 2018/07/27 09:50:44 http: TLS handshake error from 10.244.1.1:42224: EOF level=info timestamp=2018-07-27T09:50:46.727475Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:50:54 http: TLS handshake error from 10.244.1.1:42230: EOF level=info timestamp=2018-07-27T09:51:03.140015Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:51:04 http: TLS handshake error from 10.244.1.1:42236: EOF level=info timestamp=2018-07-27T09:51:05.890579Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T09:51:07.735830Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T09:51:10.114673Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:51:14 http: TLS handshake error from 10.244.1.1:42242: EOF level=info timestamp=2018-07-27T09:51:16.781540Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:51:24 http: TLS handshake error from 10.244.1.1:42248: EOF level=info timestamp=2018-07-27T09:51:33.194652Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:51:34 http: TLS handshake error from 10.244.1.1:42254: EOF level=info timestamp=2018-07-27T09:51:35.943809Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T09:51:37.709670Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 Pod name: virt-api-bcc6b587d-k269w Pod phase: Running 2018/07/27 09:49:10 http: TLS handshake error from 10.244.0.1:46850: EOF 2018/07/27 09:49:20 http: TLS handshake error from 10.244.0.1:46910: EOF 2018/07/27 09:49:30 http: TLS handshake error from 10.244.0.1:46970: EOF 2018/07/27 09:49:40 http: TLS handshake error from 10.244.0.1:47030: EOF 2018/07/27 09:49:50 http: TLS handshake error from 10.244.0.1:47090: EOF 2018/07/27 09:50:00 http: TLS handshake error from 10.244.0.1:47150: EOF 2018/07/27 09:50:10 http: TLS handshake error from 10.244.0.1:47210: EOF 2018/07/27 09:50:20 http: TLS handshake error from 10.244.0.1:47270: EOF 2018/07/27 09:50:30 http: TLS handshake error from 10.244.0.1:47330: EOF 2018/07/27 09:50:40 http: TLS handshake error from 10.244.0.1:47390: EOF 2018/07/27 09:50:50 http: TLS handshake error from 10.244.0.1:47450: EOF 2018/07/27 09:51:00 http: TLS handshake error from 10.244.0.1:47510: EOF 2018/07/27 09:51:10 http: TLS handshake error from 10.244.0.1:47570: EOF 2018/07/27 09:51:20 http: TLS handshake error from 10.244.0.1:47630: EOF 2018/07/27 09:51:30 http: TLS handshake error from 10.244.0.1:47690: EOF Pod name: virt-controller-67dcdd8464-8pvbc Pod phase: Running level=error timestamp=2018-07-27T09:47:32.342107Z pos=vm.go:462 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2kcxvmk kind= uid=153bb114-9182-11e8-9c71-525500d15501 msg="Cant find the matching VM for VirtualMachineInstance: testvmizzd2kcxvmk" level=info timestamp=2018-07-27T09:47:32.342162Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2kcxvmk kind= uid=153bb114-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:32.342206Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2kcxvmk kind= uid=153bb114-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=error timestamp=2018-07-27T09:47:33.498089Z pos=replicaset.go:230 component=virt-controller service=http namespace=kubevirt-test-default name=replicasetptp77 kind= uid=153923c9-9182-11e8-9c71-525500d15501 reason="Operation cannot be fulfilled on virtualmachineinstancereplicasets.kubevirt.io \"replicasetptp77\": the object has been modified; please apply your changes to the latest version and try again" msg="Updating the replicaset status failed." level=info timestamp=2018-07-27T09:47:33.498178Z pos=replicaset.go:137 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstancereplicasets.kubevirt.io \"replicasetptp77\": the object has been modified; please apply your changes to the latest version and try again" msg="re-enqueuing VirtualMachineInstanceReplicaSet kubevirt-test-default/replicasetptp77" level=info timestamp=2018-07-27T09:47:33.899631Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmizzd2kcxvmk\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmizzd2kcxvmk" level=info timestamp=2018-07-27T09:47:34.501213Z pos=vm.go:459 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Looking for VirtualMachineInstance Ref" level=error timestamp=2018-07-27T09:47:34.501334Z pos=vm.go:462 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Cant find the matching VM for VirtualMachineInstance: testvmizzd2km4hwq" level=info timestamp=2018-07-27T09:47:34.501401Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:34.501539Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.120102Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:38.120713Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.314985Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmibk5s2\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmibk5s2" level=info timestamp=2018-07-27T09:48:39.420222Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:48:39.421511Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-67dcdd8464-z6bgg Pod phase: Running level=info timestamp=2018-07-27T09:07:02.572685Z pos=application.go:174 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-jxj94 Pod phase: Running level=info timestamp=2018-07-27T09:47:19.763725Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:19.769615Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:20.871997Z pos=vm.go:342 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Shutting down domain for VirtualMachineInstance with deletion timestamp." level=info timestamp=2018-07-27T09:47:20.872512Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T09:47:20.874086Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmind64hp4jxl" level=info timestamp=2018-07-27T09:47:21.092559Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.092698Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:47:21.093354Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.093451Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.093501Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.096761Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.303941Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.304085Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.304241Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.305043Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-handler-tsdfg Pod phase: Running level=info timestamp=2018-07-27T09:47:53.339519Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.339597Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:47:53.371303Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.371451Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:53.376244Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.019328Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:48:39.019562Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:48:39.019713Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027415Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.027536Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027573Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.144850Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.145110Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.181952Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.182091Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-launcher-testvmi645bh-6n8xf Pod phase: Pending ------------------------------ • Failure [180.388 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:71 should be successfully started /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 with Disk PVC [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.014s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 ------------------------------ STEP: Starting a VirtualMachineInstance STEP: Waiting until the VirtualMachineInstance will start level=info timestamp=2018-07-27T09:48:40.099129Z pos=utils.go:243 component=tests msg="Created virtual machine pod virt-launcher-testvmi645bh-6n8xf" Pod name: disks-images-provider-h2skr Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-wv5xv Pod phase: Running copy all images to host mount directory Pod name: virt-api-bcc6b587d-d5h4w Pod phase: Running 2018/07/27 09:53:44 http: TLS handshake error from 10.244.1.1:42332: EOF level=info timestamp=2018-07-27T09:53:47.051975Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:53:54 http: TLS handshake error from 10.244.1.1:42338: EOF level=info timestamp=2018-07-27T09:54:03.450643Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:54:04 http: TLS handshake error from 10.244.1.1:42344: EOF level=info timestamp=2018-07-27T09:54:06.197341Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T09:54:07.731540Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T09:54:10.241386Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:54:14 http: TLS handshake error from 10.244.1.1:42350: EOF level=info timestamp=2018-07-27T09:54:17.111010Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:54:24 http: TLS handshake error from 10.244.1.1:42356: EOF level=info timestamp=2018-07-27T09:54:33.509809Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:54:34 http: TLS handshake error from 10.244.1.1:42362: EOF level=info timestamp=2018-07-27T09:54:36.259756Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T09:54:37.769739Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 Pod name: virt-api-bcc6b587d-k269w Pod phase: Running 2018/07/27 09:52:10 http: TLS handshake error from 10.244.0.1:47930: EOF 2018/07/27 09:52:20 http: TLS handshake error from 10.244.0.1:47990: EOF 2018/07/27 09:52:30 http: TLS handshake error from 10.244.0.1:48050: EOF 2018/07/27 09:52:40 http: TLS handshake error from 10.244.0.1:48110: EOF 2018/07/27 09:52:50 http: TLS handshake error from 10.244.0.1:48170: EOF 2018/07/27 09:53:00 http: TLS handshake error from 10.244.0.1:48230: EOF 2018/07/27 09:53:10 http: TLS handshake error from 10.244.0.1:48290: EOF 2018/07/27 09:53:20 http: TLS handshake error from 10.244.0.1:48350: EOF 2018/07/27 09:53:30 http: TLS handshake error from 10.244.0.1:48410: EOF 2018/07/27 09:53:40 http: TLS handshake error from 10.244.0.1:48470: EOF 2018/07/27 09:53:50 http: TLS handshake error from 10.244.0.1:48530: EOF 2018/07/27 09:54:00 http: TLS handshake error from 10.244.0.1:48590: EOF 2018/07/27 09:54:10 http: TLS handshake error from 10.244.0.1:48650: EOF 2018/07/27 09:54:20 http: TLS handshake error from 10.244.0.1:48710: EOF 2018/07/27 09:54:30 http: TLS handshake error from 10.244.0.1:48770: EOF Pod name: virt-controller-67dcdd8464-8pvbc Pod phase: Running level=error timestamp=2018-07-27T09:47:33.498089Z pos=replicaset.go:230 component=virt-controller service=http namespace=kubevirt-test-default name=replicasetptp77 kind= uid=153923c9-9182-11e8-9c71-525500d15501 reason="Operation cannot be fulfilled on virtualmachineinstancereplicasets.kubevirt.io \"replicasetptp77\": the object has been modified; please apply your changes to the latest version and try again" msg="Updating the replicaset status failed." level=info timestamp=2018-07-27T09:47:33.498178Z pos=replicaset.go:137 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstancereplicasets.kubevirt.io \"replicasetptp77\": the object has been modified; please apply your changes to the latest version and try again" msg="re-enqueuing VirtualMachineInstanceReplicaSet kubevirt-test-default/replicasetptp77" level=info timestamp=2018-07-27T09:47:33.899631Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmizzd2kcxvmk\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmizzd2kcxvmk" level=info timestamp=2018-07-27T09:47:34.501213Z pos=vm.go:459 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Looking for VirtualMachineInstance Ref" level=error timestamp=2018-07-27T09:47:34.501334Z pos=vm.go:462 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Cant find the matching VM for VirtualMachineInstance: testvmizzd2km4hwq" level=info timestamp=2018-07-27T09:47:34.501401Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:34.501539Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.120102Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:38.120713Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.314985Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmibk5s2\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmibk5s2" level=info timestamp=2018-07-27T09:48:39.420222Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:48:39.421511Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:51:39.950534Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:51:39.951875Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:51:39.998797Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6z2qq\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6z2qq" Pod name: virt-controller-67dcdd8464-z6bgg Pod phase: Running level=info timestamp=2018-07-27T09:07:02.572685Z pos=application.go:174 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-jxj94 Pod phase: Running level=info timestamp=2018-07-27T09:47:19.763725Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:19.769615Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:20.871997Z pos=vm.go:342 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Shutting down domain for VirtualMachineInstance with deletion timestamp." level=info timestamp=2018-07-27T09:47:20.872512Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T09:47:20.874086Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmind64hp4jxl" level=info timestamp=2018-07-27T09:47:21.092559Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.092698Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:47:21.093354Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.093451Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.093501Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.096761Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.303941Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.304085Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.304241Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.305043Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-handler-tsdfg Pod phase: Running level=info timestamp=2018-07-27T09:47:53.339519Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.339597Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:47:53.371303Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.371451Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:53.376244Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.019328Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:48:39.019562Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:48:39.019713Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027415Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.027536Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027573Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.144850Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.145110Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.181952Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.182091Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-launcher-testvmi6z2qq-5577p Pod phase: Pending • Failure [180.653 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:71 should be successfully started /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 with CDRom PVC [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.007s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 ------------------------------ STEP: Starting a VirtualMachineInstance STEP: Waiting until the VirtualMachineInstance will start level=info timestamp=2018-07-27T09:51:40.555153Z pos=utils.go:243 component=tests msg="Created virtual machine pod virt-launcher-testvmi6z2qq-5577p" Pod name: disks-images-provider-h2skr Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-wv5xv Pod phase: Running copy all images to host mount directory Pod name: virt-api-bcc6b587d-d5h4w Pod phase: Running level=info timestamp=2018-07-27T09:56:47.384812Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:56:54 http: TLS handshake error from 10.244.1.1:42446: EOF level=info timestamp=2018-07-27T09:57:03.768022Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:57:04 http: TLS handshake error from 10.244.1.1:42452: EOF level=info timestamp=2018-07-27T09:57:06.512114Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T09:57:07.718346Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T09:57:10.417420Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:57:14 http: TLS handshake error from 10.244.1.1:42458: EOF level=info timestamp=2018-07-27T09:57:17.437864Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:57:24 http: TLS handshake error from 10.244.1.1:42464: EOF level=info timestamp=2018-07-27T09:57:33.824338Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:57:34 http: TLS handshake error from 10.244.1.1:42470: EOF level=info timestamp=2018-07-27T09:57:36.562336Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T09:57:37.753338Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T09:57:40.436679Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-api-bcc6b587d-k269w Pod phase: Running 2018/07/27 09:55:20 http: TLS handshake error from 10.244.0.1:49070: EOF 2018/07/27 09:55:30 http: TLS handshake error from 10.244.0.1:49130: EOF 2018/07/27 09:55:40 http: TLS handshake error from 10.244.0.1:49190: EOF 2018/07/27 09:55:50 http: TLS handshake error from 10.244.0.1:49250: EOF 2018/07/27 09:56:00 http: TLS handshake error from 10.244.0.1:49310: EOF 2018/07/27 09:56:10 http: TLS handshake error from 10.244.0.1:49370: EOF 2018/07/27 09:56:20 http: TLS handshake error from 10.244.0.1:49430: EOF 2018/07/27 09:56:30 http: TLS handshake error from 10.244.0.1:49490: EOF 2018/07/27 09:56:40 http: TLS handshake error from 10.244.0.1:49550: EOF 2018/07/27 09:56:50 http: TLS handshake error from 10.244.0.1:49610: EOF 2018/07/27 09:57:00 http: TLS handshake error from 10.244.0.1:49670: EOF 2018/07/27 09:57:10 http: TLS handshake error from 10.244.0.1:49730: EOF 2018/07/27 09:57:20 http: TLS handshake error from 10.244.0.1:49790: EOF 2018/07/27 09:57:30 http: TLS handshake error from 10.244.0.1:49850: EOF 2018/07/27 09:57:40 http: TLS handshake error from 10.244.0.1:49910: EOF Pod name: virt-controller-67dcdd8464-8pvbc Pod phase: Running level=info timestamp=2018-07-27T09:47:33.899631Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmizzd2kcxvmk\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmizzd2kcxvmk" level=info timestamp=2018-07-27T09:47:34.501213Z pos=vm.go:459 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Looking for VirtualMachineInstance Ref" level=error timestamp=2018-07-27T09:47:34.501334Z pos=vm.go:462 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Cant find the matching VM for VirtualMachineInstance: testvmizzd2km4hwq" level=info timestamp=2018-07-27T09:47:34.501401Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:34.501539Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.120102Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:38.120713Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.314985Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmibk5s2\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmibk5s2" level=info timestamp=2018-07-27T09:48:39.420222Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:48:39.421511Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:51:39.950534Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:51:39.951875Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:51:39.998797Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6z2qq\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6z2qq" level=info timestamp=2018-07-27T09:54:40.622927Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:54:40.623101Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-67dcdd8464-z6bgg Pod phase: Running level=info timestamp=2018-07-27T09:07:02.572685Z pos=application.go:174 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-jxj94 Pod phase: Running level=info timestamp=2018-07-27T09:47:19.763725Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:19.769615Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:20.871997Z pos=vm.go:342 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Shutting down domain for VirtualMachineInstance with deletion timestamp." level=info timestamp=2018-07-27T09:47:20.872512Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T09:47:20.874086Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmind64hp4jxl" level=info timestamp=2018-07-27T09:47:21.092559Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.092698Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:47:21.093354Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.093451Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.093501Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.096761Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.303941Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.304085Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.304241Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.305043Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-handler-tsdfg Pod phase: Running level=info timestamp=2018-07-27T09:47:53.339519Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.339597Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:47:53.371303Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.371451Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:53.376244Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.019328Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:48:39.019562Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:48:39.019713Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027415Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.027536Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027573Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.144850Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.145110Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.181952Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.182091Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-launcher-testvmi4mpkt-l6n9c Pod phase: Pending • Failure [180.527 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:71 should be successfully started and stopped multiple times /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 with Disk PVC [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.005s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 ------------------------------ STEP: Starting and stopping the VirtualMachineInstance number of times STEP: Starting a VirtualMachineInstance STEP: Waiting until the VirtualMachineInstance will start level=info timestamp=2018-07-27T09:54:41.245391Z pos=utils.go:243 component=tests msg="Created virtual machine pod virt-launcher-testvmi4mpkt-l6n9c" Pod name: disks-images-provider-h2skr Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-wv5xv Pod phase: Running copy all images to host mount directory Pod name: virt-api-bcc6b587d-d5h4w Pod phase: Running level=info timestamp=2018-07-27T09:59:47.704969Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 09:59:54 http: TLS handshake error from 10.244.1.1:42554: EOF level=info timestamp=2018-07-27T10:00:04.081220Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:00:04 http: TLS handshake error from 10.244.1.1:42560: EOF level=info timestamp=2018-07-27T10:00:06.824423Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:00:07.747607Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:00:10.539601Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:00:14 http: TLS handshake error from 10.244.1.1:42566: EOF level=info timestamp=2018-07-27T10:00:17.759618Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:00:24 http: TLS handshake error from 10.244.1.1:42572: EOF level=info timestamp=2018-07-27T10:00:34.133946Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:00:34 http: TLS handshake error from 10.244.1.1:42578: EOF level=info timestamp=2018-07-27T10:00:36.872803Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:00:37.721780Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:00:40.562672Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-api-bcc6b587d-k269w Pod phase: Running 2018/07/27 09:58:20 http: TLS handshake error from 10.244.0.1:50150: EOF 2018/07/27 09:58:30 http: TLS handshake error from 10.244.0.1:50210: EOF 2018/07/27 09:58:40 http: TLS handshake error from 10.244.0.1:50270: EOF 2018/07/27 09:58:50 http: TLS handshake error from 10.244.0.1:50330: EOF 2018/07/27 09:59:00 http: TLS handshake error from 10.244.0.1:50390: EOF 2018/07/27 09:59:10 http: TLS handshake error from 10.244.0.1:50450: EOF 2018/07/27 09:59:20 http: TLS handshake error from 10.244.0.1:50510: EOF 2018/07/27 09:59:30 http: TLS handshake error from 10.244.0.1:50570: EOF 2018/07/27 09:59:40 http: TLS handshake error from 10.244.0.1:50630: EOF 2018/07/27 09:59:50 http: TLS handshake error from 10.244.0.1:50690: EOF 2018/07/27 10:00:00 http: TLS handshake error from 10.244.0.1:50750: EOF 2018/07/27 10:00:10 http: TLS handshake error from 10.244.0.1:50810: EOF 2018/07/27 10:00:20 http: TLS handshake error from 10.244.0.1:50870: EOF 2018/07/27 10:00:30 http: TLS handshake error from 10.244.0.1:50930: EOF 2018/07/27 10:00:40 http: TLS handshake error from 10.244.0.1:50990: EOF Pod name: virt-controller-67dcdd8464-8pvbc Pod phase: Running level=info timestamp=2018-07-27T09:47:34.501401Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:34.501539Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmizzd2km4hwq kind= uid=16853d52-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.120102Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:47:38.120713Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:47:38.314985Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmibk5s2\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmibk5s2" level=info timestamp=2018-07-27T09:48:39.420222Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:48:39.421511Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi645bh kind= uid=3d35c87d-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:51:39.950534Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:51:39.951875Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:51:39.998797Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6z2qq\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6z2qq" level=info timestamp=2018-07-27T09:54:40.622927Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:54:40.623101Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:57:41.132211Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7fkpx kind= uid=80198ec0-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:57:41.132395Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7fkpx kind= uid=80198ec0-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:57:41.224693Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7fkpx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7fkpx" Pod name: virt-controller-67dcdd8464-z6bgg Pod phase: Running level=info timestamp=2018-07-27T09:07:02.572685Z pos=application.go:174 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-jxj94 Pod phase: Running level=info timestamp=2018-07-27T09:47:19.763725Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:19.769615Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:20.871997Z pos=vm.go:342 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Shutting down domain for VirtualMachineInstance with deletion timestamp." level=info timestamp=2018-07-27T09:47:20.872512Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T09:47:20.874086Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmind64hp4jxl" level=info timestamp=2018-07-27T09:47:21.092559Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.092698Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:47:21.093354Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.093451Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.093501Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.096761Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.303941Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.304085Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.304241Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.305043Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-handler-tsdfg Pod phase: Running level=info timestamp=2018-07-27T09:47:53.339519Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.339597Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:47:53.371303Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:53.371451Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:53.376244Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.019328Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:48:39.019562Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:48:39.019713Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027415Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.027536Z pos=vm.go:392 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="No update processing required" level=info timestamp=2018-07-27T09:48:39.027573Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.144850Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.145110Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind= uid=18a704db-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:48:39.181952Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:48:39.182091Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmibk5s2 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-launcher-testvmi7fkpx-ddbp7 Pod phase: Pending • Failure [180.443 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:71 should be successfully started and stopped multiple times /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 with CDRom PVC [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.003s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 ------------------------------ STEP: Starting and stopping the VirtualMachineInstance number of times STEP: Starting a VirtualMachineInstance STEP: Waiting until the VirtualMachineInstance will start level=info timestamp=2018-07-27T09:57:41.756700Z pos=utils.go:243 component=tests msg="Created virtual machine pod virt-launcher-testvmi7fkpx-ddbp7" 2018/07/27 06:01:31 read closing down: EOF • [SLOW TEST:50.159 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 With an emptyDisk defined /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:113 should create a writeable emptyDisk with the right capacity /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:115 ------------------------------ • [SLOW TEST:50.260 seconds] Storage 2018/07/27 06:02:22 read closing down: EOF /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 With an emptyDisk defined and a specified serial number /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:163 should create a writeable emptyDisk with the specified serial number /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:165 ------------------------------ Pod name: disks-images-provider-h2skr Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-wv5xv Pod phase: Running copy all images to host mount directory Pod name: virt-api-bcc6b587d-d5h4w Pod phase: Running level=info timestamp=2018-07-27T10:04:37.292039Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:04:37.752962Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:04:40.730558Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:04:42.730876Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:04:42.738758Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/07/27 10:04:44 http: TLS handshake error from 10.244.1.1:42736: EOF level=info timestamp=2018-07-27T10:04:48.246355Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:04:54 http: TLS handshake error from 10.244.1.1:42742: EOF level=info timestamp=2018-07-27T10:05:04.603184Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:05:04 http: TLS handshake error from 10.244.1.1:42748: EOF level=info timestamp=2018-07-27T10:05:07.340797Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:05:07.757515Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:05:10.759713Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:05:14 http: TLS handshake error from 10.244.1.1:42754: EOF level=info timestamp=2018-07-27T10:05:18.300055Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-api-bcc6b587d-k269w Pod phase: Running 2018/07/27 10:03:00 http: TLS handshake error from 10.244.0.1:51838: EOF 2018/07/27 10:03:10 http: TLS handshake error from 10.244.0.1:51898: EOF 2018/07/27 10:03:20 http: TLS handshake error from 10.244.0.1:51958: EOF 2018/07/27 10:03:30 http: TLS handshake error from 10.244.0.1:52018: EOF 2018/07/27 10:03:40 http: TLS handshake error from 10.244.0.1:52078: EOF 2018/07/27 10:03:50 http: TLS handshake error from 10.244.0.1:52138: EOF 2018/07/27 10:04:00 http: TLS handshake error from 10.244.0.1:52198: EOF 2018/07/27 10:04:10 http: TLS handshake error from 10.244.0.1:52258: EOF 2018/07/27 10:04:20 http: TLS handshake error from 10.244.0.1:1024: EOF 2018/07/27 10:04:30 http: TLS handshake error from 10.244.0.1:52378: EOF 2018/07/27 10:04:40 http: TLS handshake error from 10.244.0.1:52438: EOF 2018/07/27 10:04:50 http: TLS handshake error from 10.244.0.1:52498: EOF 2018/07/27 10:05:00 http: TLS handshake error from 10.244.0.1:52558: EOF 2018/07/27 10:05:10 http: TLS handshake error from 10.244.0.1:52618: EOF 2018/07/27 10:05:20 http: TLS handshake error from 10.244.0.1:52678: EOF Pod name: virt-controller-67dcdd8464-8pvbc Pod phase: Running level=info timestamp=2018-07-27T09:51:39.950534Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:51:39.951875Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6z2qq kind= uid=a8d191f8-9182-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:51:39.998797Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6z2qq\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6z2qq" level=info timestamp=2018-07-27T09:54:40.622927Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:54:40.623101Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:57:41.132211Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7fkpx kind= uid=80198ec0-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:57:41.132395Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7fkpx kind= uid=80198ec0-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:57:41.224693Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7fkpx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7fkpx" level=info timestamp=2018-07-27T10:00:41.613779Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit6n7m kind= uid=ebac8b8b-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:00:41.613918Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit6n7m kind= uid=ebac8b8b-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:00:41.702313Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmit6n7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmit6n7m" level=info timestamp=2018-07-27T10:01:31.739879Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:01:31.740627Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:02:22.010497Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiwt5js kind= uid=2783dadd-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:02:22.011873Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiwt5js kind= uid=2783dadd-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-67dcdd8464-z6bgg Pod phase: Running level=info timestamp=2018-07-27T09:07:02.572685Z pos=application.go:174 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-jxj94 Pod phase: Running level=info timestamp=2018-07-27T09:47:19.763725Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:19.769615Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:20.871997Z pos=vm.go:342 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Shutting down domain for VirtualMachineInstance with deletion timestamp." level=info timestamp=2018-07-27T09:47:20.872512Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T09:47:20.874086Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmind64hp4jxl" level=info timestamp=2018-07-27T09:47:21.092559Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.092698Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:47:21.093354Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.093451Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.093501Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.096761Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.303941Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.304085Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.304241Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.305043Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-handler-tsdfg Pod phase: Running level=info timestamp=2018-07-27T10:02:21.809400Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T10:02:21.809783Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmipv546" level=info timestamp=2018-07-27T10:02:22.021493Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.021599Z pos=vm.go:330 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Shutting down domain for deleted VirtualMachineInstance object." level=info timestamp=2018-07-27T10:02:22.021634Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing shutdown." level=info timestamp=2018-07-27T10:02:22.021925Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.023564Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T10:02:22.023694Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T10:02:22.023775Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.023827Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.025780Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T10:02:22.103992Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.104205Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.104442Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.104535Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-launcher-testvmiwt5js-ps5g7 Pod phase: Pending • Failure [180.453 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 With ephemeral alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:205 should be successfully started [It] /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:207 Timed out after 90.005s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 ------------------------------ STEP: Starting a VirtualMachineInstance STEP: Waiting until the VirtualMachineInstance will start level=info timestamp=2018-07-27T10:02:22.635608Z pos=utils.go:243 component=tests msg="Created virtual machine pod virt-launcher-testvmiwt5js-ps5g7" Pod name: disks-images-provider-h2skr Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-wv5xv Pod phase: Running copy all images to host mount directory Pod name: virt-api-bcc6b587d-d5h4w Pod phase: Running level=info timestamp=2018-07-27T10:07:37.605167Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:07:37.727640Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:07:40.894318Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:07:42.745157Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:07:42.746791Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/07/27 10:07:44 http: TLS handshake error from 10.244.1.1:42844: EOF level=info timestamp=2018-07-27T10:07:48.566227Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:07:54 http: TLS handshake error from 10.244.1.1:42850: EOF 2018/07/27 10:08:04 http: TLS handshake error from 10.244.1.1:42856: EOF level=info timestamp=2018-07-27T10:08:04.914785Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:08:07.660393Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:08:07.743427Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:08:10.915849Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:08:14 http: TLS handshake error from 10.244.1.1:42862: EOF level=info timestamp=2018-07-27T10:08:18.620020Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-api-bcc6b587d-k269w Pod phase: Running 2018/07/27 10:06:00 http: TLS handshake error from 10.244.0.1:52918: EOF 2018/07/27 10:06:10 http: TLS handshake error from 10.244.0.1:52978: EOF 2018/07/27 10:06:20 http: TLS handshake error from 10.244.0.1:53038: EOF 2018/07/27 10:06:30 http: TLS handshake error from 10.244.0.1:53098: EOF 2018/07/27 10:06:40 http: TLS handshake error from 10.244.0.1:53158: EOF 2018/07/27 10:06:50 http: TLS handshake error from 10.244.0.1:53218: EOF 2018/07/27 10:07:00 http: TLS handshake error from 10.244.0.1:53278: EOF 2018/07/27 10:07:10 http: TLS handshake error from 10.244.0.1:53338: EOF 2018/07/27 10:07:20 http: TLS handshake error from 10.244.0.1:53398: EOF 2018/07/27 10:07:30 http: TLS handshake error from 10.244.0.1:53458: EOF 2018/07/27 10:07:40 http: TLS handshake error from 10.244.0.1:53518: EOF 2018/07/27 10:07:50 http: TLS handshake error from 10.244.0.1:53578: EOF 2018/07/27 10:08:00 http: TLS handshake error from 10.244.0.1:53638: EOF 2018/07/27 10:08:10 http: TLS handshake error from 10.244.0.1:53698: EOF 2018/07/27 10:08:20 http: TLS handshake error from 10.244.0.1:53758: EOF Pod name: virt-controller-67dcdd8464-8pvbc Pod phase: Running level=info timestamp=2018-07-27T09:54:40.622927Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:54:40.623101Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4mpkt kind= uid=14819699-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:57:41.132211Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7fkpx kind= uid=80198ec0-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T09:57:41.132395Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7fkpx kind= uid=80198ec0-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T09:57:41.224693Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7fkpx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7fkpx" level=info timestamp=2018-07-27T10:00:41.613779Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit6n7m kind= uid=ebac8b8b-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:00:41.613918Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit6n7m kind= uid=ebac8b8b-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:00:41.702313Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmit6n7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmit6n7m" level=info timestamp=2018-07-27T10:01:31.739879Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:01:31.740627Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:02:22.010497Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiwt5js kind= uid=2783dadd-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:02:22.011873Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiwt5js kind= uid=2783dadd-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:05:22.464398Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqfcdj kind= uid=9312f188-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:05:22.465566Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqfcdj kind= uid=9312f188-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:05:22.511645Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmiqfcdj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmiqfcdj" Pod name: virt-controller-67dcdd8464-z6bgg Pod phase: Running level=info timestamp=2018-07-27T09:07:02.572685Z pos=application.go:174 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-jxj94 Pod phase: Running level=info timestamp=2018-07-27T09:47:19.763725Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:19.769615Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:20.871997Z pos=vm.go:342 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Shutting down domain for VirtualMachineInstance with deletion timestamp." level=info timestamp=2018-07-27T09:47:20.872512Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T09:47:20.874086Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmind64hp4jxl" level=info timestamp=2018-07-27T09:47:21.092559Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.092698Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:47:21.093354Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.093451Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.093501Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.096761Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.303941Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.304085Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.304241Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.305043Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-handler-tsdfg Pod phase: Running level=info timestamp=2018-07-27T10:02:21.809400Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T10:02:21.809783Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmipv546" level=info timestamp=2018-07-27T10:02:22.021493Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.021599Z pos=vm.go:330 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Shutting down domain for deleted VirtualMachineInstance object." level=info timestamp=2018-07-27T10:02:22.021634Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing shutdown." level=info timestamp=2018-07-27T10:02:22.021925Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.023564Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T10:02:22.023694Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T10:02:22.023775Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.023827Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.025780Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T10:02:22.103992Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.104205Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.104442Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.104535Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-launcher-testvmiqfcdj-9sqcj Pod phase: Pending • Failure [180.476 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 With ephemeral alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:205 should not persist data [It] /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:218 Timed out after 90.005s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 ------------------------------ STEP: Starting the VirtualMachineInstance STEP: Starting a VirtualMachineInstance STEP: Waiting until the VirtualMachineInstance will start level=info timestamp=2018-07-27T10:05:23.032276Z pos=utils.go:243 component=tests msg="Created virtual machine pod virt-launcher-testvmiqfcdj-9sqcj" Pod name: disks-images-provider-h2skr Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-wv5xv Pod phase: Running copy all images to host mount directory Pod name: virt-api-bcc6b587d-d5h4w Pod phase: Running 2018/07/27 10:11:34 http: TLS handshake error from 10.244.1.1:42982: EOF level=info timestamp=2018-07-27T10:11:35.275863Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:11:37.751585Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:11:38.017066Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:11:41.048092Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:11:44 http: TLS handshake error from 10.244.1.1:42988: EOF level=info timestamp=2018-07-27T10:11:48.997356Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:11:54 http: TLS handshake error from 10.244.1.1:42994: EOF 2018/07/27 10:12:04 http: TLS handshake error from 10.244.1.1:43000: EOF level=info timestamp=2018-07-27T10:12:05.327879Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:12:07.750611Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url=/ proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-07-27T10:12:08.070920Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-07-27T10:12:11.069658Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/07/27 10:12:14 http: TLS handshake error from 10.244.1.1:43006: EOF level=info timestamp=2018-07-27T10:12:19.051170Z pos=filter.go:46 component=virt-api remoteAddress=10.244.0.0 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-api-bcc6b587d-k269w Pod phase: Running 2018/07/27 10:10:00 http: TLS handshake error from 10.244.0.1:54358: EOF 2018/07/27 10:10:10 http: TLS handshake error from 10.244.0.1:54418: EOF 2018/07/27 10:10:20 http: TLS handshake error from 10.244.0.1:54478: EOF 2018/07/27 10:10:30 http: TLS handshake error from 10.244.0.1:54538: EOF 2018/07/27 10:10:40 http: TLS handshake error from 10.244.0.1:54598: EOF 2018/07/27 10:10:50 http: TLS handshake error from 10.244.0.1:54658: EOF 2018/07/27 10:11:00 http: TLS handshake error from 10.244.0.1:54718: EOF 2018/07/27 10:11:10 http: TLS handshake error from 10.244.0.1:54778: EOF 2018/07/27 10:11:20 http: TLS handshake error from 10.244.0.1:54838: EOF 2018/07/27 10:11:30 http: TLS handshake error from 10.244.0.1:54898: EOF 2018/07/27 10:11:40 http: TLS handshake error from 10.244.0.1:54958: EOF 2018/07/27 10:11:50 http: TLS handshake error from 10.244.0.1:55018: EOF 2018/07/27 10:12:00 http: TLS handshake error from 10.244.0.1:55078: EOF 2018/07/27 10:12:10 http: TLS handshake error from 10.244.0.1:55138: EOF 2018/07/27 10:12:20 http: TLS handshake error from 10.244.0.1:55198: EOF Pod name: virt-controller-67dcdd8464-8pvbc Pod phase: Running level=info timestamp=2018-07-27T09:57:41.224693Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7fkpx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7fkpx" level=info timestamp=2018-07-27T10:00:41.613779Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit6n7m kind= uid=ebac8b8b-9183-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:00:41.613918Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit6n7m kind= uid=ebac8b8b-9183-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:00:41.702313Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmit6n7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmit6n7m" level=info timestamp=2018-07-27T10:01:31.739879Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:01:31.740627Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:02:22.010497Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiwt5js kind= uid=2783dadd-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:02:22.011873Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiwt5js kind= uid=2783dadd-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:05:22.464398Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqfcdj kind= uid=9312f188-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:05:22.465566Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqfcdj kind= uid=9312f188-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:05:22.511645Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmiqfcdj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmiqfcdj" level=info timestamp=2018-07-27T10:08:23.020799Z pos=preset.go:139 component=virt-controller service=http namespace=kubevirt-test-default name=testvmik5g94 kind= uid=feafab92-9184-11e8-9c71-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-07-27T10:08:23.020955Z pos=preset.go:165 component=virt-controller service=http namespace=kubevirt-test-default name=testvmik5g94 kind= uid=feafab92-9184-11e8-9c71-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-07-27T10:08:23.057173Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmik5g94\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmik5g94" level=info timestamp=2018-07-27T10:08:23.091491Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmik5g94\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmik5g94" Pod name: virt-controller-67dcdd8464-z6bgg Pod phase: Running level=info timestamp=2018-07-27T09:07:02.572685Z pos=application.go:174 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-jxj94 Pod phase: Running level=info timestamp=2018-07-27T09:47:19.763725Z pos=vm.go:389 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing vmi update" level=info timestamp=2018-07-27T09:47:19.769615Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:20.871997Z pos=vm.go:342 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Shutting down domain for VirtualMachineInstance with deletion timestamp." level=info timestamp=2018-07-27T09:47:20.872512Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T09:47:20.874086Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmind64hp4jxl" level=info timestamp=2018-07-27T09:47:21.092559Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.092698Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T09:47:21.093354Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind= uid=041a00b9-9182-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.093451Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.093501Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.096761Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T09:47:21.303941Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.304085Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T09:47:21.304241Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T09:47:21.305043Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmind64hp4jxl kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-handler-tsdfg Pod phase: Running level=info timestamp=2018-07-27T10:02:21.809400Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Processing shutdown." level=info timestamp=2018-07-27T10:02:21.809783Z pos=vm.go:540 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="grace period expired, killing deleted VirtualMachineInstance testvmipv546" level=info timestamp=2018-07-27T10:02:22.021493Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind= uid=098d309d-9184-11e8-9c71-525500d15501 msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.021599Z pos=vm.go:330 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Shutting down domain for deleted VirtualMachineInstance object." level=info timestamp=2018-07-27T10:02:22.021634Z pos=vm.go:383 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing shutdown." level=info timestamp=2018-07-27T10:02:22.021925Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.023564Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T10:02:22.023694Z pos=vm.go:678 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=Domain uid= msg="Domain deleted" level=info timestamp=2018-07-27T10:02:22.023775Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.023827Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.025780Z pos=server.go:75 component=virt-handler msg="Received Domain Event of type DELETED" level=info timestamp=2018-07-27T10:02:22.103992Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.104205Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." level=info timestamp=2018-07-27T10:02:22.104442Z pos=vm.go:386 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Processing local ephemeral data cleanup for shutdown domain." level=info timestamp=2018-07-27T10:02:22.104535Z pos=vm.go:413 component=virt-handler namespace=kubevirt-test-default name=testvmipv546 kind=VirtualMachineInstance uid= msg="Synchronization loop succeeded." Pod name: virt-launcher-testvmik5g94-b9xxt Pod phase: Pending • Failure [240.554 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:46 Starting a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:70 With VirtualMachineInstance with two PVCs /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:266 should start vmi multiple times [It] /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:278 Timed out after 120.005s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 ------------------------------ STEP: Starting and stopping the VirtualMachineInstance number of times STEP: Starting a VirtualMachineInstance STEP: Waiting until the VirtualMachineInstance will start level=info timestamp=2018-07-27T10:08:23.566997Z pos=utils.go:243 component=tests msg="Created virtual machine pod virt-launcher-testvmik5g94-b9xxt" • [SLOW TEST:35.367 seconds] LeaderElection /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:43 Start a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:53 when the controller pod is not running /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:54 should success /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:55 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.006 seconds] Templates /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:42 Launching VMI from VM Template [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:60 with given Fedora Template /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:193 should succeed to generate a VM JSON file using oc-process command /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:150 Skip test that requires oc binary /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1393 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] Templates /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:42 Launching VMI from VM Template [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:60 with given Fedora Template /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:193 with given VM JSON from the Template /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:152 should succeed to create a VM using oc-create command /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:156 Skip test that requires oc binary /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1393 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] Templates /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:42 Launching VMI from VM Template [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:60 with given Fedora Template /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:193 with given VM JSON from the Template /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:152 with given VM from the VM JSON /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:158 should succeed to launch a VMI using oc-patch command /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:161 Skip test that requires oc binary /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1393 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.001 seconds] Templates /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:42 Launching VMI from VM Template [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:60 with given Fedora Template /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:193 with given VM JSON from the Template /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:152 with given VM from the VM JSON /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:158 with given VMI from the VM /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:163 should succeed to terminate the VMI using oc-patch command /root/go/src/kubevirt.io/kubevirt/tests/template_test.go:166 Skip test that requires oc binary /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1393 ------------------------------ • [SLOW TEST:17.564 seconds] HookSidecars /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:40 VMI definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:58 with SM BIOS hook sidecar /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:59 should successfully start with hook sidecar annotation /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:60 ------------------------------ • [SLOW TEST:18.378 seconds] HookSidecars /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:40 VMI definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:58 with SM BIOS hook sidecar /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:59 should call Collect and OnDefineDomain on the hook sidecar /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:67 ------------------------------ • [SLOW TEST:20.677 seconds] HookSidecars /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:40 VMI definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:58 with SM BIOS hook sidecar /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:59 should update domain XML with SM BIOS properties /root/go/src/kubevirt.io/kubevirt/tests/vmi_hook_sidecar_test.go:83 ------------------------------ • ------------------------------ • [SLOW TEST:15.722 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 should start it /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:76 ------------------------------ • [SLOW TEST:18.320 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 should attach virt-launcher to it /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:82 ------------------------------ ••••2018/07/27 06:15:22 read closing down: EOF ------------------------------ • [SLOW TEST:51.689 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 with boot order /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:170 should be able to boot from selected disk /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 Alpine as first boot /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ 2018/07/27 06:15:49 read closing down: EOF • [SLOW TEST:27.280 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 with boot order /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:170 should be able to boot from selected disk /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 Cirros as first boot /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • [SLOW TEST:14.302 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:201 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:202 should retry starting the VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:203 ------------------------------ • [SLOW TEST:16.587 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:201 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:202 should log warning and proceed once the secret is there /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:233 ------------------------------ • [SLOW TEST:44.386 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 when virt-launcher crashes /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:281 should be stopped and have Failed phase /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:282 ------------------------------ • [SLOW TEST:22.407 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 when virt-handler crashes /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:304 should recover and continue management /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:305 ------------------------------ • [SLOW TEST:8.269 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 when virt-handler is responsive /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:335 should indicate that a node is ready for vmis /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:336 ------------------------------ • [SLOW TEST:77.454 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 when virt-handler is not responsive /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:366 the node controller should react /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:405 ------------------------------ • [SLOW TEST:15.409 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 with node tainted /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:458 the vmi with tolerations should be scheduled /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:480 ------------------------------ • ------------------------------ S [SKIPPING] [0.246 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:530 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-default [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Skip log query tests for JENKINS ci test environment /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:535 ------------------------------ S [SKIPPING] [0.057 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:530 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-alternative [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Skip log query tests for JENKINS ci test environment /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:535 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.065 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 VirtualMachineInstance Emulation Mode /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:591 should enable emulation in virt-launcher [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:603 Software emulation is not enabled on this cluster /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:599 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.062 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 VirtualMachineInstance Emulation Mode /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:591 should be reflected in domain XML [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:640 Software emulation is not enabled on this cluster /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:599 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.053 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:70 VirtualMachineInstance Emulation Mode /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:591 should request a TUN device but not KVM [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:684 Software emulation is not enabled on this cluster /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:599 ------------------------------ •••• ------------------------------ • [SLOW TEST:18.961 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Delete a VirtualMachineInstance's Pod /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:836 should result in the VirtualMachineInstance moving to a finalized state /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:837 ------------------------------ • [SLOW TEST:20.498 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Delete a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:868 with an active pod. /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:869 should result in pod being terminated /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:870 ------------------------------ • [SLOW TEST:21.332 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Delete a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:868 with grace period greater than 0 /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:893 should run graceful shutdown /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:894 ------------------------------ • [SLOW TEST:30.294 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Killed VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:945 should be in Failed phase /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:946 ------------------------------ • [SLOW TEST:24.103 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:48 Killed VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:945 should be left alone by virt-handler /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:973 ------------------------------ • [SLOW TEST:49.011 seconds] 2018/07/27 06:21:55 read closing down: EOF CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:46 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:80 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:81 should have cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:82 ------------------------------ • [SLOW TEST:162.238 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:46 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:80 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:81 with injected ssh-key 2018/07/27 06:24:37 read closing down: EOF /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:92 should have ssh-key under authorized keys /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:93 ------------------------------ 2018/07/27 06:25:24 read closing down: EOF 2018/07/27 06:25:34 read closing down: EOF • [SLOW TEST:56.786 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:46 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:80 with cloudInitNoCloud userData source /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:118 should process provided cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:119 ------------------------------ 2018/07/27 06:26:21 read closing down: EOF • [SLOW TEST:47.462 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:46 A new VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:80 should take user-data from k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmi_userdata_test.go:162 ------------------------------ Waiting for namespace kubevirt-test-default to be removed, this can take a while ... Waiting for namespace kubevirt-test-alternative to be removed, this can take a while ... Summarizing 7 Failures: [Fail] Storage Starting a VirtualMachineInstance with Alpine PVC should be successfully started [It] with Disk PVC /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 [Fail] Storage Starting a VirtualMachineInstance with Alpine PVC should be successfully started [It] with CDRom PVC /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 [Fail] Storage Starting a VirtualMachineInstance with Alpine PVC should be successfully started and stopped multiple times [It] with Disk PVC /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 [Fail] Storage Starting a VirtualMachineInstance with Alpine PVC should be successfully started and stopped multiple times [It] with CDRom PVC /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 [Fail] Storage Starting a VirtualMachineInstance With ephemeral alpine PVC [It] should be successfully started /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 [Fail] Storage Starting a VirtualMachineInstance With ephemeral alpine PVC [It] should not persist data /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 [Fail] Storage Starting a VirtualMachineInstance With VirtualMachineInstance with two PVCs [It] should start vmi multiple times /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1049 Ran 132 of 148 Specs in 4700.782 seconds FAIL! -- 125 Passed | 7 Failed | 0 Pending | 16 Skipped --- FAIL: TestTests (4700.81s) FAIL make: *** [functest] Error 1 + make cluster-down ./cluster/down.sh