+ export WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release + WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release + [[ openshift-3.10-release =~ openshift-.* ]] + [[ openshift-3.10-release =~ .*-crio-.* ]] + export KUBEVIRT_PROVIDER=os-3.10.0 + KUBEVIRT_PROVIDER=os-3.10.0 + export KUBEVIRT_NUM_NODES=2 + KUBEVIRT_NUM_NODES=2 + export NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + export NAMESPACE=kube-system + NAMESPACE=kube-system + trap '{ make cluster-down; }' EXIT SIGINT SIGTERM SIGSTOP + make cluster-down ./cluster/down.sh + make cluster-up ./cluster/up.sh Downloading ....... Downloading ....... 2018/08/03 20:50:47 Waiting for host: 192.168.66.102:22 2018/08/03 20:50:50 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/08/03 20:50:58 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/08/03 20:51:06 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/08/03 20:51:14 Connected to tcp://192.168.66.102:22 + systemctl stop origin-node.service + rm -rf /etc/origin/ /etc/etcd/ /var/lib/origin /var/lib/etcd/ ++ docker ps -q + containers= + '[' -n '' ']' ++ docker ps -q -a + containers='2cfbef31c987 e183c40c07dc 861f604efed4 12902ad26342 028539b1f68b bd6f07c1906c d1f95a33a226 c43f96b6da26 e007e5cfd226 b42e2bceca6e 00531aec6f9a e4ad39ba6cef 504c3df6bbf4 eb1ec0b445ce b8955b91e8e5 f739ed8f3e59 07668d85ab3a a6045d125d7b 2ce17110e009 b45f64ab28ef 3a15945be9e1 2a0af99ae1d1 0ece927846d7 0202d5f5dfae 8ce743769d8f 2efb36567bd8 96b65c0493c5 e9ce89fa30e3' + '[' -n '2cfbef31c987 e183c40c07dc 861f604efed4 12902ad26342 028539b1f68b bd6f07c1906c d1f95a33a226 c43f96b6da26 e007e5cfd226 b42e2bceca6e 00531aec6f9a e4ad39ba6cef 504c3df6bbf4 eb1ec0b445ce b8955b91e8e5 f739ed8f3e59 07668d85ab3a a6045d125d7b 2ce17110e009 b45f64ab28ef 3a15945be9e1 2a0af99ae1d1 0ece927846d7 0202d5f5dfae 8ce743769d8f 2efb36567bd8 96b65c0493c5 e9ce89fa30e3' ']' + docker rm -f 2cfbef31c987 e183c40c07dc 861f604efed4 12902ad26342 028539b1f68b bd6f07c1906c d1f95a33a226 c43f96b6da26 e007e5cfd226 b42e2bceca6e 00531aec6f9a e4ad39ba6cef 504c3df6bbf4 eb1ec0b445ce b8955b91e8e5 f739ed8f3e59 07668d85ab3a a6045d125d7b 2ce17110e009 b45f64ab28ef 3a15945be9e1 2a0af99ae1d1 0ece927846d7 0202d5f5dfae 8ce743769d8f 2efb36567bd8 96b65c0493c5 e9ce89fa30e3 2cfbef31c987 e183c40c07dc 861f604efed4 12902ad26342 028539b1f68b bd6f07c1906c d1f95a33a226 c43f96b6da26 e007e5cfd226 b42e2bceca6e 00531aec6f9a e4ad39ba6cef 504c3df6bbf4 eb1ec0b445ce b8955b91e8e5 f739ed8f3e59 07668d85ab3a a6045d125d7b 2ce17110e009 b45f64ab28ef 3a15945be9e1 2a0af99ae1d1 0ece927846d7 0202d5f5dfae 8ce743769d8f 2efb36567bd8 96b65c0493c5 e9ce89fa30e3 2018/08/03 20:51:19 Waiting for host: 192.168.66.101:22 2018/08/03 20:51:22 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/08/03 20:51:30 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/08/03 20:51:38 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/08/03 20:51:43 Connected to tcp://192.168.66.101:22 + inventory_file=/root/inventory + openshift_ansible=/root/openshift-ansible + echo '[new_nodes]' + sed -i '/\[OSEv3:children\]/a new_nodes' /root/inventory + nodes_found=false ++ seq 2 100 + for i in '$(seq 2 100)' ++ printf node%02d 2 + node=node02 ++ printf 192.168.66.1%02d 2 + node_ip=192.168.66.102 + set +e + ping 192.168.66.102 -c 1 PING 192.168.66.102 (192.168.66.102) 56(84) bytes of data. 64 bytes from 192.168.66.102: icmp_seq=1 ttl=64 time=2.26 ms --- 192.168.66.102 ping statistics --- 1 packets transmitted, 1 received, 0% packet loss, time 0ms rtt min/avg/max/mdev = 2.262/2.262/2.262/0.000 ms Found node02. Adding it to the inventory. + '[' 0 -ne 0 ']' + nodes_found=true + set -e + echo '192.168.66.102 node02' + echo 'Found node02. Adding it to the inventory.' + echo 'node02 openshift_node_group_name="node-config-compute" openshift_schedulable=true openshift_ip=192.168.66.102' + for i in '$(seq 2 100)' ++ printf node%02d 3 + node=node03 ++ printf 192.168.66.1%02d 3 + node_ip=192.168.66.103 + set +e + ping 192.168.66.103 -c 1 PING 192.168.66.103 (192.168.66.103) 56(84) bytes of data. From 192.168.66.101 icmp_seq=1 Destination Host Unreachable --- 192.168.66.103 ping statistics --- 1 packets transmitted, 0 received, +1 errors, 100% packet loss, time 0ms + '[' 1 -ne 0 ']' + break + '[' true = true ']' + ansible-playbook -i /root/inventory /root/openshift-ansible/playbooks/openshift-node/scaleup.yml PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Evaluate oo_nodes_to_bootstrap] ****************************************** ok: [localhost] => (item=node02) TASK [Add masters to oo_nodes_to_bootstrap] ************************************ ok: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) PLAY [Ensure there are new_nodes] ********************************************** TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] PLAY [Initialization Checkpoint Start] ***************************************** TASK [Set install initialization 'In Progress'] ******************************** ok: [node01] PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Evaluate oo_nodes_to_bootstrap] ****************************************** ok: [localhost] => (item=node02) TASK [Add masters to oo_nodes_to_bootstrap] ************************************ ok: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) [WARNING]: Could not match supplied host pattern, ignoring: oo_lb_to_config PLAY [Ensure that all non-node hosts are accessible] *************************** TASK [Gathering Facts] ********************************************************* ok: [node01] PLAY [Initialize basic host facts] ********************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/deprecations.yml for node01, node02 TASK [openshift_sanitize_inventory : Check for usage of deprecated variables] *** ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : debug] ************************************ skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_stats] ******************************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Assign deprecated variables to correct counterparts] *** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_logging.yml for node01, node02 included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_metrics.yml for node01, node02 TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : Standardize on latest variable names] ***** ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : Normalize openshift_release] ************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Abort when openshift_release is invalid] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/unsupported.yml for node01, node02 TASK [openshift_sanitize_inventory : Ensure that openshift_use_dnsmasq is true] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that openshift_node_dnsmasq_install_network_manager_hook is true] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* skipping: [node01] => (item=openshift_hosted_etcd_storage_kind) skipping: [node02] => (item=openshift_hosted_etcd_storage_kind) TASK [openshift_sanitize_inventory : Ensure that dynamic provisioning is set if using dynamic storage] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure clusterid is set along with the cloudprovider] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure ansible_service_broker_remove and ansible_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure template_service_broker_remove and template_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that all requires vsphere configuration variables are set] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : ensure provider configuration variables are defined] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure removed web console extension variables are not set] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that web console port matches API server port] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : At least one master is schedulable] ******* skipping: [node01] skipping: [node02] TASK [Detecting Operating System from ostree_booted] *************************** ok: [node02] ok: [node01] TASK [set openshift_deployment_type if unset] ********************************** skipping: [node01] skipping: [node02] TASK [check for node already bootstrapped] ************************************* ok: [node02] ok: [node01] TASK [initialize_facts set fact openshift_is_bootstrapped] ********************* ok: [node01] ok: [node02] TASK [initialize_facts set fact openshift_is_atomic and openshift_is_containerized] *** ok: [node02] ok: [node01] TASK [Determine Atomic Host Docker Version] ************************************ skipping: [node01] skipping: [node02] TASK [assert atomic host docker version is 1.12 or later] ********************** skipping: [node01] skipping: [node02] PLAY [Retrieve existing master configs and validate] *************************** TASK [openshift_control_plane : stat] ****************************************** ok: [node01] TASK [openshift_control_plane : slurp] ***************************************** ok: [node01] TASK [openshift_control_plane : set_fact] ************************************** ok: [node01] TASK [openshift_control_plane : Check for file paths outside of /etc/origin/master in master's config] *** ok: [node01] TASK [openshift_control_plane : set_fact] ************************************** ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] TASK [set_fact] **************************************************************** skipping: [node01] PLAY [Initialize special first-master variables] ******************************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] PLAY [Disable web console if required] ***************************************** TASK [set_fact] **************************************************************** skipping: [node01] PLAY [Setup yum repositories for all hosts] ************************************ TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Install Red Hat Subscription manager] ******************* skipping: [node02] TASK [rhel_subscribe : Is host already registered?] **************************** skipping: [node02] TASK [rhel_subscribe : Register host] ****************************************** skipping: [node02] TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Determine if OpenShift Pool Already Attached] *********** skipping: [node02] TASK [rhel_subscribe : Attach to OpenShift Pool] ******************************* skipping: [node02] TASK [rhel_subscribe : Satellite preparation] ********************************** skipping: [node02] TASK [openshift_repos : openshift_repos detect ostree] ************************* ok: [node02] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** ok: [node02] TASK [openshift_repos : Remove openshift_additional.repo file] ***************** ok: [node02] TASK [openshift_repos : Create any additional repos that are defined] ********** TASK [openshift_repos : include_tasks] ***************************************** skipping: [node02] TASK [openshift_repos : include_tasks] ***************************************** included: /root/openshift-ansible/roles/openshift_repos/tasks/centos_repos.yml for node02 TASK [openshift_repos : Configure origin gpg keys] ***************************** ok: [node02] TASK [openshift_repos : Configure correct origin release repository] *********** ok: [node02] => (item=/root/openshift-ansible/roles/openshift_repos/templates/CentOS-OpenShift-Origin.repo.j2) TASK [openshift_repos : Ensure clean repo cache in the event repos have been changed manually] *** changed: [node02] => { "msg": "First run of openshift_repos" } TASK [openshift_repos : Record that openshift_repos already ran] *************** ok: [node02] RUNNING HANDLER [openshift_repos : refresh cache] ****************************** changed: [node02] PLAY [Install packages necessary for installer] ******************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Determine if chrony is installed] **************************************** changed: [node02] [WARNING]: Consider using the yum, dnf or zypper module rather than running rpm. If you need to use command because yum, dnf or zypper is insufficient you can add warn=False to this command task or set command_warnings=False in ansible.cfg to get rid of this message. TASK [Install ntp package] ***************************************************** skipping: [node02] TASK [Start and enable ntpd/chronyd] ******************************************* changed: [node02] TASK [Ensure openshift-ansible installer package deps are installed] *********** ok: [node02] => (item=iproute) ok: [node02] => (item=dbus-python) ok: [node02] => (item=PyYAML) ok: [node02] => (item=python-ipaddress) ok: [node02] => (item=libsemanage-python) ok: [node02] => (item=yum-utils) ok: [node02] => (item=python-docker) PLAY [Initialize cluster facts] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [get openshift_current_version] ******************************************* ok: [node02] ok: [node01] TASK [set_fact openshift_portal_net if present on masters] ********************* ok: [node02] ok: [node01] TASK [Gather Cluster facts] **************************************************** changed: [node02] changed: [node01] TASK [Set fact of no_proxy_internal_hostnames] ********************************* skipping: [node01] skipping: [node02] TASK [Initialize openshift.node.sdn_mtu] *************************************** changed: [node02] ok: [node01] PLAY [Initialize etcd host variables] ****************************************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [include_role : openshift_version] **************************************** TASK [openshift_version : Use openshift_current_version fact as version to configure if already installed] *** ok: [node01] TASK [openshift_version : Set openshift_version to openshift_release if undefined] *** skipping: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] => { "msg": "openshift_pkg_version was not defined. Falling back to -3.10.0" } TASK [openshift_version : set_fact] ******************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : assert openshift_release in openshift_image_tag] ***** ok: [node01] => { "changed": false, "msg": "All assertions passed" } TASK [openshift_version : assert openshift_release in openshift_pkg_version] *** ok: [node01] => { "changed": false, "msg": "All assertions passed" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_release": "3.10" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_image_tag": "v3.10.0-rc.0" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_pkg_version": "-3.10.0*" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_version": "3.10.0" } TASK [set openshift_version booleans (first master)] *************************** ok: [node01] PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [set_fact] **************************************************************** ok: [node02] TASK [set openshift_version booleans (masters and nodes)] ********************** ok: [node02] PLAY [Verify Requirements] ***************************************************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [Run variable sanity checks] ********************************************** ok: [node01] TASK [Validate openshift_node_groups and openshift_node_group_name] ************ ok: [node01] PLAY [Initialization Checkpoint End] ******************************************* TASK [Set install initialization 'Complete'] *********************************** ok: [node01] PLAY [Validate node hostnames] ************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Query DNS for IP address of node02] ************************************** ok: [node02] TASK [Validate openshift_hostname when defined] ******************************** skipping: [node02] TASK [Validate openshift_ip exists on node when defined] *********************** skipping: [node02] PLAY [Configure os_firewall] *************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [os_firewall : Detecting Atomic Host Operating System] ******************** ok: [node02] TASK [os_firewall : Set fact r_os_firewall_is_atomic] ************************** ok: [node02] TASK [os_firewall : Fail - Firewalld is not supported on Atomic Host] ********** skipping: [node02] TASK [os_firewall : Install firewalld packages] ******************************** skipping: [node02] TASK [os_firewall : Ensure iptables services are not enabled] ****************** skipping: [node02] => (item=iptables) skipping: [node02] => (item=ip6tables) TASK [os_firewall : Wait 10 seconds after disabling iptables] ****************** skipping: [node02] TASK [os_firewall : Start and enable firewalld service] ************************ skipping: [node02] TASK [os_firewall : need to pause here, otherwise the firewalld service starting can sometimes cause ssh to fail] *** skipping: [node02] TASK [os_firewall : Restart polkitd] ******************************************* skipping: [node02] TASK [os_firewall : Wait for polkit action to have been created] *************** skipping: [node02] TASK [os_firewall : Ensure firewalld service is not enabled] ******************* ok: [node02] TASK [os_firewall : Wait 10 seconds after disabling firewalld] ***************** skipping: [node02] TASK [os_firewall : Install iptables packages] ********************************* ok: [node02] => (item=iptables) ok: [node02] => (item=iptables-services) TASK [os_firewall : Start and enable iptables service] ************************* ok: [node02 -> node02] => (item=node02) TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [node02] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [container_runtime : Setup the docker-storage for overlay] **************** skipping: [node02] TASK [container_runtime : Create file system on extra volume device] *********** TASK [container_runtime : Create mount entry for extra volume] ***************** PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_excluder : Install docker excluder - yum] ********************** ok: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* ok: [node02] TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] TASK [container_runtime : Getting current systemd-udevd exec command] ********** skipping: [node02] TASK [container_runtime : Assure systemd-udevd.service.d directory exists] ***** skipping: [node02] TASK [container_runtime : Create systemd-udevd override file] ****************** skipping: [node02] TASK [container_runtime : Add enterprise registry, if necessary] *************** skipping: [node02] TASK [container_runtime : Add http_proxy to /etc/atomic.conf] ****************** skipping: [node02] TASK [container_runtime : Add https_proxy to /etc/atomic.conf] ***************** skipping: [node02] TASK [container_runtime : Add no_proxy to /etc/atomic.conf] ******************** skipping: [node02] TASK [container_runtime : Get current installed Docker version] **************** ok: [node02] TASK [container_runtime : Error out if Docker pre-installed but too old] ******* skipping: [node02] TASK [container_runtime : Error out if requested Docker is too old] ************ skipping: [node02] TASK [container_runtime : Install Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure docker.service.d directory exists] ************ ok: [node02] TASK [container_runtime : Configure Docker service unit file] ****************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Set registry params] ********************************* skipping: [node02] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': []}) TASK [container_runtime : Place additional/blocked/insecure registries in /etc/containers/registries.conf] *** skipping: [node02] TASK [container_runtime : Set Proxy Settings] ********************************** skipping: [node02] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [container_runtime : Set various Docker options] ************************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Configure Docker Network OPTIONS] ******************** ok: [node02] TASK [container_runtime : Detect if docker is already started] ***************** ok: [node02] TASK [container_runtime : Start the Docker service] **************************** ok: [node02] TASK [container_runtime : set_fact] ******************************************** ok: [node02] TASK [container_runtime : Check for docker_storage_path/overlay2] ************** ok: [node02] TASK [container_runtime : Fixup SELinux permissions for docker] **************** changed: [node02] TASK [container_runtime : Ensure /var/lib/containers exists] ******************* ok: [node02] TASK [container_runtime : Fix SELinux Permissions on /var/lib/containers] ****** ok: [node02] TASK [container_runtime : Check for credentials file for registry auth] ******** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth] ***** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth (alternative)] *** skipping: [node02] TASK [container_runtime : stat the docker data dir] **************************** ok: [node02] TASK [container_runtime : stop the current running docker] ********************* skipping: [node02] TASK [container_runtime : copy "/var/lib/docker" to "/var/lib/containers/docker"] *** skipping: [node02] TASK [container_runtime : Set the selinux context on /var/lib/containers/docker] *** skipping: [node02] TASK [container_runtime : restorecon the /var/lib/containers/docker] *********** skipping: [node02] TASK [container_runtime : Remove the old docker location] ********************** skipping: [node02] TASK [container_runtime : Setup the link] ************************************** skipping: [node02] TASK [container_runtime : start docker] **************************************** skipping: [node02] TASK [container_runtime : Fail if Atomic Host since this is an rpm request] **** skipping: [node02] TASK [container_runtime : Getting current systemd-udevd exec command] ********** skipping: [node02] TASK [container_runtime : Assure systemd-udevd.service.d directory exists] ***** skipping: [node02] TASK [container_runtime : Create systemd-udevd override file] ****************** skipping: [node02] TASK [container_runtime : Add enterprise registry, if necessary] *************** skipping: [node02] TASK [container_runtime : Check that overlay is in the kernel] ***************** skipping: [node02] TASK [container_runtime : Add overlay to modprobe.d] *************************** skipping: [node02] TASK [container_runtime : Manually modprobe overlay into the kernel] *********** skipping: [node02] TASK [container_runtime : Enable and start systemd-modules-load] *************** skipping: [node02] TASK [container_runtime : Install cri-o] *************************************** skipping: [node02] TASK [container_runtime : Remove CRI-O default configuration files] ************ skipping: [node02] => (item=/etc/cni/net.d/200-loopback.conf) skipping: [node02] => (item=/etc/cni/net.d/100-crio-bridge.conf) TASK [container_runtime : Create the CRI-O configuration] ********************** skipping: [node02] TASK [container_runtime : Ensure CNI configuration directory exists] *********** skipping: [node02] TASK [container_runtime : Add iptables allow rules] **************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove iptables rules] ******************************* TASK [container_runtime : Add firewalld allow rules] *************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove firewalld allow rules] ************************ TASK [container_runtime : Configure the CNI network] *************************** skipping: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-network] ****************** skipping: [node02] TASK [container_runtime : Start the CRI-O service] ***************************** skipping: [node02] TASK [container_runtime : Ensure /var/lib/containers exists] ******************* skipping: [node02] TASK [container_runtime : Fix SELinux Permissions on /var/lib/containers] ****** skipping: [node02] TASK [container_runtime : Check for credentials file for registry auth] ******** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth] ***** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth (alternative)] *** skipping: [node02] TASK [container_runtime : stat the docker data dir] **************************** skipping: [node02] TASK [container_runtime : stop the current running docker] ********************* skipping: [node02] TASK [container_runtime : copy "/var/lib/docker" to "/var/lib/containers/docker"] *** skipping: [node02] TASK [container_runtime : Set the selinux context on /var/lib/containers/docker] *** skipping: [node02] TASK [container_runtime : restorecon the /var/lib/containers/docker] *********** skipping: [node02] TASK [container_runtime : Remove the old docker location] ********************** skipping: [node02] TASK [container_runtime : Setup the link] ************************************** skipping: [node02] TASK [container_runtime : start docker] **************************************** skipping: [node02] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [include_role : openshift_version] **************************************** TASK [openshift_version : Use openshift_current_version fact as version to configure if already installed] *** skipping: [node01] TASK [openshift_version : Set openshift_version to openshift_release if undefined] *** skipping: [node01] TASK [openshift_version : debug] *********************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : debug] *********************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : assert openshift_release in openshift_image_tag] ***** ok: [node01] => { "changed": false, "msg": "All assertions passed" } TASK [openshift_version : assert openshift_release in openshift_pkg_version] *** ok: [node01] => { "changed": false, "msg": "All assertions passed" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_release": "3.10" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_image_tag": "v3.10.0-rc.0" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_pkg_version": "-3.10.0*" } TASK [openshift_version : debug] *********************************************** ok: [node01] => { "openshift_version": "3.10.0" } TASK [set openshift_version booleans (first master)] *************************** ok: [node01] PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [set_fact] **************************************************************** ok: [node02] TASK [set openshift_version booleans (masters and nodes)] ********************** ok: [node02] PLAY [Node Preparation Checkpoint Start] *************************************** TASK [Set Node preparation 'In Progress'] ************************************** ok: [node01] PLAY [Only target nodes that have not yet been bootstrapped] ******************* TASK [Gathering Facts] ********************************************************* ok: [localhost] TASK [add_host] **************************************************************** skipping: [localhost] => (item=node02) ok: [localhost] => (item=node01) PLAY [Disable excluders] ******************************************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": true } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": true } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/disable.yml for node02 TASK [openshift_excluder : Get available excluder version] ********************* skipping: [node02] TASK [openshift_excluder : Fail when excluder package is not found] ************ skipping: [node02] TASK [openshift_excluder : Set fact excluder_version] ************************** skipping: [node02] TASK [openshift_excluder : origin-docker-excluder version detected] ************ skipping: [node02] TASK [openshift_excluder : Printing upgrade target version] ******************** skipping: [node02] TASK [openshift_excluder : Check the available origin-docker-excluder version is at most of the upgrade target version] *** skipping: [node02] TASK [openshift_excluder : Get available excluder version] ********************* skipping: [node02] TASK [openshift_excluder : Fail when excluder package is not found] ************ skipping: [node02] TASK [openshift_excluder : Set fact excluder_version] ************************** skipping: [node02] TASK [openshift_excluder : origin-excluder version detected] ******************* skipping: [node02] TASK [openshift_excluder : Printing upgrade target version] ******************** skipping: [node02] TASK [openshift_excluder : Check the available origin-excluder version is at most of the upgrade target version] *** skipping: [node02] TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* changed: [node02] TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** changed: [node02] TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* changed: [node02] PLAY [Configure nodes] ********************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_cloud_provider : Set cloud provider facts] ********************* skipping: [node02] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [node02] TASK [openshift_cloud_provider : include the defined cloud provider files] ***** skipping: [node02] TASK [openshift_node : fail] *************************************************** skipping: [node02] TASK [openshift_node : Check for NetworkManager service] *********************** ok: [node02] TASK [openshift_node : Set fact using_network_manager] ************************* ok: [node02] TASK [openshift_node : Install dnsmasq] **************************************** ok: [node02] TASK [openshift_node : ensure origin/node directory exists] ******************** changed: [node02] => (item=/etc/origin) changed: [node02] => (item=/etc/origin/node) TASK [openshift_node : Install NetworkManager during node_bootstrap provisioning] *** skipping: [node02] TASK [openshift_node : Install network manager dispatch script] **************** skipping: [node02] TASK [openshift_node : Install dnsmasq configuration] ************************** ok: [node02] TASK [openshift_node : Deploy additional dnsmasq.conf] ************************* skipping: [node02] TASK [openshift_node : Enable dnsmasq] ***************************************** ok: [node02] TASK [openshift_node : Install network manager dispatch script] **************** ok: [node02] TASK [openshift_node : Add iptables allow rules] ******************************* ok: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) ok: [node02] => (item={u'port': u'10256/tcp', u'service': u'Kubernetes kube-proxy health check for service load balancers'}) ok: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) ok: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) ok: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove iptables rules] ********************************** TASK [openshift_node : Add firewalld allow rules] ****************************** skipping: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) skipping: [node02] => (item={u'port': u'10256/tcp', u'service': u'Kubernetes kube-proxy health check for service load balancers'}) skipping: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) skipping: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove firewalld allow rules] *************************** TASK [openshift_node : Checking for journald.conf] ***************************** ok: [node02] TASK [openshift_node : Create journald persistence directories] **************** ok: [node02] TASK [openshift_node : Update journald setup] ********************************** ok: [node02] => (item={u'var': u'Storage', u'val': u'persistent'}) ok: [node02] => (item={u'var': u'Compress', u'val': True}) ok: [node02] => (item={u'var': u'SyncIntervalSec', u'val': u'1s'}) ok: [node02] => (item={u'var': u'RateLimitInterval', u'val': u'1s'}) ok: [node02] => (item={u'var': u'RateLimitBurst', u'val': 10000}) ok: [node02] => (item={u'var': u'SystemMaxUse', u'val': u'8G'}) ok: [node02] => (item={u'var': u'SystemKeepFree', u'val': u'20%'}) ok: [node02] => (item={u'var': u'SystemMaxFileSize', u'val': u'10M'}) ok: [node02] => (item={u'var': u'MaxRetentionSec', u'val': u'1month'}) ok: [node02] => (item={u'var': u'MaxFileSec', u'val': u'1day'}) ok: [node02] => (item={u'var': u'ForwardToSyslog', u'val': False}) ok: [node02] => (item={u'var': u'ForwardToWall', u'val': False}) TASK [openshift_node : Restart journald] *************************************** skipping: [node02] TASK [openshift_node : Disable swap] ******************************************* ok: [node02] TASK [openshift_node : Install node, clients, and conntrack packages] ********** ok: [node02] => (item={u'name': u'origin-node-3.10.0*'}) ok: [node02] => (item={u'name': u'origin-clients-3.10.0*'}) ok: [node02] => (item={u'name': u'conntrack-tools'}) TASK [openshift_node : Restart cri-o] ****************************************** skipping: [node02] TASK [openshift_node : restart NetworkManager to ensure resolv.conf is present] *** changed: [node02] TASK [openshift_node : sysctl] ************************************************* ok: [node02] TASK [openshift_node : Check for credentials file for registry auth] *********** skipping: [node02] TASK [openshift_node : Create credentials for registry auth] ******************* skipping: [node02] TASK [openshift_node : Create credentials for registry auth (alternative)] ***** skipping: [node02] TASK [openshift_node : Setup ro mount of /root/.docker for containerized hosts] *** skipping: [node02] TASK [openshift_node : Check that node image is present] *********************** changed: [node02] TASK [openshift_node : Pre-pull node image] ************************************ skipping: [node02] TASK [openshift_node : Copy node script to the node] *************************** ok: [node02] TASK [openshift_node : Install Node service file] ****************************** ok: [node02] TASK [openshift_node : Ensure old system path is set] ************************** skipping: [node02] => (item=/etc/origin/openvswitch) skipping: [node02] => (item=/var/lib/kubelet) skipping: [node02] => (item=/opt/cni/bin) TASK [openshift_node : Check status of node image pre-pull] ******************** skipping: [node02] TASK [openshift_node : Copy node container image to ostree storage] ************ skipping: [node02] TASK [openshift_node : Install or Update node system container] **************** skipping: [node02] TASK [openshift_node : Restart network manager to ensure networking configuration is in place] *** skipping: [node02] TASK [openshift_node : Configure Node settings] ******************************** ok: [node02] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS='}) ok: [node02] => (item={u'regex': u'^DEBUG_LOGLEVEL=', u'line': u'DEBUG_LOGLEVEL=2'}) ok: [node02] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.10.0-rc.0'}) TASK [openshift_node : Configure Proxy Settings] ******************************* skipping: [node02] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) skipping: [node02] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [node02] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=[],172.30.0.0/16,10.128.0.0/14'}) TASK [openshift_node : file] *************************************************** skipping: [node02] TASK [openshift_node : Create the Node config] ********************************* changed: [node02] TASK [openshift_node : Configure Node Environment Variables] ******************* TASK [openshift_node : Ensure the node static pod directory exists] ************ changed: [node02] TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [node02] => (item=None) skipping: [node02] => (item=None) skipping: [node02] TASK [openshift_node : Check status of node image pre-pull] ******************** skipping: [node02] TASK [openshift_node : Install NFS storage plugin dependencies] **************** ok: [node02] TASK [openshift_node : Check for existence of nfs sebooleans] ****************** ok: [node02] => (item=virt_use_nfs) ok: [node02] => (item=virt_sandbox_use_nfs) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:18.128236', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.019646', '_ansible_item_label': u'virt_use_nfs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:18.108590', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:19.518541', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.010737', '_ansible_item_label': u'virt_sandbox_use_nfs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:19.507804', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:18.128236', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.019646', '_ansible_item_label': u'virt_use_nfs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:18.108590', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:19.518541', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.010737', '_ansible_item_label': u'virt_sandbox_use_nfs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:19.507804', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Install GlusterFS storage plugin dependencies] ********** ok: [node02] TASK [openshift_node : Check for existence of fusefs sebooleans] *************** ok: [node02] => (item=virt_use_fusefs) ok: [node02] => (item=virt_sandbox_use_fusefs) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:26.714447', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.011903', '_ansible_item_label': u'virt_use_fusefs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:26.702544', '_ansible_ignore_errors': None, 'failed': False}) ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:28.042624', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.013089', '_ansible_item_label': u'virt_sandbox_use_fusefs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:28.029535', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:26.714447', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.011903', '_ansible_item_label': u'virt_use_fusefs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:26.702544', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-08-03 21:00:28.042624', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.013089', '_ansible_item_label': u'virt_sandbox_use_fusefs', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'warn': True, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'argv': None, u'creates': None, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-08-03 21:00:28.029535', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Install Ceph storage plugin dependencies] *************** ok: [node02] TASK [openshift_node : Install iSCSI storage plugin dependencies] ************** ok: [node02] => (item=iscsi-initiator-utils) ok: [node02] => (item=device-mapper-multipath) TASK [openshift_node : restart services] *************************************** ok: [node02] => (item=multipathd) ok: [node02] => (item=rpcbind) ok: [node02] => (item=iscsid) TASK [openshift_node : Template multipath configuration] *********************** changed: [node02] TASK [openshift_node : Enable and start multipath] ***************************** changed: [node02] TASK [tuned : Check for tuned package] ***************************************** ok: [node02] TASK [tuned : Set tuned OpenShift variables] *********************************** ok: [node02] TASK [tuned : Ensure directory structure exists] ******************************* ok: [node02] => (item={'serole': 'object_r', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'setype': 'admin_home_t', 'state': 'directory', 'gid': 0, 'mode': '0755', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'ctime': 1531032437.8490183}) ok: [node02] => (item={'serole': 'object_r', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'setype': 'admin_home_t', 'state': 'directory', 'gid': 0, 'mode': '0755', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'ctime': 1531032437.8490183}) ok: [node02] => (item={'serole': 'object_r', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'setype': 'admin_home_t', 'state': 'directory', 'gid': 0, 'mode': '0755', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'ctime': 1531032437.8490183}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'recommend.conf', 'size': 290, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 594, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) TASK [tuned : Ensure files are populated from templates] *********************** skipping: [node02] => (item={'serole': 'object_r', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'setype': 'admin_home_t', 'state': 'directory', 'gid': 0, 'mode': '0755', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'ctime': 1531032437.8490183}) skipping: [node02] => (item={'serole': 'object_r', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'setype': 'admin_home_t', 'state': 'directory', 'gid': 0, 'mode': '0755', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'ctime': 1531032437.8490183}) skipping: [node02] => (item={'serole': 'object_r', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'setype': 'admin_home_t', 'state': 'directory', 'gid': 0, 'mode': '0755', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'ctime': 1531032437.8490183}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'recommend.conf', 'size': 290, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1531032437.8490183, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1531032437.8490183, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 594, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) TASK [tuned : Make tuned use the recommended tuned profile on restart] ********* changed: [node02] => (item=/etc/tuned/active_profile) changed: [node02] => (item=/etc/tuned/profile_mode) TASK [tuned : Restart tuned service] ******************************************* changed: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Install logrotate] ******* ok: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Setup logrotate.d scripts] *** PLAY [node bootstrap config] *************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_node : install needed rpm(s)] ********************************** ok: [node02] => (item=origin-node) ok: [node02] => (item=origin-docker-excluder) ok: [node02] => (item=ansible) ok: [node02] => (item=bash-completion) ok: [node02] => (item=docker) ok: [node02] => (item=haproxy) ok: [node02] => (item=dnsmasq) ok: [node02] => (item=ntp) ok: [node02] => (item=logrotate) ok: [node02] => (item=httpd-tools) ok: [node02] => (item=bind-utils) ok: [node02] => (item=firewalld) ok: [node02] => (item=libselinux-python) ok: [node02] => (item=conntrack-tools) ok: [node02] => (item=openssl) ok: [node02] => (item=iproute) ok: [node02] => (item=python-dbus) ok: [node02] => (item=PyYAML) ok: [node02] => (item=yum-utils) ok: [node02] => (item=glusterfs-fuse) ok: [node02] => (item=device-mapper-multipath) ok: [node02] => (item=nfs-utils) ok: [node02] => (item=cockpit-ws) ok: [node02] => (item=cockpit-system) ok: [node02] => (item=cockpit-bridge) ok: [node02] => (item=cockpit-docker) ok: [node02] => (item=iscsi-initiator-utils) ok: [node02] => (item=ceph-common) TASK [openshift_node : create the directory for node] ************************** skipping: [node02] TASK [openshift_node : laydown systemd override] ******************************* skipping: [node02] TASK [openshift_node : update the sysconfig to have necessary variables] ******* ok: [node02] => (item={u'regexp': u'^KUBECONFIG=.*', u'line': u'KUBECONFIG=/etc/origin/node/bootstrap.kubeconfig'}) TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [node02] => (item=None) skipping: [node02] => (item=None) skipping: [node02] TASK [openshift_node : disable origin-node service] **************************** changed: [node02] => (item=origin-node.service) TASK [openshift_node : Check for RPM generated config marker file .config_managed] *** ok: [node02] TASK [openshift_node : create directories for bootstrapping] ******************* ok: [node02] => (item=/root/openshift_bootstrap) changed: [node02] => (item=/var/lib/origin/openshift.local.config) changed: [node02] => (item=/var/lib/origin/openshift.local.config/node) ok: [node02] => (item=/etc/docker/certs.d/docker-registry.default.svc:5000) TASK [openshift_node : laydown the bootstrap.yml file for on boot configuration] *** ok: [node02] TASK [openshift_node : Create a symlink to the node client CA for the docker registry] *** ok: [node02] TASK [openshift_node : Remove RPM generated config files if present] *********** skipping: [node02] => (item=master) skipping: [node02] => (item=.config_managed) TASK [openshift_node : find all files in /etc/origin/node so we can remove them] *** skipping: [node02] TASK [openshift_node : Remove everything except the resolv.conf required for node] *** skipping: [node02] TASK [openshift_node_group : create node config template] ********************** changed: [node02] TASK [openshift_node_group : remove existing node config] ********************** changed: [node02] TASK [openshift_node_group : Ensure required directories are present] ********** ok: [node02] => (item=/etc/origin/node/pods) changed: [node02] => (item=/etc/origin/node/certificates) TASK [openshift_node_group : Update the sysconfig to group "node-config-compute"] *** changed: [node02] TASK [set_fact] **************************************************************** ok: [node02] PLAY [Re-enable excluder if it was previously enabled] ************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": true } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": true } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/enable.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** changed: [node02] PLAY [Node Preparation Checkpoint End] ***************************************** TASK [Set Node preparation 'Complete'] ***************************************** ok: [node01] PLAY [Distribute bootstrap and start nodes] ************************************ TASK [openshift_node : Gather node information] ******************************** changed: [node02] ok: [node01] TASK [openshift_node : Copy master bootstrap config locally] ******************* ok: [node02] TASK [openshift_node : Distribute bootstrap kubeconfig if one does not exist] *** ok: [node01] changed: [node02] TASK [openshift_node : Start and enable node for bootstrapping] **************** changed: [node01] changed: [node02] TASK [openshift_node : Get node logs] ****************************************** skipping: [node02] skipping: [node01] TASK [openshift_node : debug] ************************************************** skipping: [node02] skipping: [node01] TASK [openshift_node : fail] *************************************************** skipping: [node02] skipping: [node01] PLAY [Approve any pending CSR requests from inventory nodes] ******************* TASK [Dump all candidate bootstrap hostnames] ********************************** ok: [node01] => { "msg": [ "node02", "node01" ] } TASK [Find all hostnames for bootstrapping] ************************************ ok: [node01] TASK [Dump the bootstrap hostnames] ******************************************** ok: [node01] => { "msg": [ "node02", "node01" ] } TASK [Approve bootstrap nodes] ************************************************* changed: [node01] TASK [Get CSRs] **************************************************************** skipping: [node01] TASK [Report approval errors] ************************************************** skipping: [node01] PLAY [Ensure any inventory labels are applied to the nodes] ******************** TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [openshift_manage_node : Wait for master API to become available before proceeding] *** skipping: [node02] TASK [openshift_manage_node : Wait for Node Registration] ********************** ok: [node01 -> node01] ok: [node02 -> node01] TASK [openshift_manage_node : include_tasks] *********************************** included: /root/openshift-ansible/roles/openshift_manage_node/tasks/config.yml for node02, node01 TASK [openshift_manage_node : Set node schedulability] ************************* ok: [node02 -> node01] ok: [node01 -> node01] TASK [openshift_manage_node : include_tasks] *********************************** included: /root/openshift-ansible/roles/openshift_manage_node/tasks/set_default_node_role.yml for node02, node01 TASK [openshift_manage_node : Retrieve nodes that are marked with the infra selector or the legacy infra selector] *** ok: [node02 -> node01] TASK [openshift_manage_node : Label infra or legacy infra nodes with the new role label] *** TASK [openshift_manage_node : Retrieve non-infra, non-master nodes that are not yet labeled compute] *** ok: [node02 -> node01] TASK [openshift_manage_node : label non-master non-infra nodes compute] ******** TASK [openshift_manage_node : Label all-in-one master as a compute node] ******* skipping: [node02] PLAY RECAP ********************************************************************* localhost : ok=30 changed=0 unreachable=0 failed=0 node01 : ok=71 changed=3 unreachable=0 failed=0 node02 : ok=155 changed=33 unreachable=0 failed=0 INSTALLER STATUS *************************************************************** Initialization : Complete (0:04:04) Node Preparation : Complete (0:05:02) Sending file modes: C0755 110489328 oc Sending file modes: C0600 5649 admin.kubeconfig Cluster "node01:8443" set. Cluster "node01:8443" set. + set +e + kubectl get nodes --no-headers + cluster/kubectl.sh get nodes --no-headers node01 Ready compute,infra,master 26d v1.10.0+b81c8f8 node02 Ready compute 1m v1.10.0+b81c8f8 + kubectl_rc=0 + '[' 0 -ne 0 ']' ++ kubectl get nodes --no-headers ++ cluster/kubectl.sh get nodes --no-headers ++ grep NotReady + '[' -n '' ']' + set -e + echo 'Nodes are ready:' Nodes are ready: + kubectl get nodes + cluster/kubectl.sh get nodes NAME STATUS ROLES AGE VERSION node01 Ready compute,infra,master 26d v1.10.0+b81c8f8 node02 Ready compute 1m v1.10.0+b81c8f8 + make cluster-sync ./cluster/build.sh Building ... Untagged: localhost:33198/kubevirt/virt-controller:devel Untagged: localhost:33198/kubevirt/virt-controller@sha256:9bf90106d81994bcf6803d68ed9c1a9ac5e9b656e3710d07edc66098b33984f0 Deleted: sha256:4a201c00311143384879dffb8f3cb2a58445fce6f72e92333ec99806f049db37 Deleted: sha256:00128c27df18b297216caf432b5229e9ed4c3f3a45a81dbd5ed49266913f69da Deleted: sha256:26e47e2eed9acea603b71d06e7eea1f1986c88ca77fa6196661adb8aa2d8aa4b Deleted: sha256:2c5263f2db2a3719d7833571f14b14a7902203c8d9ad112719f45ec45993701a Untagged: localhost:33198/kubevirt/virt-launcher:devel Untagged: localhost:33198/kubevirt/virt-launcher@sha256:a27e41edb5916c8e64c2f33c620440f8a2d4462144c7ecf5666d76d2b8f340f3 Deleted: sha256:60275eafa66c6d55801165bf18ff2f677587dd496aaf1f512f25fdfa5a5842d2 Deleted: sha256:e447a0c26991048dcc9f08eaf02fb20059d4b4ab1eb95d396388dd2d56f3230a Deleted: sha256:6d3d9fef7422b18d3fee54ddd81160d3c77734dda339dd15303f8c088f7fb7c4 Deleted: sha256:51a6d569e1457fa0ef12fc13714ba141c35a5aab6e87ed665572969832dfd421 Deleted: sha256:0f116f97a756b2df505fb0db2a1016935b2d4757ae97d80a1cf4bfbe2bb6796a Deleted: sha256:d7b47bbf991c755d825684c96ea6c903851837bc282a434e3f48fdf5f5ee557a Deleted: sha256:41fb49daafbba6fd85dc280dddf512b50a69c2bc957e3d071dc7af2368f46257 Deleted: sha256:8cdbab1aef3fb495c89ae9335a9b5a79e238f8759eee521f8c830f09037e3474 Deleted: sha256:277fd58fb83804ab8f4abf184763fe271618e84917de3851c9d047b1e816dfc4 Deleted: sha256:ce000497b174b7565f15e6db69951d77f7e45c12e5dddc40e49c0bc0ab146e8c Deleted: sha256:e4fbfd7dd6f40461d8a2394f1e979ad4cc89598cb10a59cd4443bd9d4a55b12c Deleted: sha256:742cf69ff591f3b31c628b6af10e778069ff6b9847963e5fc4748aac4cd0a089 Untagged: localhost:33198/kubevirt/virt-handler:devel Untagged: localhost:33198/kubevirt/virt-handler@sha256:991f1a2b62b64a9f62fc5f501b5086cf8e4b1c810450fefcaf5c6de0b1c5131a Deleted: sha256:77969069065221b72c6bcea7aaa4ea414e7a367c82e7a2555b8244a82b2b0369 Deleted: sha256:2ab400acac6f5f80dc01e29fa85801f3c11e923ce53f19ddfe43500959ebf215 Deleted: sha256:ebf34d3b2e34d9e47ce7dafd54dac792d3b86f491bc454fba3ef1588a44347ae Deleted: sha256:f68e411f12d618367c26108505d64bb3386c8b1672a1a1a31a68242b1de65e03 Untagged: localhost:33198/kubevirt/virt-api:devel Untagged: localhost:33198/kubevirt/virt-api@sha256:2a290ce264af408b6cc2942c58333359941ae1e846149472230163bbfdf4b578 Deleted: sha256:e0de336e3ebe00879bd39a5b78323e5ab341c1f8498f36a442213c65484aad0a Deleted: sha256:cbd2bf56e7971b4065ebc3f5ee1c001e1b88ac77fc3ac09b131d3557a76efbb6 Deleted: sha256:fdbdfdd8a9252377efbba3ace1b3c04449fb8bb9bd4ec2cf8139f836e8760dc3 Deleted: sha256:8191a145762585d1df855300d425bca120e82aec6e0712a47fc55808f14943fc Untagged: localhost:33198/kubevirt/subresource-access-test:devel Untagged: localhost:33198/kubevirt/subresource-access-test@sha256:66442c9dc9c238fc5c1a167727dceb7238b378f1fb9bef83b6006b6dda08fd95 Deleted: sha256:0016c01777f237a282408d9c91280f69f9103c427c987beabe9730a4277c69e7 Deleted: sha256:c748c4ca2d4bf662e6d241629687c83265e07ce8ac34d323ca4356e6272ec698 Deleted: sha256:5834911395de42cc42a356de4ba082b0df5ef5617fcb5281effb5e4de23b8c59 Deleted: sha256:6c10726cffdf21cbef6c9d5d32a91ec88fac929362b9c3c101846cdf78f68721 Untagged: localhost:33198/kubevirt/example-hook-sidecar:devel Untagged: localhost:33198/kubevirt/example-hook-sidecar@sha256:241863e982884cd48b3c556ac6eba67a3dd41c13f8a119f07a2d606a9d979886 Deleted: sha256:28b17f4ae30455972341e0e0b22f65bb16e7b9892467b48c651d42bea029fe79 Deleted: sha256:ed2ddb41eafe36b605013fa91c93f375b35b94c0a5f6deb2d26a4173365ea069 Deleted: sha256:d649ae9c07b2df797b07925768a684af2d51709884621940fce4572b6a3dc84e Deleted: sha256:cbd75576fc64cd675fc25e56d66c82ac9d9bb9b26ea7dce10b4f91c2bf6cc948 sha256:dcf2b21fa2ed11dcf9dbba21b1cca0ee3fad521a0e9aee61c06d0b0b66a4b200 go version go1.10 linux/amd64 go version go1.10 linux/amd64 make[1]: Entering directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt' hack/dockerized "./hack/check.sh && KUBEVIRT_VERSION= ./hack/build-go.sh install " && ./hack/build-copy-artifacts.sh sha256:dcf2b21fa2ed11dcf9dbba21b1cca0ee3fad521a0e9aee61c06d0b0b66a4b200 go version go1.10 linux/amd64 go version go1.10 linux/amd64 find: '/root/go/src/kubevirt.io/kubevirt/_out/cmd': No such file or directory Compiling tests... compiled tests.test hack/build-docker.sh build Sending build context to Docker daemon 40.39 MB Step 1/8 : FROM fedora:28 ---> cc510acfcd70 Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> bfe77d5699ed Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-controller ---> Using cache ---> b00c84523b53 Step 4/8 : WORKDIR /home/virt-controller ---> Using cache ---> b76b8bd8cd39 Step 5/8 : USER 1001 ---> Using cache ---> b6d9ad9ed232 Step 6/8 : COPY virt-controller /usr/bin/virt-controller ---> Using cache ---> 5200bf704485 Step 7/8 : ENTRYPOINT /usr/bin/virt-controller ---> Using cache ---> 2247cc3d94c8 Step 8/8 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "virt-controller" '' ---> Running in f9cb6921a338 ---> d1b82d88135c Removing intermediate container f9cb6921a338 Successfully built d1b82d88135c Sending build context to Docker daemon 43.32 MB Step 1/10 : FROM kubevirt/libvirt:4.2.0 ---> 5f0bfe81a3e0 Step 2/10 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 945996802736 Step 3/10 : RUN dnf -y install socat genisoimage util-linux libcgroup-tools ethtool net-tools sudo && dnf -y clean all && test $(id -u qemu) = 107 # make sure that the qemu user really is 107 ---> Using cache ---> 672f9ab56316 Step 4/10 : COPY virt-launcher /usr/bin/virt-launcher ---> Using cache ---> b4a6bafe7250 Step 5/10 : COPY kubevirt-sudo /etc/sudoers.d/kubevirt ---> Using cache ---> 7674bdd0cc0f Step 6/10 : RUN setcap CAP_NET_BIND_SERVICE=+eip /usr/bin/qemu-system-x86_64 ---> Using cache ---> 6ef8645dfd21 Step 7/10 : RUN mkdir -p /usr/share/kubevirt/virt-launcher ---> Using cache ---> 5c3e64c57221 Step 8/10 : COPY entrypoint.sh libvirtd.sh sock-connector /usr/share/kubevirt/virt-launcher/ ---> Using cache ---> 67fa77bc6dec Step 9/10 : ENTRYPOINT /usr/share/kubevirt/virt-launcher/entrypoint.sh ---> Using cache ---> 793f21bce80c Step 10/10 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "virt-launcher" '' ---> Running in 855c95e9626e ---> a0760a9235e6 Removing intermediate container 855c95e9626e Successfully built a0760a9235e6 Sending build context to Docker daemon 38.45 MB Step 1/5 : FROM fedora:28 ---> cc510acfcd70 Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> bfe77d5699ed Step 3/5 : COPY virt-handler /usr/bin/virt-handler ---> Using cache ---> c1e368e6066a Step 4/5 : ENTRYPOINT /usr/bin/virt-handler ---> Using cache ---> 820cf2e803de Step 5/5 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "virt-handler" '' ---> Running in adba39760339 ---> 977c563d1d18 Removing intermediate container adba39760339 Successfully built 977c563d1d18 Sending build context to Docker daemon 38.81 MB Step 1/8 : FROM fedora:28 ---> cc510acfcd70 Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> bfe77d5699ed Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-api ---> Using cache ---> ed1ebf600ee1 Step 4/8 : WORKDIR /home/virt-api ---> Using cache ---> 0769dad023e5 Step 5/8 : USER 1001 ---> Using cache ---> 0cb65afb0c2b Step 6/8 : COPY virt-api /usr/bin/virt-api ---> Using cache ---> 492f4f182385 Step 7/8 : ENTRYPOINT /usr/bin/virt-api ---> Using cache ---> 5e9d82d28268 Step 8/8 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "virt-api" '' ---> Running in e3da25fa1d35 ---> 984956aa9b95 Removing intermediate container e3da25fa1d35 Successfully built 984956aa9b95 Sending build context to Docker daemon 4.096 kB Step 1/7 : FROM fedora:28 ---> cc510acfcd70 Step 2/7 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> bfe77d5699ed Step 3/7 : ENV container docker ---> Using cache ---> 62847a2a1fa8 Step 4/7 : RUN mkdir -p /images/custom /images/alpine && truncate -s 64M /images/custom/disk.img && curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /images/alpine/disk.img ---> Using cache ---> 02134835a6aa Step 5/7 : ADD entrypoint.sh / ---> Using cache ---> ec0843818da7 Step 6/7 : CMD /entrypoint.sh ---> Using cache ---> 754029bb4bd2 Step 7/7 : LABEL "disks-images-provider" '' "kubevirt-functional-tests-openshift-3.10-release1" '' ---> Using cache ---> 6327b8256318 Successfully built 6327b8256318 Sending build context to Docker daemon 2.56 kB Step 1/5 : FROM fedora:28 ---> cc510acfcd70 Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> bfe77d5699ed Step 3/5 : ENV container docker ---> Using cache ---> 62847a2a1fa8 Step 4/5 : RUN dnf -y install procps-ng nmap-ncat && dnf -y clean all ---> Using cache ---> 207487abe7b2 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "vm-killer" '' ---> Using cache ---> 27cf5472530f Successfully built 27cf5472530f Sending build context to Docker daemon 5.12 kB Step 1/7 : FROM debian:sid ---> 68f33cf86aab Step 2/7 : MAINTAINER "David Vossel" \ ---> Using cache ---> 5734d749eb5c Step 3/7 : ENV container docker ---> Using cache ---> f8775a77966f Step 4/7 : RUN apt-get update && apt-get install -y bash curl bzip2 qemu-utils && mkdir -p /disk && rm -rf /var/lib/apt/lists/* ---> Using cache ---> 1a40cf222a61 Step 5/7 : ADD entry-point.sh / ---> Using cache ---> 77b545d92fe7 Step 6/7 : CMD /entry-point.sh ---> Using cache ---> dfe20d463305 Step 7/7 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "registry-disk-v1alpha" '' ---> Using cache ---> 5efdf368e732 Successfully built 5efdf368e732 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33592/kubevirt/registry-disk-v1alpha:devel ---> 5efdf368e732 Step 2/4 : MAINTAINER "David Vossel" \ ---> Using cache ---> 386f7e924456 Step 3/4 : RUN curl https://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img > /disk/cirros.img ---> Using cache ---> f473a86e4d6a Step 4/4 : LABEL "cirros-registry-disk-demo" '' "kubevirt-functional-tests-openshift-3.10-release1" '' ---> Using cache ---> a4ca4c67d45c Successfully built a4ca4c67d45c Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33592/kubevirt/registry-disk-v1alpha:devel ---> 5efdf368e732 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 369bca39dcc2 Step 3/4 : RUN curl -g -L https://download.fedoraproject.org/pub/fedora/linux/releases/27/CloudImages/x86_64/images/Fedora-Cloud-Base-27-1.6.x86_64.qcow2 > /disk/fedora.qcow2 ---> Using cache ---> de1e81f43a59 Step 4/4 : LABEL "fedora-cloud-registry-disk-demo" '' "kubevirt-functional-tests-openshift-3.10-release1" '' ---> Using cache ---> a5867eac6e05 Successfully built a5867eac6e05 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33592/kubevirt/registry-disk-v1alpha:devel ---> 5efdf368e732 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 369bca39dcc2 Step 3/4 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /disk/alpine.iso ---> Using cache ---> 1083d820f9c8 Step 4/4 : LABEL "alpine-registry-disk-demo" '' "kubevirt-functional-tests-openshift-3.10-release1" '' ---> Using cache ---> 11512d828b9c Successfully built 11512d828b9c Sending build context to Docker daemon 35.59 MB Step 1/8 : FROM fedora:28 ---> cc510acfcd70 Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> bfe77d5699ed Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virtctl ---> Using cache ---> 985fe391c056 Step 4/8 : WORKDIR /home/virtctl ---> Using cache ---> 3b2cae8ac543 Step 5/8 : USER 1001 ---> Using cache ---> 0c06e5b4a900 Step 6/8 : COPY subresource-access-test /subresource-access-test ---> Using cache ---> b8db0646f024 Step 7/8 : ENTRYPOINT /subresource-access-test ---> Using cache ---> d524c4957628 Step 8/8 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "subresource-access-test" '' ---> Running in 9f6675d9b6fc ---> 0729d7637f8d Removing intermediate container 9f6675d9b6fc Successfully built 0729d7637f8d Sending build context to Docker daemon 3.072 kB Step 1/9 : FROM fedora:28 ---> cc510acfcd70 Step 2/9 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> bfe77d5699ed Step 3/9 : ENV container docker ---> Using cache ---> 62847a2a1fa8 Step 4/9 : RUN dnf -y install make git gcc && dnf -y clean all ---> Using cache ---> d3456b1644b1 Step 5/9 : ENV GIMME_GO_VERSION 1.9.2 ---> Using cache ---> 0ba81fddbba1 Step 6/9 : RUN mkdir -p /gimme && curl -sL https://raw.githubusercontent.com/travis-ci/gimme/master/gimme | HOME=/gimme bash >> /etc/profile.d/gimme.sh ---> Using cache ---> 5d33abe3f819 Step 7/9 : ENV GOPATH "/go" GOBIN "/usr/bin" ---> Using cache ---> 783826523be1 Step 8/9 : RUN mkdir -p /go && source /etc/profile.d/gimme.sh && go get github.com/masterzen/winrm-cli ---> Using cache ---> 711bc8d15952 Step 9/9 : LABEL "kubevirt-functional-tests-openshift-3.10-release1" '' "winrmcli" '' ---> Using cache ---> fe40426b785b Successfully built fe40426b785b Sending build context to Docker daemon 36.8 MB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> e3238544ad97 Step 3/5 : COPY example-hook-sidecar /example-hook-sidecar ---> Using cache ---> 8ff50de665fa Step 4/5 : ENTRYPOINT /example-hook-sidecar ---> Using cache ---> 3e877e8e3efc Step 5/5 : LABEL "example-hook-sidecar" '' "kubevirt-functional-tests-openshift-3.10-release1" '' ---> Running in 484817dd4b0a ---> a5d817fbf3d9 Removing intermediate container 484817dd4b0a Successfully built a5d817fbf3d9 hack/build-docker.sh push The push refers to a repository [localhost:33592/kubevirt/virt-controller] 6a1377fa7d58: Preparing aa89340cf7a8: Preparing 891e1e4ef82a: Preparing aa89340cf7a8: Pushed 6a1377fa7d58: Pushed 891e1e4ef82a: Pushed devel: digest: sha256:5a6557a417e9f41684abaa479611346279e8dd400703cf809e0dbb0c95b1c057 size: 949 The push refers to a repository [localhost:33592/kubevirt/virt-launcher] 5d42131cfa75: Preparing b944abd209df: Preparing c0cc9e036393: Preparing 926ccafbbd0b: Preparing 5766fa5e1828: Preparing 633427c64a24: Preparing da38cf808aa5: Preparing b83399358a92: Preparing 186d8b3e4fd8: Preparing fa6154170bf5: Preparing 5eefb9960a36: Preparing 891e1e4ef82a: Preparing b83399358a92: Waiting 186d8b3e4fd8: Waiting da38cf808aa5: Waiting fa6154170bf5: Waiting 891e1e4ef82a: Waiting 633427c64a24: Waiting 5eefb9960a36: Waiting b944abd209df: Pushed 5d42131cfa75: Pushed 926ccafbbd0b: Pushed b83399358a92: Pushed da38cf808aa5: Pushed fa6154170bf5: Pushed 186d8b3e4fd8: Pushed 891e1e4ef82a: Mounted from kubevirt/virt-controller c0cc9e036393: Pushed 633427c64a24: Pushed 5766fa5e1828: Pushed 5eefb9960a36: Pushed devel: digest: sha256:10a1ed6b63841f35951de1d4a7de555c1ab3376e8d2bd77b8f7aa937aff07df3 size: 2828 The push refers to a repository [localhost:33592/kubevirt/virt-handler] 72f5f339d668: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/virt-launcher 72f5f339d668: Pushed devel: digest: sha256:850f306aa968563ae68d706c96ba0e89ef9613c52bf7401eb82c60adbf1ace40 size: 740 The push refers to a repository [localhost:33592/kubevirt/virt-api] 65cc3ca80f17: Preparing 82fc744c99b4: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/virt-handler 82fc744c99b4: Pushed 65cc3ca80f17: Pushed devel: digest: sha256:5c30c5148e41a794df2af917223c3588dac4307d3ec96b8eddcf2607ed0ab0a7 size: 948 The push refers to a repository [localhost:33592/kubevirt/disks-images-provider] 71ad31feb2c5: Preparing 21d4b721776e: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/virt-api 71ad31feb2c5: Pushed 21d4b721776e: Pushed devel: digest: sha256:5dc088106df85eb01f2ad0566624239b95b34986820107944e36d309183fd4cd size: 948 The push refers to a repository [localhost:33592/kubevirt/vm-killer] c4cfadeeaf5f: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/disks-images-provider c4cfadeeaf5f: Pushed devel: digest: sha256:39b817b79b1fbce75dbb476bc261b2752fd6466bf98d373208d5144579da22b0 size: 740 The push refers to a repository [localhost:33592/kubevirt/registry-disk-v1alpha] 661cce8d8e52: Preparing 41e0baba3077: Preparing 25edbec0eaea: Preparing 661cce8d8e52: Pushed 41e0baba3077: Pushed 25edbec0eaea: Pushed devel: digest: sha256:0df707a55243af8792380fba68a76307017494c503e0e9071ed55d7d3c3611d4 size: 948 The push refers to a repository [localhost:33592/kubevirt/cirros-registry-disk-demo] f9f97de3966a: Preparing 661cce8d8e52: Preparing 41e0baba3077: Preparing 25edbec0eaea: Preparing 25edbec0eaea: Mounted from kubevirt/registry-disk-v1alpha 41e0baba3077: Mounted from kubevirt/registry-disk-v1alpha 661cce8d8e52: Mounted from kubevirt/registry-disk-v1alpha f9f97de3966a: Pushed devel: digest: sha256:3f818f67105a36bdc42bdbfad87fc29d0028e39a0dceef92d12efbcf8e16e5ed size: 1160 The push refers to a repository [localhost:33592/kubevirt/fedora-cloud-registry-disk-demo] 24cdf3b545f2: Preparing 661cce8d8e52: Preparing 41e0baba3077: Preparing 25edbec0eaea: Preparing 41e0baba3077: Mounted from kubevirt/cirros-registry-disk-demo 25edbec0eaea: Mounted from kubevirt/cirros-registry-disk-demo 661cce8d8e52: Mounted from kubevirt/cirros-registry-disk-demo 24cdf3b545f2: Pushed devel: digest: sha256:a6a571626690141c7da4cf0e1eb4fd75e5dd9ae427d5070c2729214cfbd6a192 size: 1161 The push refers to a repository [localhost:33592/kubevirt/alpine-registry-disk-demo] d8e356e905f4: Preparing 661cce8d8e52: Preparing 41e0baba3077: Preparing 25edbec0eaea: Preparing 25edbec0eaea: Mounted from kubevirt/fedora-cloud-registry-disk-demo 661cce8d8e52: Mounted from kubevirt/fedora-cloud-registry-disk-demo 41e0baba3077: Mounted from kubevirt/fedora-cloud-registry-disk-demo d8e356e905f4: Pushed devel: digest: sha256:c27568048aa8e031860d98cdced0370763745ad80581e62432568dac45abf1fb size: 1160 The push refers to a repository [localhost:33592/kubevirt/subresource-access-test] 558e43b6c1b1: Preparing 25cb73590a9d: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/vm-killer 25cb73590a9d: Pushed 558e43b6c1b1: Pushed devel: digest: sha256:ea28e3e376d1a455d39c425ac05c911d82a3435f81c44945d89fe917944b5ff2 size: 948 The push refers to a repository [localhost:33592/kubevirt/winrmcli] f8083e002d0b: Preparing 53c709abc882: Preparing 9ca98a0f492b: Preparing 891e1e4ef82a: Preparing 891e1e4ef82a: Mounted from kubevirt/subresource-access-test f8083e002d0b: Pushed 9ca98a0f492b: Pushed 53c709abc882: Pushed devel: digest: sha256:4fe6c9666a841b61b962d7fb73ccb7cb0dabc3b56e1657cfdfd9005e1a36d38c size: 1165 The push refers to a repository [localhost:33592/kubevirt/example-hook-sidecar] 23a50bf7c861: Preparing 39bae602f753: Preparing 23a50bf7c861: Pushed 39bae602f753: Pushed devel: digest: sha256:b32199970db95db1a96edf5cdba47005f707e0eae073bddc6adf04be271f2a30 size: 740 make[1]: Leaving directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt' Done ./cluster/clean.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ MANIFEST_TEMPLATES_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/templates/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ KUBEVIRT_PROVIDER=os-3.10.0 ++ KUBEVIRT_NUM_NODES=2 ++ '[' -z kubevirt-functional-tests-openshift-3.10-release ']' ++ provider_prefix=kubevirt-functional-tests-openshift-3.10-release1 ++ job_prefix=kubevirt-functional-tests-openshift-3.10-release1 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.7.0-189-g96d0165 ++ KUBEVIRT_VERSION=v0.7.0-189-g96d0165 + source cluster/os-3.10.0/provider.sh ++ set -e ++ image=os-3.10.0@sha256:50a4b8ee3e07d592e7e4fbf3eb1401980a5947499dfdc3d847c085b5775aaa9a ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --net=host --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/gocli@sha256:aa7f295a7908fa333ab5e98ef3af0bfafbabfd3cee2b83f9af47f722e3000f6a' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ KUBEVIRT_PROVIDER=os-3.10.0 ++ source hack/config-default.sh source hack/config-os-3.10.0.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test cmd/example-hook-sidecar' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/disks-images-provider images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli cmd/example-hook-sidecar' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ namespace=kube-system ++ test -f hack/config-provider-os-3.10.0.sh ++ source hack/config-provider-os-3.10.0.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/cluster/os-3.10.0/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/cluster/os-3.10.0/.kubectl +++ docker_prefix=localhost:33592/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Cleaning up ...' Cleaning up ... + cluster/kubectl.sh get vmis --all-namespaces -o=custom-columns=NAME:.metadata.name,NAMESPACE:.metadata.namespace,FINALIZERS:.metadata.finalizers --no-headers + grep foregroundDeleteVirtualMachine + read p error: the server doesn't have a resource type "vmis" + _kubectl delete ds -l kubevirt.io -n kube-system --cascade=false --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=libvirt --force --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=virt-handler --force --grace-period 0 No resources found + namespaces=(default ${namespace}) + for i in '${namespaces[@]}' + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete deployment -l kubevirt.io No resources found + _kubectl -n default delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete rs -l kubevirt.io No resources found + _kubectl -n default delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete services -l kubevirt.io No resources found + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n default delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete secrets -l kubevirt.io No resources found + _kubectl -n default delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete pv -l kubevirt.io No resources found + _kubectl -n default delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete pvc -l kubevirt.io No resources found + _kubectl -n default delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete ds -l kubevirt.io No resources found + _kubectl -n default delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n default delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete pods -l kubevirt.io No resources found + _kubectl -n default delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n default delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete rolebinding -l kubevirt.io No resources found + _kubectl -n default delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete roles -l kubevirt.io No resources found + _kubectl -n default delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete clusterroles -l kubevirt.io No resources found + _kubectl -n default delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n default delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n default get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.10.0/.kubeconfig ++ KUBECONFIG=cluster/os-3.10.0/.kubeconfig ++ cluster/os-3.10.0/.kubectl -n default get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + for i in '${namespaces[@]}' + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete deployment -l kubevirt.io No resources found + _kubectl -n kube-system delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete rs -l kubevirt.io No resources found + _kubectl -n kube-system delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete services -l kubevirt.io No resources found + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n kube-system delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete secrets -l kubevirt.io No resources found + _kubectl -n kube-system delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete pv -l kubevirt.io No resources found + _kubectl -n kube-system delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete pvc -l kubevirt.io No resources found + _kubectl -n kube-system delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete ds -l kubevirt.io No resources found + _kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n kube-system delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete pods -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete rolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete roles -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete clusterroles -l kubevirt.io No resources found + _kubectl -n kube-system delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl -n kube-system delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.10.0/.kubeconfig ++ KUBECONFIG=cluster/os-3.10.0/.kubeconfig ++ cluster/os-3.10.0/.kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + sleep 2 + echo Done Done ./cluster/deploy.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ MANIFEST_TEMPLATES_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/templates/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ KUBEVIRT_PROVIDER=os-3.10.0 ++ KUBEVIRT_NUM_NODES=2 ++ '[' -z kubevirt-functional-tests-openshift-3.10-release ']' ++ provider_prefix=kubevirt-functional-tests-openshift-3.10-release1 ++ job_prefix=kubevirt-functional-tests-openshift-3.10-release1 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.7.0-189-g96d0165 ++ KUBEVIRT_VERSION=v0.7.0-189-g96d0165 + source cluster/os-3.10.0/provider.sh ++ set -e ++ image=os-3.10.0@sha256:50a4b8ee3e07d592e7e4fbf3eb1401980a5947499dfdc3d847c085b5775aaa9a ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --net=host --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/gocli@sha256:aa7f295a7908fa333ab5e98ef3af0bfafbabfd3cee2b83f9af47f722e3000f6a' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ KUBEVIRT_PROVIDER=os-3.10.0 ++ source hack/config-default.sh source hack/config-os-3.10.0.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test cmd/example-hook-sidecar' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/disks-images-provider images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli cmd/example-hook-sidecar' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ namespace=kube-system ++ test -f hack/config-provider-os-3.10.0.sh ++ source hack/config-provider-os-3.10.0.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/cluster/os-3.10.0/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/cluster/os-3.10.0/.kubectl +++ docker_prefix=localhost:33592/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Deploying ...' Deploying ... + [[ -z openshift-3.10-release ]] + [[ openshift-3.10-release =~ .*-dev ]] + [[ openshift-3.10-release =~ .*-release ]] + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/demo-content.yaml =~ .*demo.* ]] + continue + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml =~ .*demo.* ]] + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml clusterrole.rbac.authorization.k8s.io "kubevirt.io:admin" created clusterrole.rbac.authorization.k8s.io "kubevirt.io:edit" created clusterrole.rbac.authorization.k8s.io "kubevirt.io:view" created serviceaccount "kubevirt-apiserver" created clusterrolebinding.rbac.authorization.k8s.io "kubevirt-apiserver" created clusterrolebinding.rbac.authorization.k8s.io "kubevirt-apiserver-auth-delegator" created rolebinding.rbac.authorization.k8s.io "kubevirt-apiserver" created role.rbac.authorization.k8s.io "kubevirt-apiserver" created clusterrole.rbac.authorization.k8s.io "kubevirt-apiserver" created clusterrole.rbac.authorization.k8s.io "kubevirt-controller" created serviceaccount "kubevirt-controller" created serviceaccount "kubevirt-privileged" created clusterrolebinding.rbac.authorization.k8s.io "kubevirt-controller" created clusterrolebinding.rbac.authorization.k8s.io "kubevirt-controller-cluster-admin" created clusterrolebinding.rbac.authorization.k8s.io "kubevirt-privileged-cluster-admin" created clusterrole.rbac.authorization.k8s.io "kubevirt.io:default" created clusterrolebinding.rbac.authorization.k8s.io "kubevirt.io:default" created service "virt-api" created deployment.extensions "virt-api" created deployment.extensions "virt-controller" created daemonset.extensions "virt-handler" created customresourcedefinition.apiextensions.k8s.io "virtualmachineinstances.kubevirt.io" created customresourcedefinition.apiextensions.k8s.io "virtualmachineinstancereplicasets.kubevirt.io" created customresourcedefinition.apiextensions.k8s.io "virtualmachineinstancepresets.kubevirt.io" created customresourcedefinition.apiextensions.k8s.io "virtualmachines.kubevirt.io" created + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R persistentvolumeclaim "disk-alpine" created persistentvolume "host-path-disk-alpine" created persistentvolumeclaim "disk-custom" created persistentvolume "host-path-disk-custom" created daemonset.extensions "disks-images-provider" created serviceaccount "kubevirt-testing" created clusterrolebinding.rbac.authorization.k8s.io "kubevirt-testing-cluster-admin" created + [[ os-3.10.0 =~ os-* ]] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-controller"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-testing"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-privileged"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-apiserver"] + _kubectl adm policy add-scc-to-user privileged admin + export KUBECONFIG=cluster/os-3.10.0/.kubeconfig + KUBECONFIG=cluster/os-3.10.0/.kubeconfig + cluster/os-3.10.0/.kubectl adm policy add-scc-to-user privileged admin scc "privileged" added to: ["admin"] + echo Done Done + namespaces=(kube-system default) + [[ kube-system != \k\u\b\e\-\s\y\s\t\e\m ]] + timeout=300 + sample=30 + for i in '${namespaces[@]}' + current_time=0 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'disks-images-provider-hwh68 0/1 ContainerCreating 0 3s disks-images-provider-mmwgg 0/1 ContainerCreating 0 3s virt-api-7d79764579-rk48t 0/1 ContainerCreating 0 6s virt-api-7d79764579-swtbv 0/1 ContainerCreating 0 6s virt-controller-7d57d96b65-fp8kf 0/1 ContainerCreating 0 6s virt-controller-7d57d96b65-r4stl 0/1 ContainerCreating 0 6s virt-handler-7fcr8 0/1 ContainerCreating 0 6s virt-handler-9rw2k 0/1 ContainerCreating 0 6s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running disks-images-provider-hwh68 0/1 ContainerCreating 0 4s disks-images-provider-mmwgg 0/1 ContainerCreating 0 4s virt-api-7d79764579-rk48t 0/1 ContainerCreating 0 7s virt-api-7d79764579-swtbv 0/1 ContainerCreating 0 7s virt-controller-7d57d96b65-fp8kf 0/1 ContainerCreating 0 7s virt-controller-7d57d96b65-r4stl 0/1 ContainerCreating 0 7s virt-handler-7fcr8 0/1 ContainerCreating 0 7s virt-handler-9rw2k 0/1 ContainerCreating 0 7s + sleep 30 + current_time=30 + '[' 30 -gt 300 ']' ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n '' ']' + current_time=0 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers ++ grep false + '[' -n '' ']' + kubectl get pods -n kube-system + cluster/kubectl.sh get pods -n kube-system NAME READY STATUS RESTARTS AGE disks-images-provider-hwh68 1/1 Running 0 37s disks-images-provider-mmwgg 1/1 Running 0 37s master-api-node01 1/1 Running 1 26d master-controllers-node01 1/1 Running 1 26d master-etcd-node01 1/1 Running 1 26d virt-api-7d79764579-rk48t 1/1 Running 1 40s virt-api-7d79764579-swtbv 1/1 Running 0 40s virt-controller-7d57d96b65-fp8kf 1/1 Running 0 40s virt-controller-7d57d96b65-r4stl 1/1 Running 0 40s virt-handler-7fcr8 1/1 Running 0 40s virt-handler-9rw2k 1/1 Running 0 40s + for i in '${namespaces[@]}' + current_time=0 ++ kubectl get pods -n default --no-headers ++ cluster/kubectl.sh get pods -n default --no-headers ++ grep -v Running + '[' -n '' ']' + current_time=0 ++ kubectl get pods -n default '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers ++ grep false ++ cluster/kubectl.sh get pods -n default '-ocustom-columns=status:status.containerStatuses[*].ready' --no-headers + '[' -n '' ']' + kubectl get pods -n default + cluster/kubectl.sh get pods -n default NAME READY STATUS RESTARTS AGE docker-registry-1-rl562 1/1 Running 1 26d registry-console-1-rw9zf 1/1 Running 1 26d router-1-6cch9 1/1 Running 1 26d + kubectl version + cluster/kubectl.sh version oc v3.10.0-rc.0+c20e215 kubernetes v1.10.0+b81c8f8 features: Basic-Auth GSSAPI Kerberos SPNEGO Server https://127.0.0.1:33589 openshift v3.10.0-rc.0+c20e215 kubernetes v1.10.0+b81c8f8 + ginko_params='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/junit.xml' + [[ openshift-3.10-release =~ windows.* ]] + FUNC_TEST_ARGS='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-3.10-release/junit.xml' + make functest hack/dockerized "hack/build-func-tests.sh" sha256:dcf2b21fa2ed11dcf9dbba21b1cca0ee3fad521a0e9aee61c06d0b0b66a4b200 go version go1.10 linux/amd64 go version go1.10 linux/amd64 Compiling tests... compiled tests.test hack/functests.sh Running Suite: Tests Suite ========================== Random Seed: 1533331101 Will run 148 of 148 specs Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:19:33 http: TLS handshake error from 10.129.0.1:44372: EOF 2018/08/03 21:19:43 http: TLS handshake error from 10.129.0.1:44384: EOF level=info timestamp=2018-08-03T21:19:49.610678Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:19:53 http: TLS handshake error from 10.129.0.1:44396: EOF 2018/08/03 21:20:03 http: TLS handshake error from 10.129.0.1:44420: EOF 2018/08/03 21:20:13 http: TLS handshake error from 10.129.0.1:44432: EOF 2018/08/03 21:20:23 http: TLS handshake error from 10.129.0.1:44444: EOF 2018/08/03 21:20:33 http: TLS handshake error from 10.129.0.1:44456: EOF 2018/08/03 21:20:43 http: TLS handshake error from 10.129.0.1:44468: EOF level=info timestamp=2018-08-03T21:20:49.624941Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:20:53 http: TLS handshake error from 10.129.0.1:44480: EOF 2018/08/03 21:21:03 http: TLS handshake error from 10.129.0.1:44492: EOF 2018/08/03 21:21:13 http: TLS handshake error from 10.129.0.1:44504: EOF level=info timestamp=2018-08-03T21:21:19.596390Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:21:23 http: TLS handshake error from 10.129.0.1:44516: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running 2018/08/03 21:20:45 http: TLS handshake error from 10.129.0.1:51050: EOF level=info timestamp=2018-08-03T21:20:47.912751Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:20:48.742333Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:20:55 http: TLS handshake error from 10.129.0.1:51062: EOF level=info timestamp=2018-08-03T21:20:58.600200Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:20:58.966058Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:21:05 http: TLS handshake error from 10.129.0.1:51074: EOF level=info timestamp=2018-08-03T21:21:05.659442Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:21:09.263366Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:21:15 http: TLS handshake error from 10.129.0.1:51086: EOF level=info timestamp=2018-08-03T21:21:18.137641Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:21:19.394579Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:21:25.048866Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:21:25.052533Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 21:21:25 http: TLS handshake error from 10.129.0.1:51098: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:17:28.200171Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:28.200216Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer vmInformer" level=info timestamp=2018-08-03T21:17:28.200238Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer limitrangeInformer" level=info timestamp=2018-08-03T21:17:28.200256Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer vmiInformer" level=info timestamp=2018-08-03T21:17:28.200273Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer kubeVirtPodInformer" level=info timestamp=2018-08-03T21:17:28.200289Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer kubeVirtNodeInformer" level=info timestamp=2018-08-03T21:17:28.200305Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer vmiPresetInformer" level=info timestamp=2018-08-03T21:17:28.200351Z pos=vm.go:85 component=virt-controller service=http msg="Starting VirtualMachine controller." level=info timestamp=2018-08-03T21:17:28.207633Z pos=node.go:104 component=virt-controller service=http msg="Starting node controller." level=info timestamp=2018-08-03T21:17:28.207720Z pos=vmi.go:129 component=virt-controller service=http msg="Starting vmi controller." level=info timestamp=2018-08-03T21:17:28.207760Z pos=replicaset.go:111 component=virt-controller service=http msg="Starting VirtualMachineInstanceReplicaSet controller." level=info timestamp=2018-08-03T21:17:28.207799Z pos=preset.go:74 component=virt-controller service=http msg="Starting Virtual Machine Initializer." level=info timestamp=2018-08-03T21:18:25.130242Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi46z7j kind= uid=c1a29e9f-9762-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:18:25.130864Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi46z7j kind= uid=c1a29e9f-9762-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:18:26.485632Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi46z7j\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi46z7j" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=info timestamp=2018-08-03T21:17:36.056369Z pos=virt-handler.go:89 component=virt-handler hostname=node02 level=info timestamp=2018-08-03T21:17:36.061203Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:36.361779Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:36.362119Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:36.363390Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:24.464403Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=info timestamp=2018-08-03T21:17:30.265794Z pos=virt-handler.go:89 component=virt-handler hostname=node01 level=info timestamp=2018-08-03T21:17:30.286009Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:30.487255Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:30.487883Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:30.489165Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:18.454344Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi46z7j-tmphn Pod phase: Pending • Failure [182.724 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 VirtualMachineInstance definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:55 with 3 CPU cores /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:56 should report 3 cpu cores under guest OS [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:62 Timed out after 90.011s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Starting a VirtualMachineInstance level=info timestamp=2018-08-03T21:18:26.293711Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi46z7j kind=VirtualMachineInstance uid=c1a29e9f-9762-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi46z7j-tmphn" • Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:22:43 http: TLS handshake error from 10.129.0.1:44614: EOF level=info timestamp=2018-08-03T21:22:49.036007Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:22:49.712488Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:22:53 http: TLS handshake error from 10.129.0.1:44626: EOF 2018/08/03 21:23:03 http: TLS handshake error from 10.129.0.1:44638: EOF 2018/08/03 21:23:13 http: TLS handshake error from 10.129.0.1:44650: EOF level=info timestamp=2018-08-03T21:23:19.592678Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:23:23 http: TLS handshake error from 10.129.0.1:44662: EOF 2018/08/03 21:23:33 http: TLS handshake error from 10.129.0.1:44674: EOF 2018/08/03 21:23:43 http: TLS handshake error from 10.129.0.1:44686: EOF 2018/08/03 21:23:53 http: TLS handshake error from 10.129.0.1:44698: EOF 2018/08/03 21:24:03 http: TLS handshake error from 10.129.0.1:44710: EOF 2018/08/03 21:24:13 http: TLS handshake error from 10.129.0.1:44722: EOF level=info timestamp=2018-08-03T21:24:19.613661Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:24:23 http: TLS handshake error from 10.129.0.1:44734: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:23:49.628592Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:23:49.753120Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:23:52.664006Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:23:55 http: TLS handshake error from 10.129.0.1:51280: EOF level=info timestamp=2018-08-03T21:24:00.246844Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:24:02.873512Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:24:05 http: TLS handshake error from 10.129.0.1:51292: EOF level=info timestamp=2018-08-03T21:24:07.399533Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:24:13.116213Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:24:14.402513Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:24:14.406873Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 21:24:15 http: TLS handshake error from 10.129.0.1:51304: EOF level=info timestamp=2018-08-03T21:24:20.088553Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:24:23.310054Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:24:25 http: TLS handshake error from 10.129.0.1:51316: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:17:28.200273Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer kubeVirtPodInformer" level=info timestamp=2018-08-03T21:17:28.200289Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer kubeVirtNodeInformer" level=info timestamp=2018-08-03T21:17:28.200305Z pos=virtinformers.go:107 component=virt-controller service=http msg="STARTING informer vmiPresetInformer" level=info timestamp=2018-08-03T21:17:28.200351Z pos=vm.go:85 component=virt-controller service=http msg="Starting VirtualMachine controller." level=info timestamp=2018-08-03T21:17:28.207633Z pos=node.go:104 component=virt-controller service=http msg="Starting node controller." level=info timestamp=2018-08-03T21:17:28.207720Z pos=vmi.go:129 component=virt-controller service=http msg="Starting vmi controller." level=info timestamp=2018-08-03T21:17:28.207760Z pos=replicaset.go:111 component=virt-controller service=http msg="Starting VirtualMachineInstanceReplicaSet controller." level=info timestamp=2018-08-03T21:17:28.207799Z pos=preset.go:74 component=virt-controller service=http msg="Starting Virtual Machine Initializer." level=info timestamp=2018-08-03T21:18:25.130242Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi46z7j kind= uid=c1a29e9f-9762-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:18:25.130864Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi46z7j kind= uid=c1a29e9f-9762-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:18:26.485632Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi46z7j\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi46z7j" level=info timestamp=2018-08-03T21:21:26.531480Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmif7lrt kind= uid=2dd0070a-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:21:26.534180Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmif7lrt kind= uid=2dd0070a-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:21:27.989803Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmir2hkz kind= uid=2eb0a90c-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:21:27.990444Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmir2hkz kind= uid=2eb0a90c-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=info timestamp=2018-08-03T21:17:36.056369Z pos=virt-handler.go:89 component=virt-handler hostname=node02 level=info timestamp=2018-08-03T21:17:36.061203Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:36.361779Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:36.362119Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:36.363390Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:24.464403Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:40.551681Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=info timestamp=2018-08-03T21:17:30.265794Z pos=virt-handler.go:89 component=virt-handler hostname=node01 level=info timestamp=2018-08-03T21:17:30.286009Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:30.487255Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:30.487883Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:30.489165Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:18.454344Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:34.860408Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmir2hkz-qpnqb Pod phase: Pending ------------------------------ • Failure [180.722 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 VirtualMachineInstance definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:55 with hugepages /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:164 should consume hugepages /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 hugepages-2Mi [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.009s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Starting a VM level=info timestamp=2018-08-03T21:21:27.884044Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmir2hkz kind=VirtualMachineInstance uid=2eb0a90c-9763-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmir2hkz-qpnqb" S [SKIPPING] [0.415 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 VirtualMachineInstance definition /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:55 with hugepages /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:164 should consume hugepages /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 hugepages-1Gi [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 No node with hugepages hugepages-1Gi capacity /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:216 ------------------------------ • Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:25:23 http: TLS handshake error from 10.129.0.1:44808: EOF 2018/08/03 21:25:33 http: TLS handshake error from 10.129.0.1:44820: EOF 2018/08/03 21:25:43 http: TLS handshake error from 10.129.0.1:44832: EOF 2018/08/03 21:25:53 http: TLS handshake error from 10.129.0.1:44844: EOF 2018/08/03 21:26:03 http: TLS handshake error from 10.129.0.1:44856: EOF 2018/08/03 21:26:13 http: TLS handshake error from 10.129.0.1:44868: EOF level=info timestamp=2018-08-03T21:26:19.690795Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:26:23 http: TLS handshake error from 10.129.0.1:44880: EOF 2018/08/03 21:26:33 http: TLS handshake error from 10.129.0.1:44892: EOF 2018/08/03 21:26:43 http: TLS handshake error from 10.129.0.1:44904: EOF level=info timestamp=2018-08-03T21:26:49.565203Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:26:53 http: TLS handshake error from 10.129.0.1:44916: EOF 2018/08/03 21:27:03 http: TLS handshake error from 10.129.0.1:44928: EOF 2018/08/03 21:27:13 http: TLS handshake error from 10.129.0.1:44940: EOF 2018/08/03 21:27:23 http: TLS handshake error from 10.129.0.1:44952: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running 2018/08/03 21:26:45 http: TLS handshake error from 10.129.0.1:51486: EOF level=info timestamp=2018-08-03T21:26:46.257590Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:26:51.427211Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:26:55 http: TLS handshake error from 10.129.0.1:51498: EOF level=info timestamp=2018-08-03T21:26:56.471077Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:27:02.061528Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:27:05 http: TLS handshake error from 10.129.0.1:51510: EOF level=info timestamp=2018-08-03T21:27:06.690951Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:27:08.930620Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:27:15 http: TLS handshake error from 10.129.0.1:51522: EOF level=info timestamp=2018-08-03T21:27:16.902785Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:27:19.581471Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:27:21.545492Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:27:25 http: TLS handshake error from 10.129.0.1:51534: EOF level=info timestamp=2018-08-03T21:27:27.163354Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:18:26.485632Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi46z7j\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi46z7j" level=info timestamp=2018-08-03T21:21:26.531480Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmif7lrt kind= uid=2dd0070a-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:21:26.534180Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmif7lrt kind= uid=2dd0070a-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:21:27.989803Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmir2hkz kind= uid=2eb0a90c-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:21:27.990444Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmir2hkz kind= uid=2eb0a90c-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:29.009890Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6vzk8 kind= uid=9a96fb9e-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:24:29.012304Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6vzk8 kind= uid=9a96fb9e-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:29.134703Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:29.185808Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:30.263433Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi6vzk8, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 9a96fb9e-9763-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:30.475121Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:24:30.475516Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:30.661579Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.862892Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.904436Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=info timestamp=2018-08-03T21:17:36.056369Z pos=virt-handler.go:89 component=virt-handler hostname=node02 level=info timestamp=2018-08-03T21:17:36.061203Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:36.361779Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:36.362119Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:36.363390Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:24.464403Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:40.551681Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=info timestamp=2018-08-03T21:17:30.265794Z pos=virt-handler.go:89 component=virt-handler hostname=node01 level=info timestamp=2018-08-03T21:17:30.286009Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:30.487255Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:30.487883Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:30.489165Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:18.454344Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:34.860408Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi4pl84-rqdcw Pod phase: Pending ------------------------------ • Failure in Spec Setup (BeforeEach) [181.082 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 with CPU spec /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:294 when CPU model defined [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:340 should report defined CPU model /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:341 Timed out after 90.016s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:24:30.426385Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi4pl84 kind=VirtualMachineInstance uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi4pl84-rqdcw" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:28:33 http: TLS handshake error from 10.129.0.1:45038: EOF 2018/08/03 21:28:43 http: TLS handshake error from 10.129.0.1:45050: EOF level=info timestamp=2018-08-03T21:28:49.649516Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:28:53 http: TLS handshake error from 10.129.0.1:45062: EOF 2018/08/03 21:29:03 http: TLS handshake error from 10.129.0.1:45074: EOF 2018/08/03 21:29:13 http: TLS handshake error from 10.129.0.1:45086: EOF level=info timestamp=2018-08-03T21:29:19.784051Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:29:23 http: TLS handshake error from 10.129.0.1:45098: EOF 2018/08/03 21:29:33 http: TLS handshake error from 10.129.0.1:45110: EOF 2018/08/03 21:29:43 http: TLS handshake error from 10.129.0.1:45122: EOF 2018/08/03 21:29:53 http: TLS handshake error from 10.129.0.1:45134: EOF 2018/08/03 21:30:03 http: TLS handshake error from 10.129.0.1:45146: EOF 2018/08/03 21:30:13 http: TLS handshake error from 10.129.0.1:45158: EOF level=info timestamp=2018-08-03T21:30:19.542647Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:30:23 http: TLS handshake error from 10.129.0.1:45170: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:29:47.124746Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=10s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:29:49.620914Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:29:49.845795Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:29:52.718458Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:29:55 http: TLS handshake error from 10.129.0.1:51716: EOF level=info timestamp=2018-08-03T21:30:00.116006Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:30:03.255845Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:30:05 http: TLS handshake error from 10.129.0.1:51728: EOF level=info timestamp=2018-08-03T21:30:10.337121Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:30:10.512372Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:30:15 http: TLS handshake error from 10.129.0.1:51740: EOF level=info timestamp=2018-08-03T21:30:20.546660Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:30:22.836439Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:30:25 http: TLS handshake error from 10.129.0.1:51752: EOF level=info timestamp=2018-08-03T21:30:30.693368Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:21:27.989803Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmir2hkz kind= uid=2eb0a90c-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:21:27.990444Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmir2hkz kind= uid=2eb0a90c-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:29.009890Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6vzk8 kind= uid=9a96fb9e-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:24:29.012304Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6vzk8 kind= uid=9a96fb9e-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:29.134703Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:29.185808Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:30.263433Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi6vzk8, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 9a96fb9e-9763-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:30.475121Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:24:30.475516Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:30.661579Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.862892Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.904436Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:27:31.466434Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:27:31.467467Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:27:31.590110Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmigt4mn\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmigt4mn" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=info timestamp=2018-08-03T21:17:36.056369Z pos=virt-handler.go:89 component=virt-handler hostname=node02 level=info timestamp=2018-08-03T21:17:36.061203Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:36.361779Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:36.362119Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:36.363390Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:24.464403Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:40.551681Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:45.116321Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:08.055346Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=info timestamp=2018-08-03T21:17:30.265794Z pos=virt-handler.go:89 component=virt-handler hostname=node01 level=info timestamp=2018-08-03T21:17:30.286009Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:30.487255Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:30.487883Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:30.489165Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:18.454344Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:34.860408Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:39.724203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:02.696667Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmigt4mn-qgskv Pod phase: Pending • Failure in Spec Setup (BeforeEach) [180.936 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 with CPU spec /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:294 when CPU model equals to passthrough [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:368 should report exactly the same model as node CPU /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:369 Timed out after 90.018s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:27:31.342889Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmigt4mn kind=VirtualMachineInstance uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmigt4mn-qgskv" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:31:13 http: TLS handshake error from 10.129.0.1:45232: EOF 2018/08/03 21:31:23 http: TLS handshake error from 10.129.0.1:45244: EOF 2018/08/03 21:31:33 http: TLS handshake error from 10.129.0.1:45256: EOF 2018/08/03 21:31:43 http: TLS handshake error from 10.129.0.1:45268: EOF 2018/08/03 21:31:53 http: TLS handshake error from 10.129.0.1:45280: EOF 2018/08/03 21:32:03 http: TLS handshake error from 10.129.0.1:45292: EOF 2018/08/03 21:32:13 http: TLS handshake error from 10.129.0.1:45304: EOF 2018/08/03 21:32:23 http: TLS handshake error from 10.129.0.1:45316: EOF 2018/08/03 21:32:33 http: TLS handshake error from 10.129.0.1:45328: EOF 2018/08/03 21:32:43 http: TLS handshake error from 10.129.0.1:45340: EOF level=info timestamp=2018-08-03T21:32:47.211895Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:32:53 http: TLS handshake error from 10.129.0.1:45352: EOF 2018/08/03 21:33:03 http: TLS handshake error from 10.129.0.1:45364: EOF 2018/08/03 21:33:13 http: TLS handshake error from 10.129.0.1:45376: EOF 2018/08/03 21:33:23 http: TLS handshake error from 10.129.0.1:45388: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:32:53.867383Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:32:54.082215Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:32:55 http: TLS handshake error from 10.129.0.1:51934: EOF level=info timestamp=2018-08-03T21:33:04.499710Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:33:05.059866Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:33:05 http: TLS handshake error from 10.129.0.1:51946: EOF level=info timestamp=2018-08-03T21:33:11.024985Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:33:11.052707Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:33:12.547703Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:33:14.673014Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:33:15 http: TLS handshake error from 10.129.0.1:51958: EOF level=info timestamp=2018-08-03T21:33:19.572367Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:33:24.370959Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:33:24.828745Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:33:25 http: TLS handshake error from 10.129.0.1:51970: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:24:29.012304Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6vzk8 kind= uid=9a96fb9e-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:29.134703Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:29.185808Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:30.263433Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi6vzk8, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 9a96fb9e-9763-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:30.475121Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:24:30.475516Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:30.661579Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.862892Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.904436Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:27:31.466434Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:27:31.467467Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:27:31.590110Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmigt4mn\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmigt4mn" level=info timestamp=2018-08-03T21:30:32.514315Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9bh2x kind= uid=73414e2b-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:30:32.516390Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9bh2x kind= uid=73414e2b-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:30:32.698635Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9bh2x\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9bh2x" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=info timestamp=2018-08-03T21:17:36.061203Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:36.361779Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:36.362119Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:36.363390Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:24.464403Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:40.551681Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:45.116321Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:08.055346Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:41.899869Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:32:02.318803Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=info timestamp=2018-08-03T21:17:30.286009Z pos=virtinformers.go:107 component=virt-handler msg="STARTING informer configMapInformer" level=info timestamp=2018-08-03T21:17:30.487255Z pos=vm.go:208 component=virt-handler msg="Starting virt-handler controller." level=error timestamp=2018-08-03T21:17:30.487883Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:30.489165Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:18.454344Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:34.860408Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:39.724203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:02.696667Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:36.547122Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:31:57.038514Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi9bh2x-rjtjr Pod phase: Pending • Failure in Spec Setup (BeforeEach) [180.878 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 with CPU spec /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:294 when CPU model not defined [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:392 should report CPU model from libvirt capabilities /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:393 Timed out after 90.015s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:30:32.448450Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi9bh2x kind=VirtualMachineInstance uid=73414e2b-9764-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi9bh2x-rjtjr" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:34:23 http: TLS handshake error from 10.129.0.1:45462: EOF 2018/08/03 21:34:33 http: TLS handshake error from 10.129.0.1:45474: EOF 2018/08/03 21:34:43 http: TLS handshake error from 10.129.0.1:45486: EOF 2018/08/03 21:34:53 http: TLS handshake error from 10.129.0.1:45498: EOF 2018/08/03 21:35:03 http: TLS handshake error from 10.129.0.1:45510: EOF 2018/08/03 21:35:13 http: TLS handshake error from 10.129.0.1:45522: EOF level=info timestamp=2018-08-03T21:35:19.541288Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:35:23 http: TLS handshake error from 10.129.0.1:45534: EOF 2018/08/03 21:35:33 http: TLS handshake error from 10.129.0.1:45546: EOF 2018/08/03 21:35:43 http: TLS handshake error from 10.129.0.1:45558: EOF 2018/08/03 21:35:53 http: TLS handshake error from 10.129.0.1:45570: EOF 2018/08/03 21:36:03 http: TLS handshake error from 10.129.0.1:45582: EOF 2018/08/03 21:36:13 http: TLS handshake error from 10.129.0.1:45594: EOF 2018/08/03 21:36:23 http: TLS handshake error from 10.129.0.1:45606: EOF 2018/08/03 21:36:33 http: TLS handshake error from 10.129.0.1:45618: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running 2018/08/03 21:35:55 http: TLS handshake error from 10.129.0.1:52152: EOF level=info timestamp=2018-08-03T21:35:55.850642Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:35:58.676387Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:36:05 http: TLS handshake error from 10.129.0.1:52164: EOF level=info timestamp=2018-08-03T21:36:06.824876Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:36:08.888051Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:36:11.090505Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:36:11.095751Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:36:14.072706Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:36:15 http: TLS handshake error from 10.129.0.1:52176: EOF level=info timestamp=2018-08-03T21:36:19.134328Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:36:19.709943Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:36:25 http: TLS handshake error from 10.129.0.1:52188: EOF level=info timestamp=2018-08-03T21:36:26.239473Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:36:29.538351Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:24:30.263433Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6vzk8\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi6vzk8, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 9a96fb9e-9763-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6vzk8" level=info timestamp=2018-08-03T21:24:30.475121Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:24:30.475516Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4pl84 kind= uid=9b7737f6-9763-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:24:30.661579Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.862892Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:24:30.904436Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4pl84\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4pl84" level=info timestamp=2018-08-03T21:27:31.466434Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:27:31.467467Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:27:31.590110Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmigt4mn\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmigt4mn" level=info timestamp=2018-08-03T21:30:32.514315Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9bh2x kind= uid=73414e2b-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:30:32.516390Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9bh2x kind= uid=73414e2b-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:30:32.698635Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9bh2x\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9bh2x" level=info timestamp=2018-08-03T21:33:33.190856Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9bh2x\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi9bh2x, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 73414e2b-9764-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9bh2x" level=info timestamp=2018-08-03T21:33:33.292422Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidpqjn kind= uid=df02dbf7-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:33:33.293919Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidpqjn kind= uid=df02dbf7-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:17:36.362119Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:36.363390Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:24.464403Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:40.551681Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:45.116321Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:08.055346Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:41.899869Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:32:02.318803Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:17:30.487883Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=info timestamp=2018-08-03T21:17:30.489165Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:18.454344Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:34.860408Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:39.724203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:02.696667Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:36.547122Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:31:57.038514Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmidpqjn-2rp6h Pod phase: Pending • Failure [180.795 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:44 New VirtualMachineInstance with all supported drives /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:413 should have all the device nodes [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_configuration_test.go:436 Timed out after 90.012s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:33:33.202682Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmidpqjn kind=VirtualMachineInstance uid=df02dbf7-9764-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmidpqjn-2rp6h" • Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:37:23 http: TLS handshake error from 10.129.0.1:45680: EOF 2018/08/03 21:37:33 http: TLS handshake error from 10.129.0.1:45692: EOF 2018/08/03 21:37:43 http: TLS handshake error from 10.129.0.1:45704: EOF 2018/08/03 21:37:53 http: TLS handshake error from 10.129.0.1:45716: EOF 2018/08/03 21:38:03 http: TLS handshake error from 10.129.0.1:45728: EOF 2018/08/03 21:38:13 http: TLS handshake error from 10.129.0.1:45740: EOF 2018/08/03 21:38:23 http: TLS handshake error from 10.129.0.1:45752: EOF 2018/08/03 21:38:33 http: TLS handshake error from 10.129.0.1:45764: EOF 2018/08/03 21:38:43 http: TLS handshake error from 10.129.0.1:45776: EOF 2018/08/03 21:38:53 http: TLS handshake error from 10.129.0.1:45788: EOF 2018/08/03 21:39:03 http: TLS handshake error from 10.129.0.1:45800: EOF 2018/08/03 21:39:13 http: TLS handshake error from 10.129.0.1:45812: EOF level=info timestamp=2018-08-03T21:39:19.595054Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:39:23 http: TLS handshake error from 10.129.0.1:45824: EOF 2018/08/03 21:39:33 http: TLS handshake error from 10.129.0.1:45836: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:38:52.089895Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:38:55 http: TLS handshake error from 10.129.0.1:52370: EOF level=info timestamp=2018-08-03T21:38:57.312282Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:39:02.292067Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:39:05 http: TLS handshake error from 10.129.0.1:52382: EOF level=info timestamp=2018-08-03T21:39:08.468874Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:39:12.776734Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:39:13.675056Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:39:13.679056Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 21:39:15 http: TLS handshake error from 10.129.0.1:52394: EOF level=info timestamp=2018-08-03T21:39:15.816642Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:39:22.888461Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:39:25 http: TLS handshake error from 10.129.0.1:52406: EOF level=info timestamp=2018-08-03T21:39:27.590525Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:39:33.054573Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:27:31.466434Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:27:31.467467Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmigt4mn kind= uid=0758dfd5-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:27:31.590110Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmigt4mn\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmigt4mn" level=info timestamp=2018-08-03T21:30:32.514315Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9bh2x kind= uid=73414e2b-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:30:32.516390Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9bh2x kind= uid=73414e2b-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:30:32.698635Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9bh2x\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9bh2x" level=info timestamp=2018-08-03T21:33:33.190856Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9bh2x\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi9bh2x, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 73414e2b-9764-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9bh2x" level=info timestamp=2018-08-03T21:33:33.292422Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidpqjn kind= uid=df02dbf7-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:33:33.293919Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidpqjn kind= uid=df02dbf7-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:36:34.252658Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxjjk kind= uid=4adf90db-9765-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:36:34.253551Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxjjk kind= uid=4adf90db-9765-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:36:34.363751Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmikxjjk\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmikxjjk" level=info timestamp=2018-08-03T21:36:34.418729Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmikxjjk\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmikxjjk, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 4adf90db-9765-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmikxjjk" level=info timestamp=2018-08-03T21:36:34.490579Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixppwq kind= uid=4b0464be-9765-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:36:34.490943Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixppwq kind= uid=4b0464be-9765-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=info timestamp=2018-08-03T21:17:36.363390Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:24.464403Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:40.551681Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:45.116321Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:08.055346Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:41.899869Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:32:02.318803Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=info timestamp=2018-08-03T21:17:30.489165Z pos=cache.go:151 component=virt-handler msg="Synchronizing domains" level=error timestamp=2018-08-03T21:19:18.454344Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:34.860408Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:39.724203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:02.696667Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:36.547122Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:31:57.038514Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmixppwq-6hnqd Pod phase: Pending ------------------------------ • Failure [181.134 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 should start it [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:78 Timed out after 90.018s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:36:34.319469Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmixppwq kind=VirtualMachineInstance uid=4b0464be-9765-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmixppwq-6hnqd" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:40:33 http: TLS handshake error from 10.129.0.1:45910: EOF 2018/08/03 21:40:43 http: TLS handshake error from 10.129.0.1:45922: EOF 2018/08/03 21:40:53 http: TLS handshake error from 10.129.0.1:45934: EOF 2018/08/03 21:41:03 http: TLS handshake error from 10.129.0.1:45946: EOF 2018/08/03 21:41:13 http: TLS handshake error from 10.129.0.1:45958: EOF 2018/08/03 21:41:23 http: TLS handshake error from 10.129.0.1:45970: EOF 2018/08/03 21:41:33 http: TLS handshake error from 10.129.0.1:45982: EOF 2018/08/03 21:41:43 http: TLS handshake error from 10.129.0.1:45994: EOF level=info timestamp=2018-08-03T21:41:49.622777Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:41:53 http: TLS handshake error from 10.129.0.1:46006: EOF 2018/08/03 21:42:03 http: TLS handshake error from 10.129.0.1:46018: EOF 2018/08/03 21:42:13 http: TLS handshake error from 10.129.0.1:46030: EOF level=info timestamp=2018-08-03T21:42:19.600569Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:42:23 http: TLS handshake error from 10.129.0.1:46042: EOF 2018/08/03 21:42:33 http: TLS handshake error from 10.129.0.1:46054: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:41:51.202610Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=10s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:41:51.218582Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=10s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:41:55 http: TLS handshake error from 10.129.0.1:52588: EOF level=info timestamp=2018-08-03T21:41:56.126623Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:41:59.028577Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:42:05 http: TLS handshake error from 10.129.0.1:52600: EOF level=info timestamp=2018-08-03T21:42:06.363029Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:42:10.191766Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:42:15 http: TLS handshake error from 10.129.0.1:52612: EOF level=info timestamp=2018-08-03T21:42:16.664239Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:42:17.722500Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:42:25 http: TLS handshake error from 10.129.0.1:52624: EOF level=info timestamp=2018-08-03T21:42:26.924382Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:42:29.248905Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:42:35 http: TLS handshake error from 10.129.0.1:52636: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:33:33.190856Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9bh2x\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi9bh2x, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 73414e2b-9764-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9bh2x" level=info timestamp=2018-08-03T21:33:33.292422Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidpqjn kind= uid=df02dbf7-9764-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:33:33.293919Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidpqjn kind= uid=df02dbf7-9764-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:36:34.252658Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxjjk kind= uid=4adf90db-9765-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:36:34.253551Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxjjk kind= uid=4adf90db-9765-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:36:34.363751Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmikxjjk\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmikxjjk" level=info timestamp=2018-08-03T21:36:34.418729Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmikxjjk\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmikxjjk, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 4adf90db-9765-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmikxjjk" level=info timestamp=2018-08-03T21:36:34.490579Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixppwq kind= uid=4b0464be-9765-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:36:34.490943Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixppwq kind= uid=4b0464be-9765-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:39:35.642736Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixppwq\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmixppwq, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 4b0464be-9765-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixppwq" level=info timestamp=2018-08-03T21:39:35.824530Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4h8f6 kind= uid=b716e407-9765-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:39:35.825868Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4h8f6 kind= uid=b716e407-9765-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:39:36.184122Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4h8f6\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4h8f6" level=info timestamp=2018-08-03T21:39:36.241942Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4h8f6\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4h8f6" level=info timestamp=2018-08-03T21:39:36.274221Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4h8f6\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4h8f6" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:20:29.252241Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:40.551681Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:45.116321Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:08.055346Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:41.899869Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:32:02.318803Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:20:23.465505Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:21:34.860408Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:39.724203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:02.696667Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:36.547122Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:31:57.038514Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi4h8f6-mrfsj Pod phase: Pending • Failure [180.961 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 should attach virt-launcher to it [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:84 Timed out after 90.010s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:39:35.898536Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi4h8f6 kind=VirtualMachineInstance uid=b716e407-9765-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi4h8f6-mrfsj" •••• Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:43:43 http: TLS handshake error from 10.129.0.1:46140: EOF 2018/08/03 21:43:53 http: TLS handshake error from 10.129.0.1:46152: EOF 2018/08/03 21:44:03 http: TLS handshake error from 10.129.0.1:46164: EOF 2018/08/03 21:44:13 http: TLS handshake error from 10.129.0.1:46176: EOF level=info timestamp=2018-08-03T21:44:19.635005Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:44:23 http: TLS handshake error from 10.129.0.1:46188: EOF 2018/08/03 21:44:33 http: TLS handshake error from 10.129.0.1:46200: EOF 2018/08/03 21:44:43 http: TLS handshake error from 10.129.0.1:46212: EOF level=info timestamp=2018-08-03T21:44:49.705727Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:44:53 http: TLS handshake error from 10.129.0.1:46224: EOF 2018/08/03 21:45:03 http: TLS handshake error from 10.129.0.1:46236: EOF 2018/08/03 21:45:13 http: TLS handshake error from 10.129.0.1:46248: EOF level=info timestamp=2018-08-03T21:45:19.688692Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:45:23 http: TLS handshake error from 10.129.0.1:46260: EOF 2018/08/03 21:45:33 http: TLS handshake error from 10.129.0.1:46272: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:44:52.523083Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=10s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:44:52.542509Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=10s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:44:55 http: TLS handshake error from 10.129.0.1:52806: EOF level=info timestamp=2018-08-03T21:45:00.015062Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:45:00.257084Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:45:05 http: TLS handshake error from 10.129.0.1:52818: EOF level=info timestamp=2018-08-03T21:45:10.411255Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:45:11.944594Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:45:15 http: TLS handshake error from 10.129.0.1:52830: EOF level=info timestamp=2018-08-03T21:45:19.272856Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:45:20.715569Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:45:25 http: TLS handshake error from 10.129.0.1:52842: EOF level=info timestamp=2018-08-03T21:45:30.149888Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:45:30.920683Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:45:35 http: TLS handshake error from 10.129.0.1:52854: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:39:36.274221Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4h8f6\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4h8f6" level=info timestamp=2018-08-03T21:42:36.549810Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4h8f6\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi4h8f6, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: b716e407-9765-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4h8f6" level=info timestamp=2018-08-03T21:42:37.147979Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7dnrj kind= uid=232b33e3-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:42:37.148627Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7dnrj kind= uid=232b33e3-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:42:37.279520Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7dnrj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7dnrj" level=info timestamp=2018-08-03T21:42:37.343581Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7dnrj\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi7dnrj, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 232b33e3-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7dnrj" level=info timestamp=2018-08-03T21:42:37.459467Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip7zwr kind= uid=235c674c-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:42:37.459799Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip7zwr kind= uid=235c674c-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:42:37.586036Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmip7zwr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmip7zwr" level=info timestamp=2018-08-03T21:42:37.625420Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmip7zwr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmip7zwr" level=info timestamp=2018-08-03T21:42:37.701853Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmip7zwr\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmip7zwr, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 235c674c-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmip7zwr" level=info timestamp=2018-08-03T21:42:37.774778Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixck7m kind= uid=238d115c-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:42:37.775228Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixck7m kind= uid=238d115c-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:42:37.892716Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixck7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixck7m" level=info timestamp=2018-08-03T21:42:37.955324Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixck7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixck7m" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:22:47.606227Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:09.357397Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:45.116321Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:08.055346Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:41.899869Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:32:02.318803Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:39.613831Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:22:41.943195Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:24:03.747460Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:39.724203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:02.696667Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:36.547122Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:31:57.038514Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmixck7m-prt9c Pod phase: Pending ------------------------------ • Failure [180.771 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 with boot order /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:172 should be able to boot from selected disk /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 Alpine as first boot [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.013s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: defining a VirtualMachineInstance with an Alpine disk STEP: adding a Cirros Disk STEP: setting boot order STEP: starting VirtualMachineInstance STEP: Waiting the VirtualMachineInstance start level=info timestamp=2018-08-03T21:42:37.662283Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmixck7m kind=VirtualMachineInstance uid=238d115c-9766-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmixck7m-prt9c" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:46:23 http: TLS handshake error from 10.129.0.1:46334: EOF 2018/08/03 21:46:33 http: TLS handshake error from 10.129.0.1:46346: EOF 2018/08/03 21:46:43 http: TLS handshake error from 10.129.0.1:46358: EOF level=info timestamp=2018-08-03T21:46:49.626202Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:46:53 http: TLS handshake error from 10.129.0.1:46370: EOF 2018/08/03 21:47:03 http: TLS handshake error from 10.129.0.1:46382: EOF 2018/08/03 21:47:13 http: TLS handshake error from 10.129.0.1:46394: EOF 2018/08/03 21:47:23 http: TLS handshake error from 10.129.0.1:46406: EOF 2018/08/03 21:47:33 http: TLS handshake error from 10.129.0.1:46418: EOF 2018/08/03 21:47:43 http: TLS handshake error from 10.129.0.1:46430: EOF 2018/08/03 21:47:53 http: TLS handshake error from 10.129.0.1:46442: EOF 2018/08/03 21:48:03 http: TLS handshake error from 10.129.0.1:46454: EOF 2018/08/03 21:48:13 http: TLS handshake error from 10.129.0.1:46466: EOF 2018/08/03 21:48:23 http: TLS handshake error from 10.129.0.1:46478: EOF 2018/08/03 21:48:33 http: TLS handshake error from 10.129.0.1:46490: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:48:01.126051Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:48:03.817402Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:48:05 http: TLS handshake error from 10.129.0.1:53036: EOF level=info timestamp=2018-08-03T21:48:13.293215Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:48:13.728696Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:48:13.732520Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:48:14.268834Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:48:15 http: TLS handshake error from 10.129.0.1:53048: EOF level=info timestamp=2018-08-03T21:48:19.640598Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:48:20.544273Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:48:24.493914Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:48:25 http: TLS handshake error from 10.129.0.1:53060: EOF level=info timestamp=2018-08-03T21:48:31.508843Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:48:34.586208Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:48:35 http: TLS handshake error from 10.129.0.1:53072: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:42:37.148627Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi7dnrj kind= uid=232b33e3-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:42:37.279520Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7dnrj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7dnrj" level=info timestamp=2018-08-03T21:42:37.343581Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi7dnrj\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi7dnrj, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 232b33e3-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi7dnrj" level=info timestamp=2018-08-03T21:42:37.459467Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip7zwr kind= uid=235c674c-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:42:37.459799Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip7zwr kind= uid=235c674c-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:42:37.586036Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmip7zwr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmip7zwr" level=info timestamp=2018-08-03T21:42:37.625420Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmip7zwr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmip7zwr" level=info timestamp=2018-08-03T21:42:37.701853Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmip7zwr\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmip7zwr, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 235c674c-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmip7zwr" level=info timestamp=2018-08-03T21:42:37.774778Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixck7m kind= uid=238d115c-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:42:37.775228Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixck7m kind= uid=238d115c-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:42:37.892716Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixck7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixck7m" level=info timestamp=2018-08-03T21:42:37.955324Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixck7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixck7m" level=info timestamp=2018-08-03T21:45:38.452432Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixck7m\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmixck7m, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 238d115c-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixck7m" level=info timestamp=2018-08-03T21:45:38.595902Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipzxzt kind= uid=8f53cc62-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:45:38.596486Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipzxzt kind= uid=8f53cc62-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:25:46.495463Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:45.116321Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:08.055346Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:41.899869Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:32:02.318803Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:39.613831Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:25:41.035844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:27:39.724203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:29:02.696667Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:30:36.547122Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:31:57.038514Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmipzxzt-9h69q Pod phase: Pending • Failure [180.823 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 with boot order /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:172 should be able to boot from selected disk /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 Cirros as first boot [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.010s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: defining a VirtualMachineInstance with an Alpine disk STEP: adding a Cirros Disk STEP: setting boot order STEP: starting VirtualMachineInstance STEP: Waiting the VirtualMachineInstance start level=info timestamp=2018-08-03T21:45:38.477121Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmipzxzt kind=VirtualMachineInstance uid=8f53cc62-9766-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmipzxzt-9h69q" • [SLOW TEST:60.392 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:203 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:204 should retry starting the VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:205 ------------------------------ •! Panic [60.306 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:203 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:204 should log warning and proceed once the secret is there [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:235 Test Panicked runtime error: invalid memory address or nil pointer dereference /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/panic.go:505 Full Stack Trace /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/panic.go:505 +0x229 kubevirt.io/kubevirt/tests_test.glob..func16.3.9.1.2() /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:256 +0x431 kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/leafnodes.(*runner).runSync(0xc42012a8a0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /root/go/src/kubevirt.io/kubevirt/tests/tests_suite_test.go:43 +0xaa testing.tRunner(0xc42067e5a0, 0x1339680) /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:777 +0xd0 created by testing.(*T).Run /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:824 +0x2e0 ------------------------------ STEP: Starting a VirtualMachineInstance STEP: Checking that VirtualMachineInstance start failed level=info timestamp=2018-08-03T21:49:39.715213Z pos=utils.go:255 component=tests namespace=kubevirt-test-default name=testvmin2nkg kind=VirtualMachineInstance uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmin2nkg-6pmrl" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:52:03 http: TLS handshake error from 10.129.0.1:46746: EOF 2018/08/03 21:52:13 http: TLS handshake error from 10.129.0.1:46758: EOF level=info timestamp=2018-08-03T21:52:19.644790Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:52:23 http: TLS handshake error from 10.129.0.1:46770: EOF 2018/08/03 21:52:33 http: TLS handshake error from 10.129.0.1:46782: EOF 2018/08/03 21:52:43 http: TLS handshake error from 10.129.0.1:46794: EOF level=info timestamp=2018-08-03T21:52:47.206493Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:52:49.028564Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:52:49.589841Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:52:53 http: TLS handshake error from 10.129.0.1:46806: EOF 2018/08/03 21:53:03 http: TLS handshake error from 10.129.0.1:46818: EOF 2018/08/03 21:53:13 http: TLS handshake error from 10.129.0.1:46830: EOF level=info timestamp=2018-08-03T21:53:19.576919Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:53:23 http: TLS handshake error from 10.129.0.1:46842: EOF 2018/08/03 21:53:33 http: TLS handshake error from 10.129.0.1:46854: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running 2018/08/03 21:52:55 http: TLS handshake error from 10.129.0.1:53388: EOF level=info timestamp=2018-08-03T21:53:00.643263Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:53:04.384290Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:53:05 http: TLS handshake error from 10.129.0.1:53400: EOF level=info timestamp=2018-08-03T21:53:11.269621Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:53:13.263549Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:53:13.271685Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 21:53:15 http: TLS handshake error from 10.129.0.1:53412: EOF level=info timestamp=2018-08-03T21:53:16.229427Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:53:21.554674Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:53:23.037911Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:53:25 http: TLS handshake error from 10.129.0.1:53424: EOF level=info timestamp=2018-08-03T21:53:31.784095Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:53:34.504265Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:53:35 http: TLS handshake error from 10.129.0.1:53436: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:42:37.955324Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixck7m\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixck7m" level=info timestamp=2018-08-03T21:45:38.452432Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixck7m\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmixck7m, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 238d115c-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixck7m" level=info timestamp=2018-08-03T21:45:38.595902Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipzxzt kind= uid=8f53cc62-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:45:38.596486Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmipzxzt kind= uid=8f53cc62-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:48:39.271430Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmipzxzt\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmipzxzt, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 8f53cc62-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmipzxzt" level=info timestamp=2018-08-03T21:48:39.480607Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9hprn kind= uid=fb244d57-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:48:39.481300Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9hprn kind= uid=fb244d57-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:39.620931Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9hprn\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi9hprn, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: fb244d57-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9hprn" level=info timestamp=2018-08-03T21:49:39.782333Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:49:39.783106Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:40.005460Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.111282Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmin2nkg, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1f14a30d-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.294831Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:50:40.295560Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:50:40.714244Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:30:41.899869Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:32:02.318803Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:39.613831Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:31:57.038514Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmifvnf9-dwr9w Pod phase: Pending • Failure [180.970 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 when virt-launcher crashes /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:283 should be stopped and have Failed phase [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:284 Timed out after 90.013s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:50:40.352164Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmifvnf9 kind=VirtualMachineInstance uid=43265a57-9767-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmifvnf9-dwr9w" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:54:23 http: TLS handshake error from 10.129.0.1:46914: EOF 2018/08/03 21:54:33 http: TLS handshake error from 10.129.0.1:46926: EOF 2018/08/03 21:54:43 http: TLS handshake error from 10.129.0.1:46938: EOF 2018/08/03 21:54:53 http: TLS handshake error from 10.129.0.1:46952: EOF 2018/08/03 21:55:03 http: TLS handshake error from 10.129.0.1:46964: EOF 2018/08/03 21:55:13 http: TLS handshake error from 10.129.0.1:46976: EOF level=info timestamp=2018-08-03T21:55:19.593869Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:55:23 http: TLS handshake error from 10.129.0.1:46988: EOF 2018/08/03 21:55:33 http: TLS handshake error from 10.129.0.1:47000: EOF 2018/08/03 21:55:43 http: TLS handshake error from 10.129.0.1:47012: EOF 2018/08/03 21:55:53 http: TLS handshake error from 10.129.0.1:47024: EOF 2018/08/03 21:56:03 http: TLS handshake error from 10.129.0.1:47036: EOF 2018/08/03 21:56:13 http: TLS handshake error from 10.129.0.1:47048: EOF 2018/08/03 21:56:23 http: TLS handshake error from 10.129.0.1:47060: EOF 2018/08/03 21:56:33 http: TLS handshake error from 10.129.0.1:47072: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:56:04.973344Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:05 http: TLS handshake error from 10.129.0.1:53618: EOF level=info timestamp=2018-08-03T21:56:05.971696Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:56:13.294593Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:56:13.298495Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:56:15.195258Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:15 http: TLS handshake error from 10.129.0.1:53630: EOF level=info timestamp=2018-08-03T21:56:17.951855Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:56:19.737714Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:56:24.668025Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:25 http: TLS handshake error from 10.129.0.1:53642: EOF level=info timestamp=2018-08-03T21:56:25.456065Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:35 http: TLS handshake error from 10.129.0.1:53654: EOF level=info timestamp=2018-08-03T21:56:35.653603Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:56:36.224662Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:48:39.271430Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmipzxzt\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmipzxzt, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 8f53cc62-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmipzxzt" level=info timestamp=2018-08-03T21:48:39.480607Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9hprn kind= uid=fb244d57-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:48:39.481300Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9hprn kind= uid=fb244d57-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:39.620931Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9hprn\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi9hprn, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: fb244d57-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9hprn" level=info timestamp=2018-08-03T21:49:39.782333Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:49:39.783106Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:40.005460Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.111282Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmin2nkg, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1f14a30d-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.294831Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:50:40.295560Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:50:40.714244Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.028178Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmifvnf9, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 43265a57-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.192668Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:53:41.193273Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:53:41.299488Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5l7kj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5l7kj" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:39.613831Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi5l7kj-59tlt Pod phase: Pending • Failure [180.821 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 when virt-handler crashes /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:307 should recover and continue management [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:308 Timed out after 90.012s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:53:41.016632Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi5l7kj kind=VirtualMachineInstance uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi5l7kj-59tlt" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:54:23 http: TLS handshake error from 10.129.0.1:46914: EOF 2018/08/03 21:54:33 http: TLS handshake error from 10.129.0.1:46926: EOF 2018/08/03 21:54:43 http: TLS handshake error from 10.129.0.1:46938: EOF 2018/08/03 21:54:53 http: TLS handshake error from 10.129.0.1:46952: EOF 2018/08/03 21:55:03 http: TLS handshake error from 10.129.0.1:46964: EOF 2018/08/03 21:55:13 http: TLS handshake error from 10.129.0.1:46976: EOF level=info timestamp=2018-08-03T21:55:19.593869Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:55:23 http: TLS handshake error from 10.129.0.1:46988: EOF 2018/08/03 21:55:33 http: TLS handshake error from 10.129.0.1:47000: EOF 2018/08/03 21:55:43 http: TLS handshake error from 10.129.0.1:47012: EOF 2018/08/03 21:55:53 http: TLS handshake error from 10.129.0.1:47024: EOF 2018/08/03 21:56:03 http: TLS handshake error from 10.129.0.1:47036: EOF 2018/08/03 21:56:13 http: TLS handshake error from 10.129.0.1:47048: EOF 2018/08/03 21:56:23 http: TLS handshake error from 10.129.0.1:47060: EOF 2018/08/03 21:56:33 http: TLS handshake error from 10.129.0.1:47072: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:56:04.973344Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:05 http: TLS handshake error from 10.129.0.1:53618: EOF level=info timestamp=2018-08-03T21:56:05.971696Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:56:13.294593Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:56:13.298495Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:56:15.195258Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:15 http: TLS handshake error from 10.129.0.1:53630: EOF level=info timestamp=2018-08-03T21:56:17.951855Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:56:19.737714Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:56:24.668025Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:25 http: TLS handshake error from 10.129.0.1:53642: EOF level=info timestamp=2018-08-03T21:56:25.456065Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:56:35 http: TLS handshake error from 10.129.0.1:53654: EOF level=info timestamp=2018-08-03T21:56:35.653603Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:56:36.224662Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:48:39.271430Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmipzxzt\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmipzxzt, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 8f53cc62-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmipzxzt" level=info timestamp=2018-08-03T21:48:39.480607Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9hprn kind= uid=fb244d57-9766-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:48:39.481300Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9hprn kind= uid=fb244d57-9766-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:39.620931Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9hprn\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi9hprn, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: fb244d57-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9hprn" level=info timestamp=2018-08-03T21:49:39.782333Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:49:39.783106Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:40.005460Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.111282Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmin2nkg, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1f14a30d-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.294831Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:50:40.295560Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:50:40.714244Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.028178Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmifvnf9, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 43265a57-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.192668Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:53:41.193273Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:53:41.299488Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5l7kj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5l7kj" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:33:23.504047Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:38.190675Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:39.613831Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:33:18.236120Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:34:32.949635Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:36:14.108332Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" • Failure [0.559 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 when virt-handler is responsive /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:338 should indicate that a node is ready for vmis [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:339 Expected <[]v1.Node | len:0, cap:0>: [] not to be empty /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:344 ------------------------------ STEP: adding a heartbeat annotation and a schedulable label to the node Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 21:57:33 http: TLS handshake error from 10.129.0.1:47144: EOF 2018/08/03 21:57:43 http: TLS handshake error from 10.129.0.1:47156: EOF level=info timestamp=2018-08-03T21:57:49.682941Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:57:53 http: TLS handshake error from 10.129.0.1:47168: EOF 2018/08/03 21:58:03 http: TLS handshake error from 10.129.0.1:47182: EOF 2018/08/03 21:58:13 http: TLS handshake error from 10.129.0.1:47194: EOF 2018/08/03 21:58:23 http: TLS handshake error from 10.129.0.1:47206: EOF 2018/08/03 21:58:33 http: TLS handshake error from 10.129.0.1:47218: EOF 2018/08/03 21:58:43 http: TLS handshake error from 10.129.0.1:47230: EOF 2018/08/03 21:58:53 http: TLS handshake error from 10.129.0.1:47242: EOF 2018/08/03 21:59:03 http: TLS handshake error from 10.129.0.1:47254: EOF 2018/08/03 21:59:13 http: TLS handshake error from 10.129.0.1:47266: EOF level=info timestamp=2018-08-03T21:59:19.705063Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 21:59:23 http: TLS handshake error from 10.129.0.1:47278: EOF 2018/08/03 21:59:33 http: TLS handshake error from 10.129.0.1:47290: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T21:58:58.677235Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:59:05 http: TLS handshake error from 10.129.0.1:53836: EOF level=info timestamp=2018-08-03T21:59:07.862322Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:59:08.921780Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:59:11.975885Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T21:59:11.981212Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 21:59:15 http: TLS handshake error from 10.129.0.1:53848: EOF level=info timestamp=2018-08-03T21:59:19.138089Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:59:19.435545Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:59:25 http: TLS handshake error from 10.129.0.1:53860: EOF level=info timestamp=2018-08-03T21:59:26.197588Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:59:29.434917Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 21:59:35 http: TLS handshake error from 10.129.0.1:53872: EOF level=info timestamp=2018-08-03T21:59:38.106954Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T21:59:39.548639Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:49:39.620931Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi9hprn\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi9hprn, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: fb244d57-9766-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi9hprn" level=info timestamp=2018-08-03T21:49:39.782333Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:49:39.783106Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:40.005460Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.111282Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmin2nkg, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1f14a30d-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.294831Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:50:40.295560Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:50:40.714244Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.028178Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmifvnf9, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 43265a57-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.192668Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:53:41.193273Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:53:41.299488Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5l7kj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5l7kj" level=info timestamp=2018-08-03T21:56:42.513234Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiszgnx kind= uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:56:42.514178Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiszgnx kind= uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:56:42.652332Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmiszgnx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmiszgnx" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:36:19.354673Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:38:21.496844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:39.613831Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:58.262918Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:38:16.255056Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmiszgnx-xwqkj Pod phase: Pending • Failure in Spec Setup (BeforeEach) [180.925 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 when virt-handler is not responsive /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:369 the node controller should react [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:408 Timed out after 90.127s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:56:42.374107Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmiszgnx kind=VirtualMachineInstance uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmiszgnx-xwqkj" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running level=info timestamp=2018-08-03T22:00:49.686057Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:00:53 http: TLS handshake error from 10.129.0.1:47386: EOF 2018/08/03 22:01:03 http: TLS handshake error from 10.129.0.1:47400: EOF 2018/08/03 22:01:13 http: TLS handshake error from 10.129.0.1:47412: EOF level=info timestamp=2018-08-03T22:01:19.659966Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:01:23 http: TLS handshake error from 10.129.0.1:47424: EOF 2018/08/03 22:01:33 http: TLS handshake error from 10.129.0.1:47436: EOF 2018/08/03 22:01:43 http: TLS handshake error from 10.129.0.1:47448: EOF level=info timestamp=2018-08-03T22:01:49.646457Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:01:53 http: TLS handshake error from 10.129.0.1:47460: EOF 2018/08/03 22:02:03 http: TLS handshake error from 10.129.0.1:47472: EOF 2018/08/03 22:02:13 http: TLS handshake error from 10.129.0.1:47484: EOF 2018/08/03 22:02:23 http: TLS handshake error from 10.129.0.1:47496: EOF 2018/08/03 22:02:33 http: TLS handshake error from 10.129.0.1:47508: EOF 2018/08/03 22:02:43 http: TLS handshake error from 10.129.0.1:47520: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running 2018/08/03 22:01:55 http: TLS handshake error from 10.129.0.1:54042: EOF level=info timestamp=2018-08-03T22:01:57.789601Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:02:02.928880Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:02:05 http: TLS handshake error from 10.129.0.1:54054: EOF level=info timestamp=2018-08-03T22:02:09.324741Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:02:13.114111Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:02:15 http: TLS handshake error from 10.129.0.1:54066: EOF level=info timestamp=2018-08-03T22:02:19.679984Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:02:21.295800Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:02:23.316962Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:02:25 http: TLS handshake error from 10.129.0.1:54078: EOF level=info timestamp=2018-08-03T22:02:28.079012Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:02:33.604321Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:02:35 http: TLS handshake error from 10.129.0.1:54090: EOF level=info timestamp=2018-08-03T22:02:39.424508Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:49:39.783106Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmin2nkg kind= uid=1f14a30d-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:49:40.005460Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.111282Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmin2nkg\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmin2nkg, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1f14a30d-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmin2nkg" level=info timestamp=2018-08-03T21:50:40.294831Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:50:40.295560Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifvnf9 kind= uid=43265a57-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:50:40.714244Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.028178Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifvnf9\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmifvnf9, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 43265a57-9767-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifvnf9" level=info timestamp=2018-08-03T21:53:41.192668Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:53:41.193273Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:53:41.299488Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5l7kj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5l7kj" level=info timestamp=2018-08-03T21:56:42.513234Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiszgnx kind= uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:56:42.514178Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiszgnx kind= uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:56:42.652332Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmiszgnx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmiszgnx" level=info timestamp=2018-08-03T21:59:43.492452Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimbjf6 kind= uid=86ea68c2-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:59:43.493333Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimbjf6 kind= uid=86ea68c2-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:39:43.005499Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:37.274595Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:39.613831Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:58.262918Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:54.835797Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:40:10.497362Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:41:25.488113Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmimbjf6-bxmzh Pod phase: Pending • Failure [180.920 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 with node tainted /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:461 the vmi with tolerations should be scheduled [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:483 Timed out after 90.018s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T21:59:43.562099Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmimbjf6 kind=VirtualMachineInstance uid=86ea68c2-9768-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmimbjf6-bxmzh" • Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running level=info timestamp=2018-08-03T22:03:49.682204Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:03:53 http: TLS handshake error from 10.129.0.1:47604: EOF 2018/08/03 22:04:03 http: TLS handshake error from 10.129.0.1:47618: EOF 2018/08/03 22:04:13 http: TLS handshake error from 10.129.0.1:47630: EOF 2018/08/03 22:04:23 http: TLS handshake error from 10.129.0.1:47642: EOF 2018/08/03 22:04:33 http: TLS handshake error from 10.129.0.1:47654: EOF 2018/08/03 22:04:43 http: TLS handshake error from 10.129.0.1:47666: EOF level=info timestamp=2018-08-03T22:04:49.650770Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:04:53 http: TLS handshake error from 10.129.0.1:47678: EOF 2018/08/03 22:05:03 http: TLS handshake error from 10.129.0.1:47690: EOF 2018/08/03 22:05:13 http: TLS handshake error from 10.129.0.1:47702: EOF level=info timestamp=2018-08-03T22:05:19.713094Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:05:23 http: TLS handshake error from 10.129.0.1:47714: EOF 2018/08/03 22:05:33 http: TLS handshake error from 10.129.0.1:47726: EOF 2018/08/03 22:05:43 http: TLS handshake error from 10.129.0.1:47738: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:04:56.834403Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:04:59.818761Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:05:05 http: TLS handshake error from 10.129.0.1:54272: EOF level=info timestamp=2018-08-03T22:05:07.096464Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:05:10.744254Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:05:15 http: TLS handshake error from 10.129.0.1:54284: EOF level=info timestamp=2018-08-03T22:05:17.306482Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:05:23.009742Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:05:25 http: TLS handshake error from 10.129.0.1:54296: EOF level=info timestamp=2018-08-03T22:05:27.636246Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:05:30.141665Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:05:35 http: TLS handshake error from 10.129.0.1:54308: EOF level=info timestamp=2018-08-03T22:05:37.825729Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:05:40.999505Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:05:45 http: TLS handshake error from 10.129.0.1:54320: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:53:41.193273Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5l7kj kind= uid=aef9fc6a-9767-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:53:41.299488Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5l7kj\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5l7kj" level=info timestamp=2018-08-03T21:56:42.513234Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiszgnx kind= uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:56:42.514178Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiszgnx kind= uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:56:42.652332Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmiszgnx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmiszgnx" level=info timestamp=2018-08-03T21:59:43.492452Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimbjf6 kind= uid=86ea68c2-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:59:43.493333Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimbjf6 kind= uid=86ea68c2-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:02:44.410831Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmimbjf6\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmimbjf6, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 86ea68c2-9768-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmimbjf6" level=info timestamp=2018-08-03T22:02:44.635218Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2x77f kind= uid=f2e0ff19-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:02:44.635856Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2x77f kind= uid=f2e0ff19-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:02:44.779674Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi2x77f\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi2x77f" level=info timestamp=2018-08-03T22:02:46.344579Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmix5zqd kind= uid=f3ea48ff-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:02:46.344915Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmix5zqd kind= uid=f3ea48ff-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:02:46.475875Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmix5zqd\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmix5zqd" level=info timestamp=2018-08-03T22:02:46.518838Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmix5zqd\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmix5zqd" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:45:29.843700Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:58.262918Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:54.835797Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:03:12.077940Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:43:15.737287Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:44:53.524888Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmix5zqd-gmvks Pod phase: Pending ------------------------------ • Failure [181.056 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:533 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-default [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.011s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Creating a VirtualMachineInstance with different namespace level=info timestamp=2018-08-03T22:02:46.200048Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmix5zqd kind=VirtualMachineInstance uid=f3ea48ff-9768-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmix5zqd-gmvks" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running level=info timestamp=2018-08-03T22:06:49.773848Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:06:53 http: TLS handshake error from 10.129.0.1:47822: EOF 2018/08/03 22:07:03 http: TLS handshake error from 10.129.0.1:47836: EOF 2018/08/03 22:07:13 http: TLS handshake error from 10.129.0.1:47848: EOF level=info timestamp=2018-08-03T22:07:19.737713Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:07:23 http: TLS handshake error from 10.129.0.1:47860: EOF 2018/08/03 22:07:33 http: TLS handshake error from 10.129.0.1:47872: EOF 2018/08/03 22:07:43 http: TLS handshake error from 10.129.0.1:47884: EOF level=info timestamp=2018-08-03T22:07:49.638839Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:07:53 http: TLS handshake error from 10.129.0.1:47896: EOF 2018/08/03 22:08:03 http: TLS handshake error from 10.129.0.1:47908: EOF 2018/08/03 22:08:13 http: TLS handshake error from 10.129.0.1:47920: EOF 2018/08/03 22:08:23 http: TLS handshake error from 10.129.0.1:47932: EOF 2018/08/03 22:08:33 http: TLS handshake error from 10.129.0.1:47944: EOF 2018/08/03 22:08:43 http: TLS handshake error from 10.129.0.1:47956: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:08:10.724960Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:08:11.504983Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:08:11.508828Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:08:12.391985Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:08:15 http: TLS handshake error from 10.129.0.1:54502: EOF level=info timestamp=2018-08-03T22:08:19.577753Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:08:20.923963Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:08:24.694784Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:08:25 http: TLS handshake error from 10.129.0.1:54514: EOF level=info timestamp=2018-08-03T22:08:31.157101Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:08:31.720957Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:08:35 http: TLS handshake error from 10.129.0.1:54526: EOF level=info timestamp=2018-08-03T22:08:41.531191Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:08:42.657660Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:08:45 http: TLS handshake error from 10.129.0.1:54538: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T21:56:42.514178Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiszgnx kind= uid=1b082c6e-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T21:56:42.652332Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmiszgnx\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmiszgnx" level=info timestamp=2018-08-03T21:59:43.492452Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimbjf6 kind= uid=86ea68c2-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T21:59:43.493333Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimbjf6 kind= uid=86ea68c2-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:02:44.410831Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmimbjf6\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmimbjf6, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 86ea68c2-9768-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmimbjf6" level=info timestamp=2018-08-03T22:02:44.635218Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2x77f kind= uid=f2e0ff19-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:02:44.635856Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2x77f kind= uid=f2e0ff19-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:02:44.779674Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi2x77f\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi2x77f" level=info timestamp=2018-08-03T22:02:46.344579Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmix5zqd kind= uid=f3ea48ff-9768-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:02:46.344915Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmix5zqd kind= uid=f3ea48ff-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:02:46.475875Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmix5zqd\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmix5zqd" level=info timestamp=2018-08-03T22:02:46.518838Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmix5zqd\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmix5zqd" level=info timestamp=2018-08-03T22:05:47.296652Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-alternative name=testvmihvwzv kind= uid=5fc13a30-9769-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:05:47.297524Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-alternative name=testvmihvwzv kind= uid=5fc13a30-9769-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:05:47.584565Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmihvwzv\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-alternative/testvmihvwzv" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:46:41.316710Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:48:25.128863Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:58.262918Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:54.835797Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:03:12.077940Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:46:05.056989Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:47:48.860840Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" • Failure [180.970 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:533 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-alternative [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.020s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Creating a VirtualMachineInstance with different namespace level=info timestamp=2018-08-03T22:05:47.275799Z pos=utils.go:244 component=tests namespace=kubevirt-test-alternative name=testvmihvwzv kind=VirtualMachineInstance uid=5fc13a30-9769-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmihvwzv-s49cn" S [SKIPPING] in Spec Setup (BeforeEach) [0.400 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 VirtualMachineInstance Emulation Mode /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:590 should enable emulation in virt-launcher [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:602 Software emulation is not enabled on this cluster /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:598 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.190 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Creating a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:72 VirtualMachineInstance Emulation Mode /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:590 should be reflected in domain XML [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:639 Software emulation is not enabled on this cluster /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:598 ------------------------------ •• Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:09:43 http: TLS handshake error from 10.129.0.1:48028: EOF 2018/08/03 22:09:53 http: TLS handshake error from 10.129.0.1:48040: EOF 2018/08/03 22:10:03 http: TLS handshake error from 10.129.0.1:48054: EOF 2018/08/03 22:10:13 http: TLS handshake error from 10.129.0.1:48066: EOF level=info timestamp=2018-08-03T22:10:19.731656Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:10:23 http: TLS handshake error from 10.129.0.1:48078: EOF 2018/08/03 22:10:33 http: TLS handshake error from 10.129.0.1:48090: EOF 2018/08/03 22:10:43 http: TLS handshake error from 10.129.0.1:48102: EOF 2018/08/03 22:10:53 http: TLS handshake error from 10.129.0.1:48114: EOF 2018/08/03 22:11:03 http: TLS handshake error from 10.129.0.1:48126: EOF 2018/08/03 22:11:13 http: TLS handshake error from 10.129.0.1:48138: EOF 2018/08/03 22:11:23 http: TLS handshake error from 10.129.0.1:48150: EOF 2018/08/03 22:11:33 http: TLS handshake error from 10.129.0.1:48162: EOF 2018/08/03 22:11:43 http: TLS handshake error from 10.129.0.1:48174: EOF level=info timestamp=2018-08-03T22:11:49.637315Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:11:11.539764Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:11:11.546210Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:11:13.802116Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:11:14.709725Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:11:15 http: TLS handshake error from 10.129.0.1:54720: EOF level=info timestamp=2018-08-03T22:11:19.588526Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:11:24.923485Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:11:25 http: TLS handshake error from 10.129.0.1:54732: EOF level=info timestamp=2018-08-03T22:11:26.371232Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:11:33.205793Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:11:35.155694Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:11:35 http: TLS handshake error from 10.129.0.1:54744: EOF level=info timestamp=2018-08-03T22:11:44.088066Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:11:45 http: TLS handshake error from 10.129.0.1:54756: EOF level=info timestamp=2018-08-03T22:11:45.431204Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:02:46.344915Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmix5zqd kind= uid=f3ea48ff-9768-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:02:46.475875Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmix5zqd\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmix5zqd" level=info timestamp=2018-08-03T22:02:46.518838Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmix5zqd\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmix5zqd" level=info timestamp=2018-08-03T22:05:47.296652Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-alternative name=testvmihvwzv kind= uid=5fc13a30-9769-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:05:47.297524Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-alternative name=testvmihvwzv kind= uid=5fc13a30-9769-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:05:47.584565Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmihvwzv\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-alternative/testvmihvwzv" level=info timestamp=2018-08-03T22:08:48.094593Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmihvwzv\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-alternative/testvmihvwzv, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 5fc13a30-9769-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-alternative/testvmihvwzv" level=info timestamp=2018-08-03T22:08:48.563901Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilg5j7 kind= uid=cbce4678-9769-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:08:48.564609Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilg5j7 kind= uid=cbce4678-9769-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:08:48.722592Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmilg5j7\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmilg5j7" level=info timestamp=2018-08-03T22:08:49.294966Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmilg5j7\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmilg5j7, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: cbce4678-9769-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmilg5j7" level=info timestamp=2018-08-03T22:08:49.809369Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixm52r kind= uid=cc8d9a25-9769-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:08:49.810772Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixm52r kind= uid=cc8d9a25-9769-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:08:50.239298Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:08:50.292465Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:49:30.918320Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:13.806251Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:58.262918Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:54.835797Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:03:12.077940Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:49:59.125018Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:51:03.426164Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmixm52r-dxg6g Pod phase: Pending ------------------------------ • Failure [180.772 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Delete a VirtualMachineInstance's Pod /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:748 should result in the VirtualMachineInstance moving to a finalized state [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:749 Timed out after 90.009s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Creating the VirtualMachineInstance level=info timestamp=2018-08-03T22:08:49.909441Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmixm52r kind=VirtualMachineInstance uid=cc8d9a25-9769-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmixm52r-dxg6g" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:12:53 http: TLS handshake error from 10.129.0.1:48258: EOF 2018/08/03 22:13:03 http: TLS handshake error from 10.129.0.1:48272: EOF 2018/08/03 22:13:13 http: TLS handshake error from 10.129.0.1:48284: EOF 2018/08/03 22:13:23 http: TLS handshake error from 10.129.0.1:48296: EOF 2018/08/03 22:13:33 http: TLS handshake error from 10.129.0.1:48308: EOF 2018/08/03 22:13:43 http: TLS handshake error from 10.129.0.1:48320: EOF level=info timestamp=2018-08-03T22:13:49.714034Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:13:53 http: TLS handshake error from 10.129.0.1:48332: EOF 2018/08/03 22:14:03 http: TLS handshake error from 10.129.0.1:48344: EOF 2018/08/03 22:14:13 http: TLS handshake error from 10.129.0.1:48356: EOF level=info timestamp=2018-08-03T22:14:19.634668Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:14:23 http: TLS handshake error from 10.129.0.1:48368: EOF 2018/08/03 22:14:33 http: TLS handshake error from 10.129.0.1:48380: EOF 2018/08/03 22:14:43 http: TLS handshake error from 10.129.0.1:48392: EOF level=info timestamp=2018-08-03T22:14:49.925073Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:14:08.573209Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:14:10.996007Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:14:11.003632Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 22:14:15 http: TLS handshake error from 10.129.0.1:54938: EOF level=info timestamp=2018-08-03T22:14:15.592782Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:14:18.846470Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:14:25 http: TLS handshake error from 10.129.0.1:54950: EOF level=info timestamp=2018-08-03T22:14:28.352881Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:14:29.204222Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:14:35 http: TLS handshake error from 10.129.0.1:54962: EOF level=info timestamp=2018-08-03T22:14:35.263946Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:14:39.451734Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:14:45 http: TLS handshake error from 10.129.0.1:54974: EOF level=info timestamp=2018-08-03T22:14:45.908451Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:14:49.901054Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:08:48.094593Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmihvwzv\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-alternative/testvmihvwzv, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 5fc13a30-9769-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-alternative/testvmihvwzv" level=info timestamp=2018-08-03T22:08:48.563901Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilg5j7 kind= uid=cbce4678-9769-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:08:48.564609Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilg5j7 kind= uid=cbce4678-9769-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:08:48.722592Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmilg5j7\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmilg5j7" level=info timestamp=2018-08-03T22:08:49.294966Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmilg5j7\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmilg5j7, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: cbce4678-9769-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmilg5j7" level=info timestamp=2018-08-03T22:08:49.809369Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixm52r kind= uid=cc8d9a25-9769-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:08:49.810772Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixm52r kind= uid=cc8d9a25-9769-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:08:50.239298Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:08:50.292465Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:11:50.486781Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmixm52r, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: cc8d9a25-9769-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:11:50.652396Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmib5xqr kind= uid=385791f1-976a-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:11:50.653088Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmib5xqr kind= uid=385791f1-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:11:50.851064Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.911881Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.953293Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:53:03.682626Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:54:25.588844Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:58.262918Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:54.835797Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:03:12.077940Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:52:25.233852Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmib5xqr-pnp28 Pod phase: Pending • Failure [180.928 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Delete a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:780 with an active pod. /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:781 should result in pod being terminated [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:782 Timed out after 90.018s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Creating the VirtualMachineInstance level=info timestamp=2018-08-03T22:11:50.540949Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmib5xqr kind=VirtualMachineInstance uid=385791f1-976a-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmib5xqr-pnp28" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:15:53 http: TLS handshake error from 10.129.0.1:48476: EOF 2018/08/03 22:16:03 http: TLS handshake error from 10.129.0.1:48490: EOF 2018/08/03 22:16:13 http: TLS handshake error from 10.129.0.1:48502: EOF level=info timestamp=2018-08-03T22:16:19.974256Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:16:23 http: TLS handshake error from 10.129.0.1:48514: EOF 2018/08/03 22:16:33 http: TLS handshake error from 10.129.0.1:48526: EOF 2018/08/03 22:16:43 http: TLS handshake error from 10.129.0.1:48538: EOF 2018/08/03 22:16:53 http: TLS handshake error from 10.129.0.1:48550: EOF 2018/08/03 22:17:03 http: TLS handshake error from 10.129.0.1:48562: EOF 2018/08/03 22:17:13 http: TLS handshake error from 10.129.0.1:48574: EOF level=info timestamp=2018-08-03T22:17:19.730483Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:17:23 http: TLS handshake error from 10.129.0.1:48586: EOF 2018/08/03 22:17:33 http: TLS handshake error from 10.129.0.1:48598: EOF 2018/08/03 22:17:43 http: TLS handshake error from 10.129.0.1:48610: EOF level=info timestamp=2018-08-03T22:17:49.789262Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:17:03.105181Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:17:05 http: TLS handshake error from 10.129.0.1:55144: EOF level=info timestamp=2018-08-03T22:17:06.478443Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:17:13.280372Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:17:15 http: TLS handshake error from 10.129.0.1:55156: EOF level=info timestamp=2018-08-03T22:17:17.226444Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:17:23.504797Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:17:25 http: TLS handshake error from 10.129.0.1:55168: EOF level=info timestamp=2018-08-03T22:17:29.807764Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:17:33.731939Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:17:35 http: TLS handshake error from 10.129.0.1:55180: EOF level=info timestamp=2018-08-03T22:17:36.719941Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:17:43.988865Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:17:45 http: TLS handshake error from 10.129.0.1:55192: EOF level=info timestamp=2018-08-03T22:17:47.625103Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:08:49.809369Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixm52r kind= uid=cc8d9a25-9769-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:08:49.810772Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmixm52r kind= uid=cc8d9a25-9769-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:08:50.239298Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:08:50.292465Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:11:50.486781Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmixm52r, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: cc8d9a25-9769-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:11:50.652396Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmib5xqr kind= uid=385791f1-976a-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:11:50.653088Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmib5xqr kind= uid=385791f1-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:11:50.851064Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.911881Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.953293Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:14:51.491374Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmib5xqr, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 385791f1-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:14:51.742809Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:14:51.743507Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:14:51.967439Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:14:52.009075Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:55:38.142625Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:58.262918Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:54.835797Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:03:12.077940Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:53:37.796234Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:55:16.853766Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi4rdbf-fmdvr Pod phase: Pending • Failure [180.957 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Delete a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:780 with ACPI and 0 grace period seconds /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:806 should result in vmi status failed [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:807 Timed out after 90.014s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Creating the VirtualMachineInstance level=info timestamp=2018-08-03T22:14:51.628977Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi4rdbf kind=VirtualMachineInstance uid=a4482203-976a-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi4rdbf-fmdvr" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:18:43 http: TLS handshake error from 10.129.0.1:48682: EOF 2018/08/03 22:18:53 http: TLS handshake error from 10.129.0.1:48694: EOF 2018/08/03 22:19:03 http: TLS handshake error from 10.129.0.1:48708: EOF 2018/08/03 22:19:13 http: TLS handshake error from 10.129.0.1:48720: EOF 2018/08/03 22:19:23 http: TLS handshake error from 10.129.0.1:48732: EOF 2018/08/03 22:19:33 http: TLS handshake error from 10.129.0.1:48744: EOF 2018/08/03 22:19:43 http: TLS handshake error from 10.129.0.1:48756: EOF 2018/08/03 22:19:53 http: TLS handshake error from 10.129.0.1:48768: EOF 2018/08/03 22:20:03 http: TLS handshake error from 10.129.0.1:48780: EOF 2018/08/03 22:20:13 http: TLS handshake error from 10.129.0.1:48792: EOF 2018/08/03 22:20:23 http: TLS handshake error from 10.129.0.1:48804: EOF 2018/08/03 22:20:33 http: TLS handshake error from 10.129.0.1:48816: EOF 2018/08/03 22:20:43 http: TLS handshake error from 10.129.0.1:48828: EOF level=info timestamp=2018-08-03T22:20:49.636310Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:20:53 http: TLS handshake error from 10.129.0.1:48840: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:20:07.257588Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:20:08.174488Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:20:15 http: TLS handshake error from 10.129.0.1:55374: EOF level=info timestamp=2018-08-03T22:20:17.488919Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:20:18.808904Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:20:19.625458Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:20:25 http: TLS handshake error from 10.129.0.1:55386: EOF level=info timestamp=2018-08-03T22:20:27.814681Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:20:31.707380Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:20:35 http: TLS handshake error from 10.129.0.1:55398: EOF level=info timestamp=2018-08-03T22:20:37.991449Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:20:38.354596Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:20:45 http: TLS handshake error from 10.129.0.1:55410: EOF level=info timestamp=2018-08-03T22:20:48.212779Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:20:49.130450Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:11:50.486781Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmixm52r\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmixm52r, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: cc8d9a25-9769-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmixm52r" level=info timestamp=2018-08-03T22:11:50.652396Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmib5xqr kind= uid=385791f1-976a-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:11:50.653088Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmib5xqr kind= uid=385791f1-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:11:50.851064Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.911881Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.953293Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:14:51.491374Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmib5xqr, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 385791f1-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:14:51.742809Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:14:51.743507Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:14:51.967439Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:14:52.009075Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:52.740449Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi4rdbf, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: a4482203-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:53.125720Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:17:53.127897Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:17:53.252896Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvminsv6r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvminsv6r" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T21:58:36.506060Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:54.835797Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:03:12.077940Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:56:56.081412Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvminsv6r-hj624 Pod phase: Pending • Failure [181.416 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Delete a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:780 with ACPI and some grace period seconds /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:831 should result in vmi status succeeded [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:832 Timed out after 90.013s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Creating the VirtualMachineInstance level=info timestamp=2018-08-03T22:17:52.921104Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvminsv6r kind=VirtualMachineInstance uid=1065b036-976b-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvminsv6r-hj624" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:22:13 http: TLS handshake error from 10.129.0.1:48938: EOF 2018/08/03 22:22:23 http: TLS handshake error from 10.129.0.1:48950: EOF 2018/08/03 22:22:33 http: TLS handshake error from 10.129.0.1:48962: EOF 2018/08/03 22:22:43 http: TLS handshake error from 10.129.0.1:48974: EOF level=info timestamp=2018-08-03T22:22:47.210306Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:22:49.169372Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:22:49.739586Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:22:53 http: TLS handshake error from 10.129.0.1:48986: EOF 2018/08/03 22:23:03 http: TLS handshake error from 10.129.0.1:48998: EOF 2018/08/03 22:23:13 http: TLS handshake error from 10.129.0.1:49010: EOF level=info timestamp=2018-08-03T22:23:19.696090Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:23:23 http: TLS handshake error from 10.129.0.1:49022: EOF 2018/08/03 22:23:33 http: TLS handshake error from 10.129.0.1:49034: EOF 2018/08/03 22:23:43 http: TLS handshake error from 10.129.0.1:49046: EOF 2018/08/03 22:23:53 http: TLS handshake error from 10.129.0.1:49058: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:23:14.028356Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:23:14.031948Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 22:23:15 http: TLS handshake error from 10.129.0.1:55592: EOF level=info timestamp=2018-08-03T22:23:20.141392Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:23:20.878058Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:23:25 http: TLS handshake error from 10.129.0.1:55604: EOF level=info timestamp=2018-08-03T22:23:31.147830Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:23:33.326705Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:23:35 http: TLS handshake error from 10.129.0.1:55616: EOF level=info timestamp=2018-08-03T22:23:40.095905Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:23:41.350625Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:23:45 http: TLS handshake error from 10.129.0.1:55628: EOF level=info timestamp=2018-08-03T22:23:49.796998Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:23:50.392176Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:23:51.583525Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:11:50.851064Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.911881Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:11:50.953293Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:14:51.491374Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmib5xqr, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 385791f1-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:14:51.742809Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:14:51.743507Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:14:51.967439Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:14:52.009075Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:52.740449Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi4rdbf, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: a4482203-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:53.125720Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:17:53.127897Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:17:53.252896Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvminsv6r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvminsv6r" level=info timestamp=2018-08-03T22:20:53.799360Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvminsv6r\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvminsv6r, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1065b036-976b-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvminsv6r" level=info timestamp=2018-08-03T22:20:54.065932Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5xqrc kind= uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:20:54.066525Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5xqrc kind= uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:01:15.220297Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:03:12.077940Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T21:58:16.166773Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T21:59:34.496282Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi5xqrc-vqh29 Pod phase: Pending • Failure [180.829 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Delete a VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:780 with grace period greater than 0 /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:856 should run graceful shutdown [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:857 Timed out after 90.014s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Setting a VirtualMachineInstance termination grace period to 5 STEP: Creating the VirtualMachineInstance level=info timestamp=2018-08-03T22:20:53.993349Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi5xqrc kind=VirtualMachineInstance uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi5xqrc-vqh29" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:24:53 http: TLS handshake error from 10.129.0.1:49130: EOF 2018/08/03 22:25:03 http: TLS handshake error from 10.129.0.1:49142: EOF 2018/08/03 22:25:13 http: TLS handshake error from 10.129.0.1:49156: EOF 2018/08/03 22:25:23 http: TLS handshake error from 10.129.0.1:49168: EOF 2018/08/03 22:25:33 http: TLS handshake error from 10.129.0.1:49180: EOF 2018/08/03 22:25:43 http: TLS handshake error from 10.129.0.1:49192: EOF level=info timestamp=2018-08-03T22:25:49.825422Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:25:53 http: TLS handshake error from 10.129.0.1:49204: EOF 2018/08/03 22:26:03 http: TLS handshake error from 10.129.0.1:49216: EOF 2018/08/03 22:26:13 http: TLS handshake error from 10.129.0.1:49228: EOF 2018/08/03 22:26:23 http: TLS handshake error from 10.129.0.1:49240: EOF 2018/08/03 22:26:33 http: TLS handshake error from 10.129.0.1:49252: EOF 2018/08/03 22:26:43 http: TLS handshake error from 10.129.0.1:49264: EOF level=info timestamp=2018-08-03T22:26:49.744523Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:26:53 http: TLS handshake error from 10.129.0.1:49276: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:26:14.047961Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/openapi/v2 proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:26:14.052555Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:26:14.545916Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:26:15 http: TLS handshake error from 10.129.0.1:55810: EOF level=info timestamp=2018-08-03T22:26:19.677613Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:26:21.689902Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:26:24.745091Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:26:25 http: TLS handshake error from 10.129.0.1:55822: EOF level=info timestamp=2018-08-03T22:26:34.755472Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:26:34.940440Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:26:35 http: TLS handshake error from 10.129.0.1:55834: EOF level=info timestamp=2018-08-03T22:26:41.950013Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:26:45.151567Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:26:45 http: TLS handshake error from 10.129.0.1:55846: EOF level=info timestamp=2018-08-03T22:26:51.966791Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:14:51.491374Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmib5xqr\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmib5xqr, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 385791f1-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmib5xqr" level=info timestamp=2018-08-03T22:14:51.742809Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:14:51.743507Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:14:51.967439Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:14:52.009075Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:52.740449Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi4rdbf, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: a4482203-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:53.125720Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:17:53.127897Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:17:53.252896Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvminsv6r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvminsv6r" level=info timestamp=2018-08-03T22:20:53.799360Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvminsv6r\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvminsv6r, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1065b036-976b-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvminsv6r" level=info timestamp=2018-08-03T22:20:54.065932Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5xqrc kind= uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:20:54.066525Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5xqrc kind= uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:23:54.570994Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5xqrc\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi5xqrc, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 7c3dddb1-976b-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5xqrc" level=info timestamp=2018-08-03T22:23:54.672601Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2stjg kind= uid=e7e51527-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:23:54.674990Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2stjg kind= uid=e7e51527-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:04:33.559909Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:01:31.354698Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:02:57.505643Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi2stjg-ffp5x Pod phase: Pending • Failure [180.661 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Killed VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:908 should be in Failed phase [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:909 Timed out after 90.008s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Starting a VirtualMachineInstance level=info timestamp=2018-08-03T22:23:54.493571Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmi2stjg kind=VirtualMachineInstance uid=e7e51527-976b-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmi2stjg-ffp5x" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:28:03 http: TLS handshake error from 10.129.0.1:49360: EOF 2018/08/03 22:28:13 http: TLS handshake error from 10.129.0.1:49374: EOF level=info timestamp=2018-08-03T22:28:19.691120Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:28:23 http: TLS handshake error from 10.129.0.1:49386: EOF 2018/08/03 22:28:33 http: TLS handshake error from 10.129.0.1:49398: EOF 2018/08/03 22:28:43 http: TLS handshake error from 10.129.0.1:49410: EOF 2018/08/03 22:28:53 http: TLS handshake error from 10.129.0.1:49422: EOF 2018/08/03 22:29:03 http: TLS handshake error from 10.129.0.1:49434: EOF 2018/08/03 22:29:13 http: TLS handshake error from 10.129.0.1:49446: EOF level=info timestamp=2018-08-03T22:29:19.649464Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:29:23 http: TLS handshake error from 10.129.0.1:49458: EOF 2018/08/03 22:29:33 http: TLS handshake error from 10.129.0.1:49470: EOF 2018/08/03 22:29:43 http: TLS handshake error from 10.129.0.1:49482: EOF level=info timestamp=2018-08-03T22:29:49.762703Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:29:53 http: TLS handshake error from 10.129.0.1:49494: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:29:12.332917Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:29:12.881565Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:29:15 http: TLS handshake error from 10.129.0.1:56028: EOF level=info timestamp=2018-08-03T22:29:18.501863Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:29:22.864676Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:29:25 http: TLS handshake error from 10.129.0.1:56040: EOF level=info timestamp=2018-08-03T22:29:28.717000Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:29:35 http: TLS handshake error from 10.129.0.1:56052: EOF level=info timestamp=2018-08-03T22:29:36.494962Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:29:38.917684Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:29:43.023097Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:29:45 http: TLS handshake error from 10.129.0.1:56064: EOF level=info timestamp=2018-08-03T22:29:49.209956Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:29:53.132731Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:29:55 http: TLS handshake error from 10.129.0.1:56076: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:14:51.743507Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4rdbf kind= uid=a4482203-976a-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:14:51.967439Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:14:52.009075Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:52.740449Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4rdbf\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi4rdbf, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: a4482203-976a-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4rdbf" level=info timestamp=2018-08-03T22:17:53.125720Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:17:53.127897Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvminsv6r kind= uid=1065b036-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:17:53.252896Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvminsv6r\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvminsv6r" level=info timestamp=2018-08-03T22:20:53.799360Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvminsv6r\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvminsv6r, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1065b036-976b-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvminsv6r" level=info timestamp=2018-08-03T22:20:54.065932Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5xqrc kind= uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:20:54.066525Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5xqrc kind= uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:23:54.570994Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5xqrc\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi5xqrc, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 7c3dddb1-976b-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5xqrc" level=info timestamp=2018-08-03T22:23:54.672601Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2stjg kind= uid=e7e51527-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:23:54.674990Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2stjg kind= uid=e7e51527-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:26:55.457299Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmig8jss kind= uid=53a6b6fe-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:26:55.458243Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmig8jss kind= uid=53a6b6fe-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:05:40.682843Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:07:51.095796Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:28:32.348619Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:05:01.116945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:06:22.592477Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:27:28.345378Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:29:37.221014Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmig8jss-74hhp Pod phase: Pending • Failure [180.804 seconds] VMIlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:51 Killed VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:908 should be left alone by virt-handler [It] /root/go/src/kubevirt.io/kubevirt/tests/vmi_lifecycle_test.go:936 Timed out after 90.012s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ STEP: Starting a VirtualMachineInstance level=info timestamp=2018-08-03T22:26:55.256720Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmig8jss kind=VirtualMachineInstance uid=53a6b6fe-976c-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmig8jss-74hhp" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:31:13 http: TLS handshake error from 10.129.0.1:49592: EOF 2018/08/03 22:31:23 http: TLS handshake error from 10.129.0.1:49604: EOF 2018/08/03 22:31:33 http: TLS handshake error from 10.129.0.1:49616: EOF 2018/08/03 22:31:43 http: TLS handshake error from 10.129.0.1:49628: EOF level=info timestamp=2018-08-03T22:31:49.655109Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:31:53 http: TLS handshake error from 10.129.0.1:49640: EOF 2018/08/03 22:32:03 http: TLS handshake error from 10.129.0.1:49652: EOF 2018/08/03 22:32:13 http: TLS handshake error from 10.129.0.1:49664: EOF level=info timestamp=2018-08-03T22:32:19.731585Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:32:23 http: TLS handshake error from 10.129.0.1:49676: EOF 2018/08/03 22:32:33 http: TLS handshake error from 10.129.0.1:49688: EOF 2018/08/03 22:32:43 http: TLS handshake error from 10.129.0.1:49700: EOF level=info timestamp=2018-08-03T22:32:47.216935Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:32:50.029369Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:32:53 http: TLS handshake error from 10.129.0.1:49712: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:32:14.494934Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:32:15 http: TLS handshake error from 10.129.0.1:56246: EOF level=info timestamp=2018-08-03T22:32:22.737244Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:32:24.496525Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:32:25 http: TLS handshake error from 10.129.0.1:56258: EOF level=info timestamp=2018-08-03T22:32:32.849094Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:32:35 http: TLS handshake error from 10.129.0.1:56270: EOF level=info timestamp=2018-08-03T22:32:38.107727Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:32:43.202536Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:32:44.743700Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:32:45 http: TLS handshake error from 10.129.0.1:56282: EOF level=info timestamp=2018-08-03T22:32:49.020777Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:32:53.373025Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:32:54.624022Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:32:55 http: TLS handshake error from 10.129.0.1:56294: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:20:54.066525Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi5xqrc kind= uid=7c3dddb1-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:23:54.570994Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi5xqrc\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi5xqrc, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 7c3dddb1-976b-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi5xqrc" level=info timestamp=2018-08-03T22:23:54.672601Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2stjg kind= uid=e7e51527-976b-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:23:54.674990Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2stjg kind= uid=e7e51527-976b-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:26:55.457299Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmig8jss kind= uid=53a6b6fe-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:26:55.458243Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmig8jss kind= uid=53a6b6fe-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:29:56.186986Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmig8jss\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmig8jss, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 53a6b6fe-976c-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmig8jss" level=info timestamp=2018-08-03T22:29:56.290972Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilgpx7 kind= uid=bf6e6cf6-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:29:56.291635Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilgpx7 kind= uid=bf6e6cf6-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:29:56.319839Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9qqqm kind= uid=bf728b5b-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:29:56.320221Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9qqqm kind= uid=bf728b5b-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:29:56.333656Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxwdh kind= uid=bf75da14-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:29:56.333873Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxwdh kind= uid=bf75da14-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:29:56.357681Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit47sd kind= uid=bf79c938-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:29:56.358010Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit47sd kind= uid=bf79c938-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:08:56.490301Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:12.559575Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:28:32.348619Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:30:41.174615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:06.317948Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:08:33.006089Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:27:28.345378Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:29:37.221014Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:31:02.362336Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:52.141249Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi9qqqm-2pbxs Pod phase: Pending Pod name: virt-launcher-testvmikxwdh-nxlh4 Pod phase: Pending Pod name: virt-launcher-testvmilgpx7-k2tb5 Pod phase: Pending Pod name: virt-launcher-testvmit47sd-sh2sj Pod phase: Pending • Failure in Spec Setup (BeforeEach) [181.322 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be able to reach [BeforeEach] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 the Inbound VirtualMachineInstance /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.012s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T22:29:56.238075Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmilgpx7 kind=VirtualMachineInstance uid=bf6e6cf6-976c-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmilgpx7-k2tb5" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:33:43 http: TLS handshake error from 10.129.0.1:49772: EOF 2018/08/03 22:33:53 http: TLS handshake error from 10.129.0.1:49784: EOF 2018/08/03 22:34:03 http: TLS handshake error from 10.129.0.1:49796: EOF 2018/08/03 22:34:13 http: TLS handshake error from 10.129.0.1:49810: EOF level=info timestamp=2018-08-03T22:34:19.761322Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:34:23 http: TLS handshake error from 10.129.0.1:49822: EOF 2018/08/03 22:34:33 http: TLS handshake error from 10.129.0.1:49834: EOF 2018/08/03 22:34:43 http: TLS handshake error from 10.129.0.1:49846: EOF 2018/08/03 22:34:53 http: TLS handshake error from 10.129.0.1:49858: EOF 2018/08/03 22:35:03 http: TLS handshake error from 10.129.0.1:49870: EOF 2018/08/03 22:35:13 http: TLS handshake error from 10.129.0.1:49882: EOF 2018/08/03 22:35:23 http: TLS handshake error from 10.129.0.1:49894: EOF 2018/08/03 22:35:33 http: TLS handshake error from 10.129.0.1:49906: EOF 2018/08/03 22:35:43 http: TLS handshake error from 10.129.0.1:49918: EOF 2018/08/03 22:35:53 http: TLS handshake error from 10.129.0.1:49930: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:35:16.490380Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:35:19.667352Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:35:25 http: TLS handshake error from 10.129.0.1:56476: EOF level=info timestamp=2018-08-03T22:35:25.786847Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:35:26.690378Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:35:35 http: TLS handshake error from 10.129.0.1:56488: EOF level=info timestamp=2018-08-03T22:35:36.907424Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:35:39.948823Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:35:45 http: TLS handshake error from 10.129.0.1:56500: EOF level=info timestamp=2018-08-03T22:35:46.849635Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:35:47.026016Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:35:49.740913Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:35:55 http: TLS handshake error from 10.129.0.1:56512: EOF level=info timestamp=2018-08-03T22:35:56.074911Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:35:57.288340Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:29:56.319839Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9qqqm kind= uid=bf728b5b-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:29:56.320221Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi9qqqm kind= uid=bf728b5b-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:29:56.333656Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxwdh kind= uid=bf75da14-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:29:56.333873Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikxwdh kind= uid=bf75da14-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:29:56.357681Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit47sd kind= uid=bf79c938-976c-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:29:56.358010Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmit47sd kind= uid=bf79c938-976c-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:32:58.087694Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifs7cw kind= uid=2bc93391-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:32:58.088688Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmifs7cw kind= uid=2bc93391-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:32:58.118011Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimb9vm kind= uid=2bcf5b9c-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:32:58.118422Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimb9vm kind= uid=2bcf5b9c-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:32:58.149915Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmisjs9l kind= uid=2bd3ebc8-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:32:58.150331Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmisjs9l kind= uid=2bd3ebc8-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:32:58.222102Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4t74s kind= uid=2bda54cc-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:32:58.222648Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4t74s kind= uid=2bda54cc-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:32:58.428427Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifs7cw\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifs7cw" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:12:01.685606Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:13:19.119983Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:28:32.348619Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:30:41.174615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:06.317948Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:33:53.118704Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:35:32.905219Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:09:38.432239Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:10:54.513984Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:27:28.345378Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:29:37.221014Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:31:02.362336Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:52.141249Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:34:39.259890Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi4t74s-v5wqh Pod phase: Pending Pod name: virt-launcher-testvmifs7cw-zl285 Pod phase: Pending Pod name: virt-launcher-testvmimb9vm-xw7sx Pod phase: Pending Pod name: virt-launcher-testvmisjs9l-cv224 Pod phase: Pending • Failure in Spec Setup (BeforeEach) [181.904 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be able to reach [BeforeEach] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 the Inbound VirtualMachineInstance with pod network connectivity explicitly set /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.014s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T22:32:58.025852Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmifs7cw kind=VirtualMachineInstance uid=2bc93391-976d-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmifs7cw-zl285" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:37:13 http: TLS handshake error from 10.129.0.1:50028: EOF level=info timestamp=2018-08-03T22:37:19.772266Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:37:23 http: TLS handshake error from 10.129.0.1:50040: EOF 2018/08/03 22:37:33 http: TLS handshake error from 10.129.0.1:50052: EOF 2018/08/03 22:37:43 http: TLS handshake error from 10.129.0.1:50064: EOF level=info timestamp=2018-08-03T22:37:49.778409Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:37:53 http: TLS handshake error from 10.129.0.1:50076: EOF 2018/08/03 22:38:03 http: TLS handshake error from 10.129.0.1:50088: EOF 2018/08/03 22:38:13 http: TLS handshake error from 10.129.0.1:50100: EOF level=info timestamp=2018-08-03T22:38:19.796327Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:38:23 http: TLS handshake error from 10.129.0.1:50112: EOF 2018/08/03 22:38:33 http: TLS handshake error from 10.129.0.1:50124: EOF 2018/08/03 22:38:43 http: TLS handshake error from 10.129.0.1:50136: EOF level=info timestamp=2018-08-03T22:38:49.735069Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:38:53 http: TLS handshake error from 10.129.0.1:50148: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:38:14.005066Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/swagger.json proto=HTTP/2.0 statusCode=404 contentLength=19 2018/08/03 22:38:15 http: TLS handshake error from 10.129.0.1:56682: EOF level=info timestamp=2018-08-03T22:38:18.003904Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:38:21.126483Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:38:25 http: TLS handshake error from 10.129.0.1:56694: EOF level=info timestamp=2018-08-03T22:38:27.354031Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:38:31.299946Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:38:35 http: TLS handshake error from 10.129.0.1:56706: EOF level=info timestamp=2018-08-03T22:38:41.883816Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:38:42.023741Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:38:45 http: TLS handshake error from 10.129.0.1:56718: EOF level=info timestamp=2018-08-03T22:38:48.363869Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:38:52.173075Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:38:55 http: TLS handshake error from 10.129.0.1:56730: EOF level=info timestamp=2018-08-03T22:38:57.768398Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:32:58.149915Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmisjs9l kind= uid=2bd3ebc8-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:32:58.150331Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmisjs9l kind= uid=2bd3ebc8-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:32:58.222102Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4t74s kind= uid=2bda54cc-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:32:58.222648Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi4t74s kind= uid=2bda54cc-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:32:58.428427Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmifs7cw\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmifs7cw" level=info timestamp=2018-08-03T22:35:59.507426Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi4t74s\": StorageError: invalid object, Code: 4, Key: /kubernetes.io/kubevirt.io/virtualmachineinstances/kubevirt-test-default/testvmi4t74s, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 2bda54cc-976d-11e8-948d-525500d15501, UID in object meta: " msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi4t74s" level=info timestamp=2018-08-03T22:36:00.012792Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmivlsxw kind= uid=98391a07-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:36:00.013623Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmivlsxw kind= uid=98391a07-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:36:00.080659Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidvm6t kind= uid=98439d8c-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:36:00.080980Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidvm6t kind= uid=98439d8c-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:36:00.142483Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqh9cg kind= uid=9848afe0-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:36:00.142732Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqh9cg kind= uid=9848afe0-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:36:00.205680Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip8c6j kind= uid=9852b3db-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:36:00.205904Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip8c6j kind= uid=9852b3db-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:36:00.300766Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmidvm6t\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmidvm6t" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:15:12.639051Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:28:32.348619Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:30:41.174615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:06.317948Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:33:53.118704Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:35:32.905219Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:37:27.367654Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:56.563945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:12:16.986072Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:14:24.207716Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:27:28.345378Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:29:37.221014Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:31:02.362336Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:52.141249Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:34:39.259890Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:36:33.718395Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:02.867765Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmidvm6t-6fxdp Pod phase: Pending Pod name: virt-launcher-testvmip8c6j-4nc6b Pod phase: Pending Pod name: virt-launcher-testvmiqh9cg-b8wj4 Pod phase: Pending Pod name: virt-launcher-testvmivlsxw-6bbzz Pod phase: Pending • Failure in Spec Setup (BeforeEach) [181.967 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be able to reach [BeforeEach] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 the Inbound VirtualMachineInstance with custom MAC address /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.014s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T22:36:00.054773Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmivlsxw kind=VirtualMachineInstance uid=98391a07-976d-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmivlsxw-6bbzz" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:40:03 http: TLS handshake error from 10.129.0.1:50232: EOF 2018/08/03 22:40:13 http: TLS handshake error from 10.129.0.1:50246: EOF level=info timestamp=2018-08-03T22:40:19.818355Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:40:23 http: TLS handshake error from 10.129.0.1:50258: EOF 2018/08/03 22:40:33 http: TLS handshake error from 10.129.0.1:50270: EOF 2018/08/03 22:40:43 http: TLS handshake error from 10.129.0.1:50282: EOF 2018/08/03 22:40:53 http: TLS handshake error from 10.129.0.1:50294: EOF 2018/08/03 22:41:03 http: TLS handshake error from 10.129.0.1:50306: EOF 2018/08/03 22:41:13 http: TLS handshake error from 10.129.0.1:50318: EOF level=info timestamp=2018-08-03T22:41:19.717581Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:41:23 http: TLS handshake error from 10.129.0.1:50330: EOF 2018/08/03 22:41:33 http: TLS handshake error from 10.129.0.1:50342: EOF 2018/08/03 22:41:43 http: TLS handshake error from 10.129.0.1:50354: EOF level=info timestamp=2018-08-03T22:41:49.984897Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:41:53 http: TLS handshake error from 10.129.0.1:50366: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:41:15.083734Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:41:15 http: TLS handshake error from 10.129.0.1:56900: EOF level=info timestamp=2018-08-03T22:41:19.529096Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:41:25 http: TLS handshake error from 10.129.0.1:56912: EOF level=info timestamp=2018-08-03T22:41:25.317102Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:41:28.745087Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:41:35 http: TLS handshake error from 10.129.0.1:56924: EOF level=info timestamp=2018-08-03T22:41:35.590230Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:41:43.143463Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:41:45 http: TLS handshake error from 10.129.0.1:56936: EOF level=info timestamp=2018-08-03T22:41:45.812353Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:41:50.043698Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:41:55 http: TLS handshake error from 10.129.0.1:56948: EOF level=info timestamp=2018-08-03T22:41:56.098607Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:41:59.041662Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:36:00.080659Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidvm6t kind= uid=98439d8c-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:36:00.080980Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmidvm6t kind= uid=98439d8c-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:36:00.142483Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqh9cg kind= uid=9848afe0-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:36:00.142732Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqh9cg kind= uid=9848afe0-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:36:00.205680Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip8c6j kind= uid=9852b3db-976d-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:36:00.205904Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmip8c6j kind= uid=9852b3db-976d-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:36:00.300766Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmidvm6t\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmidvm6t" level=info timestamp=2018-08-03T22:39:01.823687Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilnq7q kind= uid=04996dda-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:39:01.826853Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilnq7q kind= uid=04996dda-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:39:01.840624Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2szhf kind= uid=049c4da1-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:39:01.841845Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2szhf kind= uid=049c4da1-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:39:01.870984Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikwfrl kind= uid=049f3f51-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:39:01.872332Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikwfrl kind= uid=049f3f51-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:39:01.904911Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6glm5 kind= uid=04a4ce13-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:39:01.905117Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6glm5 kind= uid=04a4ce13-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:17:05.313245Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:18.595019Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:19:49.474203Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:28:32.348619Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:30:41.174615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:06.317948Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:33:53.118704Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:35:32.905219Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:37:27.367654Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:56.563945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:40:06.043026Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:16:17.681063Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:18:10.359513Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:27:28.345378Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:29:37.221014Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:31:02.362336Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:52.141249Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:34:39.259890Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:36:33.718395Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:02.867765Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:39:12.412748Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:41:23.530615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmi2szhf-cjr8t Pod phase: Pending Pod name: virt-launcher-testvmi6glm5-zr94k Pod phase: Pending Pod name: virt-launcher-testvmikwfrl-m9vzx Pod phase: Pending Pod name: virt-launcher-testvmilnq7q-h4rzg Pod phase: Pending • Failure in Spec Setup (BeforeEach) [181.680 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be able to reach [BeforeEach] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 the internet /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.012s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T22:39:01.728418Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmilnq7q kind=VirtualMachineInstance uid=04996dda-976e-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmilnq7q-h4rzg" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:43:03 http: TLS handshake error from 10.129.0.1:50450: EOF 2018/08/03 22:43:13 http: TLS handshake error from 10.129.0.1:50464: EOF 2018/08/03 22:43:23 http: TLS handshake error from 10.129.0.1:50476: EOF 2018/08/03 22:43:33 http: TLS handshake error from 10.129.0.1:50488: EOF 2018/08/03 22:43:43 http: TLS handshake error from 10.129.0.1:50500: EOF level=info timestamp=2018-08-03T22:43:49.910547Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:43:53 http: TLS handshake error from 10.129.0.1:50512: EOF 2018/08/03 22:44:03 http: TLS handshake error from 10.129.0.1:50524: EOF 2018/08/03 22:44:13 http: TLS handshake error from 10.129.0.1:50536: EOF level=info timestamp=2018-08-03T22:44:19.791417Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:44:23 http: TLS handshake error from 10.129.0.1:50548: EOF 2018/08/03 22:44:33 http: TLS handshake error from 10.129.0.1:50560: EOF 2018/08/03 22:44:43 http: TLS handshake error from 10.129.0.1:50572: EOF 2018/08/03 22:44:53 http: TLS handshake error from 10.129.0.1:50584: EOF 2018/08/03 22:45:03 http: TLS handshake error from 10.129.0.1:50596: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:44:19.385451Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:44:21.230969Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:44:25 http: TLS handshake error from 10.129.0.1:57130: EOF level=info timestamp=2018-08-03T22:44:29.496708Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:44:30.922096Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:44:35 http: TLS handshake error from 10.129.0.1:57142: EOF level=info timestamp=2018-08-03T22:44:39.619305Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:44:44.889945Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:44:45 http: TLS handshake error from 10.129.0.1:57154: EOF level=info timestamp=2018-08-03T22:44:49.678686Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:44:49.839764Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:44:51.539333Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:44:55 http: TLS handshake error from 10.129.0.1:57166: EOF level=info timestamp=2018-08-03T22:44:59.957597Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:45:01.211414Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:39:01.840624Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2szhf kind= uid=049c4da1-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:39:01.841845Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi2szhf kind= uid=049c4da1-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:39:01.870984Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikwfrl kind= uid=049f3f51-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:39:01.872332Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikwfrl kind= uid=049f3f51-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:39:01.904911Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6glm5 kind= uid=04a4ce13-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:39:01.905117Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmi6glm5 kind= uid=04a4ce13-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:42:03.040250Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmi6glm5\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmi6glm5" level=info timestamp=2018-08-03T22:42:03.672245Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimwjx9 kind= uid=70fa8b4b-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:42:03.672777Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmimwjx9 kind= uid=70fa8b4b-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:42:03.772954Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmighx5p kind= uid=710249c0-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:42:03.773316Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmighx5p kind= uid=710249c0-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:42:03.879516Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiljd7g kind= uid=7112673d-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:42:03.879747Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiljd7g kind= uid=7112673d-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:42:04.029623Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqs85x kind= uid=712002a9-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:42:04.030112Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqs85x kind= uid=712002a9-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:21:54.099221Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:28:32.348619Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:30:41.174615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:06.317948Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:33:53.118704Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:35:32.905219Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:37:27.367654Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:56.563945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:40:06.043026Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:42:10.638110Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:43:57.136630Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:45:03.374813Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:20:15.062338Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:22:04.299383Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:27:28.345378Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:29:37.221014Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:31:02.362336Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:52.141249Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:34:39.259890Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:36:33.718395Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:02.867765Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:39:12.412748Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:41:23.530615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:43:10.059813Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:44:16.259098Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmighx5p-mfm7v Pod phase: Pending Pod name: virt-launcher-testvmiljd7g-ksnx7 Pod phase: Pending Pod name: virt-launcher-testvmimwjx9-psv9c Pod phase: Pending Pod name: virt-launcher-testvmiqs85x-7q96p Pod phase: Pending • Failure in Spec Setup (BeforeEach) [182.111 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be reachable via the propagated IP from a Pod [BeforeEach] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 on the same node from Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.012s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T22:42:03.803032Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmimwjx9 kind=VirtualMachineInstance uid=70fa8b4b-976e-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmimwjx9-psv9c" Pod name: disks-images-provider-hwh68 Pod phase: Running copy all images to host mount directory Pod name: disks-images-provider-mmwgg Pod phase: Running copy all images to host mount directory Pod name: virt-api-7d79764579-rk48t Pod phase: Running 2018/08/03 22:46:03 http: TLS handshake error from 10.129.0.1:50668: EOF 2018/08/03 22:46:13 http: TLS handshake error from 10.129.0.1:50680: EOF level=info timestamp=2018-08-03T22:46:19.707599Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:46:23 http: TLS handshake error from 10.129.0.1:50694: EOF 2018/08/03 22:46:33 http: TLS handshake error from 10.129.0.1:50706: EOF 2018/08/03 22:46:43 http: TLS handshake error from 10.129.0.1:50718: EOF 2018/08/03 22:46:53 http: TLS handshake error from 10.129.0.1:50730: EOF 2018/08/03 22:47:03 http: TLS handshake error from 10.129.0.1:50742: EOF 2018/08/03 22:47:13 http: TLS handshake error from 10.129.0.1:50754: EOF level=info timestamp=2018-08-03T22:47:19.739739Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 2018/08/03 22:47:23 http: TLS handshake error from 10.129.0.1:50766: EOF 2018/08/03 22:47:33 http: TLS handshake error from 10.129.0.1:50778: EOF 2018/08/03 22:47:43 http: TLS handshake error from 10.129.0.1:50790: EOF 2018/08/03 22:47:53 http: TLS handshake error from 10.129.0.1:50802: EOF 2018/08/03 22:48:03 http: TLS handshake error from 10.129.0.1:50814: EOF Pod name: virt-api-7d79764579-swtbv Pod phase: Running level=info timestamp=2018-08-03T22:47:23.286769Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:47:25 http: TLS handshake error from 10.129.0.1:57348: EOF level=info timestamp=2018-08-03T22:47:32.143068Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:47:33.515197Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:47:35 http: TLS handshake error from 10.129.0.1:57360: EOF level=info timestamp=2018-08-03T22:47:43.792774Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:47:45 http: TLS handshake error from 10.129.0.1:57372: EOF level=info timestamp=2018-08-03T22:47:46.675808Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:47:49.735534Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url=/ proto=HTTP/1.1 statusCode=404 contentLength=19 level=info timestamp=2018-08-03T22:47:53.745719Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:47:53.867908Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:47:55 http: TLS handshake error from 10.129.0.1:57384: EOF level=info timestamp=2018-08-03T22:48:02.497299Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 level=info timestamp=2018-08-03T22:48:04.081113Z pos=filter.go:46 component=virt-api remoteAddress=10.128.0.1 username=- method=GET url="/apis/subresources.kubevirt.io/v1alpha2?timeout=32s" proto=HTTP/2.0 statusCode=200 contentLength=136 2018/08/03 22:48:05 http: TLS handshake error from 10.129.0.1:57396: EOF Pod name: virt-controller-7d57d96b65-fp8kf Pod phase: Running level=info timestamp=2018-08-03T22:42:03.772954Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmighx5p kind= uid=710249c0-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:42:03.773316Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmighx5p kind= uid=710249c0-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:42:03.879516Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiljd7g kind= uid=7112673d-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:42:03.879747Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiljd7g kind= uid=7112673d-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:42:04.029623Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqs85x kind= uid=712002a9-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:42:04.030112Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmiqs85x kind= uid=712002a9-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:45:05.767840Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmix58ts kind= uid=dd84924a-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:45:05.769294Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmix58ts kind= uid=dd84924a-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:45:05.796938Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikvqw9 kind= uid=dd898b9f-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:45:05.797299Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmikvqw9 kind= uid=dd898b9f-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:45:05.839619Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilpn6d kind= uid=dd8e7ae4-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:45:05.839835Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmilpn6d kind= uid=dd8e7ae4-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:45:05.896783Z pos=preset.go:142 component=virt-controller service=http namespace=kubevirt-test-default name=testvmicjpv5 kind= uid=dd9523f7-976e-11e8-948d-525500d15501 msg="Initializing VirtualMachineInstance" level=info timestamp=2018-08-03T22:45:05.897010Z pos=preset.go:171 component=virt-controller service=http namespace=kubevirt-test-default name=testvmicjpv5 kind= uid=dd9523f7-976e-11e8-948d-525500d15501 msg="Marking VirtualMachineInstance as initialized" level=info timestamp=2018-08-03T22:45:06.320810Z pos=vmi.go:157 component=virt-controller service=http reason="Operation cannot be fulfilled on virtualmachineinstances.kubevirt.io \"testvmix58ts\": the object has been modified; please apply your changes to the latest version and try again" msg="reenqueuing VirtualMachineInstance kubevirt-test-default/testvmix58ts" Pod name: virt-controller-7d57d96b65-r4stl Pod phase: Running level=info timestamp=2018-08-03T21:17:28.454463Z pos=application.go:177 component=virt-controller service=http action=listening interface=0.0.0.0 port=8182 Pod name: virt-handler-7fcr8 Pod phase: Running level=error timestamp=2018-08-03T22:23:43.366209Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:24:49.970848Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:26:25.552238Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:28:32.348619Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:30:41.174615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:06.317948Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:33:53.118704Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:35:32.905219Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:37:27.367654Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:56.563945Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:40:06.043026Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:42:10.638110Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:43:57.136630Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:45:03.374813Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:46:51.644630Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-handler-9rw2k Pod phase: Running level=error timestamp=2018-08-03T22:24:14.853754Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:25:21.525127Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:27:28.345378Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:29:37.221014Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:31:02.362336Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:32:52.141249Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:34:39.259890Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:36:33.718395Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:38:02.867765Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:39:12.412748Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:41:23.530615Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:43:10.059813Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:44:16.259098Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:46:04.568960Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" level=error timestamp=2018-08-03T22:47:49.459070Z pos=health.go:55 component=virt-handler reason="tun device does not show up in /proc/misc, is the module loaded?" msg="Check for mandatory device /dev/net/tun failed" Pod name: virt-launcher-testvmicjpv5-rzvvk Pod phase: Pending Pod name: virt-launcher-testvmikvqw9-wflhn Pod phase: Pending Pod name: virt-launcher-testvmilpn6d-w24ws Pod phase: Pending Pod name: virt-launcher-testvmix58ts-ftf6l Pod phase: Pending • Failure in Spec Setup (BeforeEach) [181.924 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:48 should be reachable via the propagated IP from a Pod [BeforeEach] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 on a different node from Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Timed out after 90.076s. Timed out waiting for VMI to enter Running phase Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 ------------------------------ level=info timestamp=2018-08-03T22:45:05.838569Z pos=utils.go:244 component=tests namespace=kubevirt-test-default name=testvmix58ts kind=VirtualMachineInstance uid=dd84924a-976e-11e8-948d-525500d15501 msg="Created virtual machine pod virt-launcher-testvmix58ts-ftf6l" panic: test timed out after 1h30m0s goroutine 15202 [running]: testing.(*M).startAlarm.func1() /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:1240 +0xfc created by time.goFunc /gimme/.gimme/versions/go1.10.linux.amd64/src/time/sleep.go:172 +0x44 goroutine 1 [chan receive, 90 minutes]: testing.(*T).Run(0xc42067e5a0, 0x12ad0d7, 0x9, 0x1339680, 0x47fa16) /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:825 +0x301 testing.runTests.func1(0xc42067e4b0) /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:1063 +0x64 testing.tRunner(0xc42067e4b0, 0xc42059ddf8) /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:777 +0xd0 testing.runTests(0xc420686680, 0x1bc0e20, 0x1, 0x1, 0x412009) /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:1061 +0x2c4 testing.(*M).Run(0xc420688d80, 0x0) /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:978 +0x171 main.main() _testmain.go:44 +0x151 goroutine 20 [chan receive]: kubevirt.io/kubevirt/vendor/github.com/golang/glog.(*loggingT).flushDaemon(0x1be7be0) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/golang/glog/glog.go:879 +0x8b created by kubevirt.io/kubevirt/vendor/github.com/golang/glog.init.0 /root/go/src/kubevirt.io/kubevirt/vendor/github.com/golang/glog/glog.go:410 +0x203 goroutine 21 [syscall, 90 minutes]: os/signal.signal_recv(0x0) /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/sigqueue.go:139 +0xa6 os/signal.loop() /gimme/.gimme/versions/go1.10.linux.amd64/src/os/signal/signal_unix.go:22 +0x22 created by os/signal.init.0 /gimme/.gimme/versions/go1.10.linux.amd64/src/os/signal/signal_unix.go:28 +0x41 goroutine 12 [select]: kubevirt.io/kubevirt/tests.(*ObjectEventWatcher).Watch(0xc420f1f030, 0xc4204514a0) /root/go/src/kubevirt.io/kubevirt/tests/utils.go:284 +0x579 kubevirt.io/kubevirt/tests.(*ObjectEventWatcher).WaitFor(0xc420f1f030, 0x12a90a0, 0x6, 0x111b660, 0x13b0aa0, 0x0) /root/go/src/kubevirt.io/kubevirt/tests/utils.go:294 +0xba kubevirt.io/kubevirt/tests.waitForVMIStart(0x13b9580, 0xc420e0aa00, 0x5a, 0x0, 0x0, 0x0) /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1010 +0x50e kubevirt.io/kubevirt/tests.WaitForSuccessfulVMIStart(0x13b9580, 0xc420e0aa00, 0xc420f1f120, 0xa206ae) /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1047 +0x43 kubevirt.io/kubevirt/tests.WaitUntilVMIReady(0xc420e0aa00, 0x1339658, 0x1c06250) /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1052 +0x3b kubevirt.io/kubevirt/tests_test.glob..func18.4() /root/go/src/kubevirt.io/kubevirt/tests/vmi_networking_test.go:135 +0x72b kubevirt.io/kubevirt/tests.BeforeAll.func1() /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1334 +0x3f kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/leafnodes.(*runner).runSync(0xc42012b200, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/leafnodes/runner.go:113 +0x9c kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/leafnodes.(*runner).run(0xc42012b200, 0xc4203c1670, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/leafnodes/runner.go:64 +0x13e kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/leafnodes.(*SetupNode).Run(0xc42081c438, 0x13b5fc0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/leafnodes/setup_nodes.go:15 +0x7f kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/spec.(*Spec).runSample(0xc4208eed20, 0x0, 0x13b5fc0, 0xc4201014c0) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/spec/spec.go:181 +0x1f1 kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/spec.(*Spec).Run(0xc4208eed20, 0x13b5fc0, 0xc4201014c0) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/spec/spec.go:138 +0xff kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).runSpec(0xc4207e8640, 0xc4208eed20, 0x0) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:200 +0x10d kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).runSpecs(0xc4207e8640, 0x1) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:170 +0x329 kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).Run(0xc4207e8640, 0xb) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:66 +0x11b kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/suite.(*Suite).Run(0xc4200feaf0, 0x7efffee6fb70, 0xc42067e5a0, 0x12af51b, 0xb, 0xc4206866c0, 0x2, 0x2, 0x13d0ca0, 0xc4201014c0, ...) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/suite/suite.go:62 +0x27c kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo.RunSpecsWithCustomReporters(0x13b6d00, 0xc42067e5a0, 0x12af51b, 0xb, 0xc4206866a0, 0x2, 0x2, 0x2) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/ginkgo_dsl.go:221 +0x258 kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo.RunSpecsWithDefaultAndCustomReporters(0x13b6d00, 0xc42067e5a0, 0x12af51b, 0xb, 0xc4205e26e0, 0x1, 0x1, 0x1) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/ginkgo_dsl.go:209 +0xab kubevirt.io/kubevirt/tests_test.TestTests(0xc42067e5a0) /root/go/src/kubevirt.io/kubevirt/tests/tests_suite_test.go:43 +0xaa testing.tRunner(0xc42067e5a0, 0x1339680) /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:777 +0xd0 created by testing.(*T).Run /gimme/.gimme/versions/go1.10.linux.amd64/src/testing/testing.go:824 +0x2e0 goroutine 50 [chan receive, 90 minutes]: kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).registerForInterrupts(0xc4207e8640, 0xc420920000) /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:223 +0xd1 created by kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner.(*SpecRunner).Run /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/internal/specrunner/spec_runner.go:60 +0x88 goroutine 51 [select, 90 minutes, locked to thread]: runtime.gopark(0x133b4c0, 0x0, 0x12a9dde, 0x6, 0x18, 0x1) /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/proc.go:291 +0x11a runtime.selectgo(0xc420486f50, 0xc4209200c0) /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/select.go:392 +0xe50 runtime.ensureSigM.func1() /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/signal_unix.go:549 +0x1f4 runtime.goexit() /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/asm_amd64.s:2361 +0x1 goroutine 71 [IO wait]: internal/poll.runtime_pollWait(0x7efffeef8f00, 0x72, 0xc420edf850) /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/netpoll.go:173 +0x57 internal/poll.(*pollDesc).wait(0xc4208b3f18, 0x72, 0xffffffffffffff00, 0x13b7ca0, 0x1ad8640) /gimme/.gimme/versions/go1.10.linux.amd64/src/internal/poll/fd_poll_runtime.go:85 +0x9b internal/poll.(*pollDesc).waitRead(0xc4208b3f18, 0xc420d0c000, 0x8000, 0x8000) /gimme/.gimme/versions/go1.10.linux.amd64/src/internal/poll/fd_poll_runtime.go:90 +0x3d internal/poll.(*FD).Read(0xc4208b3f00, 0xc420d0c000, 0x8000, 0x8000, 0x0, 0x0, 0x0) /gimme/.gimme/versions/go1.10.linux.amd64/src/internal/poll/fd_unix.go:157 +0x17d net.(*netFD).Read(0xc4208b3f00, 0xc420d0c000, 0x8000, 0x8000, 0x0, 0x8, 0x7ffb) /gimme/.gimme/versions/go1.10.linux.amd64/src/net/fd_unix.go:202 +0x4f net.(*conn).Read(0xc42081c478, 0xc420d0c000, 0x8000, 0x8000, 0x0, 0x0, 0x0) /gimme/.gimme/versions/go1.10.linux.amd64/src/net/net.go:176 +0x6a crypto/tls.(*block).readFromUntil(0xc4208b1950, 0x7efffc3a71c8, 0xc42081c478, 0x5, 0xc42081c478, 0x0) /gimme/.gimme/versions/go1.10.linux.amd64/src/crypto/tls/conn.go:493 +0x96 crypto/tls.(*Conn).readRecord(0xc420582380, 0x133b617, 0xc4205824a0, 0x20) /gimme/.gimme/versions/go1.10.linux.amd64/src/crypto/tls/conn.go:595 +0xe0 crypto/tls.(*Conn).Read(0xc420582380, 0xc4205d6000, 0x1000, 0x1000, 0x0, 0x0, 0x0) /gimme/.gimme/versions/go1.10.linux.amd64/src/crypto/tls/conn.go:1156 +0x100 bufio.(*Reader).Read(0xc42075eb40, 0xc42083c498, 0x9, 0x9, 0xc420f2bda8, 0x1, 0x0) /gimme/.gimme/versions/go1.10.linux.amd64/src/bufio/bufio.go:216 +0x238 io.ReadAtLeast(0x13b4e00, 0xc42075eb40, 0xc42083c498, 0x9, 0x9, 0x9, 0xc420d2c2a0, 0x43f2c1, 0xc420c3b980) /gimme/.gimme/versions/go1.10.linux.amd64/src/io/io.go:309 +0x86 io.ReadFull(0x13b4e00, 0xc42075eb40, 0xc42083c498, 0x9, 0x9, 0x133b608, 0xc420edfd10, 0x462d33) /gimme/.gimme/versions/go1.10.linux.amd64/src/io/io.go:327 +0x58 kubevirt.io/kubevirt/vendor/golang.org/x/net/http2.readFrameHeader(0xc42083c498, 0x9, 0x9, 0x13b4e00, 0xc42075eb40, 0x0, 0xc400000000, 0x7babc0, 0xc420f6c528) /root/go/src/kubevirt.io/kubevirt/vendor/golang.org/x/net/http2/frame.go:237 +0x7b kubevirt.io/kubevirt/vendor/golang.org/x/net/http2.(*Framer).ReadFrame(0xc42083c460, 0xc420a08000, 0x0, 0x0, 0x0) /root/go/src/kubevirt.io/kubevirt/vendor/golang.org/x/net/http2/frame.go:492 +0xa4 kubevirt.io/kubevirt/vendor/golang.org/x/net/http2.(*clientConnReadLoop).run(0xc420edffb0, 0x133a5d8, 0xc420483fb0) /root/go/src/kubevirt.io/kubevirt/vendor/golang.org/x/net/http2/transport.go:1428 +0x8e kubevirt.io/kubevirt/vendor/golang.org/x/net/http2.(*ClientConn).readLoop(0xc42010dba0) /root/go/src/kubevirt.io/kubevirt/vendor/golang.org/x/net/http2/transport.go:1354 +0x76 created by kubevirt.io/kubevirt/vendor/golang.org/x/net/http2.(*Transport).newClientConn /root/go/src/kubevirt.io/kubevirt/vendor/golang.org/x/net/http2/transport.go:579 +0x651 goroutine 15173 [semacquire]: sync.runtime_notifyListWait(0xc420f6c540, 0xc400000001) /gimme/.gimme/versions/go1.10.linux.amd64/src/runtime/sema.go:510 +0x10b sync.(*Cond).Wait(0xc420f6c530) /gimme/.gimme/versions/go1.10.linux.amd64/src/sync/cond.go:56 +0x80 kubevirt.io/kubevirt/vendor/golang.org/x/net/http2.(*pipe).Read(0xc420f6c528, 0xc420dee001, 0x5ff, 0x5ff, 0x0, 0x0, 0x0) /root/go/src/kubevirt.io/kubevirt/vendor/golang.org/x/net/http2/pipe.go:64 +0x8f kubevirt.io/kubevirt/vendor/golang.org/x/net/http2.transportResponseBody.Read(0xc420f6c500, 0xc420dee001, 0x5ff, 0x5ff, 0x0, 0x0, 0x0) /root/go/src/kubevirt.io/kubevirt/vendor/golang.org/x/net/http2/transport.go:1674 +0xa1 encoding/json.(*Decoder).refill(0xc420dfc000, 0x7fb90a, 0x9) /gimme/.gimme/versions/go1.10.linux.amd64/src/encoding/json/stream.go:159 +0x132 encoding/json.(*Decoder).readValue(0xc420dfc000, 0x0, 0x0, 0x112bf00) /gimme/.gimme/versions/go1.10.linux.amd64/src/encoding/json/stream.go:134 +0x23d encoding/json.(*Decoder).Decode(0xc420dfc000, 0x1145800, 0xc4206e41c0, 0x13bc280, 0xc420c98000) /gimme/.gimme/versions/go1.10.linux.amd64/src/encoding/json/stream.go:63 +0x78 kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/util/framer.(*jsonFrameReader).Read(0xc42084a0f0, 0xc420e24400, 0x400, 0x400, 0xc4207d0140, 0x40, 0x38) /root/go/src/kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/util/framer/framer.go:150 +0x295 kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/runtime/serializer/streaming.(*decoder).Decode(0xc420d81900, 0x0, 0x13bdf40, 0xc4207d0140, 0x380, 0x13bc280, 0xc420921678, 0x456ae0, 0xc420921620) /root/go/src/kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/runtime/serializer/streaming/streaming.go:77 +0x95 kubevirt.io/kubevirt/vendor/k8s.io/client-go/rest/watch.(*Decoder).Decode(0xc420749600, 0xc420eddfa8, 0x5, 0x13bc280, 0xc420c98000, 0x0, 0x0) /root/go/src/kubevirt.io/kubevirt/vendor/k8s.io/client-go/rest/watch/decoder.go:49 +0x7c kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/watch.(*StreamWatcher).receive(0xc42084a120) /root/go/src/kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/watch/streamwatcher.go:93 +0x12e created by kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/watch.NewStreamWatcher /root/go/src/kubevirt.io/kubevirt/vendor/k8s.io/apimachinery/pkg/watch/streamwatcher.go:60 +0xa8 goroutine 15174 [chan receive]: kubevirt.io/kubevirt/tests.(*ObjectEventWatcher).Watch.func3(0x13be4c0, 0xc42084a120, 0xc4207360a0, 0xc420921680) /root/go/src/kubevirt.io/kubevirt/tests/utils.go:275 +0x93 created by kubevirt.io/kubevirt/tests.(*ObjectEventWatcher).Watch /root/go/src/kubevirt.io/kubevirt/tests/utils.go:273 +0x4ae make: *** [functest] Error 2 + make cluster-down ./cluster/down.sh