+ export WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio + WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio + [[ openshift-release-crio =~ openshift-.* ]] + export PROVIDER=os-3.9.0 + PROVIDER=os-3.9.0 + export VAGRANT_NUM_NODES=1 + VAGRANT_NUM_NODES=1 + export NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + export NAMESPACE=kube-system + NAMESPACE=kube-system + trap '{ make cluster-down; }' EXIT + make cluster-down ./cluster/down.sh + make cluster-up ./cluster/up.sh kubevirt-functional-tests-openshift-release-crio0_registry WARNING: You're not using the default seccomp profile kubevirt-functional-tests-openshift-release-crio0-node02 2018/04/30 18:06:03 Waiting for host: 192.168.66.102:22 2018/04/30 18:06:06 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/30 18:06:14 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/30 18:06:22 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/30 18:06:27 Connected to tcp://192.168.66.102:22 Removed symlink /etc/systemd/system/multi-user.target.wants/origin-master-api.service. Removed symlink /etc/systemd/system/origin-node.service.wants/origin-master-api.service. Removed symlink /etc/systemd/system/multi-user.target.wants/origin-master-controllers.service. kubevirt-functional-tests-openshift-release-crio0-node01 2018/04/30 18:06:34 Waiting for host: 192.168.66.101:22 2018/04/30 18:06:37 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/30 18:06:45 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/30 18:06:53 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/30 18:06:58 Connected to tcp://192.168.66.101:22 The connection to the server node01:8443 was refused - did you specify the right host or port? NAME STATUS ROLES AGE VERSION node01 Ready master 11d v1.9.1+a0ce1bc657 PING node02 (192.168.66.102) 56(84) bytes of data. 64 bytes from node02 (192.168.66.102): icmp_seq=1 ttl=64 time=0.880 ms --- node02 ping statistics --- 1 packets transmitted, 1 received, 0% packet loss, time 0ms rtt min/avg/max/mdev = 0.880/0.880/0.880/0.000 ms Found node02. Adding it to the inventory. ping: node03: Name or service not known PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) PLAY [Ensure there are new_nodes] ********************************************** TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] PLAY [Initialization Checkpoint Start] ***************************************** TASK [Set install initialization 'In Progress'] ******************************** ok: [node01] PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) [WARNING]: Could not match supplied host pattern, ignoring: oo_lb_to_config PLAY [Ensure that all non-node hosts are accessible] *************************** TASK [Gathering Facts] ********************************************************* ok: [node01] PLAY [Initialize basic host facts] ********************************************* TASK [Gathering Facts] ********************************************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/deprecations.yml for node01, node02 TASK [openshift_sanitize_inventory : Check for usage of deprecated variables] *** ok: [node02] ok: [node01] TASK [openshift_sanitize_inventory : debug] ************************************ skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_stats] ******************************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Assign deprecated variables to correct counterparts] *** included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_logging.yml for node01, node02 included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_metrics.yml for node01, node02 TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node02] ok: [node01] TASK [openshift_sanitize_inventory : Standardize on latest variable names] ***** ok: [node02] ok: [node01] TASK [openshift_sanitize_inventory : Normalize openshift_release] ************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Abort when openshift_release is invalid] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/unsupported.yml for node01, node02 TASK [openshift_sanitize_inventory : Ensure that openshift_use_dnsmasq is true] *** skipping: [node02] skipping: [node01] TASK [openshift_sanitize_inventory : Ensure that openshift_node_dnsmasq_install_network_manager_hook is true] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* skipping: [node01] => (item=None) skipping: [node02] => (item=None) TASK [openshift_sanitize_inventory : Ensure that dynamic provisioning is set if using dynamic storage] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure clusterid is set along with the cloudprovider] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure ansible_service_broker_remove and ansible_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure template_service_broker_remove and template_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that all requires vsphere configuration variables are set] *** skipping: [node01] skipping: [node02] TASK [Detecting Operating System from ostree_booted] *************************** ok: [node01] ok: [node02] TASK [set openshift_deployment_type if unset] ********************************** skipping: [node01] skipping: [node02] TASK [initialize_facts set fact openshift_is_atomic and openshift_is_containerized] *** ok: [node01] ok: [node02] TASK [Determine Atomic Host Docker Version] ************************************ skipping: [node01] skipping: [node02] TASK [assert atomic host docker version is 1.12 or later] ********************** skipping: [node01] skipping: [node02] PLAY [Initialize special first-master variables] ******************************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] PLAY [Disable web console if required] ***************************************** TASK [set_fact] **************************************************************** skipping: [node01] PLAY [Install packages necessary for installer] ******************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Ensure openshift-ansible installer package deps are installed] *********** ok: [node02] => (item=iproute) ok: [node02] => (item=dbus-python) ok: [node02] => (item=PyYAML) ok: [node02] => (item=python-ipaddress) ok: [node02] => (item=yum-utils) TASK [Ensure various deps for running system containers are installed] ********* skipping: [node02] => (item=atomic) skipping: [node02] => (item=ostree) skipping: [node02] => (item=runc) PLAY [Initialize cluster facts] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [Gather Cluster facts] **************************************************** changed: [node02] ok: [node01] TASK [Set fact of no_proxy_internal_hostnames] ********************************* skipping: [node01] skipping: [node02] TASK [Initialize openshift.node.sdn_mtu] *************************************** ok: [node02] ok: [node01] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* skipping: [node01] TASK [include_role] ************************************************************ skipping: [node01] TASK [debug] ******************************************************************* skipping: [node01] PLAY [Set openshift_version for etcd, node, and master hosts] ****************** skipping: no hosts matched PLAY [Ensure the requested version packages are available.] ******************** skipping: no hosts matched PLAY [Verify Requirements] ***************************************************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [Run variable sanity checks] ********************************************** ok: [node01] PLAY [Initialization Checkpoint End] ******************************************* TASK [Set install initialization 'Complete'] *********************************** ok: [node01] PLAY [Validate node hostnames] ************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Query DNS for IP address of node02] ************************************** ok: [node02] TASK [Validate openshift_hostname when defined] ******************************** skipping: [node02] TASK [Validate openshift_ip exists on node when defined] *********************** skipping: [node02] PLAY [Setup yum repositories for all hosts] ************************************ TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Install Red Hat Subscription manager] ******************* skipping: [node02] TASK [rhel_subscribe : Is host already registered?] **************************** skipping: [node02] TASK [rhel_subscribe : Register host] ****************************************** skipping: [node02] TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Determine if OpenShift Pool Already Attached] *********** skipping: [node02] TASK [rhel_subscribe : Attach to OpenShift Pool] ******************************* skipping: [node02] TASK [rhel_subscribe : include_tasks] ****************************************** skipping: [node02] TASK [openshift_repos : openshift_repos detect ostree] ************************* ok: [node02] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** ok: [node02] TASK [openshift_repos : Remove openshift_additional.repo file] ***************** ok: [node02] TASK [openshift_repos : Create any additional repos that are defined] ********** TASK [openshift_repos : include_tasks] ***************************************** skipping: [node02] TASK [openshift_repos : include_tasks] ***************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_repos/tasks/centos_repos.yml for node02 TASK [openshift_repos : Configure origin gpg keys] ***************************** ok: [node02] TASK [openshift_repos : Configure correct origin release repository] *********** ok: [node02] => (item=/usr/share/ansible/openshift-ansible/roles/openshift_repos/templates/CentOS-OpenShift-Origin.repo.j2) TASK [openshift_repos : Ensure clean repo cache in the event repos have been changed manually] *** changed: [node02] => { "msg": "First run of openshift_repos" } TASK [openshift_repos : Record that openshift_repos already ran] *************** ok: [node02] RUNNING HANDLER [openshift_repos : refresh cache] ****************************** changed: [node02] PLAY [Configure os_firewall] *************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [os_firewall : Detecting Atomic Host Operating System] ******************** ok: [node02] TASK [os_firewall : Set fact r_os_firewall_is_atomic] ************************** ok: [node02] TASK [os_firewall : include_tasks] ********************************************* skipping: [node02] TASK [os_firewall : include_tasks] ********************************************* included: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/iptables.yml for node02 TASK [os_firewall : Ensure firewalld service is not enabled] ******************* ok: [node02] TASK [os_firewall : Wait 10 seconds after disabling firewalld] ***************** skipping: [node02] TASK [os_firewall : Install iptables packages] ********************************* ok: [node02] => (item=iptables) ok: [node02] => (item=iptables-services) TASK [os_firewall : Start and enable iptables service] ************************* ok: [node02 -> node02] => (item=node02) TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [container_runtime : Setup the docker-storage for overlay] **************** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_excluder : Install excluders] ********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** ok: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* ok: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/pre.yml for node02 TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Add enterprise registry, if necessary] *************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Get current installed Docker version] **************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/docker_sanity.yml for node02 TASK [container_runtime : Error out if Docker pre-installed but too old] ******* skipping: [node02] TASK [container_runtime : Error out if requested Docker is too old] ************ skipping: [node02] TASK [container_runtime : Fail if Docker version requested but downgrade is required] *** skipping: [node02] TASK [container_runtime : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [node02] TASK [container_runtime : Install Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure docker.service.d directory exists] ************ ok: [node02] TASK [container_runtime : Configure Docker service unit file] ****************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Set registry params] ********************************* skipping: [node02] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': []}) TASK [container_runtime : Place additional/blocked/insecure registries in /etc/containers/registries.conf] *** skipping: [node02] TASK [container_runtime : Set Proxy Settings] ********************************** skipping: [node02] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [container_runtime : Set various Docker options] ************************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Configure Docker Network OPTIONS] ******************** ok: [node02] TASK [container_runtime : Detect if docker is already started] ***************** ok: [node02] TASK [container_runtime : Start the Docker service] **************************** ok: [node02] TASK [container_runtime : set_fact] ******************************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/post.yml for node02 TASK [container_runtime : Ensure /var/lib/containers exists] ******************* ok: [node02] TASK [container_runtime : Fix SELinux Permissions on /var/lib/containers] ****** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/registry_auth.yml for node02 TASK [container_runtime : Check for credentials file for registry auth] ******** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth] ***** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth (alternative)] *** skipping: [node02] TASK [container_runtime : stat the docker data dir] **************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Fail quickly if openshift_docker_options are set] **** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Install Docker so we can use the client] ************* skipping: [node02] TASK [container_runtime : Disable Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** skipping: [node02] TASK [container_runtime : debug] *********************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Pre-pull Container Engine System Container image] **** skipping: [node02] TASK [container_runtime : Ensure container-engine.service.d directory exists] *** skipping: [node02] TASK [container_runtime : Ensure /etc/docker directory exists] ***************** skipping: [node02] TASK [container_runtime : Install Container Engine System Container] *********** skipping: [node02] TASK [container_runtime : Configure Container Engine Service File] ************* skipping: [node02] TASK [container_runtime : Configure Container Engine] ************************** skipping: [node02] TASK [container_runtime : Start the Container Engine service] ****************** skipping: [node02] TASK [container_runtime : set_fact] ******************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Check we are not using node as a Docker container with CRI-O] *** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Check that overlay is in the kernel] ***************** skipping: [node02] TASK [container_runtime : Add overlay to modprobe.d] *************************** skipping: [node02] TASK [container_runtime : Manually modprobe overlay into the kernel] *********** skipping: [node02] TASK [container_runtime : Enable and start systemd-modules-load] *************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** skipping: [node02] TASK [container_runtime : debug] *********************************************** skipping: [node02] TASK [container_runtime : Pre-pull CRI-O System Container image] *************** skipping: [node02] TASK [container_runtime : Install CRI-O System Container] ********************** skipping: [node02] TASK [container_runtime : Remove CRI-O default configuration files] ************ skipping: [node02] => (item=/etc/cni/net.d/200-loopback.conf) skipping: [node02] => (item=/etc/cni/net.d/100-crio-bridge.conf) TASK [container_runtime : Create the CRI-O configuration] ********************** skipping: [node02] TASK [container_runtime : Ensure CNI configuration directory exists] *********** skipping: [node02] TASK [container_runtime : Add iptables allow rules] **************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove iptables rules] ******************************* TASK [container_runtime : Add firewalld allow rules] *************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove firewalld allow rules] ************************ TASK [container_runtime : Configure the CNI network] *************************** skipping: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-storage] ****************** skipping: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-network] ****************** skipping: [node02] TASK [container_runtime : Start the CRI-O service] ***************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [include_role] ************************************************************ TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** ok: [node01] TASK [openshift_version : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_version/tasks/first_master_rpm_version.yml for node01 TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [node01] TASK [openshift_version : Set openshift_version for rpm installation] ********** included: /usr/share/ansible/openshift-ansible/roles/openshift_version/tasks/check_available_rpms.yml for node01 TASK [openshift_version : Get available origin version] ************************ ok: [node01] TASK [openshift_version : fail] ************************************************ skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : set_fact] ******************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [debug] ******************************************************************* ok: [node01] => { "msg": "openshift_pkg_version set to -3.9.0" } PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [set_fact] **************************************************************** ok: [node02] PLAY [Ensure the requested version packages are available.] ******************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [include_role] ************************************************************ TASK [openshift_version : Check openshift_version for rpm installation] ******** included: /usr/share/ansible/openshift-ansible/roles/openshift_version/tasks/check_available_rpms.yml for node02 TASK [openshift_version : Get available origin version] ************************ ok: [node02] TASK [openshift_version : fail] ************************************************ skipping: [node02] TASK [openshift_version : Fail if rpm version and docker image version are different] *** skipping: [node02] TASK [openshift_version : For an RPM install, abort when the release requested does not match the available version.] *** skipping: [node02] TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_image_tag": "v3.9.0" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_pkg_version": "-3.9.0" } PLAY [Node Install Checkpoint Start] ******************************************* TASK [Set Node install 'In Progress'] ****************************************** ok: [node01] PLAY [Create OpenShift certificates for node hosts] **************************** TASK [openshift_node_certificates : Ensure CA certificate exists on openshift_ca_host] *** ok: [node02 -> node01] TASK [openshift_node_certificates : fail] ************************************** skipping: [node02] TASK [openshift_node_certificates : Check status of node certificates] ********* ok: [node02] => (item=system:node:node02.crt) ok: [node02] => (item=system:node:node02.key) ok: [node02] => (item=system:node:node02.kubeconfig) ok: [node02] => (item=ca.crt) ok: [node02] => (item=server.key) ok: [node02] => (item=server.crt) TASK [openshift_node_certificates : set_fact] ********************************** ok: [node02] TASK [openshift_node_certificates : Create openshift_generated_configs_dir if it does not exist] *** ok: [node02 -> node01] TASK [openshift_node_certificates : find] ************************************** ok: [node02 -> node01] TASK [openshift_node_certificates : Generate the node client config] *********** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Generate the node server certificate] ****** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Create a tarball of the node config directories] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Retrieve the node config tarballs from the master] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Ensure certificate directory exists] ******* ok: [node02] TASK [openshift_node_certificates : Unarchive the tarball on the node] ********* changed: [node02] TASK [openshift_node_certificates : Delete local temp directory] *************** ok: [node02 -> localhost] TASK [openshift_node_certificates : Copy OpenShift CA to system CA trust] ****** ok: [node02] => (item={u'cert': u'/etc/origin/node/ca.crt', u'id': u'openshift'}) PLAY [Disable excluders] ******************************************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": true } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": true } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/disable.yml for node02 TASK [openshift_excluder : Include verify_upgrade.yml when upgrading] ********** skipping: [node02] TASK [openshift_excluder : Disable excluders before the upgrade to remove older excluding expressions] *** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* changed: [node02] TASK [openshift_excluder : Include install.yml] ******************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Include exclude.yml] ******************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** changed: [node02] TASK [openshift_excluder : Include unexclude.yml] ****************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* changed: [node02] PLAY [Evaluate node groups] **************************************************** TASK [Gathering Facts] ********************************************************* ok: [localhost] TASK [Evaluate oo_containerized_master_nodes] ********************************** skipping: [localhost] => (item=node02) [WARNING]: Could not match supplied host pattern, ignoring: oo_containerized_master_nodes PLAY [Configure containerized nodes] ******************************************* skipping: no hosts matched PLAY [Configure nodes] ********************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_clock : Determine if chrony is installed] ********************** [WARNING]: Consider using yum, dnf or zypper module rather than running rpm changed: [node02] TASK [openshift_clock : Install ntp package] *********************************** skipping: [node02] TASK [openshift_clock : Start and enable ntpd/chronyd] ************************* changed: [node02] TASK [openshift_cloud_provider : Set cloud provider facts] ********************* skipping: [node02] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [node02] TASK [openshift_cloud_provider : include the defined cloud provider files] ***** skipping: [node02] TASK [openshift_node : fail] *************************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/dnsmasq_install.yml for node02 TASK [openshift_node : Check for NetworkManager service] *********************** ok: [node02] TASK [openshift_node : Set fact using_network_manager] ************************* ok: [node02] TASK [openshift_node : Install dnsmasq] **************************************** ok: [node02] TASK [openshift_node : ensure origin/node directory exists] ******************** ok: [node02] => (item=/etc/origin) changed: [node02] => (item=/etc/origin/node) TASK [openshift_node : Install node-dnsmasq.conf] ****************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/dnsmasq.yml for node02 TASK [openshift_node : Install dnsmasq configuration] ************************** ok: [node02] TASK [openshift_node : Deploy additional dnsmasq.conf] ************************* skipping: [node02] TASK [openshift_node : Enable dnsmasq] ***************************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/dnsmasq/network-manager.yml for node02 TASK [openshift_node : Install network manager dispatch script] **************** ok: [node02] TASK [openshift_node : Add iptables allow rules] ******************************* ok: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) ok: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) ok: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) ok: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove iptables rules] ********************************** TASK [openshift_node : Add firewalld allow rules] ****************************** skipping: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) skipping: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) skipping: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove firewalld allow rules] *************************** TASK [openshift_node : Update journald config] ********************************* included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/journald.yml for node02 TASK [openshift_node : Checking for journald.conf] ***************************** ok: [node02] TASK [openshift_node : Create journald persistence directories] **************** ok: [node02] TASK [openshift_node : Update journald setup] ********************************** ok: [node02] => (item={u'var': u'Storage', u'val': u'persistent'}) ok: [node02] => (item={u'var': u'Compress', u'val': True}) ok: [node02] => (item={u'var': u'SyncIntervalSec', u'val': u'1s'}) ok: [node02] => (item={u'var': u'RateLimitInterval', u'val': u'1s'}) ok: [node02] => (item={u'var': u'RateLimitBurst', u'val': 10000}) ok: [node02] => (item={u'var': u'SystemMaxUse', u'val': u'8G'}) ok: [node02] => (item={u'var': u'SystemKeepFree', u'val': u'20%'}) ok: [node02] => (item={u'var': u'SystemMaxFileSize', u'val': u'10M'}) ok: [node02] => (item={u'var': u'MaxRetentionSec', u'val': u'1month'}) ok: [node02] => (item={u'var': u'MaxFileSec', u'val': u'1day'}) ok: [node02] => (item={u'var': u'ForwardToSyslog', u'val': False}) ok: [node02] => (item={u'var': u'ForwardToWall', u'val': False}) TASK [openshift_node : Restart journald] *************************************** skipping: [node02] TASK [openshift_node : Disable swap] ******************************************* ok: [node02] TASK [openshift_node : include node installer] ********************************* included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/install.yml for node02 TASK [openshift_node : Install Node package, sdn-ovs, conntrack packages] ****** ok: [node02] => (item={u'name': u'origin-node-3.9.0'}) ok: [node02] => (item={u'name': u'origin-sdn-ovs-3.9.0', u'install': True}) ok: [node02] => (item={u'name': u'conntrack-tools'}) TASK [openshift_node : Pre-pull node image when containerized] ***************** skipping: [node02] TASK [openshift_node : Restart cri-o] ****************************************** skipping: [node02] TASK [openshift_node : restart NetworkManager to ensure resolv.conf is present] *** skipping: [node02] TASK [openshift_node : sysctl] ************************************************* ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/registry_auth.yml for node02 TASK [openshift_node : Check for credentials file for registry auth] *********** skipping: [node02] TASK [openshift_node : Create credentials for registry auth] ******************* skipping: [node02] TASK [openshift_node : Create credentials for registry auth (alternative)] ***** skipping: [node02] TASK [openshift_node : Setup ro mount of /root/.docker for containerized hosts] *** skipping: [node02] TASK [openshift_node : include standard node config] *************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config.yml for node02 TASK [openshift_node : Install the systemd units] ****************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/systemd_units.yml for node02 TASK [openshift_node : Install Node service file] ****************************** ok: [node02] TASK [openshift_node : include node deps docker service file] ****************** skipping: [node02] TASK [openshift_node : include ovs service environment file] ******************* skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config/configure-node-settings.yml for node02 TASK [openshift_node : Configure Node settings] ******************************** ok: [node02] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS=--loglevel=2 '}) ok: [node02] => (item={u'regex': u'^CONFIG_FILE=', u'line': u'CONFIG_FILE=/etc/origin/node/node-config.yaml'}) ok: [node02] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.9.0'}) TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config/configure-proxy-settings.yml for node02 TASK [openshift_node : Configure Proxy Settings] ******************************* skipping: [node02] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) skipping: [node02] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [node02] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=[],172.30.0.0/16,10.128.0.0/14'}) TASK [openshift_node : Pull container images] ********************************** skipping: [node02] TASK [openshift_node : Start and enable openvswitch service] ******************* skipping: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : file] *************************************************** skipping: [node02] TASK [openshift_node : Create the Node config] ********************************* changed: [node02] TASK [openshift_node : Configure Node Environment Variables] ******************* TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [node02] => (item=None) skipping: [node02] => (item=None) TASK [openshift_node : Wait for master API to become available before proceeding] *** skipping: [node02] TASK [openshift_node : Start and enable node dep] ****************************** skipping: [node02] TASK [openshift_node : Start and enable node] ********************************** ok: [node02] TASK [openshift_node : Dump logs from node service if it failed] *************** skipping: [node02] TASK [openshift_node : Abort if node failed to start] ************************** skipping: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : NFS storage plugin configuration] *********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/nfs.yml for node02 TASK [openshift_node : Install NFS storage plugin dependencies] **************** ok: [node02] TASK [openshift_node : Check for existence of nfs sebooleans] ****************** ok: [node02] => (item=virt_use_nfs) ok: [node02] => (item=virt_sandbox_use_nfs) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:40.083652', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.011408', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:40.072244', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:41.274082', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.010539', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:41.263543', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:40.083652', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.011408', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:40.072244', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:41.274082', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.010539', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:41.263543', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : GlusterFS storage plugin configuration] ***************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/glusterfs.yml for node02 TASK [openshift_node : Install GlusterFS storage plugin dependencies] ********** ok: [node02] TASK [openshift_node : Check for existence of fusefs sebooleans] *************** ok: [node02] => (item=virt_use_fusefs) ok: [node02] => (item=virt_sandbox_use_fusefs) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:48.081448', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.010232', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:48.071216', '_ansible_ignore_errors': None, 'failed': False}) ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:49.686978', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.020426', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:49.666552', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:48.081448', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.010232', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:48.071216', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-30 18:13:49.686978', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.020426', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-30 18:13:49.666552', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Ceph storage plugin configuration] ********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/ceph.yml for node02 TASK [openshift_node : Install Ceph storage plugin dependencies] *************** ok: [node02] TASK [openshift_node : iSCSI storage plugin configuration] ********************* included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/iscsi.yml for node02 TASK [openshift_node : Install iSCSI storage plugin dependencies] ************** ok: [node02] => (item=iscsi-initiator-utils) ok: [node02] => (item=device-mapper-multipath) TASK [openshift_node : restart services] *************************************** ok: [node02] => (item=multipathd) ok: [node02] => (item=rpcbind) TASK [openshift_node : Template multipath configuration] *********************** changed: [node02] TASK [openshift_node : Enable multipath] *************************************** changed: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config/workaround-bz1331590-ovs-oom-fix.yml for node02 TASK [openshift_node : Create OpenvSwitch service.d directory] ***************** ok: [node02] TASK [openshift_node : Install OpenvSwitch service OOM fix] ******************** ok: [node02] TASK [tuned : Check for tuned package] ***************************************** ok: [node02] TASK [tuned : Set tuned OpenShift variables] *********************************** ok: [node02] TASK [tuned : Ensure directory structure exists] ******************************* ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524143912.172371, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524143912.172371, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524143912.172371, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524143912.172371, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524143912.172371, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524143912.172371, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524143912.172371, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524143912.172371, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': u's0', 'seuser': u'system_u', 'serole': u'object_r', 'ctime': 1524143912.172371, 'state': u'file', 'gid': 0, 'mode': u'0644', 'mtime': 1523891799.0, 'owner': u'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': u'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': u's0', 'seuser': u'system_u', 'serole': u'object_r', 'ctime': 1524143912.172371, 'state': u'file', 'gid': 0, 'mode': u'0644', 'mtime': 1523891799.0, 'owner': u'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': u'usr_t'}) TASK [tuned : Ensure files are populated from templates] *********************** skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524143912.172371, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524143912.172371, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524143912.172371, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524143912.172371, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524143912.172371, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524143912.172371, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524143912.172371, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524143912.172371, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524143912.172371, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524143912.172371, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) TASK [tuned : Make tuned use the recommended tuned profile on restart] ********* changed: [node02] => (item=/etc/tuned/active_profile) ok: [node02] => (item=/etc/tuned/profile_mode) TASK [tuned : Restart tuned service] ******************************************* changed: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Install logrotate] ******* ok: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Setup logrotate.d scripts] *** RUNNING HANDLER [openshift_node : restart node] ******************************** changed: [node02] PLAY [create additional node network plugin groups] **************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] PLAY [etcd_client node config] ************************************************* skipping: no hosts matched [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_flannel [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_calico [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_contiv [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_kuryr PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** skipping: no hosts matched [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_nuage PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Configure Contiv masters] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node01] PLAY [Configure rest of Contiv nodes] ****************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] PLAY [Configure Kuryr node] **************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_manage_node : Wait for master API to become available before proceeding] *** skipping: [node02] TASK [openshift_manage_node : Wait for Node Registration] ********************** ok: [node02 -> node01] TASK [openshift_manage_node : include_tasks] *********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_manage_node/tasks/config.yml for node02 TASK [openshift_manage_node : Set node schedulability] ************************* ok: [node02 -> node01] TASK [openshift_manage_node : Label nodes] ************************************* ok: [node02 -> node01] TASK [Create group for deployment type] **************************************** ok: [node02] PLAY [Re-enable excluder if it was previously enabled] ************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": true } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": true } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/enable.yml for node02 TASK [openshift_excluder : Install excluders] ********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** changed: [node02] PLAY [Node Install Checkpoint End] ********************************************* TASK [Set Node install 'Complete'] ********************************************* ok: [node01] PLAY RECAP ********************************************************************* localhost : ok=25 changed=0 unreachable=0 failed=0 node01 : ok=42 changed=0 unreachable=0 failed=0 node02 : ok=184 changed=25 unreachable=0 failed=0 INSTALLER STATUS *************************************************************** Initialization : Complete (0:01:17) Node Install : Complete (0:03:48) PLAY [new_nodes] *************************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Restart openvswitch service] ********************************************* changed: [node02] PLAY RECAP ********************************************************************* node02 : ok=2 changed=1 unreachable=0 failed=0 2018/04/30 18:15:29 Waiting for host: 192.168.66.101:22 2018/04/30 18:15:29 Connected to tcp://192.168.66.101:22 2018/04/30 18:15:37 Waiting for host: 192.168.66.101:22 2018/04/30 18:15:37 Connected to tcp://192.168.66.101:22 Warning: Permanently added '[127.0.0.1]:33113' (ECDSA) to the list of known hosts. Warning: Permanently added '[127.0.0.1]:33113' (ECDSA) to the list of known hosts. Cluster "node01:8443" set. Cluster "node01:8443" set. ++ kubectl get nodes --no-headers ++ cluster/kubectl.sh get nodes --no-headers ++ grep -v Ready + '[' -n '' ']' + echo 'Nodes are ready:' Nodes are ready: + kubectl get nodes + cluster/kubectl.sh get nodes NAME STATUS ROLES AGE VERSION node01 Ready master 11d v1.9.1+a0ce1bc657 node02 Ready 1m v1.9.1+a0ce1bc657 + make cluster-sync ./cluster/build.sh Building ... sha256:0e817e41f9750e44335dde1be5cb34809abe48c8add43baf165907418e2e75ce go version go1.10 linux/amd64 go version go1.10 linux/amd64 make[1]: Entering directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt' hack/dockerized "./hack/check.sh && KUBEVIRT_VERSION= ./hack/build-go.sh install " sha256:0e817e41f9750e44335dde1be5cb34809abe48c8add43baf165907418e2e75ce go version go1.10 linux/amd64 go version go1.10 linux/amd64 Compiling tests... compiled tests.test hack/build-docker.sh build sending incremental file list ./ Dockerfile kubernetes.repo sent 854 bytes received 53 bytes 1814.00 bytes/sec total size is 1167 speedup is 1.29 Sending build context to Docker daemon 36.12 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> dde0df1b6fe4 Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-controller ---> Using cache ---> 65d6d48cdb35 Step 4/8 : WORKDIR /home/virt-controller ---> Using cache ---> e1ade8663337 Step 5/8 : USER 1001 ---> Using cache ---> 2ce44d6f372a Step 6/8 : COPY virt-controller /virt-controller ---> f950a8a6f358 Removing intermediate container 69a7ba534322 Step 7/8 : ENTRYPOINT /virt-controller ---> Running in bc17a58be2ac ---> 338ff1c1b6cc Removing intermediate container bc17a58be2ac Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-controller" '' ---> Running in ae24c67fd4ad ---> aa58f8cacd99 Removing intermediate container ae24c67fd4ad Successfully built aa58f8cacd99 sending incremental file list ./ Dockerfile entrypoint.sh kubevirt-sudo libvirtd.sh sh.sh sock-connector sent 3502 bytes received 129 bytes 7262.00 bytes/sec total size is 5953 speedup is 1.64 Sending build context to Docker daemon 38.06 MB Step 1/14 : FROM kubevirt/libvirt:3.7.0 ---> 60c80c8f7523 Step 2/14 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> d4ddb23dff45 Step 3/14 : RUN dnf -y install socat genisoimage util-linux libcgroup-tools ethtool sudo && dnf -y clean all && test $(id -u qemu) = 107 # make sure that the qemu user really is 107 ---> Using cache ---> 142a2ba860cf Step 4/14 : COPY sock-connector /sock-connector ---> Using cache ---> 02569da61faa Step 5/14 : COPY sh.sh /sh.sh ---> Using cache ---> 47d4a51575e2 Step 6/14 : COPY virt-launcher /virt-launcher ---> a90ff04224a0 Removing intermediate container 8b1abe24ba83 Step 7/14 : COPY kubevirt-sudo /etc/sudoers.d/kubevirt ---> 31868996ea32 Removing intermediate container a42fca473968 Step 8/14 : RUN chmod 0640 /etc/sudoers.d/kubevirt ---> Running in a11776a16cf5  ---> 7f41cf1c8751 Removing intermediate container a11776a16cf5 Step 9/14 : RUN rm -f /libvirtd.sh ---> Running in 1108a464986e  ---> 488de06b2213 Removing intermediate container 1108a464986e Step 10/14 : COPY libvirtd.sh /libvirtd.sh ---> 0a2844054648 Removing intermediate container aeb24e71cf3d Step 11/14 : RUN chmod a+x /libvirtd.sh ---> Running in fa7c2265e064  ---> 20c5c2b13c8b Removing intermediate container fa7c2265e064 Step 12/14 : COPY entrypoint.sh /entrypoint.sh ---> 2a80822d5a00 Removing intermediate container 13b674baed2e Step 13/14 : ENTRYPOINT /entrypoint.sh ---> Running in 93707e0ad118 ---> 3c3386edac95 Removing intermediate container 93707e0ad118 Step 14/14 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-launcher" '' ---> Running in 814e3cc68f36 ---> 4e6c8b31893b Removing intermediate container 814e3cc68f36 Successfully built 4e6c8b31893b sending incremental file list ./ Dockerfile sent 585 bytes received 34 bytes 1238.00 bytes/sec total size is 775 speedup is 1.25 Sending build context to Docker daemon 36.68 MB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> dde0df1b6fe4 Step 3/5 : COPY virt-handler /virt-handler ---> df81cd1329bd Removing intermediate container 7f5cf4c5e06e Step 4/5 : ENTRYPOINT /virt-handler ---> Running in cbaf604586e8 ---> cab66ad84c5d Removing intermediate container cbaf604586e8 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-handler" '' ---> Running in b9c75bfd1818 ---> 11c80bb0756f Removing intermediate container b9c75bfd1818 Successfully built 11c80bb0756f sending incremental file list ./ Dockerfile sent 646 bytes received 34 bytes 1360.00 bytes/sec total size is 876 speedup is 1.29 Sending build context to Docker daemon 36.82 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> dde0df1b6fe4 Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-api ---> Using cache ---> 2eeb55f39191 Step 4/8 : WORKDIR /home/virt-api ---> Using cache ---> 56cea32a45d4 Step 5/8 : USER 1001 ---> Using cache ---> d121920c238b Step 6/8 : COPY virt-api /virt-api ---> a8ce3afc50c6 Removing intermediate container 96aa51e0d527 Step 7/8 : ENTRYPOINT /virt-api ---> Running in e330646dd0ac ---> 84f9e2a426dc Removing intermediate container e330646dd0ac Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-api" '' ---> Running in 613e97c01a62 ---> 6f8358a627ff Removing intermediate container 613e97c01a62 Successfully built 6f8358a627ff sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/iscsi-demo-target-tgtd ./ Dockerfile run-tgt.sh sent 2185 bytes received 53 bytes 4476.00 bytes/sec total size is 3992 speedup is 1.78 Sending build context to Docker daemon 6.656 kB Step 1/10 : FROM fedora:27 ---> 9110ae7f579f Step 2/10 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> dde0df1b6fe4 Step 3/10 : ENV container docker ---> Using cache ---> 32cab959eac8 Step 4/10 : RUN dnf -y install scsi-target-utils bzip2 e2fsprogs ---> Using cache ---> c2339817cfe0 Step 5/10 : RUN mkdir -p /images ---> Using cache ---> a19645b68794 Step 6/10 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /images/1-alpine.img ---> Using cache ---> 3f0fa7f50785 Step 7/10 : ADD run-tgt.sh / ---> Using cache ---> 35ac6b299ab7 Step 8/10 : EXPOSE 3260 ---> Using cache ---> 259db1618b21 Step 9/10 : CMD /run-tgt.sh ---> Using cache ---> 4c9f18dec05a Step 10/10 : LABEL "iscsi-demo-target-tgtd" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Running in f46f13175ae3 ---> 25edb896f321 Removing intermediate container f46f13175ae3 Successfully built 25edb896f321 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/vm-killer ./ Dockerfile sent 610 bytes received 34 bytes 1288.00 bytes/sec total size is 797 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> dde0df1b6fe4 Step 3/5 : ENV container docker ---> Using cache ---> 32cab959eac8 Step 4/5 : RUN dnf -y install procps-ng nmap-ncat && dnf -y clean all ---> Using cache ---> 391fa00b27f9 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "vm-killer" '' ---> Running in 7660d9e17bb8 ---> f9805451b568 Removing intermediate container 7660d9e17bb8 Successfully built f9805451b568 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/registry-disk-v1alpha ./ Dockerfile entry-point.sh sent 1566 bytes received 53 bytes 3238.00 bytes/sec total size is 2542 speedup is 1.57 Sending build context to Docker daemon 5.12 kB Step 1/7 : FROM debian:sid ---> bcec0ae8107e Step 2/7 : MAINTAINER "David Vossel" \ ---> Using cache ---> 6696837acee7 Step 3/7 : ENV container docker ---> Using cache ---> 2dd2b1a02be6 Step 4/7 : RUN apt-get update && apt-get install -y bash curl bzip2 qemu-utils && mkdir -p /disk && rm -rf /var/lib/apt/lists/* ---> Using cache ---> dd3c4950b5c8 Step 5/7 : ADD entry-point.sh / ---> Using cache ---> d221e0eb5770 Step 6/7 : CMD /entry-point.sh ---> Using cache ---> 6506e61a9f41 Step 7/7 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "registry-disk-v1alpha" '' ---> Running in 2c283c29ed95 ---> 94e5ff4127fa Removing intermediate container 2c283c29ed95 Successfully built 94e5ff4127fa sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/cirros-registry-disk-demo ./ Dockerfile sent 630 bytes received 34 bytes 1328.00 bytes/sec total size is 825 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33112/kubevirt/registry-disk-v1alpha:devel ---> 94e5ff4127fa Step 2/4 : MAINTAINER "David Vossel" \ ---> Running in 45749400757b ---> c67637b04a65 Removing intermediate container 45749400757b Step 3/4 : RUN curl https://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img > /disk/cirros.img ---> Running in 8000fa284561  % Total % Received % Xferd Average Speed Time  Time Time Current Dload Upload Total Spent Left Speed 0 0 0  0 0 0  0 0 --:--:-- --:--:-- --:--:--  0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0  1 12.1M 1 176k 0 0 118k 0  0:01:44 0:00:01 0:01:43 118k 12 12.1M 12 1568k 0 0 637k 0 0:00:19 0:00:02 0:00:17 637k 59 12.1M 59 7376k 0 0 2133k 0 0:00:05 0:00:03 0:00:02 2132k 100 12.1M 100 12.1M 0 0 3129k 0 0:00:03 0:00:03 --:--:-- 3128k  ---> 050c4b5794c8 Removing intermediate container 8000fa284561 Step 4/4 : LABEL "cirros-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Running in 19e96555e468 ---> 3dce8fabbc93 Removing intermediate container 19e96555e468 Successfully built 3dce8fabbc93 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/fedora-cloud-registry-disk-demo ./ Dockerfile sent 677 bytes received 34 bytes 1422.00 bytes/sec total size is 926 speedup is 1.30 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33112/kubevirt/registry-disk-v1alpha:devel ---> 94e5ff4127fa Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Running in c9aac3cfbce5 ---> 185ce7200bb5 Removing intermediate container c9aac3cfbce5 Step 3/4 : RUN curl -g -L https://download.fedoraproject.org/pub/fedora/linux/releases/27/CloudImages/x86_64/images/Fedora-Cloud-Base-27-1.6.x86_64.qcow2 > /disk/fedora.qcow2 ---> Running in 51c05a6dee29  % Total % Received % Xferd Average Speed Time Time Time Current  Dload Upload Total Spent Left Speed 0 0 0 0 0 0  0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0  0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0 1 221M 1 3105k 0 0 1322k 0 0:02:51 0:00:02 0:02:49 3238k 4 221M 4 9081k 0 0 2760k 0 0:01:22 0:00:03 0:01:19 4777k 5 221M 5 12.7M 0 0 3034k 0 0:01:14 0:00:04 0:01:10 4486k 7 221M 7 17.0M 0 0 3289k 0 0:01:08 0:00:05 0:01:03 4454k 9 221M 9 21.3M 0 0 3469k 0 0:01:05 0:00:06 0:00:59 4453k 11 221M 11 24.6M 0 0 3457k 0 0:01:05 0:00:07 0:00:58 4472k 12 221M 12 27.1M 0 0 3351k 0 0:01:07 0:00:08 0:00:59 3739k 13 221M 13 29.4M 0 0 3248k 0 0:01:09 0:00:09 0:01:00 3432k 14 221M 14 31.5M 0 0 3140k 0 0:01:12 0:00:10 0:01:02 2982k 15 221M 15 33.8M 0 0 3065k 0 0:01:14 0:00:11 0:01:03 2558k 16 221M 16 35.6M 0 0 2973k 0 0:01:16 0:00:12 0:01:04 2267k 16 221M 16 37.1M 0 0 2861k 0 0:01:19 0:00:13 0:01:06 2044k 17 221M 17 38.2M 0 0 2741k 0 0:01:22 0:00:14 0:01:08 1801k 17 221M 17 39.0M 0 0 2614k 0 0:01:26 0:00:15 0:01:11 1528k 17 221M 17 39.4M 0 0 2475k 0 0:01:31 0:00:16 0:01:15 1138k 17 221M 17 39.8M 0 0 2356k 0 0:01:36 0:00:17 0:01:19 844k 18 221M 18 40.3M 0 0 2255k 0 0:01:40 0:00:18 0:01:22 649k 18 221M 18 40.9M 0 0 2170k 0 0:01:44 0:00:19 0:01:25 534k 18 221M 18 41.5M 0 0 2096k 0 0:01:48 0:00:20 0:01:28 516k 19 221M 19 42.2M 0 0 2029k 0 0:01:51 0:00:21 0:01:30 578k 19 221M 19 42.8M 0 0 1968k 0 0:01:55 0:00:22 0:01:33 621k 19 221M 19 43.5M 0 0 1915k 0 0:01:58 0:00:23 0:01:35 668k 20 221M 20 44.3M 0 0 1867k 0 0:02:01 0:00:24 0:01:37 701k 20 221M 20 45.1M 0 0 1826k 0 0:02:04 0:00:25 0:01:39 730k 20 221M 20 46.0M 0 0 1791k 0 0:02:06 0:00:26 0:01:40 775k 21 221M 21 46.9M 0 0 1760k 0 0:02:08 0:00:27 0:01:41 834k 21 221M 21 47.9M 0 0 1737k 0 0:02:10 0:00:28 0:01:42 903k 22 221M 22 49.1M 0 0 1717k 0 0:02:12 0:00:29 0:01:43 983k 22 221M 22 50.6M 0 0 1712k 0 0:02:12 0:00:30 0:01:42 1132k 23 221M 23 52.4M 0 0 1715k 0 0:02:12 0:00:31 0:01:41 1315k 24 221M 24 54.0M 0 0 1715k 0 0:02:12 0:00:32 0:01:40 1468k 25 221M 25 55.6M 0 0 1712k 0 0:02:12 0:00:33 0:01:39 1572k 25 221M 25 57.1M 0 0 1707k 0 0:02:12 0:00:34 0:01:38 1652k 26 221M 26 58.7M 0 0 1705k 0 0:02:13 0:00:35 0:01:38 1665k 27 221M 27 60.4M 0 0 1703k 0 0:02:13 0:00:36 0:01:37 1629k 27 221M 27 62.0M 0 0 1702k 0 0:02:13 0:00:37 0:01:36 1619k 28 221M 28 63.6M 0 0 1701k 0 0:02:13 0:00:38 0:01:35 1628k 29 221M 29 64.9M 0 0 1692k 0 0:02:14 0:00:39 0:01:35 1586k 29 221M 29 66.1M 0 0 1680k 0 0:02:15 0:00:40 0:01:35 1504k 30 221M 30 67.1M 0 0 1665k 0 0:02:16 0:00:41 0:01:35 1390k 30 221M 30 67.9M 0 0 1645k 0 0:02:17 0:00:42 0:01:35 1217k 31 221M 31 68.7M 0 0 1626k 0 0:02:19 0:00:43 0:01:36 1053k 31 221M 31 69.6M 0 0 1608k 0 0:02:21 0:00:44 0:01:37 954k 31 221M 31 70.5M 0 0 1594k 0 0:02:22 0:00:45 0:01:37 904k 32 221M 32 71.5M 0 0 1582k 0 0:02:23 0:00:46 0:01:37 897k 32 221M 32 72.6M 0 0 1572k 0 0:02:24 0:00:47 0:01:37 960k 33 221M 33 73.8M 0 0 1564k 0 0:02:24 0:00:48 0:01:36 1031k 33 221M 33 75.0M 0 0 1559k 0 0:02:25 0:00:49 0:01:36 1122k 34 221M  34 76.6M 0 0 1559k 0 0:02:25 0:00:50 0:01:35 1239k 35 221M 35 78.0M 0 0 1556k 0 0:02:25 0:00:51 0:01:34 1310k 35 221M 35 79.0M 0 0 1548k 0 0:02:26 0:00:52 0:01:34 1323k 36 221M 36 79.9M 0 0 1534k 0 0:02:27 0:00:53 0:01:34 1246k 36 221M 36 80.6M 0 0 1521k 0 0:02:29 0:00:54 0:01:35 1139k 36 221M 36 81.4M 0 0 1508k 0 0:02:30 0:00:55 0:01:35 991k 37 221M 37 82.1M 0 0 1493k 0 0:02:31 0:00:56 0:01:35 843k 37 221M 37 82.9M 0 0 1481k 0 0:02:33 0:00:57 0:01:36 780k 37 221M 37 83.7M 0 0 1470k 0 0:02:34 0:00:58 0:01:36 784k 38 221M 38 84.5M 0 0 1460k 0 0:02:35 0:00:59 0:01:36 802k 38 221M 38 85.4M 0 0 1451k 0 0:02:36 0:01:00 0:01:36 829k 39 221M 39 86.5M 0 0 1444k 0 0:02:37 0:01:01 0:01:36 901k 39 221M 39 87.3M 0 0 1436k 0 0:02:37 0:01:02 0:01:35 913k 39 221M 39 88.3M 0 0 1428k 0 0:02:38 0:01:03 0:01:35 940k 40 221M 40 89.3M 0 0 1422k 0 0:02:39 0:01:04 0:01:35 980k 40 221M 40 90.4M 0 0 1417k 0 0:02:40 0:01:05 0:01:35 1007k 41 221M 41 91.4M 0 0 1412k 0 0:02:40 0:01:06 0:01:34 1017k 41 221M 41 92.6M 0 0 1410k 0 0:02:40 0:01:07 0:01:33 1084k 42 221M 42 93.3M 0 0 1399k 0 0:02:42 0:01:08 0:01:34 1032k 42 221M 42 94.0M 0 0 1390k 0 0:02:43 0:01:09 0:01:34 963k 42 221M 42 94.8M 0 0 1381k 0 0:02:44 0:01:10 0:01:34 903k 43 221M 43 95.5M 0 0 1372k 0 0:02:45 0:01:11 0:01:34 839k 43 221M 43 96.3M 0 0 1363k 0 0:02:46 0:01:12 0:01:34 745k 43 221M 43 96.9M 0 0 1354k 0 0:02:47 0:01:13 0:01:34 736k 44 221M 44 97.6M 0 0 1345k 0 0:02:48 0:01:14 0:01:34 728k 44 221M 44 98.3M 0 0 1337k 0 0:02:49 0:01:15  0:01:34 720k 44 221M 44 99.1M 0 0 1330k 0 0:02:50 0:01:16 0:01:34 734k 45 221M 45 99.8M 0 0 1323k 0 0:02:51 0:01:17 0:01:34 732k 45 221M 45 100M 0 0 1316k 0 0:02:52 0:01:18 0:01:34 754k 45 221M 45 101M 0 0 1310k 0 0:02:53 0:01:19 0:01:34 791k 46 221M 46 102M 0 0 1304k 0 0:02:53 0:01:20 0:01:33 811k 46 221M 46 102M 0 0 1295k 0 0:02:55 0:01:21 0:01:34 749k 46 221M 46 103M 0 0 1286k 0 0:02:56 0:01:22 0:01:34 719k 46 221M 46 104M 0 0 1279k 0 0:02:57 0:01:23 0:01:34 699k 47 221M 47 104M 0 0 1272k 0 0:02:58 0:01:24 0:01:34 671k 47 221M 47 105M 0 0 1267k 0 0:02:59 0:01:25 0:01:34 671k 48 221M 48 106M 0 0 1263k 0 0:02:59 0:01:26 0:01:33 752k 48 221M 48 107M 0 0 1260k 0 0:03:00 0:01:27 0:01:33 829k 49 221M 49 108M 0 0 1259k 0 0:03:00 0:01:28 0:01:32 932k 49 221M 49 110M 0 0 1263k 0 0:02:59 0:01:29 0:01:30 1110k 50 221M 50 111M 0 0 1267k 0 0:02:59 0:01:30 0:01:29 1261k 51 221M 51 113M 0 0 1272k 0 0:02:58 0:01:31 0:01:27 1421k 52 221M 52 115M 0 0 1279k 0 0:02:57 0:01:32 0:01:25 1612k 52 221M 52 117M 0 0 1286k 0 0:02:56 0:01:33 0:01:23 1769k 53 221M 53 119M 0 0 1294k 0 0:02:55 0:01:34 0:01:21 1842k 54 221M 54 121M 0 0 1301k 0 0:02:54 0:01:35 0:01:19 1916k 55 221M 55 123M 0 0 1308k 0 0:02:53 0:01:36 0:01:17 1974k 56 221M 56 124M 0 0 1315k 0 0:02:52 0:01:37 0:01:15 1977k 57 221M 57 126M 0 0 1319k 0 0:02:52 0:01:38 0:01:14 1922k 58 221M 58 128M 0 0 1325k 0 0:02:51 0:01:39 0:01:12 1910k 58 221M 58 130M 0 0 1332k 0 0:02:50 0:01:40 0:01:10 1932k 59 221M 59 132M 0 0 1340k 0 0:02:49 0:01:41 0:01:08 1946k 60 221M 60 134M 0 0 1346k 0 0:02:48 0:01:42 0:01:06 1956k 61 221M 61 136M 0 0 1350k 0 0:02:48 0:01:43 0:01:05 1968k 62 221M 62 138M 0 0 1355k 0 0:02:47 0:01:44 0:01:03 1953k 63 221M 63 139M 0 0 1357k 0 0:02:47 0:01:45 0:01:02 1864k 63 221M 63 141M 0 0 1360k 0 0:02:46 0:01:46 0:01:00 1776k 64 221M 64 142M 0 0 1363k 0 0:02:46 0:01:47 0:00:59 1712k 65 221M 65 144M 0 0 1366k 0 0:02:46 0:01:48 0:00:58 1700k 66 221M 66 146M 0 0 1370k 0 0:02:45 0:01:49 0:00:56 1673k 66 221M 66 148M 0 0 1374k 0 0:02:45 0:01:50 0:00:55 1717k 67 221M 67 149M 0 0 1378k 0 0:02:44 0:01:51 0:00:53 1769k 68 221M 68 151M 0 0 1382k 0 0:02:44 0:01:52 0:00:52 1793k 68 221M 68 152M 0 0 1381k 0 0:02:44 0:01:53 0:00:51 1707k 69 221M 69 153M 0 0 1377k 0 0:02:44 0:01:54 0:00:50 1539k 69 221M 69 154M 0 0 1373k 0 0:02:45 0:01:55 0:00:50 1360k 70 221M 70 155M 0 0 1370k 0 0:02:45 0:01:56 0:00:49 1190k 70 221M 70 156M 0 0 1368k 0 0:02:45 0:01:57 0:00:48 1043k 71 221M 71 157M 0 0 1366k 0 0:02:45 0:01:58 0:00:47 1030k 71 221M 71 159M 0 0 1365k 0 0:02:46 0:01:59 0:00:47 1102k 72 221M 72 160M 0 0 1365k 0 0:02:46 0:02:00 0:00:46 1177k 73 221M 73 161M 0 0 1367k 0 0:02:45 0:02:01 0:00:44 1282k 73 221M 73 163M 0 0 1372k 0 0:02:45 0:02:02 0:00:43 1469k 75 221M 75 166M 0 0 1382k 0 0:02:44 0:02:03 0:00:41 1762k 76 221M 76 169M 0 0 1398k 0 0:02:42 0:02:04 0:00:38 2164k 77 221M 77 172M 0 0 1407k 0 0:02:41  0:02:05 0:00:36 2423k 78 221M 78 174M 0 0 1414k 0 0:02:40 0:02:06 0:00:34 2572k 79 221M 79 176M 0 0 1422k 0 0:02:39 0:02:07 0:00:32 2649k 80 221M 80 178M 0 0 1428k 0 0:02:38 0:02:08 0:00:30 2548k 81 221M 81 180M 0 0 1432k 0 0:02:38 0:02:09 0:00:29 2290k 82 221M 82 182M 0 0 1437k 0 0:02:37 0:02:10 0:00:27 2185k 83 221M 83 185M 0 0 1443k 0 0:02:37 0:02:11 0:00:26 2161k 84 221M 84 186M 0 0 1447k 0 0:02:36 0:02:12 0:00:24 2067k 85 221M 85 188M 0 0 1449k 0 0:02:36 0:02:13 0:00:23 1989k 86 221M 86 190M 0 0 1453k 0 0:02:36 0:02:14 0:00:22 1986k 86 221M 86 192M 0 0 1457k 0 0:02:35 0:02:15 0:00:20 1982k 87 221M 87 194M 0 0 1459k 0 0:02:35 0:02:16 0:00:19 1882k 88 221M 88 195M 0 0 1461k 0 0:02:35 0:02:17 0:00:18 1840k 89 221M 89 197M 0 0 1463k 0 0:02:35 0:02:18 0:00:17 1849k 90 221M 90 199M 0 0 1466k 0 0:02:34 0:02:19 0:00:15 1816k 90 221M 90 201M 0 0 1468k 0 0:02:34 0:02:20 0:00:14 1768k 91 221M 91 203M 0 0 1471k  0 0:02:34 0:02:21 0:00:13 1816k 92 221M 92 205M 0 0 1475k 0 0:02:33 0:02:22 0:00:11 1867k 93 221M 93 207M 0 0 1480k 0 0:02:33 0:02:23 0:00:10 1931k 94 221M 94 208M 0 0 1481k 0 0:02:33 0:02:24 0:00:09 1920k 95 221M 95 210M 0 0 1485k 0 0:02:32 0:02:25 0:00:07 1942k 96 221M 96 212M 0 0 1489k 0 0:02:32 0:02:26 0:00:06 1988k 97 221M 97 214M 0 0 1494k 0 0:02:31 0:02:27 0:00:04 2018k 97 221M 97 217M 0 0 1499k 0 0:02:31 0:02:28 0:00:03 2046k 98 221M 98 219M 0 0 1503k 0  0:02:30 0:02:29 0:00:01 2137k 99 221M 99 221M 0 0 1508k 0 0:02:30 0:02:30 --:--:-- 2189k 100 221M 100 221M 0 0 1508k 0 0:02:30 0:02:30 --:--:-- 2188k  ---> d21d80685fbe Removing intermediate container 51c05a6dee29 Step 4/4 : LABEL "fedora-cloud-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Running in 272fdc881f21 ---> 74263b4a27b9 Removing intermediate container 272fdc881f21 Successfully built 74263b4a27b9 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/alpine-registry-disk-demo ./ Dockerfile sent 639 bytes received 34 bytes 1346.00 bytes/sec total size is 866 speedup is 1.29 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:33112/kubevirt/registry-disk-v1alpha:devel ---> 94e5ff4127fa Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 185ce7200bb5 Step 3/4 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /disk/alpine.iso ---> Running in 609f8e445779  % Total % Received % Xferd Average Speed Time Time  Time Current Dload Upload Total  Spent Left Speed  0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 37.0M 0 100k 0 0 92211 0 0:07:00 0:00:01 0:06:59 92128 8 37.0M 8 3290k 0 0 1555k 0 0:00:24 0:00:02 0:00:22 1555k 23 37.0M 23 9050k 0 0 2904k 0 0:00:13 0:00:03 0:00:10 2904k 44 37.0M 44 16.5M 0 0 4111k 0 0:00:09 0:00:04 0:00:05 4110k 61 37.0M 61 22.8M 0 0 4565k 0 0:00:08 0:00:05 0:00:03 4679k 75 37.0M 75 28.0M 0 0 4701k 0 0:00:08 0:00:06 0:00:02 5730k 91 37.0M 91 33.9M 0 0 4885k 0 0:00:07 0:00:07 --:--:-- 6293k 100 37.0M 100 37.0M 0 0 4891k 0 0:00:07 0:00:07 --:--:-- 6228k  ---> b353df63b41b Removing intermediate container 609f8e445779 Step 4/4 : LABEL "alpine-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Running in 85dbc1b5e6d5 ---> 659b7f5e5a20 Removing intermediate container 85dbc1b5e6d5 Successfully built 659b7f5e5a20 sending incremental file list ./ Dockerfile sent 660 bytes received 34 bytes 1388.00 bytes/sec total size is 918 speedup is 1.32 Sending build context to Docker daemon 33.96 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> dde0df1b6fe4 Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virtctl ---> Using cache ---> 6e6e1b7931e0 Step 4/8 : WORKDIR /home/virtctl ---> Using cache ---> 9d27e69a25f2 Step 5/8 : USER 1001 ---> Using cache ---> 1760a8e197af Step 6/8 : COPY subresource-access-test /subresource-access-test ---> 22506939fe6e Removing intermediate container de02cc273424 Step 7/8 : ENTRYPOINT /subresource-access-test ---> Running in 71a025731c2c ---> 538846f76dfd Removing intermediate container 71a025731c2c Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "subresource-access-test" '' ---> Running in 2e252ef5e04d ---> 1fa056a92b84 Removing intermediate container 2e252ef5e04d Successfully built 1fa056a92b84 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/winrmcli ./ Dockerfile sent 773 bytes received 34 bytes 1614.00 bytes/sec total size is 1098 speedup is 1.36 Sending build context to Docker daemon 3.072 kB Step 1/9 : FROM fedora:27 ---> 9110ae7f579f Step 2/9 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> dde0df1b6fe4 Step 3/9 : ENV container docker ---> Using cache ---> 32cab959eac8 Step 4/9 : RUN dnf -y install make git gcc && dnf -y clean all ---> Using cache ---> 8e034c77f534 Step 5/9 : ENV GIMME_GO_VERSION 1.9.2 ---> Using cache ---> 28ec1d482013 Step 6/9 : RUN mkdir -p /gimme && curl -sL https://raw.githubusercontent.com/travis-ci/gimme/master/gimme | HOME=/gimme bash >> /etc/profile.d/gimme.sh ---> Using cache ---> db78d0286f58 Step 7/9 : ENV GOPATH "/go" GOBIN "/usr/bin" ---> Using cache ---> 7ebe54e98be4 Step 8/9 : RUN mkdir -p /go && source /etc/profile.d/gimme.sh && go get github.com/masterzen/winrm-cli ---> Using cache ---> a3b04c1816f5 Step 9/9 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "winrmcli" '' ---> Running in f04bda0fce7e ---> 3be3a9ab92d8 Removing intermediate container f04bda0fce7e Successfully built 3be3a9ab92d8 hack/build-docker.sh push The push refers to a repository [localhost:33112/kubevirt/virt-controller] 783c3ccb3e6d: Preparing 52069b1f5033: Preparing 39bae602f753: Preparing 52069b1f5033: Pushed 783c3ccb3e6d: Pushed 39bae602f753: Pushed devel: digest: sha256:8015425fda936355ab838defcd9185dd838561527a102381dcaea7e8931df87d size: 948 The push refers to a repository [localhost:33112/kubevirt/virt-launcher] 580fa9702a63: Preparing 11c6c780a6ce: Preparing 11c6c780a6ce: Preparing 9e3dd85892df: Preparing fcb265d5cfce: Preparing bb05fc965bd6: Preparing 88361ab452b9: Preparing 4ebc38848be0: Preparing b9fd8c21001d: Preparing 4d2f0529ab56: Preparing 530cc55618cd: Preparing 34fa414dfdf6: Preparing b9fd8c21001d: Waiting a1359dc556dd: Preparing 490c7c373332: Preparing 88361ab452b9: Waiting a1359dc556dd: Waiting 34fa414dfdf6: Waiting 4b440db36f72: Preparing 39bae602f753: Preparing 490c7c373332: Waiting 4b440db36f72: Waiting 39bae602f753: Waiting 4ebc38848be0: Waiting 4d2f0529ab56: Waiting 9e3dd85892df: Pushed bb05fc965bd6: Pushed fcb265d5cfce: Pushed 11c6c780a6ce: Pushed 580fa9702a63: Pushed 4ebc38848be0: Pushed b9fd8c21001d: Pushed 530cc55618cd: Pushed 34fa414dfdf6: Pushed a1359dc556dd: Pushed 490c7c373332: Pushed 39bae602f753: Mounted from kubevirt/virt-controller 4d2f0529ab56: Pushed 88361ab452b9: Pushed 4b440db36f72: Pushed devel: digest: sha256:3339590157c93c1657bc052134348790f7e7a0b1bb3c89f62500225492cebe19 size: 3653 The push refers to a repository [localhost:33112/kubevirt/virt-handler] b9809e5a6570: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/virt-launcher b9809e5a6570: Pushed devel: digest: sha256:d6ebb821891039d7025630d6b6528120c4a4328444f6ae20b38a3f076bacc8f0 size: 740 The push refers to a repository [localhost:33112/kubevirt/virt-api] 334108967326: Preparing 86b4b25303b4: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/virt-handler 86b4b25303b4: Pushed 334108967326: Pushed devel: digest: sha256:a9d3bc4b2d4ab7cd8b411be9b8ef45e64ea22b7cfa9f1e26abdfdab625a60b03 size: 948 The push refers to a repository [localhost:33112/kubevirt/iscsi-demo-target-tgtd] 80220be9fed7: Preparing 89fef61f2c06: Preparing b18a27986676: Preparing db8a56c06e31: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/virt-api 80220be9fed7: Pushed b18a27986676: Pushed 89fef61f2c06: Pushed db8a56c06e31: Pushed devel: digest: sha256:ae02117718bdc3f3460c3a5c13da86a1d77d5d5c38501e5f6423525b79a30fb6 size: 1368 The push refers to a repository [localhost:33112/kubevirt/vm-killer] 040d3361950b: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/iscsi-demo-target-tgtd 040d3361950b: Pushed devel: digest: sha256:eebc85bb8efeffb2b2f33bd1b8ba6dc38456a5c7c1d2ba16133fa86e440d37e3 size: 740 The push refers to a repository [localhost:33112/kubevirt/registry-disk-v1alpha] 4cd98e29acca: Preparing 9beeb9a18439: Preparing 6709b2da72b8: Preparing 4cd98e29acca: Pushed 9beeb9a18439: Pushed 6709b2da72b8: Pushed devel: digest: sha256:7c60431a98406484670378cd51dd661dcd1bd67fd085176b40584403790568a2 size: 948 The push refers to a repository [localhost:33112/kubevirt/cirros-registry-disk-demo] 2e8fb3e202b8: Preparing 4cd98e29acca: Preparing 9beeb9a18439: Preparing 6709b2da72b8: Preparing 9beeb9a18439: Mounted from kubevirt/registry-disk-v1alpha 6709b2da72b8: Mounted from kubevirt/registry-disk-v1alpha 4cd98e29acca: Mounted from kubevirt/registry-disk-v1alpha 2e8fb3e202b8: Pushed devel: digest: sha256:c23e16b601f1b7d50207f40bbbe3963f3acaa05232da9cab3d128a0c0749ff6f size: 1160 The push refers to a repository [localhost:33112/kubevirt/fedora-cloud-registry-disk-demo] 8bb1dcaccac2: Preparing 4cd98e29acca: Preparing 9beeb9a18439: Preparing 6709b2da72b8: Preparing 4cd98e29acca: Mounted from kubevirt/cirros-registry-disk-demo 6709b2da72b8: Mounted from kubevirt/cirros-registry-disk-demo 9beeb9a18439: Mounted from kubevirt/cirros-registry-disk-demo 8bb1dcaccac2: Pushed devel: digest: sha256:49672e002009c9d1e370d4d50b06c404d9e727a04b97e7281ab7053087003c45 size: 1161 The push refers to a repository [localhost:33112/kubevirt/alpine-registry-disk-demo] 1f005641bd6f: Preparing 4cd98e29acca: Preparing 9beeb9a18439: Preparing 6709b2da72b8: Preparing 9beeb9a18439: Mounted from kubevirt/fedora-cloud-registry-disk-demo 6709b2da72b8: Mounted from kubevirt/fedora-cloud-registry-disk-demo 4cd98e29acca: Mounted from kubevirt/fedora-cloud-registry-disk-demo 1f005641bd6f: Pushed devel: digest: sha256:e5abcc08b1f058cc9edbea4eea54f78c3c78b348c01a75297751aef9ac398aef size: 1160 The push refers to a repository [localhost:33112/kubevirt/subresource-access-test] 3fa2e8a7912c: Preparing 2c4f6b64d5e3: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/vm-killer 2c4f6b64d5e3: Pushed 3fa2e8a7912c: Pushed devel: digest: sha256:e198d102b829e06503e1627ff20e16acb1a152594a1909e9dcf0d13ce28e2414 size: 948 The push refers to a repository [localhost:33112/kubevirt/winrmcli] 161ef5381259: Preparing 2bef46eb5bf3: Preparing ac5611d25ed9: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/subresource-access-test 161ef5381259: Pushed ac5611d25ed9: Pushed 2bef46eb5bf3: Pushed devel: digest: sha256:0f74a127121b8ee4fe3bd985550fc248403e426be506d51582ff57304c4b68be size: 1165 make[1]: Leaving directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt' 2018/04/30 18:32:57 Waiting for host: 192.168.66.101:22 2018/04/30 18:32:57 Connected to tcp://192.168.66.101:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer b2fd69abfadb: Pulling fs layer 12230f5d39a2: Pulling fs layer b2fd69abfadb: Verifying Checksum b2fd69abfadb: Download complete 12230f5d39a2: Verifying Checksum 12230f5d39a2: Download complete 2176639d844b: Download complete 2176639d844b: Pull complete b2fd69abfadb: Pull complete 12230f5d39a2: Pull complete Digest: sha256:8015425fda936355ab838defcd9185dd838561527a102381dcaea7e8931df87d Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer a7576ba71334: Pulling fs layer 635c67b1c9ef: Pulling fs layer c60da62d83fb: Pulling fs layer cb886310435f: Pulling fs layer 787257bbca59: Pulling fs layer 0a1886268733: Pulling fs layer 0c13972304bc: Pulling fs layer bb54003884e8: Pulling fs layer ba9a1341cc0d: Pulling fs layer c60da62d83fb: Waiting cb886310435f: Waiting 787257bbca59: Waiting 0a1886268733: Waiting 0c13972304bc: Waiting bb54003884e8: Waiting ba9a1341cc0d: Waiting 6cc174edcebf: Waiting a7576ba71334: Waiting 635c67b1c9ef: Waiting a1e80189bea5: Waiting a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete f2ef945504a7: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete 635c67b1c9ef: Verifying Checksum 635c67b1c9ef: Download complete c60da62d83fb: Verifying Checksum c60da62d83fb: Download complete cb886310435f: Verifying Checksum cb886310435f: Download complete a7576ba71334: Verifying Checksum a7576ba71334: Download complete 787257bbca59: Verifying Checksum 787257bbca59: Download complete 0a1886268733: Verifying Checksum 0a1886268733: Download complete 0c13972304bc: Verifying Checksum 0c13972304bc: Download complete bb54003884e8: Verifying Checksum bb54003884e8: Download complete ba9a1341cc0d: Verifying Checksum ba9a1341cc0d: Download complete d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete a7576ba71334: Pull complete 635c67b1c9ef: Pull complete c60da62d83fb: Pull complete cb886310435f: Pull complete 787257bbca59: Pull complete 0a1886268733: Pull complete 0c13972304bc: Pull complete bb54003884e8: Pull complete ba9a1341cc0d: Pull complete Digest: sha256:3339590157c93c1657bc052134348790f7e7a0b1bb3c89f62500225492cebe19 Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists 5713a6a29f78: Pulling fs layer 5713a6a29f78: Verifying Checksum 5713a6a29f78: Download complete 5713a6a29f78: Pull complete Digest: sha256:d6ebb821891039d7025630d6b6528120c4a4328444f6ae20b38a3f076bacc8f0 Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists ecbe4adfb5a6: Pulling fs layer 0b7e2c0345ca: Pulling fs layer ecbe4adfb5a6: Download complete 0b7e2c0345ca: Download complete ecbe4adfb5a6: Pull complete 0b7e2c0345ca: Pull complete Digest: sha256:a9d3bc4b2d4ab7cd8b411be9b8ef45e64ea22b7cfa9f1e26abdfdab625a60b03 Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists e41ccbba2812: Pulling fs layer 1525a0b70164: Pulling fs layer f69087ebfcf1: Pulling fs layer 4180f6dc22d7: Pulling fs layer 4180f6dc22d7: Waiting 1525a0b70164: Verifying Checksum 1525a0b70164: Download complete 4180f6dc22d7: Verifying Checksum 4180f6dc22d7: Download complete f69087ebfcf1: Verifying Checksum f69087ebfcf1: Download complete e41ccbba2812: Download complete e41ccbba2812: Pull complete 1525a0b70164: Pull complete f69087ebfcf1: Pull complete 4180f6dc22d7: Pull complete Digest: sha256:ae02117718bdc3f3460c3a5c13da86a1d77d5d5c38501e5f6423525b79a30fb6 Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists 138296e7088e: Pulling fs layer 138296e7088e: Download complete 138296e7088e: Pull complete Digest: sha256:eebc85bb8efeffb2b2f33bd1b8ba6dc38456a5c7c1d2ba16133fa86e440d37e3 Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer 1d498b3a9c67: Pulling fs layer 542e5d603739: Pulling fs layer 542e5d603739: Download complete 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 1d498b3a9c67: Verifying Checksum 1d498b3a9c67: Download complete 2115d46e7396: Pull complete 1d498b3a9c67: Pull complete 542e5d603739: Pull complete Digest: sha256:7c60431a98406484670378cd51dd661dcd1bd67fd085176b40584403790568a2 Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists 1d498b3a9c67: Already exists 542e5d603739: Already exists af0d9695f534: Pulling fs layer af0d9695f534: Download complete af0d9695f534: Pull complete Digest: sha256:c23e16b601f1b7d50207f40bbbe3963f3acaa05232da9cab3d128a0c0749ff6f Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists 1d498b3a9c67: Already exists 542e5d603739: Already exists 78d1f78e4718: Pulling fs layer 78d1f78e4718: Download complete 78d1f78e4718: Pull complete Digest: sha256:49672e002009c9d1e370d4d50b06c404d9e727a04b97e7281ab7053087003c45 Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists 1d498b3a9c67: Already exists 542e5d603739: Already exists a08d92be0ca7: Pulling fs layer a08d92be0ca7: Verifying Checksum a08d92be0ca7: Download complete a08d92be0ca7: Pull complete Digest: sha256:e5abcc08b1f058cc9edbea4eea54f78c3c78b348c01a75297751aef9ac398aef Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists 3e89ff6a57d9: Pulling fs layer edbad3c6ed6f: Pulling fs layer 3e89ff6a57d9: Verifying Checksum 3e89ff6a57d9: Download complete edbad3c6ed6f: Verifying Checksum edbad3c6ed6f: Download complete 3e89ff6a57d9: Pull complete edbad3c6ed6f: Pull complete Digest: sha256:e198d102b829e06503e1627ff20e16acb1a152594a1909e9dcf0d13ce28e2414 Trying to pull repository registry:5000/kubevirt/winrmcli ... devel: Pulling from registry:5000/kubevirt/winrmcli 2176639d844b: Already exists 7c1ab5de42d5: Pulling fs layer 9391531b0959: Pulling fs layer d4e9df2eaabc: Pulling fs layer d4e9df2eaabc: Verifying Checksum d4e9df2eaabc: Download complete 7c1ab5de42d5: Verifying Checksum 7c1ab5de42d5: Download complete 9391531b0959: Verifying Checksum 9391531b0959: Download complete 7c1ab5de42d5: Pull complete 9391531b0959: Pull complete d4e9df2eaabc: Pull complete Digest: sha256:0f74a127121b8ee4fe3bd985550fc248403e426be506d51582ff57304c4b68be 2018/04/30 18:36:45 Waiting for host: 192.168.66.101:22 2018/04/30 18:36:45 Connected to tcp://192.168.66.101:22 2018/04/30 18:36:48 Waiting for host: 192.168.66.102:22 2018/04/30 18:36:48 Connected to tcp://192.168.66.102:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer b2fd69abfadb: Pulling fs layer 12230f5d39a2: Pulling fs layer b2fd69abfadb: Verifying Checksum b2fd69abfadb: Download complete 12230f5d39a2: Verifying Checksum 12230f5d39a2: Download complete 2176639d844b: Verifying Checksum 2176639d844b: Download complete 2176639d844b: Pull complete b2fd69abfadb: Pull complete 12230f5d39a2: Pull complete Digest: sha256:8015425fda936355ab838defcd9185dd838561527a102381dcaea7e8931df87d Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer a7576ba71334: Pulling fs layer 635c67b1c9ef: Pulling fs layer c60da62d83fb: Pulling fs layer cb886310435f: Pulling fs layer 787257bbca59: Pulling fs layer 0a1886268733: Pulling fs layer 0c13972304bc: Pulling fs layer bb54003884e8: Pulling fs layer ba9a1341cc0d: Pulling fs layer a1e80189bea5: Waiting 6cc174edcebf: Waiting a7576ba71334: Waiting 635c67b1c9ef: Waiting c60da62d83fb: Waiting cb886310435f: Waiting bb54003884e8: Waiting 787257bbca59: Waiting 0a1886268733: Waiting 0c13972304bc: Waiting ba9a1341cc0d: Waiting a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete f2ef945504a7: Verifying Checksum f2ef945504a7: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete 635c67b1c9ef: Verifying Checksum 635c67b1c9ef: Download complete c60da62d83fb: Verifying Checksum c60da62d83fb: Download complete cb886310435f: Verifying Checksum cb886310435f: Download complete a7576ba71334: Verifying Checksum a7576ba71334: Download complete 787257bbca59: Verifying Checksum 787257bbca59: Download complete 0a1886268733: Verifying Checksum 0a1886268733: Download complete bb54003884e8: Verifying Checksum bb54003884e8: Download complete 0c13972304bc: Verifying Checksum 0c13972304bc: Download complete ba9a1341cc0d: Verifying Checksum ba9a1341cc0d: Download complete d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete a7576ba71334: Pull complete 635c67b1c9ef: Pull complete c60da62d83fb: Pull complete cb886310435f: Pull complete 787257bbca59: Pull complete 0a1886268733: Pull complete 0c13972304bc: Pull complete bb54003884e8: Pull complete ba9a1341cc0d: Pull complete Digest: sha256:3339590157c93c1657bc052134348790f7e7a0b1bb3c89f62500225492cebe19 Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists 5713a6a29f78: Pulling fs layer 5713a6a29f78: Verifying Checksum 5713a6a29f78: Download complete 5713a6a29f78: Pull complete Digest: sha256:d6ebb821891039d7025630d6b6528120c4a4328444f6ae20b38a3f076bacc8f0 Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists ecbe4adfb5a6: Pulling fs layer 0b7e2c0345ca: Pulling fs layer ecbe4adfb5a6: Verifying Checksum ecbe4adfb5a6: Download complete 0b7e2c0345ca: Verifying Checksum 0b7e2c0345ca: Download complete ecbe4adfb5a6: Pull complete 0b7e2c0345ca: Pull complete Digest: sha256:a9d3bc4b2d4ab7cd8b411be9b8ef45e64ea22b7cfa9f1e26abdfdab625a60b03 Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists e41ccbba2812: Pulling fs layer 1525a0b70164: Pulling fs layer f69087ebfcf1: Pulling fs layer 4180f6dc22d7: Pulling fs layer 4180f6dc22d7: Waiting 1525a0b70164: Verifying Checksum 1525a0b70164: Download complete 4180f6dc22d7: Verifying Checksum 4180f6dc22d7: Download complete f69087ebfcf1: Verifying Checksum f69087ebfcf1: Download complete e41ccbba2812: Verifying Checksum e41ccbba2812: Download complete e41ccbba2812: Pull complete 1525a0b70164: Pull complete f69087ebfcf1: Pull complete 4180f6dc22d7: Pull complete Digest: sha256:ae02117718bdc3f3460c3a5c13da86a1d77d5d5c38501e5f6423525b79a30fb6 Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists 138296e7088e: Pulling fs layer 138296e7088e: Verifying Checksum 138296e7088e: Download complete 138296e7088e: Pull complete Digest: sha256:eebc85bb8efeffb2b2f33bd1b8ba6dc38456a5c7c1d2ba16133fa86e440d37e3 Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer 1d498b3a9c67: Pulling fs layer 542e5d603739: Pulling fs layer 542e5d603739: Verifying Checksum 542e5d603739: Download complete 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 1d498b3a9c67: Verifying Checksum 1d498b3a9c67: Download complete 2115d46e7396: Pull complete 1d498b3a9c67: Pull complete 542e5d603739: Pull complete Digest: sha256:7c60431a98406484670378cd51dd661dcd1bd67fd085176b40584403790568a2 Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists 1d498b3a9c67: Already exists 542e5d603739: Already exists af0d9695f534: Pulling fs layer af0d9695f534: Verifying Checksum af0d9695f534: Download complete af0d9695f534: Pull complete Digest: sha256:c23e16b601f1b7d50207f40bbbe3963f3acaa05232da9cab3d128a0c0749ff6f Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists 1d498b3a9c67: Already exists 542e5d603739: Already exists 78d1f78e4718: Pulling fs layer 78d1f78e4718: Verifying Checksum 78d1f78e4718: Download complete 78d1f78e4718: Pull complete Digest: sha256:49672e002009c9d1e370d4d50b06c404d9e727a04b97e7281ab7053087003c45 Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists 1d498b3a9c67: Already exists 542e5d603739: Already exists a08d92be0ca7: Pulling fs layer a08d92be0ca7: Verifying Checksum a08d92be0ca7: Download complete a08d92be0ca7: Pull complete Digest: sha256:e5abcc08b1f058cc9edbea4eea54f78c3c78b348c01a75297751aef9ac398aef Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists 3e89ff6a57d9: Pulling fs layer edbad3c6ed6f: Pulling fs layer 3e89ff6a57d9: Verifying Checksum 3e89ff6a57d9: Download complete edbad3c6ed6f: Verifying Checksum edbad3c6ed6f: Download complete 3e89ff6a57d9: Pull complete edbad3c6ed6f: Pull complete Digest: sha256:e198d102b829e06503e1627ff20e16acb1a152594a1909e9dcf0d13ce28e2414 Trying to pull repository registry:5000/kubevirt/winrmcli ... devel: Pulling from registry:5000/kubevirt/winrmcli 2176639d844b: Already exists 7c1ab5de42d5: Pulling fs layer 9391531b0959: Pulling fs layer d4e9df2eaabc: Pulling fs layer d4e9df2eaabc: Verifying Checksum d4e9df2eaabc: Download complete 7c1ab5de42d5: Verifying Checksum 7c1ab5de42d5: Download complete 9391531b0959: Verifying Checksum 9391531b0959: Download complete 7c1ab5de42d5: Pull complete 9391531b0959: Pull complete d4e9df2eaabc: Pull complete Digest: sha256:0f74a127121b8ee4fe3bd985550fc248403e426be506d51582ff57304c4b68be 2018/04/30 18:40:07 Waiting for host: 192.168.66.102:22 2018/04/30 18:40:07 Connected to tcp://192.168.66.102:22 Done ./cluster/clean.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0 ++ provider_prefix=kubevirt-functional-tests-openshift-release-crio0 ++ job_prefix=kubevirt-functional-tests-openshift-release-crio0 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.5.0-alpha.1-27-g5f465fb ++ KUBEVIRT_VERSION=v0.5.0-alpha.1-27-g5f465fb + source cluster/os-3.9.0/provider.sh ++ set -e ++ image=os-3.9.0@sha256:507263b67ddad581b086418d443c4fd1b2a30954e8fddff12254e0061a529410 ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0 ++ source hack/config-default.sh source hack/config-os-3.9.0.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0.sh ++ source hack/config-provider-os-3.9.0.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0/.kubectl +++ docker_prefix=localhost:33112/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Cleaning up ...' Cleaning up ... + cluster/kubectl.sh get vms --all-namespaces -o=custom-columns=NAME:.metadata.name,NAMESPACE:.metadata.namespace,FINALIZERS:.metadata.finalizers --no-headers + grep foregroundDeleteVirtualMachine + read p the server doesn't have a resource type "vms" + _kubectl delete ds -l kubevirt.io -n kube-system --cascade=false --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=libvirt --force --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=virt-handler --force --grace-period 0 No resources found + namespaces=(default ${namespace}) + for i in '${namespaces[@]}' + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete deployment -l kubevirt.io No resources found + _kubectl -n default delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete rs -l kubevirt.io No resources found + _kubectl -n default delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete services -l kubevirt.io No resources found + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n default delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete secrets -l kubevirt.io No resources found + _kubectl -n default delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete pv -l kubevirt.io No resources found + _kubectl -n default delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete pvc -l kubevirt.io No resources found + _kubectl -n default delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete ds -l kubevirt.io No resources found + _kubectl -n default delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n default delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete pods -l kubevirt.io No resources found + _kubectl -n default delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n default delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete rolebinding -l kubevirt.io No resources found + _kubectl -n default delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete roles -l kubevirt.io No resources found + _kubectl -n default delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete clusterroles -l kubevirt.io No resources found + _kubectl -n default delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n default delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n default get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0/.kubeconfig ++ cluster/os-3.9.0/.kubectl -n default get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + for i in '${namespaces[@]}' + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete deployment -l kubevirt.io No resources found + _kubectl -n kube-system delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete rs -l kubevirt.io No resources found + _kubectl -n kube-system delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete services -l kubevirt.io No resources found + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n kube-system delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete secrets -l kubevirt.io No resources found + _kubectl -n kube-system delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete pv -l kubevirt.io No resources found + _kubectl -n kube-system delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete pvc -l kubevirt.io No resources found + _kubectl -n kube-system delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete ds -l kubevirt.io No resources found + _kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n kube-system delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete pods -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete rolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete roles -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete clusterroles -l kubevirt.io No resources found + _kubectl -n kube-system delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl -n kube-system delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0/.kubeconfig ++ cluster/os-3.9.0/.kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + sleep 2 + echo Done Done ./cluster/deploy.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0 ++ provider_prefix=kubevirt-functional-tests-openshift-release-crio0 ++ job_prefix=kubevirt-functional-tests-openshift-release-crio0 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.5.0-alpha.1-27-g5f465fb ++ KUBEVIRT_VERSION=v0.5.0-alpha.1-27-g5f465fb + source cluster/os-3.9.0/provider.sh ++ set -e ++ image=os-3.9.0@sha256:507263b67ddad581b086418d443c4fd1b2a30954e8fddff12254e0061a529410 ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0 ++ source hack/config-default.sh source hack/config-os-3.9.0.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0.sh ++ source hack/config-provider-os-3.9.0.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0/.kubectl +++ docker_prefix=localhost:33112/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Deploying ...' Deploying ... + [[ -z openshift-release-crio ]] + [[ openshift-release-crio =~ .*-dev ]] + [[ openshift-release-crio =~ .*-release ]] + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/demo-content.yaml =~ .*demo.* ]] + continue + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml =~ .*demo.* ]] + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml serviceaccount "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver-auth-delegator" created rolebinding "kubevirt-apiserver" created role "kubevirt-apiserver" created clusterrole "kubevirt-apiserver" created clusterrole "kubevirt-controller" created serviceaccount "kubevirt-controller" created serviceaccount "kubevirt-privileged" created clusterrolebinding "kubevirt-controller" created clusterrolebinding "kubevirt-controller-cluster-admin" created clusterrolebinding "kubevirt-privileged-cluster-admin" created service "virt-api" created deployment "virt-api" created deployment "virt-controller" created daemonset "virt-handler" created customresourcedefinition "virtualmachines.kubevirt.io" created customresourcedefinition "virtualmachinereplicasets.kubevirt.io" created customresourcedefinition "virtualmachinepresets.kubevirt.io" created customresourcedefinition "offlinevirtualmachines.kubevirt.io" created + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R persistentvolumeclaim "disk-alpine" created persistentvolume "iscsi-disk-alpine" created persistentvolumeclaim "disk-custom" created persistentvolume "iscsi-disk-custom" created daemonset "iscsi-demo-target-tgtd" created serviceaccount "kubevirt-testing" created clusterrolebinding "kubevirt-testing-cluster-admin" created + '[' os-3.9.0 = vagrant-openshift ']' + '[' os-3.9.0 = os-3.9.0 ']' + _kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-controller"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-testing"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-privileged"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-apiserver"] + _kubectl adm policy add-scc-to-user privileged admin + export KUBECONFIG=cluster/os-3.9.0/.kubeconfig + KUBECONFIG=cluster/os-3.9.0/.kubeconfig + cluster/os-3.9.0/.kubectl adm policy add-scc-to-user privileged admin scc "privileged" added to: ["admin"] + echo Done Done ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-r9hrm 0/1 ContainerCreating 0 5s iscsi-demo-target-tgtd-wg9jx 0/1 ContainerCreating 0 5s virt-api-fd96f94b5-pg2qp 0/1 ContainerCreating 0 8s virt-api-fd96f94b5-rvxnk 0/1 ContainerCreating 0 8s virt-controller-5f7c946cc4-8bnks 0/1 ContainerCreating 0 8s virt-controller-5f7c946cc4-mgj84 0/1 ContainerCreating 0 8s virt-handler-8jtbw 0/1 ContainerCreating 0 2s virt-handler-bnbv7 0/1 ContainerCreating 0 2s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers virt-api-fd96f94b5-rvxnk 0/1 ContainerCreating 0 9s virt-controller-5f7c946cc4-8bnks 0/1 ContainerCreating 0 9s virt-handler-8jtbw 0/1 ContainerCreating 0 3s virt-handler-bnbv7 0/1 ContainerCreating 0 3s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n '' ']' ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx false virt-api-fd96f94b5-rvxnk' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx false virt-api-fd96f94b5-rvxnk + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' + '[' -n 'false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-r9hrm false iscsi-demo-target-tgtd-wg9jx + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-wg9jx' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-wg9jx + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n '' ']' ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '/virt-controller/ && /true/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ wc -l + '[' 2 -lt 1 ']' + kubectl get pods -n kube-system + cluster/kubectl.sh get pods -n kube-system NAME READY STATUS RESTARTS AGE iscsi-demo-target-tgtd-r9hrm 1/1 Running 1 1m iscsi-demo-target-tgtd-wg9jx 1/1 Running 1 1m virt-api-fd96f94b5-pg2qp 1/1 Running 0 1m virt-api-fd96f94b5-rvxnk 1/1 Running 1 1m virt-controller-5f7c946cc4-8bnks 1/1 Running 0 1m virt-controller-5f7c946cc4-mgj84 1/1 Running 0 1m virt-handler-8jtbw 1/1 Running 0 1m virt-handler-bnbv7 1/1 Running 0 1m + kubectl version + cluster/kubectl.sh version oc v3.9.0+ba7faec-1 kubernetes v1.9.1+a0ce1bc657 features: Basic-Auth GSSAPI Kerberos SPNEGO Server https://127.0.0.1:33109 openshift v3.9.0+ba7faec-1 kubernetes v1.9.1+a0ce1bc657 + ginko_params='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/junit.xml' + [[ -d /home/nfs/images/windows2016 ]] + [[ openshift-release-crio == windows ]] + FUNC_TEST_ARGS='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/junit.xml' + make functest hack/dockerized "hack/build-func-tests.sh" sha256:0e817e41f9750e44335dde1be5cb34809abe48c8add43baf165907418e2e75ce go version go1.10 linux/amd64 go version go1.10 linux/amd64 Compiling tests... compiled tests.test hack/functests.sh Running Suite: Tests Suite ========================== Random Seed: 1525113772 Will run 90 of 90 specs • [SLOW TEST:58.920 seconds] Health Monitoring /root/go/src/kubevirt.io/kubevirt/tests/vm_monitoring_test.go:37 A VM with a watchdog device /root/go/src/kubevirt.io/kubevirt/tests/vm_monitoring_test.go:56 should be shut down when the watchdog expires /root/go/src/kubevirt.io/kubevirt/tests/vm_monitoring_test.go:57 ------------------------------ • [SLOW TEST:44.880 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:48 VirtualMachine attached to the pod network /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:146 should be able to reach /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 the Inbound VM /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • ------------------------------ • [SLOW TEST:5.457 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:48 VirtualMachine attached to the pod network /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:146 should be reachable via the propagated IP from a Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 on the same node from Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • [SLOW TEST:6.184 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:48 VirtualMachine attached to the pod network /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:146 should be reachable via the propagated IP from a Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 on a different node from Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ •• ------------------------------ • [SLOW TEST:6.449 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:48 VirtualMachine attached to the pod network /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:146 with a service matching the vm exposed /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:218 should be able to reach the vm based on labels specified on the vm /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:238 ------------------------------ • ------------------------------ • [SLOW TEST:57.852 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:41 Starting and stopping the same VM /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:90 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:91 should success multiple times /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:92 ------------------------------ • [SLOW TEST:19.728 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:41 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:111 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:112 should not modify the spec on status update /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:113 ------------------------------ • [SLOW TEST:34.676 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:41 Starting multiple VMs /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:129 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:130 should success /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:131 ------------------------------ • ------------------------------ • [SLOW TEST:56.580 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:39 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:119 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:120 should be successfully started /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:121 ------------------------------ • [SLOW TEST:129.837 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:39 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:119 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:120 should be successfully started and stopped multiple times /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:141 ------------------------------ • [SLOW TEST:40.433 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:39 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:119 With an emptyDisk defined /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:173 should create a writeable emptyDisk with the right capacity /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:175 ------------------------------ • [SLOW TEST:54.643 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:39 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:119 With ephemeral alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:223 should be successfully started /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:225 ------------------------------ • Failure [282.217 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:39 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:119 With ephemeral alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:223 should not persist data [It] /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:245 Expected error: : 200000000000 expect: timer expired after 200 seconds not to have occurred /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:301 ------------------------------ STEP: Getting the ISCSI pod logs STEP: Checking that ISCSI is ready STEP: Starting the VM STEP: Starting a VM STEP: Waiting until the VM will start level=info timestamp=2018-04-30T18:51:46.181601Z pos=utils.go:224 component=tests msg="Created virtual machine pod virt-launcher-testvmsscls-k9z7r" level=info timestamp=2018-04-30T18:52:02.955395Z pos=utils.go:224 component=tests msg="Pod owner ship transfered to the node virt-launcher-testvmsscls-k9z7r" level=info timestamp=2018-04-30T18:52:04.023752Z pos=utils.go:224 component=tests msg="VM defined." level=info timestamp=2018-04-30T18:52:04.041715Z pos=utils.go:224 component=tests msg="VM started." STEP: Writing an arbitrary file to it's EFI partition STEP: Killing a VM STEP: Starting the VM again STEP: Starting a VM STEP: Waiting until the VM will start level=info timestamp=2018-04-30T18:52:39.259458Z pos=utils.go:224 component=tests msg="Created virtual machine pod virt-launcher-testvmsscls-k9z7r" level=info timestamp=2018-04-30T18:52:39.259682Z pos=utils.go:224 component=tests msg="Pod owner ship transfered to the node virt-launcher-testvmsscls-k9z7r" level=info timestamp=2018-04-30T18:52:39.260346Z pos=utils.go:224 component=tests msg="VM defined." level=info timestamp=2018-04-30T18:52:39.260515Z pos=utils.go:224 component=tests msg="VM started." STEP: Making sure that the previously written file is not present • [SLOW TEST:303.433 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:39 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:119 With VM with two PVCs /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:305 should start vm multiple times /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:317 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.027 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 should succeed to start a vm [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:131 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1100 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.029 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 should succeed to stop a running vm [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:137 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1100 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.018 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with winrm connection [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:148 should have correct UUID /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:190 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1100 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.011 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with winrm connection [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:148 should have pod IP /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:206 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1100 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.007 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with kubectl command [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:224 should succeed to start a vm /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:240 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1100 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.011 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:57 with kubectl command [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:224 should succeed to stop a vm /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:248 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1100 ------------------------------ • [SLOW TEST:5.897 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should scale /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 to three, to two and then to zero replicas /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • [SLOW TEST:13.180 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should scale /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 to five, to six and then to zero replicas /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ •• ------------------------------ • [SLOW TEST:23.150 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should update readyReplicas once VMs are up /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:157 ------------------------------ • [SLOW TEST:7.790 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should remove VMs once it is marked for deletion /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:169 ------------------------------ • ------------------------------ • [SLOW TEST:6.390 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:46 should not scale when paused and scale when resume /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:223 ------------------------------ • [SLOW TEST:40.742 seconds] LeaderElection /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:43 Start a VM /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:53 when the controller pod is not running /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:54 should success /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:55 ------------------------------ • [SLOW TEST:39.934 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:64 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a cirros image /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 should return that we are running cirros /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:67 ------------------------------ • [SLOW TEST:38.675 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:64 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a fedora image /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:76 should return that we are running fedora /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:77 ------------------------------ • [SLOW TEST:50.547 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:64 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 should be able to reconnect to console multiple times /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:86 ------------------------------ • [SLOW TEST:21.083 seconds] VNC /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:46 with VNC connection /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:47 should allow accessing the VNC device /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:48 ------------------------------ • [SLOW TEST:37.720 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:81 should have cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:82 ------------------------------ • [SLOW TEST:84.768 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:81 with injected ssh-key /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:93 should have ssh-key under authorized keys /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:94 ------------------------------ • [SLOW TEST:39.183 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 with cloudInitNoCloud userData source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:119 should process provided cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:120 ------------------------------ • [SLOW TEST:39.320 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 should take user-data from k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:163 ------------------------------ • [SLOW TEST:6.729 seconds] Subresource Api /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:37 Rbac Authorization /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:48 with correct permissions /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:49 should be allowed to access subresource endpoint /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:50 ------------------------------ • [SLOW TEST:6.090 seconds] Subresource Api /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:37 Rbac Authorization /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:48 Without permissions /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:54 should not be able to access subresource endpoint /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:55 ------------------------------ • ------------------------------ • [SLOW TEST:21.877 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 should start it /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:65 ------------------------------ • [SLOW TEST:20.109 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 should attach virt-launcher to it /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:73 ------------------------------ •••• ------------------------------ • [SLOW TEST:17.321 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:162 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:163 should retry starting the VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:164 ------------------------------ • [SLOW TEST:22.619 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:162 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:163 should log warning and proceed once the secret is there /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:195 ------------------------------ • [SLOW TEST:60.390 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 when virt-launcher crashes /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:243 should be stopped and have Failed phase /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:244 ------------------------------ • [SLOW TEST:41.981 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 when virt-handler crashes /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:273 should recover and continue management /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:274 ------------------------------ • [SLOW TEST:108.299 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 when virt-handler is responsive /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:310 should indicate that a node is ready for vms /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:311 ------------------------------ • [SLOW TEST:113.917 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 when virt-handler is not responsive /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:341 the node controller should react /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:374 ------------------------------ S [SKIPPING] [1.005 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:414 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-default [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Skip log query tests for JENKINS ci test environment /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:419 ------------------------------ S [SKIPPING] [0.879 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:59 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:414 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-alternative [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Skip log query tests for JENKINS ci test environment /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:419 ------------------------------ • ------------------------------ • [SLOW TEST:20.858 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Delete a VM's Pod /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:489 should result in the VM moving to a finalized state /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:490 ------------------------------ •... Timeout [92.219 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Delete a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:522 with an active pod. /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:523 should result in pod being terminated [It] /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:524 Timed out /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:524 ------------------------------ STEP: Creating the VM level=info timestamp=2018-04-30T19:16:34.588180Z pos=utils.go:224 component=tests msg="Created virtual machine pod virt-launcher-testvmsw59z-qwnpm" level=info timestamp=2018-04-30T19:16:52.865881Z pos=utils.go:224 component=tests msg="Pod owner ship transfered to the node virt-launcher-testvmsw59z-qwnpm" level=info timestamp=2018-04-30T19:16:53.920550Z pos=utils.go:224 component=tests msg="VM defined." level=info timestamp=2018-04-30T19:16:53.936455Z pos=utils.go:224 component=tests msg="VM started." STEP: Verifying VM's pod is active STEP: Deleting the VM STEP: Verifying VM's pod terminates • Failure [25.294 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Delete a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:522 with grace period greater than 0 /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:549 should run graceful shutdown [It] /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:550 Timed out after 75.000s. Expected : 1 to equal : 0 /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:544 ------------------------------ STEP: Setting a VM termination grace period to 5 STEP: Creating the VM level=info timestamp=2018-04-30T19:18:05.455707Z pos=utils.go:224 component=tests msg="Created virtual machine pod virt-launcher-testvm89k7l-cs6cr" level=info timestamp=2018-04-30T19:18:22.288384Z pos=utils.go:224 component=tests msg="Pod owner ship transfered to the node virt-launcher-testvm89k7l-cs6cr" level=info timestamp=2018-04-30T19:18:23.517682Z pos=utils.go:224 component=tests msg="VM defined." level=info timestamp=2018-04-30T19:18:23.612670Z pos=utils.go:224 component=tests msg="VM started." STEP: Deleting the VM level=info timestamp=2018-04-30T19:18:24.256311Z pos=utils.go:235 component=tests msg="Created virtual machine pod virt-launcher-testvm89k7l-cs6cr" level=info timestamp=2018-04-30T19:18:24.256465Z pos=utils.go:235 component=tests msg="Pod owner ship transfered to the node virt-launcher-testvm89k7l-cs6cr" level=info timestamp=2018-04-30T19:18:24.257068Z pos=utils.go:235 component=tests msg="VM defined." level=info timestamp=2018-04-30T19:18:24.257256Z pos=utils.go:235 component=tests msg="VM started." level=info timestamp=2018-04-30T19:18:24.647351Z pos=utils.go:235 component=tests msg="Signaled Graceful Shutdown" level=info timestamp=2018-04-30T19:18:24.647565Z pos=utils.go:235 component=tests msg="Deleted virtual machine pod virt-launcher-testvm89k7l-cs6cr" level=info timestamp=2018-04-30T19:18:24.648226Z pos=utils.go:235 component=tests msg="Signaled Graceful Shutdown" level=info timestamp=2018-04-30T19:18:24.648459Z pos=utils.go:235 component=tests msg="Signaled Graceful Shutdown" level=info timestamp=2018-04-30T19:18:24.648667Z pos=utils.go:235 component=tests msg="Signaled Graceful Shutdown" level=info timestamp=2018-04-30T19:18:24.887295Z pos=utils.go:235 component=tests msg="Signaled Graceful Shutdown" level=info timestamp=2018-04-30T19:18:29.183785Z pos=utils.go:235 component=tests msg="VM stopping" STEP: Checking that virt-handler logs VM graceful shutdown STEP: Checking that the VM does not exist after grace period • [SLOW TEST:36.080 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Killed VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:601 should be in Failed phase /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:602 ------------------------------ • [SLOW TEST:30.427 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:45 Killed VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:601 should be left alone by virt-handler /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:630 ------------------------------ ••••••••••volumedisk0 compute ------------------------------ • [SLOW TEST:51.213 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:39 VM definition /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:50 with 3 CPU cores /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:51 should report 3 cpu cores under guest OS /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:57 ------------------------------ • [SLOW TEST:36.978 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:39 New VM with all supported drives /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:110 should have all the device nodes /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:133 ------------------------------ •• ------------------------------ • [SLOW TEST:19.929 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 should update OfflineVirtualMachine once VMs are up /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:204 ------------------------------ • [SLOW TEST:5.429 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 should remove VM once the OVM is marked for deletion /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:213 ------------------------------ •• ------------------------------ • [SLOW TEST:52.652 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 should recreate VM if the VM's pod gets deleted /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:270 ------------------------------ • [SLOW TEST:49.613 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 should stop VM if running set to false /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:330 ------------------------------ • [SLOW TEST:149.659 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 should start and stop VM multiple times /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:338 ------------------------------ • [SLOW TEST:113.922 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 should not update the VM spec if Running /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:351 ------------------------------ • [SLOW TEST:207.949 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 should survive guest shutdown, multiple times /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:392 ------------------------------ • [SLOW TEST:19.626 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 Using virtctl interface /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:430 should start a VM once /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:431 ------------------------------ • [SLOW TEST:69.245 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:47 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:115 Using virtctl interface /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:430 should stop a VM once /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:462 ------------------------------ Waiting for namespace kubevirt-test-default to be removed, this can take a while ... Waiting for namespace kubevirt-test-alternative to be removed, this can take a while ... Summarizing 3 Failures: [Fail] Storage Starting a VM With ephemeral alpine PVC [It] should not persist data /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:301 [Timeout...] Vmlifecycle Delete a VM with an active pod. [It] should result in pod being terminated /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:524 [Fail] Vmlifecycle Delete a VM with grace period greater than 0 [It] should run graceful shutdown /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:544 Ran 82 of 90 Specs in 3015.806 seconds FAIL! -- 79 Passed | 3 Failed | 0 Pending | 8 Skipped --- FAIL: TestTests (3015.81s) FAIL make: *** [functest] Error 1 + make cluster-down ./cluster/down.sh bda74f94dca5 0cc9ae7bbe36 db034732e464 2b9f83e65b9f 6ffb12ed484f bda74f94dca5 0cc9ae7bbe36 db034732e464 2b9f83e65b9f 6ffb12ed484f kubevirt-functional-tests-openshift-release-crio0-node01 kubevirt-functional-tests-openshift-release-crio0-node02