+ export WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio + WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio + [[ openshift-release-crio =~ openshift-.* ]] + [[ openshift-release-crio =~ .*-crio ]] + export PROVIDER=os-3.9.0-crio + PROVIDER=os-3.9.0-crio + export VAGRANT_NUM_NODES=1 + VAGRANT_NUM_NODES=1 + export NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + export NAMESPACE=kube-system + NAMESPACE=kube-system + trap '{ make cluster-down; }' EXIT + make cluster-down ./cluster/down.sh + make cluster-up ./cluster/up.sh WARNING: You're not using the default seccomp profile WARNING: bridge-nf-call-iptables is disabled WARNING: bridge-nf-call-ip6tables is disabled kubevirt-functional-tests-openshift-release-crio0-node02 2018/05/01 11:55:18 Waiting for host: 192.168.66.102:22 2018/05/01 11:55:21 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:55:29 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:55:37 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:55:45 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:55:53 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:55:58 Connected to tcp://192.168.66.102:22 Removed symlink /etc/systemd/system/multi-user.target.wants/origin-master-api.service. Removed symlink /etc/systemd/system/origin-node.service.wants/origin-master-api.service. Removed symlink /etc/systemd/system/multi-user.target.wants/origin-master-controllers.service. kubevirt-functional-tests-openshift-release-crio0-node01 2018/05/01 11:56:06 Waiting for host: 192.168.66.101:22 2018/05/01 11:56:09 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:56:17 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:56:25 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/05/01 11:56:30 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: connection refused. Sleeping 5s 2018/05/01 11:56:35 Connected to tcp://192.168.66.101:22 NAME STATUS ROLES AGE VERSION node01 Ready master 5d v1.9.1+a0ce1bc657 PING node02 (192.168.66.102) 56(84) bytes of data. 64 bytes from node02 (192.168.66.102): icmp_seq=1 ttl=64 time=0.309 ms --- node02 ping statistics --- 1 packets transmitted, 1 received, 0% packet loss, time 0ms rtt min/avg/max/mdev = 0.309/0.309/0.309/0.000 ms Found node02. Adding it to the inventory. ping: node03: Name or service not known PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) PLAY [Ensure there are new_nodes] ********************************************** TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] PLAY [Initialization Checkpoint Start] ***************************************** TASK [Set install initialization 'In Progress'] ******************************** ok: [node01] PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) [WARNING]: Could not match supplied host pattern, ignoring: oo_lb_to_config PLAY [Ensure that all non-node hosts are accessible] *************************** TASK [Gathering Facts] ********************************************************* ok: [node01] PLAY [Initialize basic host facts] ********************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/deprecations.yml for node01, node02 TASK [openshift_sanitize_inventory : Check for usage of deprecated variables] *** ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : debug] ************************************ skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_stats] ******************************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Assign deprecated variables to correct counterparts] *** included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_logging.yml for node01, node02 included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_metrics.yml for node01, node02 TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : Standardize on latest variable names] ***** ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : Normalize openshift_release] ************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Abort when openshift_release is invalid] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_sanitize_inventory/tasks/unsupported.yml for node01, node02 TASK [openshift_sanitize_inventory : Ensure that openshift_use_dnsmasq is true] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that openshift_node_dnsmasq_install_network_manager_hook is true] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* skipping: [node01] => (item=None) skipping: [node02] => (item=None) TASK [openshift_sanitize_inventory : Ensure that dynamic provisioning is set if using dynamic storage] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure clusterid is set along with the cloudprovider] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure ansible_service_broker_remove and ansible_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure template_service_broker_remove and template_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that all requires vsphere configuration variables are set] *** skipping: [node01] skipping: [node02] TASK [Detecting Operating System from ostree_booted] *************************** ok: [node02] ok: [node01] TASK [set openshift_deployment_type if unset] ********************************** skipping: [node01] skipping: [node02] TASK [initialize_facts set fact openshift_is_atomic and openshift_is_containerized] *** ok: [node01] ok: [node02] TASK [Determine Atomic Host Docker Version] ************************************ skipping: [node01] skipping: [node02] TASK [assert atomic host docker version is 1.12 or later] ********************** skipping: [node01] skipping: [node02] PLAY [Initialize special first-master variables] ******************************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] PLAY [Disable web console if required] ***************************************** TASK [set_fact] **************************************************************** skipping: [node01] PLAY [Install packages necessary for installer] ******************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Ensure openshift-ansible installer package deps are installed] *********** ok: [node02] => (item=iproute) ok: [node02] => (item=dbus-python) ok: [node02] => (item=PyYAML) ok: [node02] => (item=python-ipaddress) ok: [node02] => (item=yum-utils) TASK [Ensure various deps for running system containers are installed] ********* skipping: [node02] => (item=atomic) skipping: [node02] => (item=ostree) skipping: [node02] => (item=runc) PLAY [Initialize cluster facts] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [Gather Cluster facts] **************************************************** changed: [node02] ok: [node01] TASK [Set fact of no_proxy_internal_hostnames] ********************************* skipping: [node01] skipping: [node02] TASK [Initialize openshift.node.sdn_mtu] *************************************** ok: [node02] ok: [node01] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* skipping: [node01] TASK [include_role] ************************************************************ skipping: [node01] TASK [debug] ******************************************************************* skipping: [node01] PLAY [Set openshift_version for etcd, node, and master hosts] ****************** skipping: no hosts matched PLAY [Ensure the requested version packages are available.] ******************** skipping: no hosts matched PLAY [Verify Requirements] ***************************************************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [Run variable sanity checks] ********************************************** ok: [node01] PLAY [Initialization Checkpoint End] ******************************************* TASK [Set install initialization 'Complete'] *********************************** ok: [node01] PLAY [Validate node hostnames] ************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Query DNS for IP address of node02] ************************************** ok: [node02] TASK [Validate openshift_hostname when defined] ******************************** skipping: [node02] TASK [Validate openshift_ip exists on node when defined] *********************** skipping: [node02] PLAY [Setup yum repositories for all hosts] ************************************ TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Install Red Hat Subscription manager] ******************* skipping: [node02] TASK [rhel_subscribe : Is host already registered?] **************************** skipping: [node02] TASK [rhel_subscribe : Register host] ****************************************** skipping: [node02] TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Determine if OpenShift Pool Already Attached] *********** skipping: [node02] TASK [rhel_subscribe : Attach to OpenShift Pool] ******************************* skipping: [node02] TASK [rhel_subscribe : include_tasks] ****************************************** skipping: [node02] TASK [openshift_repos : openshift_repos detect ostree] ************************* ok: [node02] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** ok: [node02] TASK [openshift_repos : Remove openshift_additional.repo file] ***************** ok: [node02] TASK [openshift_repos : Create any additional repos that are defined] ********** TASK [openshift_repos : include_tasks] ***************************************** skipping: [node02] TASK [openshift_repos : include_tasks] ***************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_repos/tasks/centos_repos.yml for node02 TASK [openshift_repos : Configure origin gpg keys] ***************************** ok: [node02] TASK [openshift_repos : Configure correct origin release repository] *********** ok: [node02] => (item=/usr/share/ansible/openshift-ansible/roles/openshift_repos/templates/CentOS-OpenShift-Origin.repo.j2) TASK [openshift_repos : Ensure clean repo cache in the event repos have been changed manually] *** changed: [node02] => { "msg": "First run of openshift_repos" } TASK [openshift_repos : Record that openshift_repos already ran] *************** ok: [node02] RUNNING HANDLER [openshift_repos : refresh cache] ****************************** changed: [node02] PLAY [Configure os_firewall] *************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [os_firewall : Detecting Atomic Host Operating System] ******************** ok: [node02] TASK [os_firewall : Set fact r_os_firewall_is_atomic] ************************** ok: [node02] TASK [os_firewall : include_tasks] ********************************************* skipping: [node02] TASK [os_firewall : include_tasks] ********************************************* included: /usr/share/ansible/openshift-ansible/roles/os_firewall/tasks/iptables.yml for node02 TASK [os_firewall : Ensure firewalld service is not enabled] ******************* ok: [node02] TASK [os_firewall : Wait 10 seconds after disabling firewalld] ***************** skipping: [node02] TASK [os_firewall : Install iptables packages] ********************************* ok: [node02] => (item=iptables) ok: [node02] => (item=iptables-services) TASK [os_firewall : Start and enable iptables service] ************************* ok: [node02 -> node02] => (item=node02) TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [container_runtime : Setup the docker-storage for overlay] **************** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_excluder : Install excluders] ********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** ok: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* ok: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/pre.yml for node02 TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Add enterprise registry, if necessary] *************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Get current installed Docker version] **************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/docker_sanity.yml for node02 TASK [container_runtime : Error out if Docker pre-installed but too old] ******* skipping: [node02] TASK [container_runtime : Error out if requested Docker is too old] ************ skipping: [node02] TASK [container_runtime : Fail if Docker version requested but downgrade is required] *** skipping: [node02] TASK [container_runtime : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [node02] TASK [container_runtime : Install Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure docker.service.d directory exists] ************ ok: [node02] TASK [container_runtime : Configure Docker service unit file] ****************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Set registry params] ********************************* skipping: [node02] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': []}) TASK [container_runtime : Place additional/blocked/insecure registries in /etc/containers/registries.conf] *** skipping: [node02] TASK [container_runtime : Set Proxy Settings] ********************************** skipping: [node02] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [container_runtime : Set various Docker options] ************************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Configure Docker Network OPTIONS] ******************** ok: [node02] TASK [container_runtime : Detect if docker is already started] ***************** ok: [node02] TASK [container_runtime : Start the Docker service] **************************** ok: [node02] TASK [container_runtime : set_fact] ******************************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/post.yml for node02 TASK [container_runtime : Ensure /var/lib/containers exists] ******************* ok: [node02] TASK [container_runtime : Fix SELinux Permissions on /var/lib/containers] ****** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/registry_auth.yml for node02 TASK [container_runtime : Check for credentials file for registry auth] ******** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth] ***** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth (alternative)] *** skipping: [node02] TASK [container_runtime : stat the docker data dir] **************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Fail quickly if openshift_docker_options are set] **** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Install Docker so we can use the client] ************* skipping: [node02] TASK [container_runtime : Disable Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** skipping: [node02] TASK [container_runtime : debug] *********************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Pre-pull Container Engine System Container image] **** skipping: [node02] TASK [container_runtime : Ensure container-engine.service.d directory exists] *** skipping: [node02] TASK [container_runtime : Ensure /etc/docker directory exists] ***************** skipping: [node02] TASK [container_runtime : Install Container Engine System Container] *********** skipping: [node02] TASK [container_runtime : Configure Container Engine Service File] ************* skipping: [node02] TASK [container_runtime : Configure Container Engine] ************************** skipping: [node02] TASK [container_runtime : Start the Container Engine service] ****************** skipping: [node02] TASK [container_runtime : set_fact] ******************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Check we are not using node as a Docker container with CRI-O] *** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/pre.yml for node02 TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Add enterprise registry, if necessary] *************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/syscontainer_packages.yml for node02 TASK [container_runtime : Ensure container-selinux is installed] *************** ok: [node02] TASK [container_runtime : Ensure atomic is installed] ************************** ok: [node02] TASK [container_runtime : Ensure runc is installed] **************************** ok: [node02] TASK [container_runtime : Check that overlay is in the kernel] ***************** changed: [node02] TASK [container_runtime : Add overlay to modprobe.d] *************************** skipping: [node02] TASK [container_runtime : Manually modprobe overlay into the kernel] *********** skipping: [node02] TASK [container_runtime : Enable and start systemd-modules-load] *************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/atomic_proxy.yml for node02 TASK [container_runtime : Add http_proxy to /etc/atomic.conf] ****************** skipping: [node02] TASK [container_runtime : Add https_proxy to /etc/atomic.conf] ***************** skipping: [node02] TASK [container_runtime : Add no_proxy to /etc/atomic.conf] ******************** skipping: [node02] TASK [container_runtime : debug] *********************************************** ok: [node02] => { "l_crio_image": "docker.io/kubevirtci/crio:1.9.10" } TASK [container_runtime : Pre-pull CRI-O System Container image] *************** ok: [node02] TASK [container_runtime : Install CRI-O System Container] ********************** ok: [node02] TASK [container_runtime : Remove CRI-O default configuration files] ************ ok: [node02] => (item=/etc/cni/net.d/200-loopback.conf) ok: [node02] => (item=/etc/cni/net.d/100-crio-bridge.conf) TASK [container_runtime : Create the CRI-O configuration] ********************** ok: [node02] TASK [container_runtime : Ensure CNI configuration directory exists] *********** ok: [node02] TASK [container_runtime : Add iptables allow rules] **************************** ok: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove iptables rules] ******************************* TASK [container_runtime : Add firewalld allow rules] *************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove firewalld allow rules] ************************ TASK [container_runtime : Configure the CNI network] *************************** ok: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-storage] ****************** ok: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-network] ****************** ok: [node02] TASK [container_runtime : Start the CRI-O service] ***************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/common/post.yml for node02 TASK [container_runtime : Ensure /var/lib/containers exists] ******************* ok: [node02] TASK [container_runtime : Fix SELinux Permissions on /var/lib/containers] ****** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/container_runtime/tasks/registry_auth.yml for node02 TASK [container_runtime : Check for credentials file for registry auth] ******** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth] ***** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth (alternative)] *** skipping: [node02] TASK [container_runtime : stat the docker data dir] **************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [include_role] ************************************************************ TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** ok: [node01] TASK [openshift_version : include_tasks] *************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_version/tasks/first_master_rpm_version.yml for node01 TASK [openshift_version : Set rpm version to configure if openshift_pkg_version specified] *** skipping: [node01] TASK [openshift_version : Set openshift_version for rpm installation] ********** included: /usr/share/ansible/openshift-ansible/roles/openshift_version/tasks/check_available_rpms.yml for node01 TASK [openshift_version : Get available origin version] ************************ ok: [node01] TASK [openshift_version : fail] ************************************************ skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : set_fact] ******************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] TASK [debug] ******************************************************************* ok: [node01] => { "msg": "openshift_pkg_version set to -3.9.0" } PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [set_fact] **************************************************************** ok: [node02] PLAY [Ensure the requested version packages are available.] ******************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [include_role] ************************************************************ TASK [openshift_version : Check openshift_version for rpm installation] ******** included: /usr/share/ansible/openshift-ansible/roles/openshift_version/tasks/check_available_rpms.yml for node02 TASK [openshift_version : Get available origin version] ************************ ok: [node02] TASK [openshift_version : fail] ************************************************ skipping: [node02] TASK [openshift_version : Fail if rpm version and docker image version are different] *** skipping: [node02] TASK [openshift_version : For an RPM install, abort when the release requested does not match the available version.] *** skipping: [node02] TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_image_tag": "v3.9.0" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_pkg_version": "-3.9.0" } PLAY [Node Install Checkpoint Start] ******************************************* TASK [Set Node install 'In Progress'] ****************************************** ok: [node01] PLAY [Create OpenShift certificates for node hosts] **************************** TASK [openshift_node_certificates : Ensure CA certificate exists on openshift_ca_host] *** ok: [node02 -> node01] TASK [openshift_node_certificates : fail] ************************************** skipping: [node02] TASK [openshift_node_certificates : Check status of node certificates] ********* ok: [node02] => (item=system:node:node02.crt) ok: [node02] => (item=system:node:node02.key) ok: [node02] => (item=system:node:node02.kubeconfig) ok: [node02] => (item=ca.crt) ok: [node02] => (item=server.key) ok: [node02] => (item=server.crt) TASK [openshift_node_certificates : set_fact] ********************************** ok: [node02] TASK [openshift_node_certificates : Create openshift_generated_configs_dir if it does not exist] *** ok: [node02 -> node01] TASK [openshift_node_certificates : find] ************************************** ok: [node02 -> node01] TASK [openshift_node_certificates : Generate the node client config] *********** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Generate the node server certificate] ****** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Create a tarball of the node config directories] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Retrieve the node config tarballs from the master] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Ensure certificate directory exists] ******* ok: [node02] TASK [openshift_node_certificates : Unarchive the tarball on the node] ********* changed: [node02] TASK [openshift_node_certificates : Delete local temp directory] *************** ok: [node02 -> localhost] TASK [openshift_node_certificates : Copy OpenShift CA to system CA trust] ****** ok: [node02] => (item={u'cert': u'/etc/origin/node/ca.crt', u'id': u'openshift'}) PLAY [Disable excluders] ******************************************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": true } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": true } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/disable.yml for node02 TASK [openshift_excluder : Include verify_upgrade.yml when upgrading] ********** skipping: [node02] TASK [openshift_excluder : Disable excluders before the upgrade to remove older excluding expressions] *** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* changed: [node02] TASK [openshift_excluder : Include install.yml] ******************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Include exclude.yml] ******************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** changed: [node02] TASK [openshift_excluder : Include unexclude.yml] ****************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* changed: [node02] PLAY [Evaluate node groups] **************************************************** TASK [Gathering Facts] ********************************************************* ok: [localhost] TASK [Evaluate oo_containerized_master_nodes] ********************************** skipping: [localhost] => (item=node02) [WARNING]: Could not match supplied host pattern, ignoring: oo_containerized_master_nodes PLAY [Configure containerized nodes] ******************************************* skipping: no hosts matched PLAY [Configure nodes] ********************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_clock : Determine if chrony is installed] ********************** [WARNING]: Consider using yum, dnf or zypper module rather than running rpm changed: [node02] TASK [openshift_clock : Install ntp package] *********************************** skipping: [node02] TASK [openshift_clock : Start and enable ntpd/chronyd] ************************* changed: [node02] TASK [openshift_cloud_provider : Set cloud provider facts] ********************* skipping: [node02] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [node02] TASK [openshift_cloud_provider : include the defined cloud provider files] ***** skipping: [node02] TASK [openshift_node : fail] *************************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/dnsmasq_install.yml for node02 TASK [openshift_node : Check for NetworkManager service] *********************** ok: [node02] TASK [openshift_node : Set fact using_network_manager] ************************* ok: [node02] TASK [openshift_node : Install dnsmasq] **************************************** ok: [node02] TASK [openshift_node : ensure origin/node directory exists] ******************** ok: [node02] => (item=/etc/origin) changed: [node02] => (item=/etc/origin/node) TASK [openshift_node : Install node-dnsmasq.conf] ****************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/dnsmasq.yml for node02 TASK [openshift_node : Install dnsmasq configuration] ************************** ok: [node02] TASK [openshift_node : Deploy additional dnsmasq.conf] ************************* skipping: [node02] TASK [openshift_node : Enable dnsmasq] ***************************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/dnsmasq/network-manager.yml for node02 TASK [openshift_node : Install network manager dispatch script] **************** ok: [node02] TASK [openshift_node : Add iptables allow rules] ******************************* ok: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) ok: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) ok: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) ok: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove iptables rules] ********************************** TASK [openshift_node : Add firewalld allow rules] ****************************** skipping: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) skipping: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) skipping: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove firewalld allow rules] *************************** TASK [openshift_node : Update journald config] ********************************* included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/journald.yml for node02 TASK [openshift_node : Checking for journald.conf] ***************************** ok: [node02] TASK [openshift_node : Create journald persistence directories] **************** ok: [node02] TASK [openshift_node : Update journald setup] ********************************** ok: [node02] => (item={u'var': u'Storage', u'val': u'persistent'}) ok: [node02] => (item={u'var': u'Compress', u'val': True}) ok: [node02] => (item={u'var': u'SyncIntervalSec', u'val': u'1s'}) ok: [node02] => (item={u'var': u'RateLimitInterval', u'val': u'1s'}) ok: [node02] => (item={u'var': u'RateLimitBurst', u'val': 10000}) ok: [node02] => (item={u'var': u'SystemMaxUse', u'val': u'8G'}) ok: [node02] => (item={u'var': u'SystemKeepFree', u'val': u'20%'}) ok: [node02] => (item={u'var': u'SystemMaxFileSize', u'val': u'10M'}) ok: [node02] => (item={u'var': u'MaxRetentionSec', u'val': u'1month'}) ok: [node02] => (item={u'var': u'MaxFileSec', u'val': u'1day'}) ok: [node02] => (item={u'var': u'ForwardToSyslog', u'val': False}) ok: [node02] => (item={u'var': u'ForwardToWall', u'val': False}) TASK [openshift_node : Restart journald] *************************************** skipping: [node02] TASK [openshift_node : Disable swap] ******************************************* ok: [node02] TASK [openshift_node : include node installer] ********************************* included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/install.yml for node02 TASK [openshift_node : Install Node package, sdn-ovs, conntrack packages] ****** ok: [node02] => (item={u'name': u'origin-node-3.9.0'}) ok: [node02] => (item={u'name': u'origin-sdn-ovs-3.9.0', u'install': True}) ok: [node02] => (item={u'name': u'conntrack-tools'}) TASK [openshift_node : Pre-pull node image when containerized] ***************** skipping: [node02] TASK [openshift_node : Restart cri-o] ****************************************** changed: [node02] TASK [openshift_node : restart NetworkManager to ensure resolv.conf is present] *** skipping: [node02] TASK [openshift_node : sysctl] ************************************************* ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/registry_auth.yml for node02 TASK [openshift_node : Check for credentials file for registry auth] *********** skipping: [node02] TASK [openshift_node : Create credentials for registry auth] ******************* skipping: [node02] TASK [openshift_node : Create credentials for registry auth (alternative)] ***** skipping: [node02] TASK [openshift_node : Setup ro mount of /root/.docker for containerized hosts] *** skipping: [node02] TASK [openshift_node : include standard node config] *************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config.yml for node02 TASK [openshift_node : Install the systemd units] ****************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/systemd_units.yml for node02 TASK [openshift_node : Install Node service file] ****************************** ok: [node02] TASK [openshift_node : include node deps docker service file] ****************** skipping: [node02] TASK [openshift_node : include ovs service environment file] ******************* skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config/configure-node-settings.yml for node02 TASK [openshift_node : Configure Node settings] ******************************** ok: [node02] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS=--loglevel=2 '}) ok: [node02] => (item={u'regex': u'^CONFIG_FILE=', u'line': u'CONFIG_FILE=/etc/origin/node/node-config.yaml'}) ok: [node02] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.9.0'}) TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config/configure-proxy-settings.yml for node02 TASK [openshift_node : Configure Proxy Settings] ******************************* skipping: [node02] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) skipping: [node02] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [node02] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=[],172.30.0.0/16,10.128.0.0/14'}) TASK [openshift_node : Pull container images] ********************************** skipping: [node02] TASK [openshift_node : Start and enable openvswitch service] ******************* skipping: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : file] *************************************************** skipping: [node02] TASK [openshift_node : Create the Node config] ********************************* changed: [node02] TASK [openshift_node : Configure Node Environment Variables] ******************* TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [node02] => (item=None) skipping: [node02] => (item=None) TASK [openshift_node : Wait for master API to become available before proceeding] *** skipping: [node02] TASK [openshift_node : Start and enable node dep] ****************************** skipping: [node02] TASK [openshift_node : Start and enable node] ********************************** ok: [node02] TASK [openshift_node : Dump logs from node service if it failed] *************** skipping: [node02] TASK [openshift_node : Abort if node failed to start] ************************** skipping: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : NFS storage plugin configuration] *********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/nfs.yml for node02 TASK [openshift_node : Install NFS storage plugin dependencies] **************** ok: [node02] TASK [openshift_node : Check for existence of nfs sebooleans] ****************** ok: [node02] => (item=virt_use_nfs) ok: [node02] => (item=virt_sandbox_use_nfs) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:30.745317', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.214893', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:30.530424', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:32.197427', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.018837', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:32.178590', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:30.745317', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.214893', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:30.530424', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:32.197427', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.018837', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:32.178590', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : GlusterFS storage plugin configuration] ***************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/glusterfs.yml for node02 TASK [openshift_node : Install GlusterFS storage plugin dependencies] ********** ok: [node02] TASK [openshift_node : Check for existence of fusefs sebooleans] *************** ok: [node02] => (item=virt_use_fusefs) ok: [node02] => (item=virt_sandbox_use_fusefs) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:39.767786', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.019696', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:39.748090', '_ansible_ignore_errors': None, 'failed': False}) ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:41.225914', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.020851', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:41.205063', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:39.767786', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.019696', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:39.748090', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-05-01 12:06:41.225914', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.020851', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-05-01 12:06:41.205063', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Ceph storage plugin configuration] ********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/ceph.yml for node02 TASK [openshift_node : Install Ceph storage plugin dependencies] *************** ok: [node02] TASK [openshift_node : iSCSI storage plugin configuration] ********************* included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/storage_plugins/iscsi.yml for node02 TASK [openshift_node : Install iSCSI storage plugin dependencies] ************** ok: [node02] => (item=iscsi-initiator-utils) ok: [node02] => (item=device-mapper-multipath) TASK [openshift_node : restart services] *************************************** ok: [node02] => (item=multipathd) ok: [node02] => (item=rpcbind) TASK [openshift_node : Template multipath configuration] *********************** changed: [node02] TASK [openshift_node : Enable multipath] *************************************** changed: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_node/tasks/config/workaround-bz1331590-ovs-oom-fix.yml for node02 TASK [openshift_node : Create OpenvSwitch service.d directory] ***************** ok: [node02] TASK [openshift_node : Install OpenvSwitch service OOM fix] ******************** ok: [node02] TASK [tuned : Check for tuned package] ***************************************** ok: [node02] TASK [tuned : Set tuned OpenShift variables] *********************************** ok: [node02] TASK [tuned : Ensure directory structure exists] ******************************* ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524731898.0952866, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524731898.0952866, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524731898.0942867, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524731898.0942867, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524731898.0942867, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524731898.0942867, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0952866, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0952866, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0942867, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0942867, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) TASK [tuned : Ensure files are populated from templates] *********************** skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524731898.0952866, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524731898.0952866, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524731898.0942867, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524731898.0942867, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'state': 'directory', 'ctime': 1524731898.0942867, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1524731898.0942867, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0952866, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0952866, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0942867, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) ok: [node02] => (item={'src': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'system_u', 'serole': 'object_r', 'ctime': 1524731898.0942867, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1523891799.0, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/usr/share/ansible/openshift-ansible/roles/tuned/templates', 'setype': 'usr_t'}) TASK [tuned : Make tuned use the recommended tuned profile on restart] ********* changed: [node02] => (item=/etc/tuned/active_profile) ok: [node02] => (item=/etc/tuned/profile_mode) TASK [tuned : Restart tuned service] ******************************************* changed: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Install logrotate] ******* ok: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Setup logrotate.d scripts] *** RUNNING HANDLER [openshift_node : restart node] ******************************** changed: [node02] PLAY [create additional node network plugin groups] **************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_flannel [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_calico [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_contiv [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_kuryr PLAY [etcd_client node config] ************************************************* skipping: no hosts matched PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** skipping: no hosts matched [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_nuage PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Configure Contiv masters] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node01] PLAY [Configure rest of Contiv nodes] ****************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] PLAY [Configure Kuryr node] **************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_manage_node : Wait for master API to become available before proceeding] *** skipping: [node02] TASK [openshift_manage_node : Wait for Node Registration] ********************** ok: [node02 -> node01] TASK [openshift_manage_node : include_tasks] *********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_manage_node/tasks/config.yml for node02 TASK [openshift_manage_node : Set node schedulability] ************************* ok: [node02 -> node01] TASK [openshift_manage_node : Label nodes] ************************************* changed: [node02 -> node01] TASK [Create group for deployment type] **************************************** ok: [node02] PLAY [Re-enable excluder if it was previously enabled] ************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": true } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": true } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/enable.yml for node02 TASK [openshift_excluder : Install excluders] ********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /usr/share/ansible/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** changed: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** changed: [node02] PLAY [Node Install Checkpoint End] ********************************************* TASK [Set Node install 'Complete'] ********************************************* ok: [node01] PLAY RECAP ********************************************************************* localhost : ok=25 changed=0 unreachable=0 failed=0 node01 : ok=42 changed=0 unreachable=0 failed=0 node02 : ok=208 changed=28 unreachable=0 failed=0 INSTALLER STATUS *************************************************************** Initialization : Complete (0:01:46) Node Install : Complete (0:05:05) crio=true path: /etc/crio/crio.conf path: /etc/crio/crio.conf when: crio ansible-playbook -i $inventory_file post_deployment_configuration --extra-vars="crio=${crio}" 2018/05/01 12:08:37 Waiting for host: 192.168.66.101:22 2018/05/01 12:08:37 Connected to tcp://192.168.66.101:22 2018/05/01 12:08:38 Waiting for host: 192.168.66.101:22 2018/05/01 12:08:38 Connected to tcp://192.168.66.101:22 Warning: Permanently added '[127.0.0.1]:34929' (ECDSA) to the list of known hosts. Warning: Permanently added '[127.0.0.1]:34929' (ECDSA) to the list of known hosts. Cluster "node01:8443" set. Cluster "node01:8443" set. ++ kubectl get nodes --no-headers ++ cluster/kubectl.sh get nodes --no-headers ++ grep -v Ready + '[' -n '' ']' + echo 'Nodes are ready:' Nodes are ready: + kubectl get nodes + cluster/kubectl.sh get nodes NAME STATUS ROLES AGE VERSION node01 Ready master 5d v1.9.1+a0ce1bc657 node02 Ready 1m v1.9.1+a0ce1bc657 + make cluster-sync ./cluster/build.sh Building ... sha256:3af4a6dc0b8e534faf514a9ece4d19bd3ceaf57a714488488273ef4db938d9a4 go version go1.10 linux/amd64 go version go1.10 linux/amd64 make[1]: Entering directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt' hack/dockerized "./hack/check.sh && KUBEVIRT_VERSION= ./hack/build-go.sh install " sha256:3af4a6dc0b8e534faf514a9ece4d19bd3ceaf57a714488488273ef4db938d9a4 go version go1.10 linux/amd64 go version go1.10 linux/amd64 Compiling tests... compiled tests.test hack/build-docker.sh build sending incremental file list ./ Dockerfile kubernetes.repo sent 854 bytes received 53 bytes 604.67 bytes/sec total size is 1167 speedup is 1.29 Sending build context to Docker daemon 36.12 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 0c81c3a7ddef Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-controller ---> Using cache ---> 26d08b2f873c Step 4/8 : WORKDIR /home/virt-controller ---> Using cache ---> d73a973f334b Step 5/8 : USER 1001 ---> Using cache ---> 92d14ee3fb2e Step 6/8 : COPY virt-controller /virt-controller ---> 525eaea7232c Removing intermediate container ca46d18ab076 Step 7/8 : ENTRYPOINT /virt-controller ---> Running in c403a332606b ---> 5b937b801038 Removing intermediate container c403a332606b Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-controller" '' ---> Running in da4d8283d923 ---> eb6b08f5b855 Removing intermediate container da4d8283d923 Successfully built eb6b08f5b855 sending incremental file list ./ Dockerfile entrypoint.sh kubevirt-sudo libvirtd.sh sh.sh sock-connector sent 3502 bytes received 129 bytes 7262.00 bytes/sec total size is 5953 speedup is 1.64 Sending build context to Docker daemon 38.06 MB Step 1/14 : FROM kubevirt/libvirt:3.7.0 ---> 60c80c8f7523 Step 2/14 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> fdd57f83e446 Step 3/14 : RUN dnf -y install socat genisoimage util-linux libcgroup-tools ethtool sudo && dnf -y clean all && test $(id -u qemu) = 107 # make sure that the qemu user really is 107 ---> Using cache ---> b824b882c94e Step 4/14 : COPY sock-connector /sock-connector ---> Using cache ---> 8cbe8006a6c1 Step 5/14 : COPY sh.sh /sh.sh ---> Using cache ---> 2e47a3c4a4f3 Step 6/14 : COPY virt-launcher /virt-launcher ---> 969d96e0297e Removing intermediate container 3e1c8e0a15f9 Step 7/14 : COPY kubevirt-sudo /etc/sudoers.d/kubevirt ---> 87625ec6f38b Removing intermediate container 73e58bff905a Step 8/14 : RUN chmod 0640 /etc/sudoers.d/kubevirt ---> Running in a63e1aa1d953  ---> 3945a2b29cd7 Removing intermediate container a63e1aa1d953 Step 9/14 : RUN rm -f /libvirtd.sh ---> Running in fc924ada7793  ---> c6642b05380e Removing intermediate container fc924ada7793 Step 10/14 : COPY libvirtd.sh /libvirtd.sh ---> 0c390344f4cf Removing intermediate container 6126bbc7ef98 Step 11/14 : RUN chmod a+x /libvirtd.sh ---> Running in ca873e65fa00  ---> 3de787dd5111 Removing intermediate container ca873e65fa00 Step 12/14 : COPY entrypoint.sh /entrypoint.sh ---> 12ee0b38026d Removing intermediate container 54e5d98e758a Step 13/14 : ENTRYPOINT /entrypoint.sh ---> Running in 53dc534ab1ad ---> 42ffae0b657b Removing intermediate container 53dc534ab1ad Step 14/14 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-launcher" '' ---> Running in e1b549a2dcb2 ---> b496a2d26c6d Removing intermediate container e1b549a2dcb2 Successfully built b496a2d26c6d sending incremental file list ./ Dockerfile sent 585 bytes received 34 bytes 1238.00 bytes/sec total size is 775 speedup is 1.25 Sending build context to Docker daemon 36.68 MB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 0c81c3a7ddef Step 3/5 : COPY virt-handler /virt-handler ---> 833dcf9ec5de Removing intermediate container df6d97a48c45 Step 4/5 : ENTRYPOINT /virt-handler ---> Running in b3f32825bc99 ---> c1adcdf67d2e Removing intermediate container b3f32825bc99 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-handler" '' ---> Running in efb3b2e21f9f ---> 6701c21d857f Removing intermediate container efb3b2e21f9f Successfully built 6701c21d857f sending incremental file list ./ Dockerfile sent 646 bytes received 34 bytes 1360.00 bytes/sec total size is 876 speedup is 1.29 Sending build context to Docker daemon 36.81 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 0c81c3a7ddef Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-api ---> Using cache ---> 1fa3abd9386c Step 4/8 : WORKDIR /home/virt-api ---> Using cache ---> 18a3eb497210 Step 5/8 : USER 1001 ---> Using cache ---> a0af12a5aa54 Step 6/8 : COPY virt-api /virt-api ---> 797cab2823b6 Removing intermediate container e9974e502137 Step 7/8 : ENTRYPOINT /virt-api ---> Running in 84ea685f580b ---> 831f1e7da4d1 Removing intermediate container 84ea685f580b Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "virt-api" '' ---> Running in 5a5a9e592ee6 ---> b60ddac04453 Removing intermediate container 5a5a9e592ee6 Successfully built b60ddac04453 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/iscsi-demo-target-tgtd ./ Dockerfile run-tgt.sh sent 2185 bytes received 53 bytes 4476.00 bytes/sec total size is 3992 speedup is 1.78 Sending build context to Docker daemon 6.656 kB Step 1/10 : FROM fedora:27 ---> 9110ae7f579f Step 2/10 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 0c81c3a7ddef Step 3/10 : ENV container docker ---> Using cache ---> d0b0dc01cb5d Step 4/10 : RUN dnf -y install scsi-target-utils bzip2 e2fsprogs ---> Using cache ---> 35c00214c275 Step 5/10 : RUN mkdir -p /images ---> Using cache ---> e3e179183ea6 Step 6/10 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /images/1-alpine.img ---> Using cache ---> e86b61826c05 Step 7/10 : ADD run-tgt.sh / ---> Using cache ---> db2dc53efd9e Step 8/10 : EXPOSE 3260 ---> Using cache ---> f2767bc543c9 Step 9/10 : CMD /run-tgt.sh ---> Using cache ---> c066b080f396 Step 10/10 : LABEL "iscsi-demo-target-tgtd" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Using cache ---> 86906a32123c Successfully built 86906a32123c sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/vm-killer ./ Dockerfile sent 610 bytes received 34 bytes 1288.00 bytes/sec total size is 797 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 0c81c3a7ddef Step 3/5 : ENV container docker ---> Using cache ---> d0b0dc01cb5d Step 4/5 : RUN dnf -y install procps-ng nmap-ncat && dnf -y clean all ---> Using cache ---> 1e46adaec886 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "vm-killer" '' ---> Using cache ---> 21a035599556 Successfully built 21a035599556 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/registry-disk-v1alpha ./ Dockerfile entry-point.sh sent 1566 bytes received 53 bytes 3238.00 bytes/sec total size is 2542 speedup is 1.57 Sending build context to Docker daemon 5.12 kB Step 1/7 : FROM debian:sid ---> bcec0ae8107e Step 2/7 : MAINTAINER "David Vossel" \ ---> Using cache ---> f20a31819d03 Step 3/7 : ENV container docker ---> Using cache ---> 96277a0619bb Step 4/7 : RUN apt-get update && apt-get install -y bash curl bzip2 qemu-utils && mkdir -p /disk && rm -rf /var/lib/apt/lists/* ---> Using cache ---> f43fc40caf89 Step 5/7 : ADD entry-point.sh / ---> Using cache ---> d5f7d14e890a Step 6/7 : CMD /entry-point.sh ---> Using cache ---> ac53f48bd5a2 Step 7/7 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "registry-disk-v1alpha" '' ---> Using cache ---> 35e2a12ef403 Successfully built 35e2a12ef403 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/cirros-registry-disk-demo ./ Dockerfile sent 630 bytes received 34 bytes 1328.00 bytes/sec total size is 825 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:34928/kubevirt/registry-disk-v1alpha:devel ---> 35e2a12ef403 Step 2/4 : MAINTAINER "David Vossel" \ ---> Using cache ---> d9af2fb8ae78 Step 3/4 : RUN curl https://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img > /disk/cirros.img ---> Using cache ---> fc2933a2d6ff Step 4/4 : LABEL "cirros-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Using cache ---> 0913ac54c2df Successfully built 0913ac54c2df sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/fedora-cloud-registry-disk-demo ./ Dockerfile sent 677 bytes received 34 bytes 1422.00 bytes/sec total size is 926 speedup is 1.30 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:34928/kubevirt/registry-disk-v1alpha:devel ---> 35e2a12ef403 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 4e9ce334200b Step 3/4 : RUN curl -g -L https://download.fedoraproject.org/pub/fedora/linux/releases/27/CloudImages/x86_64/images/Fedora-Cloud-Base-27-1.6.x86_64.qcow2 > /disk/fedora.qcow2 ---> Using cache ---> 88a5baf0d19a Step 4/4 : LABEL "fedora-cloud-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Using cache ---> 186b120ac06b Successfully built 186b120ac06b sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/alpine-registry-disk-demo ./ Dockerfile sent 639 bytes received 34 bytes 1346.00 bytes/sec total size is 866 speedup is 1.29 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:34928/kubevirt/registry-disk-v1alpha:devel ---> 35e2a12ef403 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 4e9ce334200b Step 3/4 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /disk/alpine.iso ---> Using cache ---> 86e384db7f30 Step 4/4 : LABEL "alpine-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release-crio0" '' ---> Using cache ---> 29130fbd6bc0 Successfully built 29130fbd6bc0 sending incremental file list ./ Dockerfile sent 660 bytes received 34 bytes 1388.00 bytes/sec total size is 918 speedup is 1.32 Sending build context to Docker daemon 33.96 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 0c81c3a7ddef Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virtctl ---> Using cache ---> b2e31dc7e946 Step 4/8 : WORKDIR /home/virtctl ---> Using cache ---> 73d45796737a Step 5/8 : USER 1001 ---> Using cache ---> 1a3751300d87 Step 6/8 : COPY subresource-access-test /subresource-access-test ---> 9df2e6301fa8 Removing intermediate container d3eac8c90414 Step 7/8 : ENTRYPOINT /subresource-access-test ---> Running in ac0d0a0acaab ---> 9eddfce8b3c2 Removing intermediate container ac0d0a0acaab Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "subresource-access-test" '' ---> Running in 77e8450ccd78 ---> d66f803be140 Removing intermediate container 77e8450ccd78 Successfully built d66f803be140 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd/winrmcli ./ Dockerfile sent 773 bytes received 34 bytes 1614.00 bytes/sec total size is 1098 speedup is 1.36 Sending build context to Docker daemon 3.072 kB Step 1/9 : FROM fedora:27 ---> 9110ae7f579f Step 2/9 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 0c81c3a7ddef Step 3/9 : ENV container docker ---> Using cache ---> d0b0dc01cb5d Step 4/9 : RUN dnf -y install make git gcc && dnf -y clean all ---> Using cache ---> 5ed65400bde3 Step 5/9 : ENV GIMME_GO_VERSION 1.9.2 ---> Using cache ---> fa084e1bba91 Step 6/9 : RUN mkdir -p /gimme && curl -sL https://raw.githubusercontent.com/travis-ci/gimme/master/gimme | HOME=/gimme bash >> /etc/profile.d/gimme.sh ---> Using cache ---> 701a1c98b779 Step 7/9 : ENV GOPATH "/go" GOBIN "/usr/bin" ---> Using cache ---> d78f779f0759 Step 8/9 : RUN mkdir -p /go && source /etc/profile.d/gimme.sh && go get github.com/masterzen/winrm-cli ---> Using cache ---> 7ba98a98fa7f Step 9/9 : LABEL "kubevirt-functional-tests-openshift-release-crio0" '' "winrmcli" '' ---> Using cache ---> f56ab807bfc1 Successfully built f56ab807bfc1 hack/build-docker.sh push The push refers to a repository [localhost:34928/kubevirt/virt-controller] e9f45edf51c1: Preparing 2d2b47e1e58b: Preparing 39bae602f753: Preparing 2d2b47e1e58b: Layer already exists 39bae602f753: Layer already exists e9f45edf51c1: Pushed devel: digest: sha256:234610ee9e235de669fd56dca92d1538f1cae216c218dcab32b9ddd46e82a7db size: 948 The push refers to a repository [localhost:34928/kubevirt/virt-launcher] f82a1b49fc9f: Preparing 053afce1cd2f: Preparing 053afce1cd2f: Preparing 7206b01818bf: Preparing d7a4ecf04e3a: Preparing 19c4caedaae0: Preparing 3681043ee1b7: Preparing 569987a574d7: Preparing 9170be750ee2: Preparing 91a2c9242f65: Preparing 530cc55618cd: Preparing 34fa414dfdf6: Preparing a1359dc556dd: Preparing 490c7c373332: Preparing 4b440db36f72: Preparing 39bae602f753: Preparing 569987a574d7: Waiting 9170be750ee2: Waiting 91a2c9242f65: Waiting 530cc55618cd: Waiting a1359dc556dd: Waiting 490c7c373332: Waiting 4b440db36f72: Waiting 39bae602f753: Waiting 34fa414dfdf6: Waiting 3681043ee1b7: Waiting 19c4caedaae0: Pushed d7a4ecf04e3a: Pushed f82a1b49fc9f: Pushed 053afce1cd2f: Pushed 7206b01818bf: Pushed 569987a574d7: Layer already exists 9170be750ee2: Layer already exists 530cc55618cd: Layer already exists 91a2c9242f65: Layer already exists 34fa414dfdf6: Layer already exists a1359dc556dd: Layer already exists 490c7c373332: Layer already exists 39bae602f753: Layer already exists 4b440db36f72: Layer already exists 3681043ee1b7: Pushed devel: digest: sha256:8a04217ff01135e9bef5f94382c9da0985c3cd2595f94d4c8950cf69819041ea size: 3653 The push refers to a repository [localhost:34928/kubevirt/virt-handler] cc647891f892: Preparing 39bae602f753: Preparing 39bae602f753: Layer already exists cc647891f892: Pushed devel: digest: sha256:bc7cfbebdf4fe9a95dda49ebc493e6a6fdcb308bfb17d290db2ea968ba40af5c size: 740 The push refers to a repository [localhost:34928/kubevirt/virt-api] 07ec75301212: Preparing 3b1bf9c72a92: Preparing 39bae602f753: Preparing 39bae602f753: Layer already exists 3b1bf9c72a92: Layer already exists 07ec75301212: Pushed devel: digest: sha256:d843ea871f729ac6e73cdfbcf12713b28f116c62a4a88633fb7d6337e0f50f20 size: 948 The push refers to a repository [localhost:34928/kubevirt/iscsi-demo-target-tgtd] 2927410cd43a: Preparing b121fc13ece8: Preparing 18dd75eb79d2: Preparing 716441edb530: Preparing 39bae602f753: Preparing 39bae602f753: Layer already exists 2927410cd43a: Layer already exists 18dd75eb79d2: Layer already exists b121fc13ece8: Layer already exists 716441edb530: Layer already exists devel: digest: sha256:d66839b7515fd5b18fff92d18a2875e4424712bb960b4cf6e2f8f182ddd0c140 size: 1368 The push refers to a repository [localhost:34928/kubevirt/vm-killer] 3ad88ce8cfdd: Preparing 39bae602f753: Preparing 39bae602f753: Layer already exists 3ad88ce8cfdd: Layer already exists devel: digest: sha256:6ed7e868106342352487a5287517c3eb2b7622c4076f33789557190683525026 size: 740 The push refers to a repository [localhost:34928/kubevirt/registry-disk-v1alpha] ecd320fbeae1: Preparing a87a1c350b94: Preparing 6709b2da72b8: Preparing a87a1c350b94: Layer already exists ecd320fbeae1: Layer already exists 6709b2da72b8: Layer already exists devel: digest: sha256:f4f776457bb632c1f41db5da5707b8c81d806ec3dacf15e1f06bca5dbcf08c1b size: 948 The push refers to a repository [localhost:34928/kubevirt/cirros-registry-disk-demo] a9f7bc6d597f: Preparing ecd320fbeae1: Preparing a87a1c350b94: Preparing 6709b2da72b8: Preparing a87a1c350b94: Layer already exists ecd320fbeae1: Layer already exists a9f7bc6d597f: Layer already exists 6709b2da72b8: Layer already exists devel: digest: sha256:6a519bc706c6feb4269ad0177d83b349185531186174dfd3d10edadc75396690 size: 1160 The push refers to a repository [localhost:34928/kubevirt/fedora-cloud-registry-disk-demo] ffda1b2b1e73: Preparing ecd320fbeae1: Preparing a87a1c350b94: Preparing 6709b2da72b8: Preparing ecd320fbeae1: Layer already exists 6709b2da72b8: Layer already exists a87a1c350b94: Layer already exists ffda1b2b1e73: Layer already exists devel: digest: sha256:01d1b8fd8233a3db69a255ea655fa7b1c9762891f11c270a662193b38e7a78b1 size: 1161 The push refers to a repository [localhost:34928/kubevirt/alpine-registry-disk-demo] 1aea729bebb7: Preparing ecd320fbeae1: Preparing a87a1c350b94: Preparing 6709b2da72b8: Preparing 6709b2da72b8: Layer already exists ecd320fbeae1: Layer already exists a87a1c350b94: Layer already exists 1aea729bebb7: Layer already exists devel: digest: sha256:72d640eaa86390b04784e462f574f5b5aa902e8373049cf051c1201f34e8c5c6 size: 1160 The push refers to a repository [localhost:34928/kubevirt/subresource-access-test] dd29315b2f9b: Preparing cddcea6287d1: Preparing 39bae602f753: Preparing 39bae602f753: Layer already exists cddcea6287d1: Layer already exists dd29315b2f9b: Pushed devel: digest: sha256:159a86871ca5a91a92f19675a0fd7e56a42b320703a7135a4b70631929876db2 size: 948 The push refers to a repository [localhost:34928/kubevirt/winrmcli] e247ea19f4a7: Preparing b9a86fe2f8d2: Preparing c8eb97d247eb: Preparing 39bae602f753: Preparing 39bae602f753: Layer already exists b9a86fe2f8d2: Layer already exists e247ea19f4a7: Layer already exists c8eb97d247eb: Layer already exists devel: digest: sha256:8491ffdace02490cc30e74f985f4792317e3850b846ab38257ddea93436a73cb size: 1165 make[1]: Leaving directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt' 2018/05/01 12:11:43 Waiting for host: 192.168.66.101:22 2018/05/01 12:11:43 Connected to tcp://192.168.66.101:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer 15a3c1110beb: Pulling fs layer f8ea559eabed: Pulling fs layer 15a3c1110beb: Verifying Checksum 15a3c1110beb: Download complete f8ea559eabed: Verifying Checksum f8ea559eabed: Download complete 2176639d844b: Verifying Checksum 2176639d844b: Download complete 2176639d844b: Pull complete 15a3c1110beb: Pull complete f8ea559eabed: Pull complete Digest: sha256:234610ee9e235de669fd56dca92d1538f1cae216c218dcab32b9ddd46e82a7db Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer 1475fa3276a7: Pulling fs layer cdad54148243: Pulling fs layer 42cab4ea66c5: Pulling fs layer 4a804ac9ad5f: Pulling fs layer a9fd8252f2e8: Pulling fs layer 145b389094c4: Pulling fs layer bf066802a1da: Pulling fs layer acfbc8319e73: Pulling fs layer da9255053d01: Pulling fs layer a1e80189bea5: Waiting 6cc174edcebf: Waiting 1475fa3276a7: Waiting cdad54148243: Waiting 42cab4ea66c5: Waiting 4a804ac9ad5f: Waiting a9fd8252f2e8: Waiting 145b389094c4: Waiting bf066802a1da: Waiting acfbc8319e73: Waiting da9255053d01: Waiting a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete f2ef945504a7: Verifying Checksum f2ef945504a7: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete cdad54148243: Verifying Checksum cdad54148243: Download complete 42cab4ea66c5: Verifying Checksum 42cab4ea66c5: Download complete 4a804ac9ad5f: Verifying Checksum 4a804ac9ad5f: Download complete 1475fa3276a7: Verifying Checksum 1475fa3276a7: Download complete a9fd8252f2e8: Verifying Checksum a9fd8252f2e8: Download complete bf066802a1da: Verifying Checksum bf066802a1da: Download complete 145b389094c4: Verifying Checksum 145b389094c4: Download complete acfbc8319e73: Verifying Checksum acfbc8319e73: Download complete da9255053d01: Verifying Checksum da9255053d01: Download complete d7240bccd145: Verifying Checksum d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete 1475fa3276a7: Pull complete cdad54148243: Pull complete 42cab4ea66c5: Pull complete 4a804ac9ad5f: Pull complete a9fd8252f2e8: Pull complete 145b389094c4: Pull complete bf066802a1da: Pull complete acfbc8319e73: Pull complete da9255053d01: Pull complete Digest: sha256:8a04217ff01135e9bef5f94382c9da0985c3cd2595f94d4c8950cf69819041ea Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists f70f8cb05fab: Pulling fs layer f70f8cb05fab: Verifying Checksum f70f8cb05fab: Download complete f70f8cb05fab: Pull complete Digest: sha256:bc7cfbebdf4fe9a95dda49ebc493e6a6fdcb308bfb17d290db2ea968ba40af5c Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists ef1a3b124206: Pulling fs layer 52dbeaada58b: Pulling fs layer ef1a3b124206: Verifying Checksum ef1a3b124206: Download complete 52dbeaada58b: Verifying Checksum 52dbeaada58b: Download complete ef1a3b124206: Pull complete 52dbeaada58b: Pull complete Digest: sha256:d843ea871f729ac6e73cdfbcf12713b28f116c62a4a88633fb7d6337e0f50f20 Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists c81a49dbe8f3: Pulling fs layer 064679d2dcf5: Pulling fs layer bcaf437d3487: Pulling fs layer de7654dd0183: Pulling fs layer de7654dd0183: Waiting 064679d2dcf5: Verifying Checksum 064679d2dcf5: Download complete de7654dd0183: Verifying Checksum de7654dd0183: Download complete bcaf437d3487: Verifying Checksum bcaf437d3487: Download complete c81a49dbe8f3: Download complete c81a49dbe8f3: Pull complete 064679d2dcf5: Pull complete bcaf437d3487: Pull complete de7654dd0183: Pull complete Digest: sha256:d66839b7515fd5b18fff92d18a2875e4424712bb960b4cf6e2f8f182ddd0c140 Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists 8fc7bdcd72e5: Pulling fs layer 8fc7bdcd72e5: Verifying Checksum 8fc7bdcd72e5: Download complete 8fc7bdcd72e5: Pull complete Digest: sha256:6ed7e868106342352487a5287517c3eb2b7622c4076f33789557190683525026 Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer 82da67e25ceb: Pulling fs layer b0e6b7e8a28f: Pulling fs layer b0e6b7e8a28f: Verifying Checksum b0e6b7e8a28f: Download complete 82da67e25ceb: Verifying Checksum 82da67e25ceb: Download complete 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 2115d46e7396: Pull complete 82da67e25ceb: Pull complete b0e6b7e8a28f: Pull complete Digest: sha256:f4f776457bb632c1f41db5da5707b8c81d806ec3dacf15e1f06bca5dbcf08c1b Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists 82da67e25ceb: Already exists b0e6b7e8a28f: Already exists 283c2ac1f971: Pulling fs layer 283c2ac1f971: Verifying Checksum 283c2ac1f971: Download complete 283c2ac1f971: Pull complete Digest: sha256:6a519bc706c6feb4269ad0177d83b349185531186174dfd3d10edadc75396690 Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists 82da67e25ceb: Already exists b0e6b7e8a28f: Already exists 61ba3cfff284: Pulling fs layer 61ba3cfff284: Verifying Checksum 61ba3cfff284: Download complete 61ba3cfff284: Pull complete Digest: sha256:01d1b8fd8233a3db69a255ea655fa7b1c9762891f11c270a662193b38e7a78b1 Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists 82da67e25ceb: Already exists b0e6b7e8a28f: Already exists d62591d5818a: Pulling fs layer d62591d5818a: Verifying Checksum d62591d5818a: Download complete d62591d5818a: Pull complete Digest: sha256:72d640eaa86390b04784e462f574f5b5aa902e8373049cf051c1201f34e8c5c6 Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists fdc994a640c8: Pulling fs layer 1bbba30f5c3d: Pulling fs layer fdc994a640c8: Verifying Checksum fdc994a640c8: Download complete 1bbba30f5c3d: Verifying Checksum 1bbba30f5c3d: Download complete fdc994a640c8: Pull complete 1bbba30f5c3d: Pull complete Digest: sha256:159a86871ca5a91a92f19675a0fd7e56a42b320703a7135a4b70631929876db2 Trying to pull repository registry:5000/kubevirt/winrmcli ... devel: Pulling from registry:5000/kubevirt/winrmcli 2176639d844b: Already exists 206ce0e712a7: Pulling fs layer b7c152dd4760: Pulling fs layer 7e51cf11cbfa: Pulling fs layer 7e51cf11cbfa: Verifying Checksum 7e51cf11cbfa: Download complete 206ce0e712a7: Verifying Checksum 206ce0e712a7: Download complete b7c152dd4760: Verifying Checksum b7c152dd4760: Download complete 206ce0e712a7: Pull complete b7c152dd4760: Pull complete 7e51cf11cbfa: Pull complete Digest: sha256:8491ffdace02490cc30e74f985f4792317e3850b846ab38257ddea93436a73cb 2018/05/01 12:14:16 Waiting for host: 192.168.66.101:22 2018/05/01 12:14:16 Connected to tcp://192.168.66.101:22 2018/05/01 12:14:19 Waiting for host: 192.168.66.102:22 2018/05/01 12:14:19 Connected to tcp://192.168.66.102:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer 15a3c1110beb: Pulling fs layer f8ea559eabed: Pulling fs layer 15a3c1110beb: Verifying Checksum 15a3c1110beb: Download complete f8ea559eabed: Verifying Checksum f8ea559eabed: Download complete 2176639d844b: Download complete 2176639d844b: Pull complete 15a3c1110beb: Pull complete f8ea559eabed: Pull complete Digest: sha256:234610ee9e235de669fd56dca92d1538f1cae216c218dcab32b9ddd46e82a7db Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer 1475fa3276a7: Pulling fs layer cdad54148243: Pulling fs layer 42cab4ea66c5: Pulling fs layer 4a804ac9ad5f: Pulling fs layer a9fd8252f2e8: Pulling fs layer 145b389094c4: Pulling fs layer bf066802a1da: Pulling fs layer acfbc8319e73: Pulling fs layer da9255053d01: Pulling fs layer 4a804ac9ad5f: Waiting a9fd8252f2e8: Waiting 145b389094c4: Waiting bf066802a1da: Waiting acfbc8319e73: Waiting da9255053d01: Waiting a1e80189bea5: Waiting 6cc174edcebf: Waiting 1475fa3276a7: Waiting cdad54148243: Waiting 42cab4ea66c5: Waiting a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete f2ef945504a7: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete cdad54148243: Verifying Checksum cdad54148243: Download complete 42cab4ea66c5: Verifying Checksum 42cab4ea66c5: Download complete 4a804ac9ad5f: Verifying Checksum 4a804ac9ad5f: Download complete a9fd8252f2e8: Verifying Checksum a9fd8252f2e8: Download complete 1475fa3276a7: Verifying Checksum 1475fa3276a7: Download complete 145b389094c4: Verifying Checksum 145b389094c4: Download complete bf066802a1da: Verifying Checksum bf066802a1da: Download complete acfbc8319e73: Verifying Checksum acfbc8319e73: Download complete da9255053d01: Verifying Checksum da9255053d01: Download complete d7240bccd145: Verifying Checksum d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete 1475fa3276a7: Pull complete cdad54148243: Pull complete 42cab4ea66c5: Pull complete 4a804ac9ad5f: Pull complete a9fd8252f2e8: Pull complete 145b389094c4: Pull complete bf066802a1da: Pull complete acfbc8319e73: Pull complete da9255053d01: Pull complete Digest: sha256:8a04217ff01135e9bef5f94382c9da0985c3cd2595f94d4c8950cf69819041ea Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists f70f8cb05fab: Pulling fs layer f70f8cb05fab: Download complete f70f8cb05fab: Pull complete Digest: sha256:bc7cfbebdf4fe9a95dda49ebc493e6a6fdcb308bfb17d290db2ea968ba40af5c Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists ef1a3b124206: Pulling fs layer 52dbeaada58b: Pulling fs layer ef1a3b124206: Verifying Checksum ef1a3b124206: Download complete 52dbeaada58b: Verifying Checksum 52dbeaada58b: Download complete ef1a3b124206: Pull complete 52dbeaada58b: Pull complete Digest: sha256:d843ea871f729ac6e73cdfbcf12713b28f116c62a4a88633fb7d6337e0f50f20 Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists c81a49dbe8f3: Pulling fs layer 064679d2dcf5: Pulling fs layer bcaf437d3487: Pulling fs layer de7654dd0183: Pulling fs layer de7654dd0183: Waiting 064679d2dcf5: Download complete de7654dd0183: Verifying Checksum de7654dd0183: Download complete bcaf437d3487: Verifying Checksum bcaf437d3487: Download complete c81a49dbe8f3: Verifying Checksum c81a49dbe8f3: Download complete c81a49dbe8f3: Pull complete 064679d2dcf5: Pull complete bcaf437d3487: Pull complete de7654dd0183: Pull complete Digest: sha256:d66839b7515fd5b18fff92d18a2875e4424712bb960b4cf6e2f8f182ddd0c140 Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists 8fc7bdcd72e5: Pulling fs layer 8fc7bdcd72e5: Verifying Checksum 8fc7bdcd72e5: Download complete 8fc7bdcd72e5: Pull complete Digest: sha256:6ed7e868106342352487a5287517c3eb2b7622c4076f33789557190683525026 Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer 82da67e25ceb: Pulling fs layer b0e6b7e8a28f: Pulling fs layer b0e6b7e8a28f: Verifying Checksum b0e6b7e8a28f: Download complete 82da67e25ceb: Verifying Checksum 82da67e25ceb: Download complete 2115d46e7396: Download complete 2115d46e7396: Pull complete 82da67e25ceb: Pull complete b0e6b7e8a28f: Pull complete Digest: sha256:f4f776457bb632c1f41db5da5707b8c81d806ec3dacf15e1f06bca5dbcf08c1b Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists 82da67e25ceb: Already exists b0e6b7e8a28f: Already exists 283c2ac1f971: Pulling fs layer 283c2ac1f971: Verifying Checksum 283c2ac1f971: Download complete 283c2ac1f971: Pull complete Digest: sha256:6a519bc706c6feb4269ad0177d83b349185531186174dfd3d10edadc75396690 Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists 82da67e25ceb: Already exists b0e6b7e8a28f: Already exists 61ba3cfff284: Pulling fs layer 61ba3cfff284: Verifying Checksum 61ba3cfff284: Download complete 61ba3cfff284: Pull complete Digest: sha256:01d1b8fd8233a3db69a255ea655fa7b1c9762891f11c270a662193b38e7a78b1 Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists 82da67e25ceb: Already exists b0e6b7e8a28f: Already exists d62591d5818a: Pulling fs layer d62591d5818a: Verifying Checksum d62591d5818a: Download complete d62591d5818a: Pull complete Digest: sha256:72d640eaa86390b04784e462f574f5b5aa902e8373049cf051c1201f34e8c5c6 Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists fdc994a640c8: Pulling fs layer 1bbba30f5c3d: Pulling fs layer fdc994a640c8: Download complete 1bbba30f5c3d: Verifying Checksum 1bbba30f5c3d: Download complete fdc994a640c8: Pull complete 1bbba30f5c3d: Pull complete Digest: sha256:159a86871ca5a91a92f19675a0fd7e56a42b320703a7135a4b70631929876db2 Trying to pull repository registry:5000/kubevirt/winrmcli ... devel: Pulling from registry:5000/kubevirt/winrmcli 2176639d844b: Already exists 206ce0e712a7: Pulling fs layer b7c152dd4760: Pulling fs layer 7e51cf11cbfa: Pulling fs layer 7e51cf11cbfa: Verifying Checksum 7e51cf11cbfa: Download complete 206ce0e712a7: Verifying Checksum 206ce0e712a7: Download complete b7c152dd4760: Download complete 206ce0e712a7: Pull complete b7c152dd4760: Pull complete 7e51cf11cbfa: Pull complete Digest: sha256:8491ffdace02490cc30e74f985f4792317e3850b846ab38257ddea93436a73cb 2018/05/01 12:16:38 Waiting for host: 192.168.66.102:22 2018/05/01 12:16:38 Connected to tcp://192.168.66.102:22 Done ./cluster/clean.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0-crio ++ provider_prefix=kubevirt-functional-tests-openshift-release-crio0 ++ job_prefix=kubevirt-functional-tests-openshift-release-crio0 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.5.0-alpha.1-24-g72f8a4d ++ KUBEVIRT_VERSION=v0.5.0-alpha.1-24-g72f8a4d + source cluster/os-3.9.0-crio/provider.sh ++ set -e ++ source cluster/os-3.9.0/provider.sh +++ set -e +++ image=os-3.9.0@sha256:69f9522fbbc7415d151d2b8f1733ab90f3efe70182b749a94cdf3f2c8694cb64 +++ source cluster/ephemeral-provider-common.sh ++++ set -e ++++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' ++ image=os-3.9.0-crio@sha256:77e34d979bba9821189d8879c9fbb3a500d16058fc5ff0d015f79c18cbf0d258 + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0-crio ++ source hack/config-default.sh source hack/config-os-3.9.0-crio.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0-crio.sh ++ source hack/config-provider-os-3.9.0-crio.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-crio/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-crio/.kubectl +++ docker_prefix=localhost:34928/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Cleaning up ...' Cleaning up ... + cluster/kubectl.sh get vms --all-namespaces -o=custom-columns=NAME:.metadata.name,NAMESPACE:.metadata.namespace,FINALIZERS:.metadata.finalizers --no-headers + grep foregroundDeleteVirtualMachine + read p the server doesn't have a resource type "vms" + _kubectl delete ds -l kubevirt.io -n kube-system --cascade=false --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=libvirt --force --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=virt-handler --force --grace-period 0 No resources found + namespaces=(default ${namespace}) + for i in '${namespaces[@]}' + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete deployment -l kubevirt.io No resources found + _kubectl -n default delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete rs -l kubevirt.io No resources found + _kubectl -n default delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete services -l kubevirt.io No resources found + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n default delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete secrets -l kubevirt.io No resources found + _kubectl -n default delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete pv -l kubevirt.io No resources found + _kubectl -n default delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete pvc -l kubevirt.io No resources found + _kubectl -n default delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete ds -l kubevirt.io No resources found + _kubectl -n default delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n default delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete pods -l kubevirt.io No resources found + _kubectl -n default delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n default delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete rolebinding -l kubevirt.io No resources found + _kubectl -n default delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete roles -l kubevirt.io No resources found + _kubectl -n default delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete clusterroles -l kubevirt.io No resources found + _kubectl -n default delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n default delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n default get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig ++ cluster/os-3.9.0-crio/.kubectl -n default get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + for i in '${namespaces[@]}' + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete deployment -l kubevirt.io No resources found + _kubectl -n kube-system delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete rs -l kubevirt.io No resources found + _kubectl -n kube-system delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete services -l kubevirt.io No resources found + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete validatingwebhookconfiguration -l kubevirt.io No resources found + _kubectl -n kube-system delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete secrets -l kubevirt.io No resources found + _kubectl -n kube-system delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete pv -l kubevirt.io No resources found + _kubectl -n kube-system delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete pvc -l kubevirt.io No resources found + _kubectl -n kube-system delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete ds -l kubevirt.io No resources found + _kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n kube-system delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete pods -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete rolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete roles -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete clusterroles -l kubevirt.io No resources found + _kubectl -n kube-system delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl -n kube-system delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig ++ cluster/os-3.9.0-crio/.kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + sleep 2 + echo Done Done ./cluster/deploy.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0-crio ++ provider_prefix=kubevirt-functional-tests-openshift-release-crio0 ++ job_prefix=kubevirt-functional-tests-openshift-release-crio0 +++ kubevirt_version +++ '[' -n '' ']' +++ '[' -d /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/.git ']' ++++ git describe --always --tags +++ echo v0.5.0-alpha.1-24-g72f8a4d ++ KUBEVIRT_VERSION=v0.5.0-alpha.1-24-g72f8a4d + source cluster/os-3.9.0-crio/provider.sh ++ set -e ++ source cluster/os-3.9.0/provider.sh +++ set -e +++ image=os-3.9.0@sha256:69f9522fbbc7415d151d2b8f1733ab90f3efe70182b749a94cdf3f2c8694cb64 +++ source cluster/ephemeral-provider-common.sh ++++ set -e ++++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' ++ image=os-3.9.0-crio@sha256:77e34d979bba9821189d8879c9fbb3a500d16058fc5ff0d015f79c18cbf0d258 + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0-crio ++ source hack/config-default.sh source hack/config-os-3.9.0-crio.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0-crio.sh ++ source hack/config-provider-os-3.9.0-crio.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-crio/.kubeconfig +++ kubectl=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-crio/.kubectl +++ docker_prefix=localhost:34928/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Deploying ...' Deploying ... + [[ -z openshift-release-crio ]] + [[ openshift-release-crio =~ .*-dev ]] + [[ openshift-release-crio =~ .*-release ]] + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/demo-content.yaml =~ .*demo.* ]] + continue + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml =~ .*demo.* ]] + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml serviceaccount "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver-auth-delegator" created rolebinding "kubevirt-apiserver" created role "kubevirt-apiserver" created clusterrole "kubevirt-apiserver" created clusterrole "kubevirt-controller" created serviceaccount "kubevirt-controller" created serviceaccount "kubevirt-privileged" created clusterrolebinding "kubevirt-controller" created clusterrolebinding "kubevirt-controller-cluster-admin" created clusterrolebinding "kubevirt-privileged-cluster-admin" created service "virt-api" created deployment "virt-api" created deployment "virt-controller" created daemonset "virt-handler" created customresourcedefinition "virtualmachines.kubevirt.io" created customresourcedefinition "virtualmachinereplicasets.kubevirt.io" created customresourcedefinition "virtualmachinepresets.kubevirt.io" created customresourcedefinition "offlinevirtualmachines.kubevirt.io" created + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release-crio/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R persistentvolumeclaim "disk-alpine" created persistentvolume "iscsi-disk-alpine" created persistentvolumeclaim "disk-custom" created persistentvolume "iscsi-disk-custom" created daemonset "iscsi-demo-target-tgtd" created serviceaccount "kubevirt-testing" created clusterrolebinding "kubevirt-testing-cluster-admin" created + '[' os-3.9.0-crio = vagrant-openshift ']' + [[ os-3.9.0-crio =~ os-3.9.0.* ]] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-controller"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-testing"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-privileged"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-apiserver"] + _kubectl adm policy add-scc-to-user privileged admin + export KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-crio/.kubeconfig + cluster/os-3.9.0-crio/.kubectl adm policy add-scc-to-user privileged admin scc "privileged" added to: ["admin"] + echo Done Done ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 3s iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 3s virt-api-fd96f94b5-86pdc 0/1 ContainerCreating 0 5s virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 5s virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 5s virt-controller-5f7c946cc4-tpt5x 0/1 ContainerCreating 0 5s virt-handler-6gcqv 0/1 ContainerCreating 0 2s virt-handler-wr72t 0/1 ContainerCreating 0 2s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 4s iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4s virt-api-fd96f94b5-86pdc 0/1 ContainerCreating 0 6s virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 6s virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 6s virt-controller-5f7c946cc4-tpt5x 0/1 ContainerCreating 0 6s virt-handler-6gcqv 0/1 ContainerCreating 0 3s virt-handler-wr72t 0/1 ContainerCreating 0 3s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 15s iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 15s virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 17s virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 17s virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 17s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 17s virt-handler-6gcqv 0/1 ImagePullBackOff 0 14s virt-handler-wr72t 0/1 ImagePullBackOff 0 14s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 15s iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 15s virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 17s virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 17s virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 17s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 17s virt-handler-6gcqv 0/1 ImagePullBackOff 0 14s virt-handler-wr72t 0/1 ImagePullBackOff 0 14s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 26s iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 26s virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 28s virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 28s virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 28s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 28s virt-handler-6gcqv 0/1 ImagePullBackOff 0 25s virt-handler-wr72t 0/1 ImagePullBackOff 0 25s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 27s iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 27s virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 29s virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 29s virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 29s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 29s virt-handler-6gcqv 0/1 ImagePullBackOff 0 26s virt-handler-wr72t 0/1 ImagePullBackOff 0 26s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 37s iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 37s virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 39s virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 39s virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 39s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 39s virt-handler-6gcqv 0/1 ImagePullBackOff 0 36s virt-handler-wr72t 0/1 ImagePullBackOff 0 36s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 38s iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 38s virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 40s virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 40s virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 40s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 40s virt-handler-6gcqv 0/1 ImagePullBackOff 0 37s virt-handler-wr72t 0/1 ErrImagePull 0 37s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 48s iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 48s virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 50s virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 50s virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 50s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 50s virt-handler-6gcqv 0/1 ImagePullBackOff 0 47s virt-handler-wr72t 0/1 ImagePullBackOff 0 47s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 49s iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 49s virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 51s virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 51s virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 51s virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 51s virt-handler-6gcqv 0/1 ImagePullBackOff 0 48s virt-handler-wr72t 0/1 ImagePullBackOff 0 48s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 59s virt-handler-wr72t 0/1 ImagePullBackOff 0 59s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 59s virt-handler-wr72t 0/1 ImagePullBackOff 0 59s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + grep -v Running + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ErrImagePull 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ErrImagePull 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ErrImagePull 0 1m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 1m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ErrImagePull 0 1m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ErrImagePull 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + grep -v Running + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 1m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 1m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 1m virt-controller-5f7c946cc4-tpt5x 0/1 ErrImagePull 0 1m virt-handler-6gcqv 0/1 ImagePullBackOff 0 1m virt-handler-wr72t 0/1 ImagePullBackOff 0 1m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 2m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 2m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 2m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 2m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 2m virt-handler-6gcqv 0/1 ImagePullBackOff 0 2m virt-handler-wr72t 0/1 ImagePullBackOff 0 2m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 3m virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ErrImagePull 0 3m virt-handler-wr72t 0/1 ErrImagePull 0 3m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 3m virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ErrImagePull 0 3m virt-handler-wr72t 0/1 ErrImagePull 0 3m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 3m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ErrImagePull 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ErrImagePull 0 3m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 3m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ErrImagePull 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ErrImagePull 0 3m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + grep -v Running + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 3m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 3m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 3m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 3m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 3m virt-handler-wr72t 0/1 ImagePullBackOff 0 3m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + grep -v Running + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 4m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 4m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 4m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 4m virt-handler-6gcqv 0/1 ImagePullBackOff 0 4m virt-handler-wr72t 0/1 ImagePullBackOff 0 4m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 5m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 5m virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 5m virt-controller-5f7c946cc4-sdtdz 0/1 ErrImagePull 0 5m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 5m virt-handler-6gcqv 0/1 ImagePullBackOff 0 5m virt-handler-wr72t 0/1 ImagePullBackOff 0 5m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 6m virt-api-fd96f94b5-xqtpn 0/1 ErrImagePull 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ErrImagePull 0 6m virt-handler-6gcqv 0/1 ErrImagePull 0 6m virt-handler-wr72t 0/1 ErrImagePull 0 6m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ErrImagePull 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ErrImagePull 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ErrImagePull 0 6m virt-handler-6gcqv 0/1 ErrImagePull 0 6m virt-handler-wr72t 0/1 ErrImagePull 0 6m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ErrImagePull 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ErrImagePull 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 6m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 6m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 6m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 6m virt-handler-6gcqv 0/1 ImagePullBackOff 0 6m virt-handler-wr72t 0/1 ImagePullBackOff 0 6m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 7m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 7m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 7m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 7m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 7m virt-handler-wr72t 0/1 ImagePullBackOff 0 7m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 8m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 8m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 8m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 8m virt-handler-6gcqv 0/1 ImagePullBackOff 0 8m virt-handler-wr72t 0/1 ImagePullBackOff 0 8m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 9m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 9m virt-handler-6gcqv 0/1 ImagePullBackOff 0 9m virt-handler-wr72t 0/1 ImagePullBackOff 0 9m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 9m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 9m virt-handler-6gcqv 0/1 ImagePullBackOff 0 9m virt-handler-wr72t 0/1 ImagePullBackOff 0 9m + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 9m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 9m virt-handler-6gcqv 0/1 ImagePullBackOff 0 9m virt-handler-wr72t 0/1 ImagePullBackOff 0 9m' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-gljb2 0/1 ImagePullBackOff 0 9m iscsi-demo-target-tgtd-nbd7d 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-86pdc 0/1 ImagePullBackOff 0 9m virt-api-fd96f94b5-xqtpn 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-sdtdz 0/1 ImagePullBackOff 0 9m virt-controller-5f7c946cc4-tpt5x 0/1 ImagePullBackOff 0 9m virt-handler-6gcqv 0/1 ImagePullBackOff 0 9m virt-handler-wr72t 0/1 ImagePullBackOff 0 9m + sleep 10 automation/test.sh: line 72: 26094 Terminated sleep 10 + make cluster-down