+ export WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release + WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release + [[ openshift-release =~ openshift-.* ]] + export PROVIDER=os-3.9.0-alpha.4 + PROVIDER=os-3.9.0-alpha.4 + export VAGRANT_NUM_NODES=1 + VAGRANT_NUM_NODES=1 + export NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + export NAMESPACE=kube-system + NAMESPACE=kube-system + trap '{ make cluster-down; }' EXIT + make cluster-down ./cluster/down.sh Unable to find image 'kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' locally Trying to pull repository docker.io/kubevirtci/cli ... sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4: Pulling from docker.io/kubevirtci/cli a8ee583972c2: Pulling fs layer 9c0deba734ef: Pulling fs layer 54cd07a1b133: Pulling fs layer 54cd07a1b133: Verifying Checksum 54cd07a1b133: Download complete a8ee583972c2: Verifying Checksum a8ee583972c2: Download complete 9c0deba734ef: Verifying Checksum 9c0deba734ef: Download complete a8ee583972c2: Pull complete 9c0deba734ef: Pull complete 54cd07a1b133: Pull complete Digest: sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4 Status: Downloaded newer image for docker.io/kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4 + make cluster-up ./cluster/up.sh Unable to find image 'kubevirtci/os-3.9@sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8' locally Trying to pull repository docker.io/kubevirtci/os-3.9 ... sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8: Pulling from docker.io/kubevirtci/os-3.9 a8ee583972c2: Already exists dd50e5a4fc23: Pulling fs layer d867b8969b5b: Pulling fs layer bc770f22e8ac: Pulling fs layer d22f17305a59: Pulling fs layer 74c8d4bbaa28: Pulling fs layer a1be06ea19b0: Pulling fs layer 646f20623323: Pulling fs layer d22f17305a59: Waiting a1be06ea19b0: Waiting 74c8d4bbaa28: Waiting d867b8969b5b: Verifying Checksum d867b8969b5b: Download complete bc770f22e8ac: Verifying Checksum bc770f22e8ac: Download complete 74c8d4bbaa28: Download complete d22f17305a59: Verifying Checksum d22f17305a59: Download complete dd50e5a4fc23: Verifying Checksum dd50e5a4fc23: Download complete dd50e5a4fc23: Pull complete d867b8969b5b: Pull complete bc770f22e8ac: Pull complete d22f17305a59: Pull complete 74c8d4bbaa28: Pull complete a1be06ea19b0: Verifying Checksum a1be06ea19b0: Download complete a1be06ea19b0: Pull complete 646f20623323: Verifying Checksum 646f20623323: Download complete 646f20623323: Pull complete Digest: sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8 Status: Downloaded newer image for docker.io/kubevirtci/os-3.9@sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8 kubevirt-functional-tests-openshift-release0_registry Unable to find image 'registry:2' locally Trying to pull repository docker.io/library/registry ... 2: Pulling from docker.io/library/registry 81033e7c1d6a: Pulling fs layer b235084c2315: Pulling fs layer c692f3a6894b: Pulling fs layer ba2177f3a70e: Pulling fs layer a8d793620947: Pulling fs layer a8d793620947: Waiting b235084c2315: Verifying Checksum b235084c2315: Download complete 81033e7c1d6a: Verifying Checksum 81033e7c1d6a: Download complete c692f3a6894b: Verifying Checksum c692f3a6894b: Download complete 81033e7c1d6a: Pull complete ba2177f3a70e: Download complete a8d793620947: Verifying Checksum a8d793620947: Download complete b235084c2315: Pull complete c692f3a6894b: Pull complete ba2177f3a70e: Pull complete a8d793620947: Pull complete Digest: sha256:672d519d7fd7bbc7a448d17956ebeefe225d5eb27509d8dc5ce67ecb4a0bce54 Status: Downloaded newer image for docker.io/registry:2 WARNING: You're not using the default seccomp profile kubevirt-functional-tests-openshift-release0-node02 2018/04/06 12:43:40 Waiting for host: 192.168.66.102:22 2018/04/06 12:43:43 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:43:51 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:43:59 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:44:07 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:44:12 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: connection refused. Sleeping 5s 2018/04/06 12:44:17 Connected to tcp://192.168.66.102:22 Removed symlink /etc/systemd/system/docker.service.wants/origin-master-api.service. Removed symlink /etc/systemd/system/origin-node.service.wants/origin-master-api.service. Removed symlink /etc/systemd/system/docker.service.wants/origin-master-controllers.service. kubevirt-functional-tests-openshift-release0-node01 2018/04/06 12:44:24 Waiting for host: 192.168.66.101:22 2018/04/06 12:44:27 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:44:35 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:44:43 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:44:51 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:44:59 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/06 12:45:04 Connected to tcp://192.168.66.101:22 The connection to the server node01:8443 was refused - did you specify the right host or port? NAME STATUS ROLES AGE VERSION node01 Ready master 1d v1.9.1+a0ce1bc657 PING node02 (192.168.66.102) 56(84) bytes of data. 64 bytes from node02 (192.168.66.102): icmp_seq=1 ttl=64 time=1.14 ms --- node02 ping statistics --- 1 packets transmitted, 1 received, 0% packet loss, time 0ms rtt min/avg/max/mdev = 1.144/1.144/1.144/0.000 ms Found node02. Adding it to the inventory. ping: node03: Name or service not known PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) PLAY [Ensure there are new_nodes] ********************************************** TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] PLAY [Initialization Checkpoint Start] ***************************************** TASK [Set install initialization 'In Progress'] ******************************** ok: [node01] PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) [WARNING]: Could not match supplied host pattern, ignoring: oo_lb_to_config PLAY [Ensure that all non-node hosts are accessible] *************************** TASK [Gathering Facts] ********************************************************* ok: [node01] PLAY [Initialize basic host facts] ********************************************* TASK [Gathering Facts] ********************************************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/deprecations.yml for node01, node02 TASK [openshift_sanitize_inventory : Check for usage of deprecated variables] *** skipping: [node01] => (item=openshift_hosted_logging_deploy) skipping: [node01] => (item=openshift_hosted_logging_hostname) skipping: [node01] => (item=openshift_hosted_logging_ops_hostname) skipping: [node01] => (item=openshift_hosted_logging_master_public_url) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_cluster_size) skipping: [node02] => (item=openshift_hosted_logging_deploy) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_cluster_size) skipping: [node01] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node02] => (item=openshift_hosted_logging_hostname) skipping: [node01] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node02] => (item=openshift_hosted_logging_ops_hostname) skipping: [node01] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_master_public_url) skipping: [node01] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_cluster_size) skipping: [node01] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_cluster_size) skipping: [node02] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node02] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node01] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node02] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_journal_source) skipping: [node01] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node02] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_storage_labels) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_pvc_dynamic) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_pvc_size) skipping: [node02] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_pvc_prefix) skipping: [node02] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node02] => (item=openshift_hosted_logging_journal_source) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node01] => (item=openshift_hosted_loggingops_storage_labels) skipping: [node02] => (item=openshift_hosted_logging_storage_labels) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_dynamic) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_size) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_prefix) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_pvc_dynamic) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_pvc_size) skipping: [node01] => (item=openshift_hosted_logging_storage_kind) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_pvc_prefix) skipping: [node01] => (item=openshift_hosted_loggingops_storage_kind) skipping: [node01] => (item=openshift_hosted_logging_storage_host) skipping: [node01] => (item=openshift_hosted_loggingops_storage_host) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node01] => (item=openshift_hosted_logging_storage_nfs_directory) skipping: [node01] => (item=openshift_hosted_loggingops_storage_nfs_directory) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_storage_volume_name) skipping: [node01] => (item=openshift_hosted_loggingops_storage_volume_name) skipping: [node01] => (item=openshift_hosted_logging_storage_volume_size) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node01] => (item=openshift_hosted_loggingops_storage_volume_size) skipping: [node02] => (item=openshift_hosted_loggingops_storage_labels) skipping: [node01] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node01] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_dynamic) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_size) skipping: [node01] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_prefix) skipping: [node01] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node01] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node01] => (item=openshift_hosted_logging_ops_hostname) skipping: [node01] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node02] => (item=openshift_hosted_logging_storage_kind) skipping: [node01] => (item=openshift_hosted_logging_journal_source) skipping: [node02] => (item=openshift_hosted_loggingops_storage_kind) skipping: [node01] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node02] => (item=openshift_hosted_logging_storage_host) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node02] => (item=openshift_hosted_loggingops_storage_host) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_storage_nfs_directory) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node02] => (item=openshift_hosted_loggingops_storage_nfs_directory) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_storage_volume_name) skipping: [node01] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node02] => (item=openshift_hosted_loggingops_storage_volume_name) skipping: [node01] => (item=openshift_hosted_logging_deployer_prefix) skipping: [node02] => (item=openshift_hosted_logging_storage_volume_size) skipping: [node01] => (item=openshift_hosted_logging_deployer_version) skipping: [node02] => (item=openshift_hosted_loggingops_storage_volume_size) skipping: [node01] => (item=openshift_hosted_metrics_deploy) skipping: [node02] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node01] => (item=openshift_hosted_metrics_storage_kind) skipping: [node02] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node01] => (item=openshift_hosted_metrics_storage_access_modes) skipping: [node01] => (item=openshift_hosted_metrics_storage_host) skipping: [node02] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_metrics_storage_nfs_directory) skipping: [node02] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node01] => (item=openshift_hosted_metrics_storage_volume_name) skipping: [node01] => (item=openshift_hosted_metrics_storage_volume_size) skipping: [node02] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_metrics_storage_labels) skipping: [node02] => (item=openshift_hosted_logging_ops_hostname) skipping: [node01] => (item=openshift_hosted_metrics_deployer_prefix) skipping: [node02] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node01] => (item=openshift_hosted_metrics_deployer_version) skipping: [node02] => (item=openshift_hosted_logging_journal_source) skipping: [node02] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node02] => (item=openshift_hosted_logging_deployer_prefix) skipping: [node02] => (item=openshift_hosted_logging_deployer_version) skipping: [node02] => (item=openshift_hosted_metrics_deploy) skipping: [node02] => (item=openshift_hosted_metrics_storage_kind) skipping: [node02] => (item=openshift_hosted_metrics_storage_access_modes) skipping: [node02] => (item=openshift_hosted_metrics_storage_host) skipping: [node02] => (item=openshift_hosted_metrics_storage_nfs_directory) skipping: [node02] => (item=openshift_hosted_metrics_storage_volume_name) skipping: [node02] => (item=openshift_hosted_metrics_storage_volume_size) skipping: [node02] => (item=openshift_hosted_metrics_storage_labels) skipping: [node02] => (item=openshift_hosted_metrics_deployer_prefix) skipping: [node02] => (item=openshift_hosted_metrics_deployer_version) TASK [openshift_sanitize_inventory : debug] ************************************ skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_stats] ******************************** skipping: [node01] TASK [openshift_sanitize_inventory : Assign deprecated variables to correct counterparts] *** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_logging.yml for node01, node02 included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_metrics.yml for node01, node02 TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node02] ok: [node01] TASK [openshift_sanitize_inventory : Standardize on latest variable names] ***** ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : Normalize openshift_release] ************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Abort when openshift_release is invalid] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/unsupported.yml for node01, node02 TASK [openshift_sanitize_inventory : Ensure that openshift_use_dnsmasq is true] *** skipping: [node02] skipping: [node01] TASK [openshift_sanitize_inventory : Ensure that openshift_node_dnsmasq_install_network_manager_hook is true] *** skipping: [node02] skipping: [node01] TASK [openshift_sanitize_inventory : set_fact] ********************************* skipping: [node02] => (item=openshift_hosted_etcd_storage_kind) skipping: [node01] => (item=openshift_hosted_etcd_storage_kind) TASK [openshift_sanitize_inventory : Ensure that dynamic provisioning is set if using dynamic storage] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure clusterid is set along with the cloudprovider] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure ansible_service_broker_remove and ansible_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure template_service_broker_remove and template_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that all requires vsphere configuration variables are set] *** skipping: [node01] skipping: [node02] TASK [Detecting Operating System from ostree_booted] *************************** ok: [node02] ok: [node01] TASK [set openshift_deployment_type if unset] ********************************** skipping: [node01] skipping: [node02] TASK [initialize_facts set fact openshift_is_atomic and openshift_is_containerized] *** ok: [node01] ok: [node02] TASK [Determine Atomic Host Docker Version] ************************************ skipping: [node01] skipping: [node02] TASK [assert atomic host docker version is 1.12 or later] ********************** skipping: [node01] skipping: [node02] PLAY [Initialize special first-master variables] ******************************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] PLAY [Disable web console if required] ***************************************** TASK [set_fact] **************************************************************** skipping: [node01] PLAY [Install packages necessary for installer] ******************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Ensure openshift-ansible installer package deps are installed] *********** ok: [node02] => (item=iproute) ok: [node02] => (item=dbus-python) ok: [node02] => (item=PyYAML) ok: [node02] => (item=python-ipaddress) ok: [node02] => (item=yum-utils) TASK [Ensure various deps for running system containers are installed] ********* skipping: [node02] => (item=atomic) skipping: [node02] => (item=ostree) skipping: [node02] => (item=runc) PLAY [Initialize cluster facts] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [Gather Cluster facts] **************************************************** ok: [node01] changed: [node02] TASK [Set fact of no_proxy_internal_hostnames] ********************************* skipping: [node01] skipping: [node02] TASK [Initialize openshift.node.sdn_mtu] *************************************** ok: [node02] ok: [node01] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* skipping: [node01] TASK [include_role] ************************************************************ skipping: [node01] TASK [debug] ******************************************************************* skipping: [node01] PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* skipping: [node02] TASK [set_fact] **************************************************************** skipping: [node02] PLAY [Ensure the requested version packages are available.] ******************** TASK [Gathering Facts] ********************************************************* skipping: [node02] TASK [include_role] ************************************************************ skipping: [node02] PLAY [Verify Requirements] ***************************************************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [Run variable sanity checks] ********************************************** ok: [node01] PLAY [Initialization Checkpoint End] ******************************************* TASK [Set install initialization 'Complete'] *********************************** ok: [node01] PLAY [Validate node hostnames] ************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Query DNS for IP address of node02] ************************************** ok: [node02] TASK [Validate openshift_hostname when defined] ******************************** skipping: [node02] TASK [Validate openshift_ip exists on node when defined] *********************** skipping: [node02] PLAY [Setup yum repositories for all hosts] ************************************ TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Install Red Hat Subscription manager] ******************* skipping: [node02] TASK [rhel_subscribe : Is host already registered?] **************************** skipping: [node02] TASK [rhel_subscribe : Register host] ****************************************** skipping: [node02] TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Determine if OpenShift Pool Already Attached] *********** skipping: [node02] TASK [rhel_subscribe : Attach to OpenShift Pool] ******************************* skipping: [node02] TASK [rhel_subscribe : include_tasks] ****************************************** skipping: [node02] TASK [openshift_repos : openshift_repos detect ostree] ************************* ok: [node02] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** ok: [node02] TASK [openshift_repos : Remove openshift_additional.repo file] ***************** ok: [node02] TASK [openshift_repos : Create any additional repos that are defined] ********** TASK [openshift_repos : include_tasks] ***************************************** skipping: [node02] TASK [openshift_repos : include_tasks] ***************************************** included: /root/openshift-ansible/roles/openshift_repos/tasks/centos_repos.yml for node02 TASK [openshift_repos : Configure origin gpg keys] ***************************** ok: [node02] TASK [openshift_repos : Configure correct origin release repository] *********** ok: [node02] => (item=/root/openshift-ansible/roles/openshift_repos/templates/CentOS-OpenShift-Origin.repo.j2) TASK [openshift_repos : Ensure clean repo cache in the event repos have been changed manually] *** changed: [node02] => { "msg": "First run of openshift_repos" } TASK [openshift_repos : Record that openshift_repos already ran] *************** ok: [node02] RUNNING HANDLER [openshift_repos : refresh cache] ****************************** changed: [node02] PLAY [Configure os_firewall] *************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [os_firewall : Detecting Atomic Host Operating System] ******************** ok: [node02] TASK [os_firewall : Set fact r_os_firewall_is_atomic] ************************** ok: [node02] TASK [os_firewall : include_tasks] ********************************************* skipping: [node02] TASK [os_firewall : include_tasks] ********************************************* included: /root/openshift-ansible/roles/os_firewall/tasks/iptables.yml for node02 TASK [os_firewall : Ensure firewalld service is not enabled] ******************* ok: [node02] TASK [os_firewall : Wait 10 seconds after disabling firewalld] ***************** skipping: [node02] TASK [os_firewall : Install iptables packages] ********************************* ok: [node02] => (item=iptables) ok: [node02] => (item=iptables-services) TASK [os_firewall : Start and enable iptables service] ************************* ok: [node02 -> node02] => (item=node02) TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [container_runtime : Setup the docker-storage for overlay] **************** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_excluder : Install excluders] ********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* ok: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/common/pre.yml for node02 TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Add enterprise registry, if necessary] *************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Get current installed Docker version] **************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/docker_sanity.yml for node02 TASK [container_runtime : Error out if Docker pre-installed but too old] ******* skipping: [node02] TASK [container_runtime : Error out if requested Docker is too old] ************ skipping: [node02] TASK [container_runtime : Fail if Docker version requested but downgrade is required] *** skipping: [node02] TASK [container_runtime : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [node02] TASK [container_runtime : Install Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure docker.service.d directory exists] ************ ok: [node02] TASK [container_runtime : Configure Docker service unit file] ****************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Set registry params] ********************************* skipping: [node02] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': []}) TASK [container_runtime : Place additional/blocked/insecure registries in /etc/containers/registries.conf] *** skipping: [node02] TASK [container_runtime : Set Proxy Settings] ********************************** skipping: [node02] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [container_runtime : Set various Docker options] ************************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Configure Docker Network OPTIONS] ******************** ok: [node02] TASK [container_runtime : Detect if docker is already started] ***************** ok: [node02] TASK [container_runtime : Start the Docker service] **************************** ok: [node02] TASK [container_runtime : set_fact] ******************************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/common/post.yml for node02 TASK [container_runtime : Ensure /var/lib/containers exists] ******************* ok: [node02] TASK [container_runtime : Fix SELinux Permissions on /var/lib/containers] ****** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/registry_auth.yml for node02 TASK [container_runtime : Check for credentials file for registry auth] ******** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth] ***** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth (alternative)] *** skipping: [node02] TASK [container_runtime : stat the docker data dir] **************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Fail quickly if openshift_docker_options are set] **** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Install Docker so we can use the client] ************* skipping: [node02] TASK [container_runtime : Disable Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** skipping: [node02] TASK [container_runtime : debug] *********************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Pre-pull Container Engine System Container image] **** skipping: [node02] TASK [container_runtime : Ensure container-engine.service.d directory exists] *** skipping: [node02] TASK [container_runtime : Ensure /etc/docker directory exists] ***************** skipping: [node02] TASK [container_runtime : Install Container Engine System Container] *********** skipping: [node02] TASK [container_runtime : Configure Container Engine Service File] ************* skipping: [node02] TASK [container_runtime : Configure Container Engine] ************************** skipping: [node02] TASK [container_runtime : Start the Container Engine service] ****************** skipping: [node02] TASK [container_runtime : set_fact] ******************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Check we are not using node as a Docker container with CRI-O] *** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Check that overlay is in the kernel] ***************** skipping: [node02] TASK [container_runtime : Add overlay to modprobe.d] *************************** skipping: [node02] TASK [container_runtime : Manually modprobe overlay into the kernel] *********** skipping: [node02] TASK [container_runtime : Enable and start systemd-modules-load] *************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** skipping: [node02] TASK [container_runtime : debug] *********************************************** skipping: [node02] TASK [container_runtime : Pre-pull CRI-O System Container image] *************** skipping: [node02] TASK [container_runtime : Install CRI-O System Container] ********************** skipping: [node02] TASK [container_runtime : Remove CRI-O default configuration files] ************ skipping: [node02] => (item=/etc/cni/net.d/200-loopback.conf) skipping: [node02] => (item=/etc/cni/net.d/100-crio-bridge.conf) TASK [container_runtime : Create the CRI-O configuration] ********************** skipping: [node02] TASK [container_runtime : Ensure CNI configuration directory exists] *********** skipping: [node02] TASK [container_runtime : Add iptables allow rules] **************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove iptables rules] ******************************* TASK [container_runtime : Add firewalld allow rules] *************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove firewalld allow rules] ************************ TASK [container_runtime : Configure the CNI network] *************************** skipping: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-storage] ****************** skipping: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-network] ****************** skipping: [node02] TASK [container_runtime : Start the CRI-O service] ***************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [include_role] ************************************************************ TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** ok: [node01] TASK [openshift_version : include_tasks] *************************************** included: /root/openshift-ansible/roles/openshift_version/tasks/first_master_containerized_version.yml for node01 TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [node01] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [node01] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] => { "msg": "openshift_pkg_version was not defined. Falling back to -3.9.0" } TASK [openshift_version : set_fact] ******************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [debug] ******************************************************************* ok: [node01] => { "msg": "openshift_pkg_version set to -3.9.0" } PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [set_fact] **************************************************************** ok: [node02] PLAY [Ensure the requested version packages are available.] ******************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [include_role] ************************************************************ TASK [openshift_version : Check openshift_version for rpm installation] ******** included: /root/openshift-ansible/roles/openshift_version/tasks/check_available_rpms.yml for node02 TASK [openshift_version : Get available origin version] ************************ ok: [node02] TASK [openshift_version : fail] ************************************************ skipping: [node02] TASK [openshift_version : Fail if rpm version and docker image version are different] *** skipping: [node02] TASK [openshift_version : For an RPM install, abort when the release requested does not match the available version.] *** skipping: [node02] TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_image_tag": "v3.9.0-alpha.4" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_pkg_version": "-3.9.0" } PLAY [Node Install Checkpoint Start] ******************************************* TASK [Set Node install 'In Progress'] ****************************************** ok: [node01] PLAY [Create OpenShift certificates for node hosts] **************************** TASK [openshift_node_certificates : Ensure CA certificate exists on openshift_ca_host] *** ok: [node02 -> node01] TASK [openshift_node_certificates : fail] ************************************** skipping: [node02] TASK [openshift_node_certificates : Check status of node certificates] ********* ok: [node02] => (item=system:node:node02.crt) ok: [node02] => (item=system:node:node02.key) ok: [node02] => (item=system:node:node02.kubeconfig) ok: [node02] => (item=ca.crt) ok: [node02] => (item=server.key) ok: [node02] => (item=server.crt) TASK [openshift_node_certificates : set_fact] ********************************** ok: [node02] TASK [openshift_node_certificates : Create openshift_generated_configs_dir if it does not exist] *** ok: [node02 -> node01] TASK [openshift_node_certificates : find] ************************************** ok: [node02 -> node01] TASK [openshift_node_certificates : Generate the node client config] *********** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Generate the node server certificate] ****** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Create a tarball of the node config directories] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Retrieve the node config tarballs from the master] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Ensure certificate directory exists] ******* ok: [node02] TASK [openshift_node_certificates : Unarchive the tarball on the node] ********* changed: [node02] TASK [openshift_node_certificates : Delete local temp directory] *************** ok: [node02 -> localhost] TASK [openshift_node_certificates : Copy OpenShift CA to system CA trust] ****** ok: [node02] => (item={u'cert': u'/etc/origin/node/ca.crt', u'id': u'openshift'}) PLAY [Disable excluders] ******************************************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": "false" } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": "false" } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/disable.yml for node02 TASK [openshift_excluder : Include verify_upgrade.yml when upgrading] ********** skipping: [node02] TASK [openshift_excluder : Disable excluders before the upgrade to remove older excluding expressions] *** included: /root/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* skipping: [node02] TASK [openshift_excluder : Include install.yml] ******************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Include exclude.yml] ******************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] TASK [openshift_excluder : Include unexclude.yml] ****************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* skipping: [node02] PLAY [Evaluate node groups] **************************************************** TASK [Gathering Facts] ********************************************************* ok: [localhost] TASK [Evaluate oo_containerized_master_nodes] ********************************** skipping: [localhost] => (item=node02) [WARNING]: Could not match supplied host pattern, ignoring: oo_containerized_master_nodes PLAY [Configure containerized nodes] ******************************************* skipping: no hosts matched PLAY [Configure nodes] ********************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_clock : Determine if chrony is installed] ********************** [WARNING]: Consider using yum, dnf or zypper module rather than running rpm changed: [node02] TASK [openshift_clock : Install ntp package] *********************************** skipping: [node02] TASK [openshift_clock : Start and enable ntpd/chronyd] ************************* changed: [node02] TASK [openshift_cloud_provider : Set cloud provider facts] ********************* ok: [node02] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_node : fail] *************************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/dnsmasq_install.yml for node02 TASK [openshift_node : Check for NetworkManager service] *********************** ok: [node02] TASK [openshift_node : Set fact using_network_manager] ************************* ok: [node02] TASK [openshift_node : Install dnsmasq] **************************************** ok: [node02] TASK [openshift_node : ensure origin/node directory exists] ******************** ok: [node02] => (item=/etc/origin) changed: [node02] => (item=/etc/origin/node) TASK [openshift_node : Install node-dnsmasq.conf] ****************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/dnsmasq.yml for node02 TASK [openshift_node : Install dnsmasq configuration] ************************** ok: [node02] TASK [openshift_node : Deploy additional dnsmasq.conf] ************************* skipping: [node02] TASK [openshift_node : Enable dnsmasq] ***************************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/dnsmasq/network-manager.yml for node02 TASK [openshift_node : Install network manager dispatch script] **************** ok: [node02] TASK [openshift_node : Add iptables allow rules] ******************************* ok: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) ok: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) ok: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) ok: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove iptables rules] ********************************** TASK [openshift_node : Add firewalld allow rules] ****************************** skipping: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) skipping: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) skipping: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove firewalld allow rules] *************************** TASK [openshift_node : Disable swap] ******************************************* ok: [node02] TASK [openshift_node : include node installer] ********************************* included: /root/openshift-ansible/roles/openshift_node/tasks/install.yml for node02 TASK [openshift_node : Install Node package, sdn-ovs, conntrack packages] ****** skipping: [node02] => (item={u'name': u'origin-node-3.9.0'}) skipping: [node02] => (item={u'name': u'origin-sdn-ovs-3.9.0', u'install': True}) skipping: [node02] => (item={u'name': u'conntrack-tools'}) TASK [openshift_node : Pre-pull node image when containerized] ***************** ok: [node02] TASK [openshift_node : Restart cri-o] ****************************************** skipping: [node02] TASK [openshift_node : restart NetworkManager to ensure resolv.conf is present] *** skipping: [node02] TASK [openshift_node : sysctl] ************************************************* ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/registry_auth.yml for node02 TASK [openshift_node : Check for credentials file for registry auth] *********** skipping: [node02] TASK [openshift_node : Create credentials for registry auth] ******************* skipping: [node02] TASK [openshift_node : Create credentials for registry auth (alternative)] ***** skipping: [node02] TASK [openshift_node : Setup ro mount of /root/.docker for containerized hosts] *** skipping: [node02] TASK [openshift_node : include standard node config] *************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config.yml for node02 TASK [openshift_node : Install the systemd units] ****************************** included: /root/openshift-ansible/roles/openshift_node/tasks/systemd_units.yml for node02 TASK [openshift_node : Install Node service file] ****************************** ok: [node02] TASK [openshift_node : include node deps docker service file] ****************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/install-node-deps-docker-service-file.yml for node02 TASK [openshift_node : Install Node dependencies docker service file] ********** ok: [node02] TASK [openshift_node : include ovs service environment file] ******************* included: /root/openshift-ansible/roles/openshift_node/tasks/config/install-ovs-service-env-file.yml for node02 TASK [openshift_node : Create the openvswitch service env file] **************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/install-ovs-docker-service-file.yml for node02 TASK [openshift_node : Install OpenvSwitch docker service file] **************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/configure-node-settings.yml for node02 TASK [openshift_node : Configure Node settings] ******************************** ok: [node02] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS=--loglevel=2'}) ok: [node02] => (item={u'regex': u'^CONFIG_FILE=', u'line': u'CONFIG_FILE=/etc/origin/node/node-config.yaml'}) ok: [node02] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.9.0-alpha.4'}) TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/configure-proxy-settings.yml for node02 TASK [openshift_node : Configure Proxy Settings] ******************************* skipping: [node02] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) skipping: [node02] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [node02] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=[],172.30.0.0/16,10.128.0.0/14'}) TASK [openshift_node : Pull container images] ********************************** included: /root/openshift-ansible/roles/openshift_node/tasks/container_images.yml for node02 TASK [openshift_node : Install Node system container] ************************** skipping: [node02] TASK [openshift_node : Install OpenvSwitch system containers] ****************** skipping: [node02] TASK [openshift_node : Pre-pull openvswitch image] ***************************** ok: [node02] TASK [openshift_node : Start and enable openvswitch service] ******************* ok: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : file] *************************************************** skipping: [node02] TASK [openshift_node : Create the Node config] ********************************* changed: [node02] TASK [openshift_node : Configure Node Environment Variables] ******************* TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [node02] => (item=None) skipping: [node02] => (item=None) TASK [openshift_node : Wait for master API to become available before proceeding] *** ok: [node02] TASK [openshift_node : Start and enable node dep] ****************************** changed: [node02] TASK [openshift_node : Start and enable node] ********************************** ok: [node02] TASK [openshift_node : Dump logs from node service if it failed] *************** skipping: [node02] TASK [openshift_node : Abort if node failed to start] ************************** skipping: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : NFS storage plugin configuration] *********************** included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/nfs.yml for node02 TASK [openshift_node : Install NFS storage plugin dependencies] **************** ok: [node02] TASK [openshift_node : Check for existence of nfs sebooleans] ****************** ok: [node02] => (item=virt_use_nfs) ok: [node02] => (item=virt_sandbox_use_nfs) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:43.359522', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.028203', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:43.331319', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:44.666901', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.017246', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:44.649655', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:43.359522', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.028203', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:43.331319', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:44.666901', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.017246', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:44.649655', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : GlusterFS storage plugin configuration] ***************** included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/glusterfs.yml for node02 TASK [openshift_node : Install GlusterFS storage plugin dependencies] ********** ok: [node02] TASK [openshift_node : Check for existence of fusefs sebooleans] *************** ok: [node02] => (item=virt_use_fusefs) ok: [node02] => (item=virt_sandbox_use_fusefs) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:50.302192', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.009302', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:50.292890', '_ansible_ignore_errors': None, 'failed': False}) ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:51.211554', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.009384', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:51.202170', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:50.302192', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.009302', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:50.292890', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-06 12:51:51.211554', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.009384', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-06 12:51:51.202170', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Ceph storage plugin configuration] ********************** included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/ceph.yml for node02 TASK [openshift_node : Install Ceph storage plugin dependencies] *************** ok: [node02] TASK [openshift_node : iSCSI storage plugin configuration] ********************* included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/iscsi.yml for node02 TASK [openshift_node : Install iSCSI storage plugin dependencies] ************** ok: [node02] => (item=iscsi-initiator-utils) ok: [node02] => (item=device-mapper-multipath) TASK [openshift_node : restart services] *************************************** ok: [node02] => (item=multipathd) ok: [node02] => (item=rpcbind) TASK [openshift_node : Template multipath configuration] *********************** changed: [node02] TASK [openshift_node : Enable multipath] *************************************** changed: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/workaround-bz1331590-ovs-oom-fix.yml for node02 TASK [openshift_node : Create OpenvSwitch service.d directory] ***************** ok: [node02] TASK [openshift_node : Install OpenvSwitch service OOM fix] ******************** ok: [node02] TASK [tuned : Check for tuned package] ***************************************** ok: [node02] TASK [tuned : Set tuned OpenShift variables] *********************************** ok: [node02] TASK [tuned : Ensure directory structure exists] ******************************* ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) TASK [tuned : Ensure files are populated from templates] *********************** skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': u's0', 'seuser': u'unconfined_u', 'serole': u'object_r', 'ctime': 1522867116.424326, 'state': u'file', 'gid': 0, 'mode': u'0644', 'mtime': 1522867116.424326, 'owner': u'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': u'admin_home_t'}) TASK [tuned : Make tuned use the recommended tuned profile on restart] ********* changed: [node02] => (item=/etc/tuned/active_profile) ok: [node02] => (item=/etc/tuned/profile_mode) TASK [tuned : Restart tuned service] ******************************************* changed: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Install logrotate] ******* ok: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Setup logrotate.d scripts] *** RUNNING HANDLER [openshift_node : restart node] ******************************** changed: [node02] PLAY [create additional node network plugin groups] **************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_flannel [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_calico [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_contiv [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_kuryr PLAY [etcd_client node config] ************************************************* skipping: no hosts matched PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** skipping: no hosts matched [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_nuage PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Configure Contiv masters] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [contiv_facts : Determine if CoreOS] ************************************** skipping: [node01] TASK [contiv_facts : Init the contiv_is_coreos fact] *************************** skipping: [node01] TASK [contiv_facts : Set the contiv_is_coreos fact] **************************** skipping: [node01] TASK [contiv_facts : Set the bin directory path for CoreOS] ******************** skipping: [node01] TASK [contiv_facts : Create the directory used to store binaries] ************** skipping: [node01] TASK [contiv_facts : Create Ansible temp directory] **************************** skipping: [node01] TASK [contiv_facts : Determine if has rpm] ************************************* skipping: [node01] TASK [contiv_facts : Init the contiv_has_rpm fact] ***************************** skipping: [node01] TASK [contiv_facts : Set the contiv_has_rpm fact] ****************************** skipping: [node01] TASK [contiv_facts : Init the contiv_has_firewalld fact] *********************** skipping: [node01] TASK [contiv_facts : Init the contiv_has_iptables fact] ************************ skipping: [node01] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] TASK [contiv : Ensure contiv_bin_dir exists] *********************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] PLAY [Configure rest of Contiv nodes] ****************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [contiv_facts : Determine if CoreOS] ************************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_is_coreos fact] *************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Set the contiv_is_coreos fact] **************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Set the bin directory path for CoreOS] ******************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Create the directory used to store binaries] ************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Create Ansible temp directory] **************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Determine if has rpm] ************************************* skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_has_rpm fact] ***************************** skipping: [node02] skipping: [node01] TASK [contiv_facts : Set the contiv_has_rpm fact] ****************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_has_firewalld fact] *********************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_has_iptables fact] ************************ skipping: [node02] skipping: [node01] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] skipping: [node02] TASK [contiv : include_tasks] ************************************************** skipping: [node01] skipping: [node02] TASK [contiv : Ensure contiv_bin_dir exists] *********************************** skipping: [node02] skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node02] skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] skipping: [node02] TASK [contiv : include_tasks] ************************************************** skipping: [node01] skipping: [node02] PLAY [Configure Kuryr node] **************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_manage_node : Wait for master API to become available before proceeding] *** ok: [node02 -> node01] TASK [openshift_manage_node : Wait for Node Registration] ********************** ok: [node02 -> node01] TASK [openshift_manage_node : include_tasks] *********************************** included: /root/openshift-ansible/roles/openshift_manage_node/tasks/config.yml for node02 TASK [openshift_manage_node : Set node schedulability] ************************* ok: [node02 -> node01] TASK [openshift_manage_node : Label nodes] ************************************* ok: [node02 -> node01] TASK [Create group for deployment type] **************************************** ok: [node02] PLAY [Re-enable excluder if it was previously enabled] ************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": "false" } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": "false" } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/enable.yml for node02 TASK [openshift_excluder : Install excluders] ********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] PLAY [Node Install Checkpoint End] ********************************************* TASK [Set Node install 'Complete'] ********************************************* ok: [node01] PLAY RECAP ********************************************************************* localhost : ok=25 changed=0 unreachable=0 failed=0 node01 : ok=36 changed=0 unreachable=0 failed=0 node02 : ok=183 changed=18 unreachable=0 failed=0 INSTALLER STATUS *************************************************************** Initialization : Complete (0:01:16) Node Install : Complete (0:03:38) PLAY [new_nodes] *************************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Restart openvswitch service] ********************************************* changed: [node02] PLAY RECAP ********************************************************************* node02 : ok=2 changed=1 unreachable=0 failed=0 2018/04/06 12:53:43 Waiting for host: 192.168.66.101:22 2018/04/06 12:53:43 Connected to tcp://192.168.66.101:22 2018/04/06 12:53:45 Waiting for host: 192.168.66.101:22 2018/04/06 12:53:45 Connected to tcp://192.168.66.101:22 Warning: Permanently added '[127.0.0.1]:32772' (ECDSA) to the list of known hosts. Warning: Permanently added '[127.0.0.1]:32772' (ECDSA) to the list of known hosts. Cluster "node01:8443" set. Cluster "node01:8443" set. ++ kubectl get nodes --no-headers ++ cluster/kubectl.sh get nodes --no-headers ++ grep -v Ready + '[' -n '' ']' + echo 'Nodes are ready:' Nodes are ready: + kubectl get nodes + cluster/kubectl.sh get nodes NAME STATUS ROLES AGE VERSION node01 Ready master 1d v1.9.1+a0ce1bc657 node02 Ready 1m v1.9.1+a0ce1bc657 + make cluster-sync ./cluster/build.sh Building ... sha256:6a9ee4540f213c441209d0e9c25ce1699434ef857140b3f1b80f0a2741809637 go version go1.9.2 linux/amd64 rsync: read error: Connection reset by peer (104) rsync error: error in rsync protocol data stream (code 12) at io.c(764) [sender=3.0.9] Waiting for rsyncd to be ready skipping directory . go version go1.9.2 linux/amd64 bc16687328c3883b01819968f1f490759af4ac856d60175f39a33c75d8f79610 bc16687328c3883b01819968f1f490759af4ac856d60175f39a33c75d8f79610 make[1]: Entering directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt' hack/dockerized "./hack/check.sh && ./hack/build-go.sh install " sha256:6a9ee4540f213c441209d0e9c25ce1699434ef857140b3f1b80f0a2741809637 go version go1.9.2 linux/amd64 rsync: read error: Connection reset by peer (104) rsync error: error in rsync protocol data stream (code 12) at io.c(764) [sender=3.0.9] Waiting for rsyncd to be ready skipping directory . go version go1.9.2 linux/amd64 Compiling tests... compiled tests.test 60a0286df6578c12a0350d65807f2449a9fc3d42e644ce0f8caedbc1c7db5b01 60a0286df6578c12a0350d65807f2449a9fc3d42e644ce0f8caedbc1c7db5b01 hack/build-docker.sh build sending incremental file list ./ Dockerfile kubernetes.repo sent 854 bytes received 53 bytes 1814.00 bytes/sec total size is 1167 speedup is 1.29 Sending build context to Docker daemon 35.7 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Running in 2c1a125bde8d ---> 05ea8a13b62f Removing intermediate container 2c1a125bde8d Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-controller ---> Running in 8e4d6df7a48e  ---> ae8a4776ec8d Removing intermediate container 8e4d6df7a48e Step 4/8 : WORKDIR /home/virt-controller ---> 84b8ba07c3c8 Removing intermediate container 726b4de593ae Step 5/8 : USER 1001 ---> Running in 30539a0e8bcc ---> f7fdfb92b218 Removing intermediate container 30539a0e8bcc Step 6/8 : COPY virt-controller /virt-controller ---> 4b6d1f4f4ecf Removing intermediate container 3d126191a0c0 Step 7/8 : ENTRYPOINT /virt-controller ---> Running in 8e25d3972700 ---> fafe52611c51 Removing intermediate container 8e25d3972700 Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-controller" '' ---> Running in f036f3346934 ---> 2a1c7a427856 Removing intermediate container f036f3346934 Successfully built 2a1c7a427856 sending incremental file list ./ Dockerfile entrypoint.sh kubevirt-sudo libvirtd.sh sh.sh sock-connector sent 3282 bytes received 129 bytes 6822.00 bytes/sec total size is 5469 speedup is 1.60 Sending build context to Docker daemon 37.44 MB Step 1/14 : FROM kubevirt/libvirt:3.7.0 Trying to pull repository docker.io/kubevirt/libvirt ... 3.7.0: Pulling from docker.io/kubevirt/libvirt 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer a1e80189bea5: Waiting 6cc174edcebf: Waiting f2ef945504a7: Download complete a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete 6cc174edcebf: Download complete a1e80189bea5: Verifying Checksum d7240bccd145: Verifying Checksum d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete Digest: sha256:664554259b7707bd1366018f121755c3699e53a4802c5f61dfa453e67159cde6 Status: Downloaded newer image for docker.io/kubevirt/libvirt:3.7.0 ---> 60c80c8f7523 Step 2/14 : MAINTAINER "The KubeVirt Project" ---> Running in 4b6a41cc63fb ---> 0cf0507c50a3 Removing intermediate container 4b6a41cc63fb Step 3/14 : RUN dnf -y install socat genisoimage util-linux libcgroup-tools ethtool sudo && dnf -y clean all && test $(id -u qemu) = 107 # make sure that the qemu user really is 107 ---> Running in 44c353445827 Fedora 27 - x86_64 - Updates 5.4 MB/s | 22 MB 00:03 Fedora 27 - x86_64 527 kB/s | 58 MB 01:52 Last metadata expiration check: 0:00:15 ago on Fri Apr 6 13:00:35 2018. Package util-linux-2.30.2-1.fc27.x86_64 is already installed, skipping. Dependencies resolved. ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: ethtool x86_64 2:4.15-1.fc27 updates 140 k genisoimage x86_64 1.1.11-37.fc27 fedora 311 k libcgroup-tools x86_64 0.41-13.fc27 fedora 88 k socat x86_64 1.7.3.2-4.fc27 fedora 296 k sudo x86_64 1.8.21p2-1.fc27 fedora 771 k Installing dependencies: libusal x86_64 1.1.11-37.fc27 fedora 142 k Transaction Summary ================================================================================ Install 6 Packages Total download size: 1.7 M Installed size: 5.9 M Downloading Packages: [MIRROR] socat-1.7.3.2-4.fc27.x86_64.rpm: Curl error (28): Timeout was reached for http://archive.linux.duke.edu/pub/fedora/linux/releases/27/Everything/x86_64/os/Packages/s/socat-1.7.3.2-4.fc27.x86_64.rpm [Connection timed out after 30002 milliseconds] [MIRROR] genisoimage-1.1.11-37.fc27.x86_64.rpm: Curl error (28): Timeout was reached for http://archive.linux.duke.edu/pub/fedora/linux/releases/27/Everything/x86_64/os/Packages/g/genisoimage-1.1.11-37.fc27.x86_64.rpm [Connection timed out after 30001 milliseconds] [MIRROR] libcgroup-tools-0.41-13.fc27.x86_64.rpm: Curl error (28): Timeout was reached for http://archive.linux.duke.edu/pub/fedora/linux/releases/27/Everything/x86_64/os/Packages/l/libcgroup-tools-0.41-13.fc27.x86_64.rpm [Connection timed out after 30001 milliseconds] (1/6): libcgroup-tools-0.41-13.fc27.x86_64.rpm 2.3 kB/s | 88 kB 00:38 (2/6): genisoimage-1.1.11-37.fc27.x86_64.rpm 7.9 kB/s | 311 kB 00:39 (3/6): socat-1.7.3.2-4.fc27.x86_64.rpm 7.5 kB/s | 296 kB 00:39 (4/6): libusal-1.1.11-37.fc27.x86_64.rpm 671 kB/s | 142 kB 00:00 (5/6): sudo-1.8.21p2-1.fc27.x86_64.rpm 1.1 MB/s | 771 kB 00:00 [MIRROR] ethtool-4.15-1.fc27.x86_64.rpm: Status code: 404 for http://ftp.ussg.iu.edu/linux/fedora/linux/updates/27/x86_64/Packages/e/ethtool-4.15-1.fc27.x86_64.rpm (6/6): ethtool-4.15-1.fc27.x86_64.rpm 23 kB/s | 140 kB 00:06 -------------------------------------------------------------------------------- Total 37 kB/s | 1.7 MB 00:46 Running transaction check Transaction check succeeded. Running transaction test Transaction test succeeded. Running transaction Preparing : 1/1 Installing : libusal-1.1.11-37.fc27.x86_64 1/6 Running scriptlet: libusal-1.1.11-37.fc27.x86_64 1/6 Installing : genisoimage-1.1.11-37.fc27.x86_64 2/6 Running scriptlet: genisoimage-1.1.11-37.fc27.x86_64 2/6 Installing : ethtool-2:4.15-1.fc27.x86_64 3/6 Installing : sudo-1.8.21p2-1.fc27.x86_64 4/6 Running scriptlet: sudo-1.8.21p2-1.fc27.x86_64 4/6 Installing : libcgroup-tools-0.41-13.fc27.x86_64 5/6 Running scriptlet: libcgroup-tools-0.41-13.fc27.x86_64 5/6 Installing : socat-1.7.3.2-4.fc27.x86_64 6/6 Running scriptlet: socat-1.7.3.2-4.fc27.x86_64 6/6Failed to connect to bus: No such file or directory  Verifying : socat-1.7.3.2-4.fc27.x86_64 1/6 Verifying : genisoimage-1.1.11-37.fc27.x86_64 2/6 Verifying : libcgroup-tools-0.41-13.fc27.x86_64 3/6 Verifying : sudo-1.8.21p2-1.fc27.x86_64 4/6 Verifying : libusal-1.1.11-37.fc27.x86_64 5/6 Verifying : ethtool-2:4.15-1.fc27.x86_64 6/6 Installed: ethtool.x86_64 2:4.15-1.fc27 genisoimage.x86_64 1.1.11-37.fc27 libcgroup-tools.x86_64 0.41-13.fc27 socat.x86_64 1.7.3.2-4.fc27 sudo.x86_64 1.8.21p2-1.fc27 libusal.x86_64 1.1.11-37.fc27 Complete! 18 files removed ---> c15e48082114 Removing intermediate container 44c353445827 Step 4/14 : COPY sock-connector /sock-connector ---> c894751092a9 Removing intermediate container 6d3b0f621cee Step 5/14 : COPY sh.sh /sh.sh ---> d798e2f29a04 Removing intermediate container d9d17118b04e Step 6/14 : COPY virt-launcher /virt-launcher ---> aa90384c99d3 Removing intermediate container 4d38c38afd7f Step 7/14 : COPY kubevirt-sudo /etc/sudoers.d/kubevirt ---> 1b078538b1a7 Removing intermediate container aabdb5a20181 Step 8/14 : RUN chmod 0640 /etc/sudoers.d/kubevirt ---> Running in c4e33372552f  ---> 191cecd4e558 Removing intermediate container c4e33372552f Step 9/14 : RUN rm -f /libvirtd.sh ---> Running in 49c3b2c48a30  ---> 6e136b976020 Removing intermediate container 49c3b2c48a30 Step 10/14 : COPY libvirtd.sh /libvirtd.sh ---> eafc1d5866d6 Removing intermediate container b9529c5de8dc Step 11/14 : RUN chmod a+x /libvirtd.sh ---> Running in 30ab6ea3585d  ---> 32ee0c689b9a Removing intermediate container 30ab6ea3585d Step 12/14 : COPY entrypoint.sh /entrypoint.sh ---> bd037df30382 Removing intermediate container 87cbd2ad203c Step 13/14 : ENTRYPOINT /entrypoint.sh ---> Running in d0e894ca940d ---> ddf3e89ccb87 Removing intermediate container d0e894ca940d Step 14/14 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-launcher" '' ---> Running in 9aabb2a4eb69 ---> 887c4f409482 Removing intermediate container 9aabb2a4eb69 Successfully built 887c4f409482 sending incremental file list ./ Dockerfile sent 585 bytes received 34 bytes 1238.00 bytes/sec total size is 775 speedup is 1.25 Sending build context to Docker daemon 36.37 MB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 05ea8a13b62f Step 3/5 : COPY virt-handler /virt-handler ---> b87a30c39407 Removing intermediate container f9a4f20e5c9d Step 4/5 : ENTRYPOINT /virt-handler ---> Running in bc1866ebaa14 ---> 50722b7f8642 Removing intermediate container bc1866ebaa14 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-handler" '' ---> Running in 8c793d0b5ab0 ---> 4e8511358b16 Removing intermediate container 8c793d0b5ab0 Successfully built 4e8511358b16 sending incremental file list ./ Dockerfile sent 864 bytes received 34 bytes 1796.00 bytes/sec total size is 1377 speedup is 1.53 Sending build context to Docker daemon 36.12 MB Step 1/9 : FROM fedora:27 ---> 9110ae7f579f Step 2/9 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 05ea8a13b62f Step 3/9 : RUN useradd -u 1001 --create-home -s /bin/bash virt-api ---> Running in c5fadbb27b21  ---> 5f85ee360661 Removing intermediate container c5fadbb27b21 Step 4/9 : WORKDIR /home/virt-api ---> b7eef8c6fdc3 Removing intermediate container 18de6150a6ed Step 5/9 : USER 1001 ---> Running in c7b90d63f99b ---> b7b39ab9a520 Removing intermediate container c7b90d63f99b Step 6/9 : RUN curl -OL https://github.com/swagger-api/swagger-ui/tarball/38f74164a7062edb5dc80ef2fdddda24f3f6eb85/swagger-ui.tar.gz && mkdir swagger-ui && tar xf swagger-ui.tar.gz -C swagger-ui --strip-components 1 && mkdir third_party && mv swagger-ui/dist third_party/swagger-ui && rm -rf swagger-ui && sed -e 's@"http://petstore.swagger.io/v2/swagger.json"@"/swaggerapi/"@' -i third_party/swagger-ui/index.html && rm swagger-ui.tar.gz && rm -rf swagger-ui ---> Running in 4fed67917332  % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 171 0 171 0 0 171 0 --:--:-- --:--:-- --:--:-- 380  0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0 100 239k 0 239k 0 0 119k 0 --:--:-- 0:00:02 --:--:-- 326k 100 682k 0 682k 0 0 227k  0 --:--:-- 0:00:03 --:--:-- 399k 100 1196k 0 1196k 0 0 299k 0 --:--:-- 0:00:04 --:--:-- 445k 100 1361k 0 1361k 0 0 340k 0 --:--:-- 0:00:04 --:--:-- 464k  ---> 22d8bebf6ab4 Removing intermediate container 4fed67917332 Step 7/9 : COPY virt-api /virt-api ---> e383fb24d45d Removing intermediate container d4531cd1e3bc Step 8/9 : ENTRYPOINT /virt-api ---> Running in d61eb333f62b ---> 047c5f46e918 Removing intermediate container d61eb333f62b Step 9/9 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-api" '' ---> Running in 5d8ed0468723 ---> 64572fa496ad Removing intermediate container 5d8ed0468723 Successfully built 64572fa496ad sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/iscsi-demo-target-tgtd ./ Dockerfile run-tgt.sh sent 2185 bytes received 53 bytes 4476.00 bytes/sec total size is 3992 speedup is 1.78 Sending build context to Docker daemon 6.656 kB Step 1/10 : FROM fedora:27 ---> 9110ae7f579f Step 2/10 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 05ea8a13b62f Step 3/10 : ENV container docker ---> Running in fd693cc2a25e ---> 42daee89c60c Removing intermediate container fd693cc2a25e Step 4/10 : RUN dnf -y install scsi-target-utils bzip2 e2fsprogs ---> Running in 21420a412882  Fedora 27 - x86_64 - Updates 1.0 MB/s | 22 MB 00:22 Fedora 27 - x86_64 1.1 MB/s | 58 MB 00:53 Last metadata expiration check: 0:00:27 ago on Fri Apr 6 13:04:59 2018. Dependencies resolved. ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: bzip2 x86_64 1.0.6-24.fc27 fedora 58 k e2fsprogs x86_64 1.43.5-2.fc27 fedora 1.0 M scsi-target-utils x86_64 1.0.70-4.fc27 fedora 215 k Installing dependencies: GeoIP-GeoLite-data noarch 2018.04-1.fc27 updates 551 k cpio x86_64 2.12-6.fc27 fedora 262 k dracut x86_64 046-8.git20180105.fc27 updates 347 k e2fsprogs-libs x86_64 1.43.5-2.fc27 fedora 197 k findutils x86_64 1:4.6.0-16.fc27 updates 523 k fuse-libs x86_64 2.9.7-9.fc27 fedora 99 k gettext x86_64 0.19.8.1-12.fc27 updates 1.1 M gettext-libs x86_64 0.19.8.1-12.fc27 updates 303 k groff-base x86_64 1.22.3-14.fc27 updates 1.0 M grub2-common noarch 1:2.02-22.fc27 updates 867 k grub2-tools-minimal x86_64 1:2.02-22.fc27 updates 174 k hostname x86_64 3.18-4.fc27 fedora 28 k hwdata noarch 0.311-1.fc27 updates 1.5 M initscripts x86_64 9.79-1.fc27 updates 394 k ipcalc x86_64 0.2.0-3.fc27 fedora 35 k iproute x86_64 4.15.0-1.fc27 updates 534 k iputils x86_64 20161105-7.fc27 fedora 157 k kbd-legacy noarch 2.0.4-4.fc27 fedora 479 k kbd-misc noarch 2.0.4-4.fc27 fedora 1.4 M kmod x86_64 25-1.fc27 updates 115 k libaio x86_64 0.3.110-9.fc27 fedora 29 k libcroco x86_64 0.6.12-3.fc27 fedora 110 k libgomp x86_64 7.3.1-5.fc27 updates 211 k libibverbs x86_64 14-4.fc27 fedora 226 k libmnl x86_64 1.0.4-4.fc27 fedora 28 k libnl3 x86_64 3.4.0-1.fc27 updates 304 k librdmacm x86_64 14-4.fc27 fedora 63 k libss x86_64 1.43.5-2.fc27 fedora 50 k libstdc++ x86_64 7.3.1-5.fc27 updates 482 k linux-atm-libs x86_64 2.5.1-19.fc27 fedora 40 k lsof x86_64 4.89-7.fc27 fedora 246 k pciutils x86_64 3.5.5-3.fc27 fedora 102 k pciutils-libs x86_64 3.5.5-3.fc27 fedora 51 k perl-Carp noarch 1.42-394.fc27 fedora 28 k perl-Config-General noarch 2.63-4.fc27 fedora 79 k perl-Encode x86_64 4:2.94-16.fc27 updates 1.5 M perl-Errno x86_64 1.28-403.fc27 updates 72 k perl-Exporter noarch 5.72-395.fc27 fedora 32 k perl-File-Path noarch 2.15-1.fc27 fedora 37 k perl-File-Temp noarch 0.230.400-394.fc27 fedora 61 k perl-Getopt-Long noarch 1:2.50-3.fc27 fedora 61 k perl-HTTP-Tiny noarch 0.070-394.fc27 fedora 56 k perl-IO x86_64 1.38-403.fc27 updates 138 k perl-MIME-Base64 x86_64 3.15-395.fc27 fedora 29 k perl-PathTools x86_64 3.74-1.fc27 updates 88 k perl-Pod-Escapes noarch 1:1.07-394.fc27 fedora 19 k perl-Pod-Perldoc noarch 3.28-395.fc27 fedora 87 k perl-Pod-Simple noarch 1:3.35-394.fc27 fedora 211 k perl-Pod-Usage noarch 4:1.69-394.fc27 fedora 33 k perl-Scalar-List-Utils x86_64 3:1.48-1.fc27 fedora 65 k perl-Socket x86_64 4:2.027-1.fc27 updates 57 k perl-Storable x86_64 1:2.62-395.fc27 fedora 84 k perl-Term-ANSIColor noarch 4.06-395.fc27 fedora 44 k perl-Term-Cap noarch 1.17-394.fc27 fedora 21 k perl-Text-ParseWords noarch 3.30-394.fc27 fedora 16 k perl-Text-Tabs+Wrap noarch 2013.0523-394.fc27 fedora 23 k perl-Time-Local noarch 1:1.250-394.fc27 fedora 30 k perl-Unicode-Normalize x86_64 1.25-395.fc27 fedora 80 k perl-constant noarch 1.33-395.fc27 fedora 24 k perl-interpreter x86_64 4:5.26.1-403.fc27 updates 6.2 M perl-libs x86_64 4:5.26.1-403.fc27 updates 1.5 M perl-macros x86_64 4:5.26.1-403.fc27 updates 68 k perl-parent noarch 1:0.236-394.fc27 fedora 18 k perl-podlators noarch 4.09-395.fc27 updates 115 k perl-threads x86_64 1:2.21-1.fc27 updates 60 k perl-threads-shared x86_64 1.58-1.fc27 updates 46 k procps-ng x86_64 3.3.10-15.fc27 fedora 395 k rdma-core x86_64 14-4.fc27 fedora 46 k sg3_utils x86_64 1.42-3.fc27 fedora 794 k sg3_utils-libs x86_64 1.42-3.fc27 fedora 75 k systemd-udev x86_64 234-10.git5f8984e.fc27 updates 1.3 M xz x86_64 5.2.3-4.fc27 fedora 150 k Installing weak dependencies: GeoIP x86_64 1.6.11-3.fc27 fedora 124 k grubby x86_64 8.40-8.fc27 updates 70 k hardlink x86_64 1:1.3-3.fc27 fedora 27 k iproute-tc x86_64 4.15.0-1.fc27 updates 389 k kbd x86_64 2.0.4-4.fc27 fedora 385 k kpartx x86_64 0.7.1-9.git847cc43.fc27 updates 87 k perl-IO-Socket-IP noarch 0.39-4.fc27 fedora 45 k perl-Mozilla-CA noarch 20160104-6.fc27 fedora 14 k pigz x86_64 2.4-1.fc27 updates 76 k systemd-bootchart x86_64 233-1.fc27 fedora 67 k Transaction Summary ================================================================================ Install 85 Packages Total download size: 28 M Installed size: 89 M Downloading Packages: (1/85): bzip2-1.0.6-24.fc27.x86_64.rpm 38 kB/s | 58 kB 00:01 (2/85): libaio-0.3.110-9.fc27.x86_64.rpm 243 kB/s | 29 kB 00:00 (3/85): scsi-target-utils-1.0.70-4.fc27.x86_64. 126 kB/s | 215 kB 00:01 (4/85): librdmacm-14-4.fc27.x86_64.rpm 545 kB/s | 63 kB 00:00 (5/85): libibverbs-14-4.fc27.x86_64.rpm 770 kB/s | 226 kB 00:00 (6/85): e2fsprogs-1.43.5-2.fc27.x86_64.rpm 519 kB/s | 1.0 MB 00:02 (7/85): lsof-4.89-7.fc27.x86_64.rpm 1.0 MB/s | 246 kB 00:00 (8/85): perl-Config-General-2.63-4.fc27.noarch. 351 kB/s | 79 kB 00:00 (9/85): perl-Getopt-Long-2.50-3.fc27.noarch.rpm 530 kB/s | 61 kB 00:00 (10/85): fuse-libs-2.9.7-9.fc27.x86_64.rpm 813 kB/s | 99 kB 00:00 (11/85): e2fsprogs-libs-1.43.5-2.fc27.x86_64.rp 1.2 MB/s | 197 kB 00:00 (12/85): sg3_utils-1.42-3.fc27.x86_64.rpm 2.3 MB/s | 794 kB 00:00 (13/85): libss-1.43.5-2.fc27.x86_64.rpm 401 kB/s | 50 kB 00:00 (14/85): rdma-core-14-4.fc27.x86_64.rpm 240 kB/s | 46 kB 00:00 (15/85): perl-Carp-1.42-394.fc27.noarch.rpm 262 kB/s | 28 kB 00:00 (16/85): perl-Exporter-5.72-395.fc27.noarch.rpm 169 kB/s | 32 kB 00:00 (17/85): perl-constant-1.33-395.fc27.noarch.rpm 221 kB/s | 24 kB 00:00 (18/85): perl-Pod-Usage-1.69-394.fc27.noarch.rp 170 kB/s | 33 kB 00:00 (19/85): perl-Text-ParseWords-3.30-394.fc27.noa 151 kB/s | 16 kB 00:00 (20/85): sg3_utils-libs-1.42-3.fc27.x86_64.rpm 382 kB/s | 75 kB 00:00 (21/85): pciutils-3.5.5-3.fc27.x86_64.rpm 822 kB/s | 102 kB 00:00 (22/85): perl-Pod-Perldoc-3.28-395.fc27.noarch. 719 kB/s | 87 kB 00:00 (23/85): pciutils-libs-3.5.5-3.fc27.x86_64.rpm 414 kB/s | 51 kB 00:00 (24/85): perl-File-Temp-0.230.400-394.fc27.noar 531 kB/s | 61 kB 00:00 (25/85): perl-HTTP-Tiny-0.070-394.fc27.noarch.r 488 kB/s | 56 kB 00:00 (26/85): perl-parent-0.236-394.fc27.noarch.rpm 175 kB/s | 18 kB 00:00 (27/85): perl-File-Path-2.15-1.fc27.noarch.rpm 332 kB/s | 37 kB 00:00 (28/85): perl-Pod-Simple-3.35-394.fc27.noarch.r 936 kB/s | 211 kB 00:00 (29/85): perl-MIME-Base64-3.15-395.fc27.x86_64. 273 kB/s | 29 kB 00:00 (30/85): perl-Scalar-List-Utils-1.48-1.fc27.x86 318 kB/s | 65 kB 00:00 (31/85): perl-Time-Local-1.250-394.fc27.noarch. 275 kB/s | 30 kB 00:00 (32/85): perl-Pod-Escapes-1.07-394.fc27.noarch. 180 kB/s | 19 kB 00:00 (33/85): perl-Text-Tabs+Wrap-2013.0523-394.fc27 209 kB/s | 23 kB 00:00 (34/85): perl-Errno-1.28-403.fc27.x86_64.rpm 101 kB/s | 72 kB 00:00 (35/85): systemd-udev-234-10.git5f8984e.fc27.x8 698 kB/s | 1.3 MB 00:01 (36/85): perl-Unicode-Normalize-1.25-395.fc27.x 721 kB/s | 80 kB 00:00 (37/85): dracut-046-8.git20180105.fc27.x86_64.r 1.1 MB/s | 347 kB 00:00 (38/85): cpio-2.12-6.fc27.x86_64.rpm 1.1 MB/s | 262 kB 00:00 (39/85): procps-ng-3.3.10-15.fc27.x86_64.rpm 2.3 MB/s | 395 kB 00:00 (40/85): xz-5.2.3-4.fc27.x86_64.rpm 1.1 MB/s | 150 kB 00:00 (41/85): perl-libs-5.26.1-403.fc27.x86_64.rpm 555 kB/s | 1.5 MB 00:02 (42/85): hostname-3.18-4.fc27.x86_64.rpm 271 kB/s | 28 kB 00:00 (43/85): ipcalc-0.2.0-3.fc27.x86_64.rpm 258 kB/s | 35 kB 00:00 (44/85): iputils-20161105-7.fc27.x86_64.rpm 1.1 MB/s | 157 kB 00:00 (45/85): initscripts-9.79-1.fc27.x86_64.rpm 693 kB/s | 394 kB 00:00 (46/85): perl-PathTools-3.74-1.fc27.x86_64.rpm 402 kB/s | 88 kB 00:00 (47/85): kmod-25-1.fc27.x86_64.rpm 381 kB/s | 115 kB 00:00 (48/85): perl-Term-ANSIColor-4.06-395.fc27.noar 425 kB/s | 44 kB 00:00 (49/85): perl-podlators-4.09-395.fc27.noarch.rp 692 kB/s | 115 kB 00:00 (50/85): perl-Term-Cap-1.17-394.fc27.noarch.rpm 159 kB/s | 21 kB 00:00 (51/85): perl-Storable-2.62-395.fc27.x86_64.rpm 595 kB/s | 84 kB 00:00 (52/85): perl-Encode-2.94-16.fc27.x86_64.rpm 1.2 MB/s | 1.5 MB 00:01 (53/85): perl-IO-1.38-403.fc27.x86_64.rpm 880 kB/s | 138 kB 00:00 (54/85): perl-Socket-2.027-1.fc27.x86_64.rpm 329 kB/s | 57 kB 00:00 (55/85): groff-base-1.22.3-14.fc27.x86_64.rpm 654 kB/s | 1.0 MB 00:01 (56/85): libnl3-3.4.0-1.fc27.x86_64.rpm 681 kB/s | 304 kB 00:00 (57/85): perl-interpreter-5.26.1-403.fc27.x86_6 1.1 MB/s | 6.2 MB 00:05 (58/85): perl-macros-5.26.1-403.fc27.x86_64.rpm 468 kB/s | 68 kB 00:00 (59/85): perl-threads-2.21-1.fc27.x86_64.rpm 407 kB/s | 60 kB 00:00 (60/85): perl-threads-shared-1.58-1.fc27.x86_64 258 kB/s | 46 kB 00:00 (61/85): hwdata-0.311-1.fc27.noarch.rpm 1.3 MB/s | 1.5 MB 00:01 (62/85): libmnl-1.0.4-4.fc27.x86_64.rpm 211 kB/s | 28 kB 00:00 (63/85): iproute-4.15.0-1.fc27.x86_64.rpm 1.0 MB/s | 534 kB 00:00 (64/85): perl-Mozilla-CA-20160104-6.fc27.noarch 110 kB/s | 14 kB 00:00 (65/85): findutils-4.6.0-16.fc27.x86_64.rpm 706 kB/s | 523 kB 00:00 (66/85): libstdc++-7.3.1-5.fc27.x86_64.rpm 999 kB/s | 482 kB 00:00 (67/85): perl-IO-Socket-IP-0.39-4.fc27.noarch.r 294 kB/s | 45 kB 00:00 (68/85): kbd-2.0.4-4.fc27.x86_64.rpm 1.9 MB/s | 385 kB 00:00 (69/85): kbd-legacy-2.0.4-4.fc27.noarch.rpm 2.4 MB/s | 479 kB 00:00 (70/85): systemd-bootchart-233-1.fc27.x86_64.rp 576 kB/s | 67 kB 00:00 (71/85): kbd-misc-2.0.4-4.fc27.noarch.rpm 4.4 MB/s | 1.4 MB 00:00 (72/85): kpartx-0.7.1-9.git847cc43.fc27.x86_64. 421 kB/s | 87 kB 00:00 (73/85): hardlink-1.3-3.fc27.x86_64.rpm 258 kB/s | 27 kB 00:00 (74/85): grubby-8.40-8.fc27.x86_64.rpm 201 kB/s | 70 kB 00:00 (75/85): GeoIP-1.6.11-3.fc27.x86_64.rpm 921 kB/s | 124 kB 00:00 (76/85): pigz-2.4-1.fc27.x86_64.rpm 201 kB/s | 76 kB 00:00 (77/85): linux-atm-libs-2.5.1-19.fc27.x86_64.rp 370 kB/s | 40 kB 00:00 (78/85): iproute-tc-4.15.0-1.fc27.x86_64.rpm 907 kB/s | 389 kB 00:00 (79/85): grub2-tools-minimal-2.02-22.fc27.x86_6 423 kB/s | 174 kB 00:00 (80/85): GeoIP-GeoLite-data-2018.04-1.fc27.noar 772 kB/s | 551 kB 00:00 (81/85): gettext-libs-0.19.8.1-12.fc27.x86_64.r 515 kB/s | 303 kB 00:00 (82/85): libcroco-0.6.12-3.fc27.x86_64.rpm 898 kB/s | 110 kB 00:00 (83/85): libgomp-7.3.1-5.fc27.x86_64.rpm 637 kB/s | 211 kB 00:00 (84/85): gettext-0.19.8.1-12.fc27.x86_64.rpm 722 kB/s | 1.1 MB 00:01 (85/85): grub2-common-2.02-22.fc27.noarch.rpm 526 kB/s | 867 kB 00:01 -------------------------------------------------------------------------------- Total 1.9 MB/s | 28 MB 00:14 Running transaction check Transaction check succeeded. Running transaction test Transaction test succeeded. Running transaction Preparing : 1/1 Installing : perl-Exporter-5.72-395.fc27.noarch 1/85 Installing : perl-libs-4:5.26.1-403.fc27.x86_64 2/85 Running scriptlet: perl-libs-4:5.26.1-403.fc27.x86_64 2/85 Installing : perl-Carp-1.42-394.fc27.noarch 3/85 Installing : perl-Scalar-List-Utils-3:1.48-1.fc27.x86_64 4/85 Installing : kmod-25-1.fc27.x86_64 5/85 Installing : systemd-udev-234-10.git5f8984e.fc27.x86_64 6/85 Running scriptlet: systemd-udev-234-10.git5f8984e.fc27.x86_64 6/85 Installing : perl-Text-ParseWords-3.30-394.fc27.noarch 7/85 Installing : libgomp-7.3.1-5.fc27.x86_64 8/85 Running scriptlet: libgomp-7.3.1-5.fc27.x86_64 8/85 Installing : libcroco-0.6.12-3.fc27.x86_64 9/85 Running scriptlet: libcroco-0.6.12-3.fc27.x86_64 9/85 Installing : libstdc++-7.3.1-5.fc27.x86_64 10/85 Running scriptlet: libstdc++-7.3.1-5.fc27.x86_64 10/85 Installing : libmnl-1.0.4-4.fc27.x86_64 11/85 Running scriptlet: libmnl-1.0.4-4.fc27.x86_64 11/85 Installing : iproute-4.15.0-1.fc27.x86_64 12/85 Installing : findutils-1:4.6.0-16.fc27.x86_64 13/85 Running scriptlet: findutils-1:4.6.0-16.fc27.x86_64 13/85 Installing : procps-ng-3.3.10-15.fc27.x86_64 14/85 Running scriptlet: procps-ng-3.3.10-15.fc27.x86_64 14/85 Installing : cpio-2.12-6.fc27.x86_64 15/85 Running scriptlet: cpio-2.12-6.fc27.x86_64 15/85 Installing : groff-base-1.22.3-14.fc27.x86_64 16/85 Installing : gettext-libs-0.19.8.1-12.fc27.x86_64 17/85 Running scriptlet: gettext-libs-0.19.8.1-12.fc27.x86_64 17/85 Installing : gettext-0.19.8.1-12.fc27.x86_64 18/85 Running scriptlet: gettext-0.19.8.1-12.fc27.x86_64 18/85 install-info: No such file or directory for /usr/share/info/gettext.info.gz Installing : perl-Term-ANSIColor-4.06-395.fc27.noarch 19/85 Installing : perl-macros-4:5.26.1-403.fc27.x86_64 20/85 Installing : perl-constant-1.33-395.fc27.noarch 21/85 Installing : perl-parent-1:0.236-394.fc27.noarch 22/85 Installing : perl-Text-Tabs+Wrap-2013.0523-394.fc27.noarch 23/85 Installing : perl-File-Path-2.15-1.fc27.noarch 24/85 Installing : perl-PathTools-3.74-1.fc27.x86_64 25/85 Installing : perl-Errno-1.28-403.fc27.x86_64 26/85 Installing : perl-Unicode-Normalize-1.25-395.fc27.x86_64 27/85 Installing : perl-IO-1.38-403.fc27.x86_64 28/85 Installing : perl-Socket-4:2.027-1.fc27.x86_64 29/85 Installing : perl-threads-1:2.21-1.fc27.x86_64 30/85 Installing : perl-threads-shared-1.58-1.fc27.x86_64 31/85 Installing : perl-interpreter-4:5.26.1-403.fc27.x86_64 32/85 Installing : perl-MIME-Base64-3.15-395.fc27.x86_64 33/85 Installing : perl-Config-General-2.63-4.fc27.noarch 34/85 Installing : perl-File-Temp-0.230.400-394.fc27.noarch 35/85 Installing : perl-Pod-Escapes-1:1.07-394.fc27.noarch 36/85 Installing : perl-Term-Cap-1.17-394.fc27.noarch 37/85 Installing : perl-Storable-1:2.62-395.fc27.x86_64 38/85 Installing : perl-Time-Local-1:1.250-394.fc27.noarch 39/85 Installing : perl-HTTP-Tiny-0.070-394.fc27.noarch 40/85 Installing : perl-Pod-Simple-1:3.35-394.fc27.noarch 41/85 Installing : perl-Pod-Usage-4:1.69-394.fc27.noarch 42/85 Installing : perl-Getopt-Long-1:2.50-3.fc27.noarch 43/85 Installing : perl-Encode-4:2.94-16.fc27.x86_64 44/85 Installing : perl-podlators-4.09-395.fc27.noarch 45/85 Installing : perl-Pod-Perldoc-3.28-395.fc27.noarch 46/85 Installing : grub2-common-1:2.02-22.fc27.noarch 47/85 Installing : grub2-tools-minimal-1:2.02-22.fc27.x86_64 48/85 Installing : GeoIP-GeoLite-data-2018.04-1.fc27.noarch 49/85 Installing : linux-atm-libs-2.5.1-19.fc27.x86_64 50/85 Running scriptlet: linux-atm-libs-2.5.1-19.fc27.x86_64 50/85 Installing : kbd-misc-2.0.4-4.fc27.noarch 51/85 Installing : kbd-legacy-2.0.4-4.fc27.noarch 52/85 Installing : libnl3-3.4.0-1.fc27.x86_64 53/85 Running scriptlet: libnl3-3.4.0-1.fc27.x86_64 53/85 Installing : hwdata-0.311-1.fc27.noarch 54/85 Installing : iputils-20161105-7.fc27.x86_64 55/85 Running scriptlet: iputils-20161105-7.fc27.x86_64 55/85 Installing : ipcalc-0.2.0-3.fc27.x86_64 56/85 Installing : hostname-3.18-4.fc27.x86_64 57/85 Installing : initscripts-9.79-1.fc27.x86_64 58/85 Running scriptlet: initscripts-9.79-1.fc27.x86_64 58/85 warning: /etc/adjtime created as /etc/adjtime.rpmnew Installing : xz-5.2.3-4.fc27.x86_64 59/85 Installing : dracut-046-8.git20180105.fc27.x86_64 60/85 Installing : pciutils-libs-3.5.5-3.fc27.x86_64 61/85 Running scriptlet: pciutils-libs-3.5.5-3.fc27.x86_64 61/85 Installing : pciutils-3.5.5-3.fc27.x86_64 62/85 Installing : rdma-core-14-4.fc27.x86_64 63/85 Installing : libibverbs-14-4.fc27.x86_64 64/85 Running scriptlet: libibverbs-14-4.fc27.x86_64 64/85 Installing : librdmacm-14-4.fc27.x86_64 65/85 Running scriptlet: librdmacm-14-4.fc27.x86_64 65/85 Installing : sg3_utils-libs-1.42-3.fc27.x86_64 66/85 Running scriptlet: sg3_utils-libs-1.42-3.fc27.x86_64 66/85 Installing : sg3_utils-1.42-3.fc27.x86_64 67/85 Installing : libss-1.43.5-2.fc27.x86_64 68/85 Running scriptlet: libss-1.43.5-2.fc27.x86_64 68/85 Installing : fuse-libs-2.9.7-9.fc27.x86_64 69/85 Running scriptlet: fuse-libs-2.9.7-9.fc27.x86_64 69/85 Installing : e2fsprogs-libs-1.43.5-2.fc27.x86_64 70/85 Running scriptlet: e2fsprogs-libs-1.43.5-2.fc27.x86_64 70/85 Installing : lsof-4.89-7.fc27.x86_64 71/85 Installing : libaio-0.3.110-9.fc27.x86_64 72/85 Running scriptlet: libaio-0.3.110-9.fc27.x86_64 72/85 Installing : scsi-target-utils-1.0.70-4.fc27.x86_64 73/85 Running scriptlet: scsi-target-utils-1.0.70-4.fc27.x86_64 73/85 Installing : e2fsprogs-1.43.5-2.fc27.x86_64 74/85 Installing : kbd-2.0.4-4.fc27.x86_64 75/85 Installing : iproute-tc-4.15.0-1.fc27.x86_64 76/85 Installing : GeoIP-1.6.11-3.fc27.x86_64 77/85 Running scriptlet: GeoIP-1.6.11-3.fc27.x86_64 77/85 Installing : grubby-8.40-8.fc27.x86_64 78/85 Installing : perl-IO-Socket-IP-0.39-4.fc27.noarch 79/85 Installing : perl-Mozilla-CA-20160104-6.fc27.noarch 80/85 Installing : hardlink-1:1.3-3.fc27.x86_64 81/85 Installing : pigz-2.4-1.fc27.x86_64 82/85 Installing : kpartx-0.7.1-9.git847cc43.fc27.x86_64 83/85 Installing : systemd-bootchart-233-1.fc27.x86_64 84/85 Running scriptlet: systemd-bootchart-233-1.fc27.x86_64 84/85 Installing : bzip2-1.0.6-24.fc27.x86_64 85/85 Running scriptlet: GeoIP-GeoLite-data-2018.04-1.fc27.noarch 85/85 Running scriptlet: bzip2-1.0.6-24.fc27.x86_64 85/85Failed to connect to bus: No such file or directory  Verifying : scsi-target-utils-1.0.70-4.fc27.x86_64 1/85 Verifying : bzip2-1.0.6-24.fc27.x86_64 2/85 Verifying : e2fsprogs-1.43.5-2.fc27.x86_64 3/85 Verifying : libaio-0.3.110-9.fc27.x86_64 4/85 Verifying : libibverbs-14-4.fc27.x86_64 5/85 Verifying : librdmacm-14-4.fc27.x86_64 6/85 Verifying : lsof-4.89-7.fc27.x86_64 7/85 Verifying : perl-Config-General-2.63-4.fc27.noarch 8/85 Verifying : perl-Getopt-Long-1:2.50-3.fc27.noarch 9/85 Verifying : sg3_utils-1.42-3.fc27.x86_64 10/85 Verifying : e2fsprogs-libs-1.43.5-2.fc27.x86_64 11/85 Verifying : fuse-libs-2.9.7-9.fc27.x86_64 12/85 Verifying : libss-1.43.5-2.fc27.x86_64 13/85 Verifying : rdma-core-14-4.fc27.x86_64 14/85 Verifying : perl-Carp-1.42-394.fc27.noarch 15/85 Verifying : perl-Exporter-5.72-395.fc27.noarch 16/85 Verifying : perl-constant-1.33-395.fc27.noarch 17/85 Verifying : perl-Pod-Usage-4:1.69-394.fc27.noarch 18/85 Verifying : perl-Text-ParseWords-3.30-394.fc27.noarch 19/85 Verifying : sg3_utils-libs-1.42-3.fc27.x86_64 20/85 Verifying : pciutils-3.5.5-3.fc27.x86_64 21/85 Verifying : perl-Pod-Perldoc-3.28-395.fc27.noarch 22/85 Verifying : pciutils-libs-3.5.5-3.fc27.x86_64 23/85 Verifying : perl-File-Temp-0.230.400-394.fc27.noarch 24/85 Verifying : perl-HTTP-Tiny-0.070-394.fc27.noarch 25/85 Verifying : perl-Pod-Simple-1:3.35-394.fc27.noarch 26/85 Verifying : perl-parent-1:0.236-394.fc27.noarch 27/85 Verifying : perl-File-Path-2.15-1.fc27.noarch 28/85 Verifying : perl-Scalar-List-Utils-3:1.48-1.fc27.x86_64 29/85 Verifying : perl-MIME-Base64-3.15-395.fc27.x86_64 30/85 Verifying : perl-Time-Local-1:1.250-394.fc27.noarch 31/85 Verifying : perl-Pod-Escapes-1:1.07-394.fc27.noarch 32/85 Verifying : perl-Text-Tabs+Wrap-2013.0523-394.fc27.noarch 33/85 Verifying : systemd-udev-234-10.git5f8984e.fc27.x86_64 34/85 Verifying : perl-Errno-1.28-403.fc27.x86_64 35/85 Verifying : perl-libs-4:5.26.1-403.fc27.x86_64 36/85 Verifying : perl-interpreter-4:5.26.1-403.fc27.x86_64 37/85 Verifying : perl-Unicode-Normalize-1.25-395.fc27.x86_64 38/85 Verifying : dracut-046-8.git20180105.fc27.x86_64 39/85 Verifying : cpio-2.12-6.fc27.x86_64 40/85 Verifying : procps-ng-3.3.10-15.fc27.x86_64 41/85 Verifying : xz-5.2.3-4.fc27.x86_64 42/85 Verifying : initscripts-9.79-1.fc27.x86_64 43/85 Verifying : hostname-3.18-4.fc27.x86_64 44/85 Verifying : ipcalc-0.2.0-3.fc27.x86_64 45/85 Verifying : iputils-20161105-7.fc27.x86_64 46/85 Verifying : kmod-25-1.fc27.x86_64 47/85 Verifying : perl-PathTools-3.74-1.fc27.x86_64 48/85 Verifying : perl-podlators-4.09-395.fc27.noarch 49/85 Verifying : perl-Term-ANSIColor-4.06-395.fc27.noarch 50/85 Verifying : perl-Term-Cap-1.17-394.fc27.noarch 51/85 Verifying : perl-Encode-4:2.94-16.fc27.x86_64 52/85 Verifying : perl-Storable-1:2.62-395.fc27.x86_64 53/85 Verifying : groff-base-1.22.3-14.fc27.x86_64 54/85 Verifying : perl-IO-1.38-403.fc27.x86_64 55/85 Verifying : perl-Socket-4:2.027-1.fc27.x86_64 56/85 Verifying : hwdata-0.311-1.fc27.noarch 57/85 Verifying : libnl3-3.4.0-1.fc27.x86_64 58/85 Verifying : perl-macros-4:5.26.1-403.fc27.x86_64 59/85 Verifying : perl-threads-1:2.21-1.fc27.x86_64 60/85 Verifying : perl-threads-shared-1.58-1.fc27.x86_64 61/85 Verifying : findutils-1:4.6.0-16.fc27.x86_64 62/85 Verifying : iproute-4.15.0-1.fc27.x86_64 63/85 Verifying : libmnl-1.0.4-4.fc27.x86_64 64/85 Verifying : libstdc++-7.3.1-5.fc27.x86_64 65/85 Verifying : perl-Mozilla-CA-20160104-6.fc27.noarch 66/85 Verifying : perl-IO-Socket-IP-0.39-4.fc27.noarch 67/85 Verifying : kbd-2.0.4-4.fc27.x86_64 68/85 Verifying : kbd-legacy-2.0.4-4.fc27.noarch 69/85 Verifying : kbd-misc-2.0.4-4.fc27.noarch 70/85 Verifying : systemd-bootchart-233-1.fc27.x86_64 71/85 Verifying : kpartx-0.7.1-9.git847cc43.fc27.x86_64 72/85 Verifying : grubby-8.40-8.fc27.x86_64 73/85 Verifying : pigz-2.4-1.fc27.x86_64 74/85 Verifying : hardlink-1:1.3-3.fc27.x86_64 75/85 Verifying : GeoIP-1.6.11-3.fc27.x86_64 76/85 Verifying : iproute-tc-4.15.0-1.fc27.x86_64 77/85 Verifying : linux-atm-libs-2.5.1-19.fc27.x86_64 78/85 Verifying : GeoIP-GeoLite-data-2018.04-1.fc27.noarch 79/85 Verifying : grub2-tools-minimal-1:2.02-22.fc27.x86_64 80/85 Verifying : grub2-common-1:2.02-22.fc27.noarch 81/85 Verifying : gettext-0.19.8.1-12.fc27.x86_64 82/85 Verifying : gettext-libs-0.19.8.1-12.fc27.x86_64 83/85 Verifying : libcroco-0.6.12-3.fc27.x86_64 84/85 Verifying : libgomp-7.3.1-5.fc27.x86_64 85/85 Installed: bzip2.x86_64 1.0.6-24.fc27 e2fsprogs.x86_64 1.43.5-2.fc27 scsi-target-utils.x86_64 1.0.70-4.fc27 GeoIP.x86_64 1.6.11-3.fc27 grubby.x86_64 8.40-8.fc27 hardlink.x86_64 1:1.3-3.fc27 iproute-tc.x86_64 4.15.0-1.fc27 kbd.x86_64 2.0.4-4.fc27 kpartx.x86_64 0.7.1-9.git847cc43.fc27 perl-IO-Socket-IP.noarch 0.39-4.fc27 perl-Mozilla-CA.noarch 20160104-6.fc27 pigz.x86_64 2.4-1.fc27 systemd-bootchart.x86_64 233-1.fc27 GeoIP-GeoLite-data.noarch 2018.04-1.fc27 cpio.x86_64 2.12-6.fc27 dracut.x86_64 046-8.git20180105.fc27 e2fsprogs-libs.x86_64 1.43.5-2.fc27 findutils.x86_64 1:4.6.0-16.fc27 fuse-libs.x86_64 2.9.7-9.fc27 gettext.x86_64 0.19.8.1-12.fc27 gettext-libs.x86_64 0.19.8.1-12.fc27 groff-base.x86_64 1.22.3-14.fc27 grub2-common.noarch 1:2.02-22.fc27 grub2-tools-minimal.x86_64 1:2.02-22.fc27 hostname.x86_64 3.18-4.fc27 hwdata.noarch 0.311-1.fc27 initscripts.x86_64 9.79-1.fc27 ipcalc.x86_64 0.2.0-3.fc27 iproute.x86_64 4.15.0-1.fc27 iputils.x86_64 20161105-7.fc27 kbd-legacy.noarch 2.0.4-4.fc27 kbd-misc.noarch 2.0.4-4.fc27 kmod.x86_64 25-1.fc27 libaio.x86_64 0.3.110-9.fc27 libcroco.x86_64 0.6.12-3.fc27 libgomp.x86_64 7.3.1-5.fc27 libibverbs.x86_64 14-4.fc27 libmnl.x86_64 1.0.4-4.fc27 libnl3.x86_64 3.4.0-1.fc27 librdmacm.x86_64 14-4.fc27 libss.x86_64 1.43.5-2.fc27 libstdc++.x86_64 7.3.1-5.fc27 linux-atm-libs.x86_64 2.5.1-19.fc27 lsof.x86_64 4.89-7.fc27 pciutils.x86_64 3.5.5-3.fc27 pciutils-libs.x86_64 3.5.5-3.fc27 perl-Carp.noarch 1.42-394.fc27 perl-Config-General.noarch 2.63-4.fc27 perl-Encode.x86_64 4:2.94-16.fc27 perl-Errno.x86_64 1.28-403.fc27 perl-Exporter.noarch 5.72-395.fc27 perl-File-Path.noarch 2.15-1.fc27 perl-File-Temp.noarch 0.230.400-394.fc27 perl-Getopt-Long.noarch 1:2.50-3.fc27 perl-HTTP-Tiny.noarch 0.070-394.fc27 perl-IO.x86_64 1.38-403.fc27 perl-MIME-Base64.x86_64 3.15-395.fc27 perl-PathTools.x86_64 3.74-1.fc27 perl-Pod-Escapes.noarch 1:1.07-394.fc27 perl-Pod-Perldoc.noarch 3.28-395.fc27 perl-Pod-Simple.noarch 1:3.35-394.fc27 perl-Pod-Usage.noarch 4:1.69-394.fc27 perl-Scalar-List-Utils.x86_64 3:1.48-1.fc27 perl-Socket.x86_64 4:2.027-1.fc27 perl-Storable.x86_64 1:2.62-395.fc27 perl-Term-ANSIColor.noarch 4.06-395.fc27 perl-Term-Cap.noarch 1.17-394.fc27 perl-Text-ParseWords.noarch 3.30-394.fc27 perl-Text-Tabs+Wrap.noarch 2013.0523-394.fc27 perl-Time-Local.noarch 1:1.250-394.fc27 perl-Unicode-Normalize.x86_64 1.25-395.fc27 perl-constant.noarch 1.33-395.fc27 perl-interpreter.x86_64 4:5.26.1-403.fc27 perl-libs.x86_64 4:5.26.1-403.fc27 perl-macros.x86_64 4:5.26.1-403.fc27 perl-parent.noarch 1:0.236-394.fc27 perl-podlators.noarch 4.09-395.fc27 perl-threads.x86_64 1:2.21-1.fc27 perl-threads-shared.x86_64 1.58-1.fc27 procps-ng.x86_64 3.3.10-15.fc27 rdma-core.x86_64 14-4.fc27 sg3_utils.x86_64 1.42-3.fc27 sg3_utils-libs.x86_64 1.42-3.fc27 systemd-udev.x86_64 234-10.git5f8984e.fc27 xz.x86_64 5.2.3-4.fc27 Complete! ---> 7201f06ab604 Removing intermediate container 21420a412882 Step 5/10 : RUN mkdir -p /images ---> Running in 96a5b4080954  ---> 8e3fc5e02ef5 Removing intermediate container 96a5b4080954 Step 6/10 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /images/1-alpine.img ---> Running in 44e5b5dcf1aa  % Total % Received % Xferd Average Speed Time  Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- 0:00:02 --:--:-- 0 0 37.0M 0 316k 0 0 105k 0 0:05:58 0:00:03 0:05:55 93258 2 37.0M 2 766k 0 0 191k 0 0:03:17 0:00:04 0:03:13 171k 3 37.0M 3 1308k 0 0 261k 0 0:02:24 0:00:05 0:02:19 262k 4 37.0M 4 1856k 0 0 309k  0 0:02:02 0:00:06 0:01:56 373k 6 37.0M 6 2435k 0 0 347k 0 0:01:48 0:00:07 0:01:41 491k 7 37.0M 7 2996k 0 0 374k 0 0:01:41 0:00:08 0:01:33 535k 9 37.0M 9 3593k 0 0 399k 0 0:01:34 0:00:09 0:01:25 565k 11 37.0M 11 4297k 0 0 429k 0 0:01:28 0:00:10 0:01:18 601k 13 37.0M 13 5098k 0 0 463k 0 0:01:21 0:00:11 0:01:10 649k 15 37.0M 15 5965k 0 0 497k 0 0:01:16 0:00:12 0:01:04 706k 18 37.0M 18 6928k 0 0 532k 0 0:01:11 0:00:13 0:00:58 786k 21 37.0M 21 7968k 0 0  569k 0 0:01:06 0:00:14 0:00:52 875k 23 37.0M 23 8871k 0 0 591k 0 0:01:04 0:00:15 0:00:49 914k 25 37.0M 25 9778k 0 0 611k 0 0:01:01 0:00:16 0:00:45 935k 27 37.0M 27 10.3M 0 0 622k  0 0:01:00 0:00:17 0:00:43 910k 29 37.0M 29 10.9M 0 0 623k 0 0:01:00 0:00:18 0:00:42 862k 31 37.0M 31 11.6M 0 0 626k 0 0:01:00 0:00:19 0:00:41 786k 33 37.0M 33 12.2M 0 0 627k 0 0:01:00 0:00:20 0:00:40 735k 34 37.0M 34 12.8M 0 0 627k 0 0:01:00 0:00:21 0:00:39 679k 36 37.0M 36 13.3M 0 0 623k 0 0:01:00 0:00:22 0:00:38 635k 37 37.0M 37 13.8M 0 0 617k 0 0:01:01 0:00:23 0:00:38 593k 38 37.0M 38 14.3M 0 0 612k 0 0:01:01 0:00:24 0:00:37 560k 40 37.0M 40 14.8M 0 0 608k 0 0:01:02 0:00:25 0:00:37 531k 41 37.0M 41 15.2M 0 0 601k 0 0:01:02 0:00:26 0:00:36 494k 42 37.0M 42 15.6M 0 0 594k 0 0:01:03 0:00:27 0:00:36 469k 43 37.0M 43 16.0M 0 0 588k 0 0:01:04 0:00:28 0:00:36 457k 44 37.0M 44 16.5M 0 0 583k 0 0:01:04 0:00:29 0:00:35 440k 45 37.0M 45 16.9M 0 0 578k 0 0:01:05 0:00:30 0:00:35 430k 47 37.0M 47 17.4M 0 0 575k 0 0:01:05 0:00:31 0:00:34 440k 48 37.0M 48 17.9M 0 0 575k 0 0:01:05 0:00:32 0:00:33 468k 50 37.0M 50 18.6M 0 0 578k 0 0:01:05 0:00:33 0:00:32 523k 52 37.0M 52 19.5M 0 0 590k 0 0:01:04 0:00:34 0:00:30 631k 56 37.0M 56 20.9M 0 0 613k 0 0:01:01 0:00:35 0:00:26 825k 61 37.0M 61 22.9M 0 0 651k 0 0:00:58 0:00:36 0:00:22 1125k 66 37.0M 66 24.7M 0 0 683k 0 0:00:55 0:00:37 0:00:18 1303k 68 37.0M 68 25.2M 0 0 679k 0 0:00:55 0:00:38 0:00:17 1339k 70 37.0M 70 26.1M 0 0 685k 0 0:00:55 0:00:39 0:00:16 1338k 74 37.0M 74 27.4M 0 0 703k 0 0:00:53 0:00:40 0:00:13 1327k 77 37.0M 77 28.6M 0 0 715k 0 0:00:52 0:00:41 0:00:11 1171k 80 37.0M 80 29.6M 0 0 722k 0 0:00:52 0:00:42 0:00:10 1072k 82 37.0M 82 30.4M 0 0 725k 0 0:00:52 0:00:43 0:00:09 1077k 84 37.0M 84 31.3M 0 0 728k 0 0:00:51 0:00:44 0:00:07 1061k 86 37.0M 86 32.1M 0 0 731k 0 0:00:51 0:00:45 0:00:06 960k 89 37.0M 89 33.0M 0 0 735k 0 0:00:51 0:00:46 0:00:05 900k 91 37.0M 91 33.8M 0 0 738k 0 0:00:51 0:00:47 0:00:04 872k 94 37.0M 94 34.8M 0 0 742k 0 0:00:50 0:00:48 0:00:02 894k 96 37.0M 96 35.8M 0 0 748k 0 0:00:50 0:00:49 0:00:01 925k 100 37.0M 100 37.0M 0 0 757k 0 0:00:50 0:00:50 --:--:-- 999k  ---> e917bd0d2f3f Removing intermediate container 44e5b5dcf1aa Step 7/10 : ADD run-tgt.sh / ---> 4b7b56c4dfb0 Removing intermediate container 7f22982bc5f7 Step 8/10 : EXPOSE 3260 ---> Running in 306e7ebccfe4 ---> f97abc900465 Removing intermediate container 306e7ebccfe4 Step 9/10 : CMD /run-tgt.sh ---> Running in 0775016d0793 ---> 63d7de24abbc Removing intermediate container 0775016d0793 Step 10/10 : LABEL "iscsi-demo-target-tgtd" '' "kubevirt-functional-tests-openshift-release0" '' ---> Running in e911a2b9ed8e ---> 530dcdbed8b8 Removing intermediate container e911a2b9ed8e Successfully built 530dcdbed8b8 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/vm-killer ./ Dockerfile sent 602 bytes received 34 bytes 1272.00 bytes/sec total size is 787 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 05ea8a13b62f Step 3/5 : ENV container docker ---> Using cache ---> 42daee89c60c Step 4/5 : RUN dnf -y install procps-ng && dnf -y clean all ---> Running in ca07084e5473 Fedora 27 - x86_64 - Updates 5.7 MB/s | 22 MB 00:03 Fedora 27 - x86_64 1.8 MB/s | 58 MB 00:32 Last metadata expiration check: 0:00:14 ago on Fri Apr 6 13:08:57 2018. Dependencies resolved. ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: procps-ng x86_64 3.3.10-15.fc27 fedora 395 k Transaction Summary ================================================================================ Install 1 Package Total download size: 395 k Installed size: 874 k Downloading Packages: procps-ng-3.3.10-15.fc27.x86_64.rpm 558 kB/s | 395 kB 00:00 -------------------------------------------------------------------------------- Total 276 kB/s | 395 kB 00:01 Running transaction check Transaction check succeeded. Running transaction test Transaction test succeeded. Running transaction Preparing : 1/1 Installing : procps-ng-3.3.10-15.fc27.x86_64 1/1 Running scriptlet: procps-ng-3.3.10-15.fc27.x86_64 1/1 Verifying : procps-ng-3.3.10-15.fc27.x86_64 1/1 Installed: procps-ng.x86_64 3.3.10-15.fc27 Complete! 18 files removed ---> 75dcb2ffd2bd Removing intermediate container ca07084e5473 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release0" '' "vm-killer" '' ---> Running in 585060ac4bb0 ---> 19403899df8b Removing intermediate container 585060ac4bb0 Successfully built 19403899df8b sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/registry-disk-v1alpha ./ Dockerfile entry-point.sh sent 1529 bytes received 53 bytes 3164.00 bytes/sec total size is 2482 speedup is 1.57 Sending build context to Docker daemon 5.12 kB Step 1/7 : FROM debian:sid Trying to pull repository docker.io/library/debian ... sid: Pulling from docker.io/library/debian 2115d46e7396: Pulling fs layer 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 2115d46e7396: Pull complete Digest: sha256:300fe2f0931dbe6d49b473e60660aeec71e2c2f779794894f97f51d72cad0e9c Status: Downloaded newer image for docker.io/debian:sid ---> bcec0ae8107e Step 2/7 : MAINTAINER "David Vossel" \ ---> Running in 91f7266f6fe7 ---> b1f5718be6dd Removing intermediate container 91f7266f6fe7 Step 3/7 : ENV container docker ---> Running in 5bd0e4aeae2a ---> dd20c8f6c0fd Removing intermediate container 5bd0e4aeae2a Step 4/7 : RUN apt-get update && apt-get install -y bash curl bzip2 qemu-utils && mkdir -p /disk && rm -rf /var/lib/apt/lists/* ---> Running in 7a3fce975c41  Get:1 http://deb.debian.org/debian sid InRelease [242 kB] Get:2 http://deb.debian.org/debian sid/main amd64 Packages [10.7 MB] Fetched 11.0 MB in 6s (1804 kB/s) Reading package lists... Reading package lists... Building dependency tree... Reading state information... The following additional packages will be installed: ca-certificates krb5-locales libaio1 libcurl3 libglib2.0-0 libglib2.0-data libgssapi-krb5-2 libicu57 libk5crypto3 libkeyutils1 libkrb5-3 libkrb5support0 libldap-2.4-2 libldap-common libnghttp2-14 libpsl5 librtmp1 libsasl2-2 libsasl2-modules libsasl2-modules-db libssh2-1 libssl1.0.2 libssl1.1 libxml2 openssl publicsuffix shared-mime-info xdg-user-dirs Suggested packages: bash-doc bzip2-doc krb5-doc krb5-user libsasl2-modules-gssapi-mit | libsasl2-modules-gssapi-heimdal libsasl2-modules-ldap libsasl2-modules-otp libsasl2-modules-sql debootstrap qemu-block-extra Recommended packages: bash-completion The following NEW packages will be installed: bzip2 ca-certificates curl krb5-locales libaio1 libcurl3 libglib2.0-0 libglib2.0-data libgssapi-krb5-2 libicu57 libk5crypto3 libkeyutils1 libkrb5-3 libkrb5support0 libldap-2.4-2 libldap-common libnghttp2-14 libpsl5 librtmp1 libsasl2-2 libsasl2-modules libsasl2-modules-db libssh2-1 libssl1.0.2 libssl1.1 libxml2 openssl publicsuffix qemu-utils shared-mime-info xdg-user-dirs The following packages will be upgraded: bash 1 upgraded, 31 newly installed, 0 to remove and 19 not upgraded. Need to get 23.1 MB of archives. After this operation, 73.8 MB of additional disk space will be used. Get:1 http://deb.debian.org/debian sid/main amd64 bash amd64 4.4.18-2 [1415 kB] Get:2 http://deb.debian.org/debian sid/main amd64 bzip2 amd64 1.0.6-8.1 [47.5 kB] Get:3 http://deb.debian.org/debian sid/main amd64 krb5-locales all 1.16-2 [94.7 kB] Get:4 http://deb.debian.org/debian sid/main amd64 libssl1.1 amd64 1.1.0h-2 [1352 kB] Get:5 http://deb.debian.org/debian sid/main amd64 openssl amd64 1.1.0h-2 [744 kB] Get:6 http://deb.debian.org/debian sid/main amd64 ca-certificates all 20170717 [178 kB] Get:7 http://deb.debian.org/debian sid/main amd64 libkeyutils1 amd64 1.5.9-9.2 [12.9 kB] Get:8 http://deb.debian.org/debian sid/main amd64 libkrb5support0 amd64 1.16-2 [62.8 kB] Get:9 http://deb.debian.org/debian sid/main amd64 libk5crypto3 amd64 1.16-2 [121 kB] Get:10 http://deb.debian.org/debian sid/main amd64 libkrb5-3 amd64 1.16-2 [316 kB] Get:11 http://deb.debian.org/debian sid/main amd64 libgssapi-krb5-2 amd64 1.16-2 [158 kB] Get:12 http://deb.debian.org/debian sid/main amd64 libsasl2-modules-db amd64 2.1.27~101-g0780600+dfsg-3 [68.2 kB] Get:13 http://deb.debian.org/debian sid/main amd64 libsasl2-2 amd64 2.1.27~101-g0780600+dfsg-3 [105 kB] Get:14 http://deb.debian.org/debian sid/main amd64 libldap-common all 2.4.45+dfsg-1 [87.0 kB] Get:15 http://deb.debian.org/debian sid/main amd64 libldap-2.4-2 amd64 2.4.45+dfsg-1 [223 kB] Get:16 http://deb.debian.org/debian sid/main amd64 libnghttp2-14 amd64 1.31.0-1 [81.4 kB] Get:17 http://deb.debian.org/debian sid/main amd64 libpsl5 amd64 0.20.1-1 [52.6 kB] Get:18 http://deb.debian.org/debian sid/main amd64 librtmp1 amd64 2.4+20151223.gitfa8646d.1-1+b1 [60.4 kB] Get:19 http://deb.debian.org/debian sid/main amd64 libssh2-1 amd64 1.8.0-1 [138 kB] Get:20 http://deb.debian.org/debian sid/main amd64 libssl1.0.2 amd64 1.0.2o-1 [1296 kB] Get:21 http://deb.debian.org/debian sid/main amd64 libcurl3 amd64 7.58.0-2 [312 kB] Get:22 http://deb.debian.org/debian sid/main amd64 curl amd64 7.58.0-2 [248 kB] Get:23 http://deb.debian.org/debian sid/main amd64 libaio1 amd64 0.3.111-1 [10.5 kB] Get:24 http://deb.debian.org/debian sid/main amd64 libglib2.0-0 amd64 2.56.0-6 [2922 kB] Get:25 http://deb.debian.org/debian sid/main amd64 libglib2.0-data all 2.56.0-6 [2664 kB] Get:26 http://deb.debian.org/debian sid/main amd64 libicu57 amd64 57.1-9 [7698 kB] Get:27 http://deb.debian.org/debian sid/main amd64 libsasl2-modules amd64 2.1.27~101-g0780600+dfsg-3 [102 kB] Get:28 http://deb.debian.org/debian sid/main amd64 libxml2 amd64 2.9.4+dfsg1-6.1 [725 kB] Get:29 http://deb.debian.org/debian sid/main amd64 publicsuffix all 20180328.1055-1 [111 kB] Get:30 http://deb.debian.org/debian sid/main amd64 qemu-utils amd64 1:2.11+dfsg-1 [924 kB] Get:31 http://deb.debian.org/debian sid/main amd64 shared-mime-info amd64 1.9-2 [753 kB] Get:32 http://deb.debian.org/debian sid/main amd64 xdg-user-dirs amd64 0.17-1 [53.5 kB] debconf: delaying package configuration, since apt-utils is not installed Fetched 23.1 MB in 3s (6992 kB/s) (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 6540 files and directories currently installed.) Preparing to unpack .../bash_4.4.18-2_amd64.deb ... Unpacking bash (4.4.18-2) over (4.4.18-1.1) ... Setting up bash (4.4.18-2) ... update-alternatives: using /usr/share/man/man7/bash-builtins.7.gz to provide /usr/share/man/man7/builtins.7.gz (builtins.7.gz) in auto mode Selecting previously unselected package bzip2. (Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 6540 files and directories currently installed.) Preparing to unpack .../00-bzip2_1.0.6-8.1_amd64.deb ... Unpacking bzip2 (1.0.6-8.1) ... Selecting previously unselected package krb5-locales. Preparing to unpack .../01-krb5-locales_1.16-2_all.deb ... Unpacking krb5-locales (1.16-2) ... Selecting previously unselected package libssl1.1:amd64. Preparing to unpack .../02-libssl1.1_1.1.0h-2_amd64.deb ... Unpacking libssl1.1:amd64 (1.1.0h-2) ... Selecting previously unselected package openssl. Preparing to unpack .../03-openssl_1.1.0h-2_amd64.deb ... Unpacking openssl (1.1.0h-2) ... Selecting previously unselected package ca-certificates. Preparing to unpack .../04-ca-certificates_20170717_all.deb ... Unpacking ca-certificates (20170717) ... Selecting previously unselected package libkeyutils1:amd64. Preparing to unpack .../05-libkeyutils1_1.5.9-9.2_amd64.deb ... Unpacking libkeyutils1:amd64 (1.5.9-9.2) ... Selecting previously unselected package libkrb5support0:amd64. Preparing to unpack .../06-libkrb5support0_1.16-2_amd64.deb ... Unpacking libkrb5support0:amd64 (1.16-2) ... Selecting previously unselected package libk5crypto3:amd64. Preparing to unpack .../07-libk5crypto3_1.16-2_amd64.deb ... Unpacking libk5crypto3:amd64 (1.16-2) ... Selecting previously unselected package libkrb5-3:amd64. Preparing to unpack .../08-libkrb5-3_1.16-2_amd64.deb ... Unpacking libkrb5-3:amd64 (1.16-2) ... Selecting previously unselected package libgssapi-krb5-2:amd64. Preparing to unpack .../09-libgssapi-krb5-2_1.16-2_amd64.deb ... Unpacking libgssapi-krb5-2:amd64 (1.16-2) ... Selecting previously unselected package libsasl2-modules-db:amd64. Preparing to unpack .../10-libsasl2-modules-db_2.1.27~101-g0780600+dfsg-3_amd64.deb ... Unpacking libsasl2-modules-db:amd64 (2.1.27~101-g0780600+dfsg-3) ... Selecting previously unselected package libsasl2-2:amd64. Preparing to unpack .../11-libsasl2-2_2.1.27~101-g0780600+dfsg-3_amd64.deb ... Unpacking libsasl2-2:amd64 (2.1.27~101-g0780600+dfsg-3) ... Selecting previously unselected package libldap-common. Preparing to unpack .../12-libldap-common_2.4.45+dfsg-1_all.deb ... Unpacking libldap-common (2.4.45+dfsg-1) ... Selecting previously unselected package libldap-2.4-2:amd64. Preparing to unpack .../13-libldap-2.4-2_2.4.45+dfsg-1_amd64.deb ... Unpacking libldap-2.4-2:amd64 (2.4.45+dfsg-1) ... Selecting previously unselected package libnghttp2-14:amd64. Preparing to unpack .../14-libnghttp2-14_1.31.0-1_amd64.deb ... Unpacking libnghttp2-14:amd64 (1.31.0-1) ... Selecting previously unselected package libpsl5:amd64. Preparing to unpack .../15-libpsl5_0.20.1-1_amd64.deb ... Unpacking libpsl5:amd64 (0.20.1-1) ... Selecting previously unselected package librtmp1:amd64. Preparing to unpack .../16-librtmp1_2.4+20151223.gitfa8646d.1-1+b1_amd64.deb ... Unpacking librtmp1:amd64 (2.4+20151223.gitfa8646d.1-1+b1) ... Selecting previously unselected package libssh2-1:amd64. Preparing to unpack .../17-libssh2-1_1.8.0-1_amd64.deb ... Unpacking libssh2-1:amd64 (1.8.0-1) ... Selecting previously unselected package libssl1.0.2:amd64. Preparing to unpack .../18-libssl1.0.2_1.0.2o-1_amd64.deb ... Unpacking libssl1.0.2:amd64 (1.0.2o-1) ... Selecting previously unselected package libcurl3:amd64. Preparing to unpack .../19-libcurl3_7.58.0-2_amd64.deb ... Unpacking libcurl3:amd64 (7.58.0-2) ... Selecting previously unselected package curl. Preparing to unpack .../20-curl_7.58.0-2_amd64.deb ... Unpacking curl (7.58.0-2) ... Selecting previously unselected package libaio1:amd64. Preparing to unpack .../21-libaio1_0.3.111-1_amd64.deb ... Unpacking libaio1:amd64 (0.3.111-1) ... Selecting previously unselected package libglib2.0-0:amd64. Preparing to unpack .../22-libglib2.0-0_2.56.0-6_amd64.deb ... Unpacking libglib2.0-0:amd64 (2.56.0-6) ... Selecting previously unselected package libglib2.0-data. Preparing to unpack .../23-libglib2.0-data_2.56.0-6_all.deb ... Unpacking libglib2.0-data (2.56.0-6) ... Selecting previously unselected package libicu57:amd64. Preparing to unpack .../24-libicu57_57.1-9_amd64.deb ... Unpacking libicu57:amd64 (57.1-9) ... Selecting previously unselected package libsasl2-modules:amd64. Preparing to unpack .../25-libsasl2-modules_2.1.27~101-g0780600+dfsg-3_amd64.deb ... Unpacking libsasl2-modules:amd64 (2.1.27~101-g0780600+dfsg-3) ... Selecting previously unselected package libxml2:amd64. Preparing to unpack .../26-libxml2_2.9.4+dfsg1-6.1_amd64.deb ... Unpacking libxml2:amd64 (2.9.4+dfsg1-6.1) ... Selecting previously unselected package publicsuffix. Preparing to unpack .../27-publicsuffix_20180328.1055-1_all.deb ... Unpacking publicsuffix (20180328.1055-1) ... Selecting previously unselected package qemu-utils. Preparing to unpack .../28-qemu-utils_1%3a2.11+dfsg-1_amd64.deb ... Unpacking qemu-utils (1:2.11+dfsg-1) ... Selecting previously unselected package shared-mime-info. Preparing to unpack .../29-shared-mime-info_1.9-2_amd64.deb ... Unpacking shared-mime-info (1.9-2) ... Selecting previously unselected package xdg-user-dirs. Preparing to unpack .../30-xdg-user-dirs_0.17-1_amd64.deb ... Unpacking xdg-user-dirs (0.17-1) ... Setting up libnghttp2-14:amd64 (1.31.0-1) ... Setting up libldap-common (2.4.45+dfsg-1) ... Setting up libpsl5:amd64 (0.20.1-1) ... Setting up libglib2.0-0:amd64 (2.56.0-6) ... No schema files found: doing nothing. Setting up libsasl2-modules-db:amd64 (2.1.27~101-g0780600+dfsg-3) ... Setting up libsasl2-2:amd64 (2.1.27~101-g0780600+dfsg-3) ... Setting up librtmp1:amd64 (2.4+20151223.gitfa8646d.1-1+b1) ... Setting up libicu57:amd64 (57.1-9) ... Setting up libxml2:amd64 (2.9.4+dfsg1-6.1) ... Setting up bzip2 (1.0.6-8.1) ... Setting up libssl1.0.2:amd64 (1.0.2o-1) ... debconf: unable to initialize frontend: Dialog debconf: (TERM is not set, so the dialog frontend is not usable.) debconf: falling back to frontend: Readline debconf: unable to initialize frontend: Readline debconf: (Can't locate Term/ReadLine.pm in @INC (you may need to install the Term::ReadLine module) (@INC contains: /etc/perl /usr/local/lib/x86_64-linux-gnu/perl/5.26.1 /usr/local/share/perl/5.26.1 /usr/lib/x86_64-linux-gnu/perl5/5.26 /usr/share/perl5 /usr/lib/x86_64-linux-gnu/perl/5.26 /usr/share/perl/5.26 /usr/local/lib/site_perl /usr/lib/x86_64-linux-gnu/perl-base) at /usr/share/perl5/Debconf/FrontEnd/Readline.pm line 7.) debconf: falling back to frontend: Teletype Setting up libssh2-1:amd64 (1.8.0-1) ... Setting up libglib2.0-data (2.56.0-6) ... Setting up krb5-locales (1.16-2) ... Processing triggers for libc-bin (2.27-2) ... Setting up libaio1:amd64 (0.3.111-1) ... Setting up publicsuffix (20180328.1055-1) ... Setting up libldap-2.4-2:amd64 (2.4.45+dfsg-1) ... Setting up libssl1.1:amd64 (1.1.0h-2) ... debconf: unable to initialize frontend: Dialog debconf: (TERM is not set, so the dialog frontend is not usable.) debconf: falling back to frontend: Readline debconf: unable to initialize frontend: Readline debconf: (Can't locate Term/ReadLine.pm in @INC (you may need to install the Term::ReadLine module) (@INC contains: /etc/perl /usr/local/lib/x86_64-linux-gnu/perl/5.26.1 /usr/local/share/perl/5.26.1 /usr/lib/x86_64-linux-gnu/perl5/5.26 /usr/share/perl5 /usr/lib/x86_64-linux-gnu/perl/5.26 /usr/share/perl/5.26 /usr/local/lib/site_perl /usr/lib/x86_64-linux-gnu/perl-base) at /usr/share/perl5/Debconf/FrontEnd/Readline.pm line 7.) debconf: falling back to frontend: Teletype Setting up openssl (1.1.0h-2) ... Setting up shared-mime-info (1.9-2) ... Setting up libkeyutils1:amd64 (1.5.9-9.2) ... Setting up libsasl2-modules:amd64 (2.1.27~101-g0780600+dfsg-3) ... Setting up ca-certificates (20170717) ... debconf: unable to initialize frontend: Dialog debconf: (TERM is not set, so the dialog frontend is not usable.) debconf: falling back to frontend: Readline debconf: unable to initialize frontend: Readline debconf: (Can't locate Term/ReadLine.pm in @INC (you may need to install the Term::ReadLine module) (@INC contains: /etc/perl /usr/local/lib/x86_64-linux-gnu/perl/5.26.1 /usr/local/share/perl/5.26.1 /usr/lib/x86_64-linux-gnu/perl5/5.26 /usr/share/perl5 /usr/lib/x86_64-linux-gnu/perl/5.26 /usr/share/perl/5.26 /usr/local/lib/site_perl /usr/lib/x86_64-linux-gnu/perl-base) at /usr/share/perl5/Debconf/FrontEnd/Readline.pm line 7.) debconf: falling back to frontend: Teletype Updating certificates in /etc/ssl/certs... 148 added, 0 removed; done. Setting up xdg-user-dirs (0.17-1) ... Setting up libkrb5support0:amd64 (1.16-2) ... Setting up qemu-utils (1:2.11+dfsg-1) ... Setting up libk5crypto3:amd64 (1.16-2) ... Setting up libkrb5-3:amd64 (1.16-2) ... Setting up libgssapi-krb5-2:amd64 (1.16-2) ... Setting up libcurl3:amd64 (7.58.0-2) ... Setting up curl (7.58.0-2) ... Processing triggers for libc-bin (2.27-2) ... Processing triggers for ca-certificates (20170717) ... Updating certificates in /etc/ssl/certs... 0 added, 0 removed; done. Running hooks in /etc/ca-certificates/update.d... done. ---> 625894bac383 Removing intermediate container 7a3fce975c41 Step 5/7 : ADD entry-point.sh / ---> b99a70a292aa Removing intermediate container 47b07310d61c Step 6/7 : CMD /entry-point.sh ---> Running in b38f59655dd0 ---> 36ef7e9fa1c2 Removing intermediate container b38f59655dd0 Step 7/7 : LABEL "kubevirt-functional-tests-openshift-release0" '' "registry-disk-v1alpha" '' ---> Running in 3613d7c77301 ---> 10b0091ccb47 Removing intermediate container 3613d7c77301 Successfully built 10b0091ccb47 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/cirros-registry-disk-demo ./ Dockerfile sent 630 bytes received 34 bytes 1328.00 bytes/sec total size is 825 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:32771/kubevirt/registry-disk-v1alpha:devel ---> 10b0091ccb47 Step 2/4 : MAINTAINER "David Vossel" \ ---> Running in aa424ef3e32a ---> 2739c4998822 Removing intermediate container aa424ef3e32a Step 3/4 : RUN curl https://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img > /disk/cirros.img ---> Running in 3fd44fff3728   % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 12.1M 0 32768 0 0 36167 0 0:05:51 --:--:-- 0:05:51 36127 3 12.1M 3 416k 0 0 218k 0  0:00:56 0:00:01 0:00:55 218k 23 12.1M 23 2880k 0 0 992k 0 0:00:12 0:00:02 0:00:10 992k 91 12.1M 91 11.0M 0 0 2914k 0 0:00:04 0:00:03 0:00:01 2914k 100 12.1M 100 12.1M 0 0 3135k 0 0:00:03 0:00:03 --:--:-- 3134k  ---> 9d5f94eed32e Removing intermediate container 3fd44fff3728 Step 4/4 : LABEL "cirros-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release0" '' ---> Running in 9dc620c88a57 ---> 06bae6631b78 Removing intermediate container 9dc620c88a57 Successfully built 06bae6631b78 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/fedora-cloud-registry-disk-demo ./ Dockerfile sent 677 bytes received 34 bytes 1422.00 bytes/sec total size is 926 speedup is 1.30 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:32771/kubevirt/registry-disk-v1alpha:devel ---> 10b0091ccb47 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Running in 9b0339b0abbd ---> a4e52638f586 Removing intermediate container 9b0339b0abbd Step 3/4 : RUN curl -g -L https://download.fedoraproject.org/pub/fedora/linux/releases/27/CloudImages/x86_64/images/Fedora-Cloud-Base-27-1.6.x86_64.qcow2 > /disk/fedora.qcow2 ---> Running in fedcd7a706e5   % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0  0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0  0 --:--:-- --:--:-- --:--:-- 0  0 221M 0 32768 0 0 18083 0 3:34:07 0:00:01 3:34:06 18083 0 221M 0 368k 0 0 147k 0 0:25:39 0:00:02 0:25:37 489k 0 221M 0 752k 0 0 209k 0 0:18:01 0:00:03 0:17:58 405k 0 221M 0 1120k 0 0 246k 0 0:15:19 0:00:04 0:15:15 398k 0 221M 0 1504k 0 0 273k 0 0:13:49 0:00:05 0:13:44 399k 0 221M 0 1936k 0 0 298k 0 0:12:41 0:00:06 0:12:35 406k 1 221M 1 2384k 0 0 317k 0  0:11:53 0:00:07 0:11:46 402k 1 221M 1 2880k 0 0 338k 0 0:11:09 0:00:08 0:11:01 433k 1 221M 1 3456k 0 0 362k 0 0:10:26 0:00:09 0:10:17 467k 1 221M 1 4000k 0 0 381k 0 0:09:55 0:00:10 0:09:45 499k 2 221M 2 4608k 0 0 400k 0 0:09:26 0:00:11 0:09:15 534k 2 221M 2 5296k 0 0 423k 0  0:08:55 0:00:12 0:08:43 582k 2 221M 2 5936k 0 0 438k 0 0:08:37 0:00:13 0:08:24 606k 2 221M 2 6576k 0 0 453k 0 0:08:20 0:00:14 0:08:06 629k 3 221M 3 7312k 0 0 471k 0 0:08:00 0:00:15 0:07:45 662k 3 221M 3 8064k 0 0 487k 0 0:07:45 0:00:16 0:07:29 686k 3 221M 3 8960k 0 0 512k 0 0:07:23 0:00:17 0:07:06 732k 4 221M 4 9.9M 0 0 551k 0 0:06:51 0:00:18 0:06:33 860k 5 221M 5 11.4M 0 0 601k 0 0:06:17 0:00:19 0:05:58 1030k 6 221M 6 13.4M 0 0 671k 0  0:05:37 0:00:20 0:05:17 1289k 7 221M 7 15.6M 0 0 745k 0 0:05:04 0:00:21 0:04:43 1605k 8 221M 8 17.7M 0 0 809k  0 0:04:40 0:00:22 0:04:18 1845k 9 221M 9 20.0M 0 0 874k 0 0:04:19 0:00:23 0:03:56 2070k 10 221M 10 22.6M 0 0 945k 0 0:04:00 0:00:24 0:03:36 2279k 11 221M 11 24.9M 0 0 1000k 0 0:03:46 0:00:25 0:03:21 2344k 11 221M 11 26.2M 0 0 1015k 0 0:03:43 0:00:26 0:03:17 2172k 12 221M 12 27.7M 0 0 1033k 0 0:03:39 0:00:27 0:03:12 2041k 13 221M 13 28.9M 0 0 1038k 0 0:03:38 0:00:28 0:03:10 1806k 13 221M 13 30.1M 0 0 1046k 0 0:03:36 0:00:29 0:03:07 1545k 14 221M 14 31.2M 0 0 1048k 0 0:03:36 0:00:30 0:03:06 1295k 14 221M 14 32.1M 0 0 1046k 0 0:03:36  0:00:31 0:03:05 1212k 14 221M 14 32.9M 0 0 1039k  0 0:03:38 0:00:32 0:03:06 1073k 15 221M 15 33.8M 0 0 1034k 0 0:03:39 0:00:33 0:03:06 1009k 15 221M 15 34.7M 0 0 1031k 0 0:03:39 0:00:34 0:03:05 943k 16 221M 16 35.6M 0 0 1027k 0 0:03:40 0:00:35 0:03:05 901k 16 221M 16 36.5M 0 0 1024k 0 0:03:41 0:00:36 0:03:05 887k 16 221M 16 37.3M 0 0 1019k 0 0:03:42 0:00:37 0:03:05 889k 17 221M 17 38.1M 0  0 1013k 0 0:03:43 0:00:38 0:03:05 871k 17 221M 17 38.7M 0 0 1005k 0 0:03:45 0:00:39 0:03:06 824k 17 221M 17 39.4M 0 0 998k 0 0:03:47 0:00:40 0:03:07 787k 18 221M 18 40.2M 0 0 992k 0 0:03:48 0:00:41 0:03:07 755k 18 221M 18 40.9M 0 0 985k 0  0:03:50 0:00:42 0:03:08 735k 18 221M 18 41.6M 0 0 981k 0 0:03:51 0:00:43 0:03:08 733k 19 221M 19 42.4M 0 0 977k 0 0:03:52 0:00:44 0:03:08 756k 19 221M 19 43.3M 0 0 974k 0 0:03:52 0:00:45 0:03:07 783k 19 221M 19 44.2M 0 0 975k 0 0:03:52  0:00:46 0:03:06 836k 20 221M 20 45.5M 0 0 981k 0 0:03:51 0:00:47 0:03:04 947k 21 221M 21 47.1M 0 0 995k 0 0:03:47 0:00:48 0:02:59 1115k 21 221M 21 48.7M 0 0 1008k 0 0:03:45 0:00:49 0:02:56 1283k 22 221M 22 50.1M 0 0 1016k 0 0:03:43 0:00:50  0:02:53 1397k 23 221M 23 51.4M 0 0 1022k 0 0:03:41 0:00:51 0:02:50 1458k 23 221M 23 52.7M 0 0 1029k 0 0:03:40 0:00:52 0:02:48 1478k 24 221M 24 54.1M 0 0 1036k  0 0:03:38 0:00:53 0:02:45 1440k 25 221M 25 55.5M 0 0 1044k 0 0:03:37 0:00:54 0:02:43 1403k 25  221M 25 56.9M 0 0 1051k 0 0:03:35 0:00:55 0:02:40 1399k 26 221M 26 58.2M 0 0 1055k 0 0:03:34 0:00:56 0:02:38 1396k 26 221M 26 59.3M 0 0 1056k 0 0:03:34 0:00:57 0:02:37 1340k 27 221M 27 59.9M 0 0 1049k 0 0:03:36 0:00:58 0:02:38 1184k 27 221M 27 60.4M 0 0 1040k 0 0:03:37 0:00:59 0:02:38 1001k 27 221M 27 61.0M 0 0 1033k 0 0:03:39 0:01:00 0:02:39 835k 27 221M 27 61.5M 0 0 1024k 0 0:03:41 0:01:01 0:02:40 676k 28 221M 28 62.0M 0 0 1016k 0 0:03:43 0:01:02 0:02:41 562k 28 221M 28 62.6M 0 0 1009k 0 0:03:44 0:01:03 0:02:41 543k 28 221M 28 63.1M 0 0 1002k 0 0:03:46 0:01:04 0:02:42 552k 28 221M 28 63.7M 0 0 996k 0 0:03:47 0:01:05 0:02:42 551k 29 221M 29 64.3M 0 0 990k 0 0:03:49 0:01:06 0:02:43 566k 29 221M 29 64.8M 0 0 983k 0 0:03:50 0:01:07 0:02:43 565k 29 221M 29 65.3M 0 0 977k 0 0:03:52 0:01:08 0:02:44 564k 29 221M 29 65.9M 0 0 971k 0 0:03:53 0:01:09 0:02:44 561k 30 221M 30 66.4M 0 0 965k 0 0:03:55  0:01:10 0:02:45 559k 30 221M 30 67.0M 0 0 958k 0 0:03:56 0:01:11 0:02:45 542k 30 221M 30 67.5M 0 0 953k 0 0:03:57 0:01:12 0:02:45 550k 30 221M 30 68.0M 0 0 947k 0 0:03:59 0:01:13 0:02:46 539k 30 221M 30 68.5M 0 0 941k 0 0:04:00  0:01:14 0:02:46 530k 31 221M 31 69.0M 0 0 936k  0 0:04:02 0:01:15 0:02:47 528k 31 221M 31 69.6M 0 0 931k 0 0:04:03 0:01:16 0:02:47 545k 31 221M 31 70.2M 0 0 928k 0 0:04:04 0:01:17 0:02:47 560k 32 221M 32 70.9M 0 0 925k 0 0:04:05 0:01:18 0:02:47 595k 32 221M 32 71.8M 0 0 924k 0 0:04:05 0:01:19 0:02:46 674k 32 221M 32 72.9M 0 0 927k 0 0:04:04 0:01:20 0:02:44 794k 33 221M 33 74.5M 0 0 936k 0 0:04:02 0:01:21 0:02:41 998k 34 221M 34 76.5M 0 0 948k 0 0:03:59 0:01:22 0:02:37 1262k 35 221M 35 78.3M 0 0 960k 0 0:03:56 0:01:23 0:02:33 1524k 35 221M 35 79.6M 0 0 965k 0 0:03:54  0:01:24 0:02:30 1613k 36 221M 36 80.8M 0 0 967k 0 0:03:54 0:01:25 0:02:29 1610k 37 221M 37 81.9M 0 0 970k 0 0:03:53 0:01:26 0:02:27 1531k 37 221M 37 82.9M 0 0 970k 0 0:03:53 0:01:27 0:02:26 1336k 37 221M 37 83.9M 0 0 971k 0 0:03:53 0:01:28 0:02:25 1144k 38 221M 38 85.0M 0 0 972k 0 0:03:53 0:01:29 0:02:24 1089k 38 221M 38 86.0M 0 0 973k 0 0:03:53 0:01:30 0:02:23 1076k 39 221M 39 87.1M 0 0 974k 0 0:03:52 0:01:31 0:02:21 1051k 39 221M 39 88.2M 0 0 976k 0 0:03:52 0:01:32 0:02:20 1079k 40 221M 40 89.3M 0 0 978k 0 0:03:51 0:01:33 0:02:18 1111k 40 221M 40 90.6M 0 0 982k 0 0:03:51 0:01:34 0:02:17 1153k 41 221M 41 92.1M 0 0 988k 0 0:03:49 0:01:35 0:02:14 1249k 42 221M 42 94.0M 0 0 997k 0 0:03:47 0:01:36 0:02:11 1415k 43 221M 43 96.3M 0 0 1011k 0 0:03:44 0:01:37 0:02:07 1662k 44 221M 44 98.4M 0 0 1023k 0 0:03:41 0:01:38 0:02:03 1865k 45 221M 45 100M 0 0 1030k 0 0:03:40 0:01:39 0:02:01 1941k 45 221M 45 101M 0 0 1032k 0 0:03:39 0:01:40 0:01:59 1879k 46 221M 46 102M 0 0 1035k 0 0:03:39 0:01:41 0:01:58 1774k 46 221M 46 104M 0 0 1039k 0 0:03:38 0:01:42 0:01:56 1594k 47 221M 47 105M 0 0 1043k 0 0:03:37 0:01:43 0:01:54 1441k 48 221M 48 106M 0 0 1046k 0 0:03:36 0:01:44 0:01:52 1364k 48 221M 48 107M 0 0 1044k 0 0:03:37 0:01:45 0:01:52 1286k 48 221M 48 108M 0 0 1042k 0 0:03:37 0:01:46 0:01:51 1183k 49 221M 49 109M 0 0 1041k 0 0:03:37 0:01:47 0:01:50 1080k 49 221M 49 110M 0 0 1040k 0 0:03:38 0:01:48 0:01:50 975k 50 221M 50 111M 0 0 1038k 0 0:03:38 0:01:49 0:01:49 871k 50 221M 50 111M 0 0 1036k 0 0:03:38 0:01:50 0:01:48 862k 50 221M 50 112M 0 0 1034k 0 0:03:39 0:01:51 0:01:48 864k 51 221M 51 113M 0 0 1033k 0 0:03:39 0:01:52 0:01:47 864k 51 221M 51 114M 0 0 1032k 0 0:03:39 0:01:53 0:01:46 866k 52 221M 52 115M 0 0 1031k 0 0:03:39 0:01:54 0:01:45 896k 52 221M 52 116M 0 0 1031k 0 0:03:39 0:01:55 0:01:44 922k 52 221M 52 117M 0 0 1031k 0 0:03:39 0:01:56 0:01:43 954k 53 221M 53 118M 0 0 1030k 0 0:03:40 0:01:57 0:01:43 959k 53 221M 53 119M 0 0 1030k 0 0:03:40 0:01:58 0:01:42 973k 54 221M 54 120M 0 0 1030k 0 0:03:40 0:01:59 0:01:41 1007k 54 221M 54 121M 0 0 1032k 0 0:03:39 0:02:00 0:01:39 1050k 55 221M 55 122M 0 0 1033k 0 0:03:39 0:02:01 0:01:38 1073k 55 221M 55 123M 0 0 1033k 0 0:03:39 0:02:02 0:01:37 1091k 56 221M 56 124M 0 0 1033k 0 0:03:39 0:02:03 0:01:36 1107k 56 221M 56 125M 0 0 1033k 0 0:03:39 0:02:04 0:01:35 1097k 57 221M 57 126M 0 0 1034k 0 0:03:39 0:02:05 0:01:34 1085k 57 221M 57 127M 0 0 1035k  0 0:03:39 0:02:06 0:01:33 1088k 58 221M 58 129M 0 0 1036k 0 0:03:38 0:02:07 0:01:31 1116k 58 221M 58 130M 0 0 1037k 0 0:03:38 0:02:08 0:01:30 1131k 59 221M 59 131M 0 0 1038k 0 0:03:38 0:02:09 0:01:29 1165k 59 221M 59 132M 0 0 1040k 0 0:03:38 0:02:10 0:01:28 1193k 60 221M 60 133M 0 0 1043k 0 0:03:37 0:02:11 0:01:26 1239k 61 221M 61 135M 0 0 1047k 0 0:03:36 0:02:12 0:01:24 1328k 62 221M 62 137M 0 0 1054k 0 0:03:35 0:02:13 0:01:22 1488k 63 221M 63 139M 0 0 1064k 0 0:03:33 0:02:14 0:01:19 1735k 64 221M 64 142M 0 0 1079k 0 0:03:30 0:02:15 0:01:15 2093k 65 221M 65 145M 0 0 1093k 0 0:03:27 0:02:16 0:01:11 2408k 67 221M 67 148M 0 0 1107k 0 0:03:24 0:02:17 0:01:07 2697k 68 221M 68 151M 0 0 1123k 0 0:03:21 0:02:18 0:01:03 2969k 70 221M 70 155M 0 0 1140k 0 0:03:18 0:02:19 0:00:59 3178k 71 221M 71 158M 0 0 1158k 0 0:03:15 0:02:20 0:00:55 3299k 72 221M 72 161M 0 0 1169k 0 0:03:13 0:02:21 0:00:52 3267k 73 221M 73 163M 0 0 1175k 0 0:03:12 0:02:22 0:00:50 3049k 74 221M 74 165M 0 0 1182k 0 0:03:11 0:02:23 0:00:48 2831k 75 221M 75 167M 0 0 1190k 0 0:03:10 0:02:24 0:00:46 2582k 76 221M 76 169M 0 0 1195k 0 0:03:09 0:02:25 0:00:44 2251k 77 221M 77 171M 0 0 1200k 0 0:03:08 0:02:26 0:00:42 2069k 78 221M 78 173M 0 0 1206k 0 0:03:08 0:02:27 0:00:41 2092k 79 221M 79 175M 0 0 1213k 0 0:03:07 0:02:28 0:00:39 2083k 80 221M 80 178M 0 0 1220k 0 0:03:05 0:02:29 0:00:36 2096k 81 221M 81 180M 0 0 1227k 0 0:03:04 0:02:30 0:00:34 2157k 82 221M 82 182M 0 0 1234k 0 0:03:03 0:02:31 0:00:32 2227k 83 221M 83 184M 0 0 1239k 0  0:03:03 0:02:32 0:00:31 2214k 84 221M 84 186M 0 0 1243k 0 0:03:02 0:02:33 0:00:29 2158k 85 221M 85 188M 0 0 1249k 0 0:03:01 0:02:34 0:00:27 2109k 86 221M 86 190M 0 0 1255k 0 0:03:00 0:02:35 0:00:25 2095k 87 221M 87 192M 0 0 1262k 0 0:02:59 0:02:36 0:00:23 2108k 88 221M 88 195M 0 0 1268k 0 0:02:58 0:02:37 0:00:21 2154k 89 221M 89 197M 0 0 1275k 0 0:02:57 0:02:38 0:00:19 2242k 90 221M 90 199M 0 0 1281k 0 0:02:56 0:02:39 0:00:17 2285k 91 221M 91 201M 0 0 1288k 0 0:02:56 0:02:40 0:00:16 2303k 92 221M 92 204M 0 0 1295k 0 0:02:55 0:02:41 0:00:14 2336k 93 221M 93 206M 0 0 1300k 0 0:02:54  0:02:42 0:00:12 2286k 94 221M 94 208M 0 0 1305k 0 0:02:53 0:02:43 0:00:10 2250k 95 221M 95 210M 0 0 1311k 0 0:02:52 0:02:44 0:00:08 2274k 96 221M 96 213M 0 0 1319k 0 0:02:52 0:02:45 0:00:07 2303k 97 221M 97 215M 0 0 1323k 0 0:02:51 0:02:46 0:00:05 2205k 97 221M 97 217M 0 0 1327k 0 0:02:50 0:02:47 0:00:03 2198k 98 221M 98 219M 0 0 1331k 0 0:02:50 0:02:48 0:00:02 2196k 99 221M 99 221M 0 0 1335k 0 0:02:49 0:02:49 --:--:-- 2097k 100 221M 100 221M 0 0 1335k 0 0:02:49 0:02:49 --:--:-- 1974k  ---> b3bda4e6dd64 Removing intermediate container fedcd7a706e5 Step 4/4 : LABEL "fedora-cloud-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release0" '' ---> Running in 7e0859ee210e ---> c55a9ecc4f6b Removing intermediate container 7e0859ee210e Successfully built c55a9ecc4f6b sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/alpine-registry-disk-demo ./ Dockerfile sent 639 bytes received 34 bytes 1346.00 bytes/sec total size is 866 speedup is 1.29 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:32771/kubevirt/registry-disk-v1alpha:devel ---> 10b0091ccb47 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> a4e52638f586 Step 3/4 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /disk/alpine.iso ---> Running in 9203defea48c  % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 37.0M 0 261k 0 0 244k 0 0:02:35 0:00:01 0:02:34 244k 2 37.0M 2 803k 0 0 394k  0 0:01:35 0:00:02 0:01:33 394k 4 37.0M 4 1636k 0 0 538k 0 0:01:10 0:00:03 0:01:07 538k 7 37.0M 7 2869k 0 0 711k 0 0:00:53 0:00:04 0:00:49 711k 12 37.0M 12 4727k 0 0 939k  0 0:00:40 0:00:05 0:00:35 950k 17 37.0M 17 6573k 0 0 1089k 0 0:00:34 0:00:06 0:00:28 1271k 21 37.0M 21 8248k 0 0 1172k 0 0:00:32 0:00:07 0:00:25 1489k 25 37.0M 25 9744k 0 0 1212k 0 0:00:31 0:00:08 0:00:23 1621k 29 37.0M 29 10.7M 0 0 1202k 0 0:00:31 0:00:09 0:00:22 1591k 31 37.0M 31 11.7M 0 0 1201k 0 0:00:31 0:00:10 0:00:21 1465k 34 37.0M 34 12.8M 0 0 1190k 0 0:00:31 0:00:11 0:00:20 1313k 37 37.0M 37 13.9M 0 0 1185k 0 0:00:31 0:00:12 0:00:19 1204k 40 37.0M 40 15.0M 0 0 1181k 0 0:00:32 0:00:13 0:00:19 1132k 43 37.0M 43 16.1M 0 0 1179k 0 0:00:32 0:00:14 0:00:18 1134k 46 37.0M 46 17.3M 0 0 1178k 0 0:00:32 0:00:15 0:00:17 1134k 49 37.0M 49 18.4M 0 0 1177k 0 0:00:32 0:00:16 0:00:16 1147k 52 37.0M 52 19.5M 0 0 1172k  0 0:00:32 0:00:17 0:00:15 1139k 56 37.0M 56 20.7M 0 0 1177k 0 0:00:32 0:00:18 0:00:14 1165k 59 37.0M 59 21.9M 0 0 1183k 0 0:00:32 0:00:19 0:00:13 1193k 62 37.0M 62 23.2M 0 0 1189k 0 0:00:31 0:00:20 0:00:11 1220k 66 37.0M 66 24.5M  0 0 1195k 0 0:00:31 0:00:21 0:00:10 1251k 69 37.0M 69 25.8M 0 0 1201k 0 0:00:31 0:00:22 0:00:09 1298k 72 37.0M 72 26.8M 0 0 1194k 0 0:00:31 0:00:23 0:00:08 1256k 74 37.0M 74 27.7M 0 0 1179k 0 0:00:32 0:00:24 0:00:08 1166k 77 37.0M 77 28.6M 0 0 1170k 0 0:00:32 0:00:25 0:00:07 1095k 79 37.0M 79 29.5M 0 0 1160k 0 0:00:32 0:00:26 0:00:06 1015k 82 37.0M 82 30.4M 0 0 1152k 0 0:00:32 0:00:27 0:00:05 938k 84 37.0M 84 31.3M 0 0 1145k 0 0:00:33 0:00:28 0:00:05 920k 87 37.0M 87 32.3M 0 0 1141k 0 0:00:33 0:00:29 0:00:04 958k 90 37.0M 90 33.3M 0 0 1134k 0 0:00:33 0:00:30 0:00:03 958k 92 37.0M 92 34.0M 0 0 1122k 0 0:00:33 0:00:31 0:00:02 928k 93 37.0M 93 34.7M 0 0 1111k 0 0:00:34 0:00:32 0:00:02 887k 95 37.0M 95 35.4M 0 0  1098k 0 0:00:34 0:00:33 0:00:01 834k 97 37.0M 97 36.0M 0 0 1083k 0 0:00:34 0:00:34 --:--:-- 750k 99 37.0M 99 36.7M 0 0 1073k 0 0:00:35 0:00:35 --:--:-- 694k 100 37.0M 100 37.0M 0 0 1068k 0 0:00:35 0:00:35 --:--:-- 680k  ---> f9e1e4f398c5 Removing intermediate container 9203defea48c Step 4/4 : LABEL "alpine-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release0" '' ---> Running in a27afa31f418 ---> 0682c3ce4082 Removing intermediate container a27afa31f418 Successfully built 0682c3ce4082 sending incremental file list ./ Dockerfile sent 660 bytes received 34 bytes 1388.00 bytes/sec total size is 918 speedup is 1.32 Sending build context to Docker daemon 33.59 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 05ea8a13b62f Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virtctl ---> Running in 07613a9b83a4  ---> 44f819f3ca2e Removing intermediate container 07613a9b83a4 Step 4/8 : WORKDIR /home/virtctl ---> c47fe5e63ba9 Removing intermediate container 46bfab4b1f51 Step 5/8 : USER 1001 ---> Running in 64c082ed4453 ---> 35c7ba95b0f9 Removing intermediate container 64c082ed4453 Step 6/8 : COPY subresource-access-test /subresource-access-test ---> 42d524df0b7e Removing intermediate container 5652ddb8d698 Step 7/8 : ENTRYPOINT /subresource-access-test ---> Running in 6369be7e79a4 ---> 76ebf31a1fde Removing intermediate container 6369be7e79a4 Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release0" '' "subresource-access-test" '' ---> Running in 09008e2cd784 ---> 2bfaaa9192f3 Removing intermediate container 09008e2cd784 Successfully built 2bfaaa9192f3 hack/build-docker.sh push The push refers to a repository [localhost:32771/kubevirt/virt-controller] 5b256ab842e5: Preparing 5acaa362ebf9: Preparing 39bae602f753: Preparing 5acaa362ebf9: Pushed 5b256ab842e5: Pushed 39bae602f753: Pushed devel: digest: sha256:139173556e88b8cac3c038be5916c21304b0774248ea6fb5f3d06537eeb927d4 size: 948 The push refers to a repository [localhost:32771/kubevirt/virt-launcher] 7aafde543145: Preparing b1d468efc76e: Preparing b1d468efc76e: Preparing b61ce7455a54: Preparing 6b437064e6d6: Preparing 0d0e03be53e0: Preparing 605245e54d03: Preparing b91448d02ad0: Preparing a6250e94a411: Preparing b5ca5f52e363: Preparing 530cc55618cd: Preparing 34fa414dfdf6: Preparing a1359dc556dd: Preparing 605245e54d03: Waiting 490c7c373332: Preparing 530cc55618cd: Waiting 34fa414dfdf6: Waiting b5ca5f52e363: Waiting 4b440db36f72: Preparing a6250e94a411: Waiting 39bae602f753: Preparing 490c7c373332: Waiting 6b437064e6d6: Pushed 7aafde543145: Pushed 0d0e03be53e0: Pushed b1d468efc76e: Pushed b61ce7455a54: Pushed b91448d02ad0: Pushed a6250e94a411: Pushed 530cc55618cd: Pushed 34fa414dfdf6: Pushed a1359dc556dd: Pushed 490c7c373332: Pushed 39bae602f753: Mounted from kubevirt/virt-controller b5ca5f52e363: Pushed 605245e54d03: Pushed 4b440db36f72: Pushed devel: digest: sha256:ab5b5262079ea48863b9878741a208d0d470aebf6486bab78176347a41c3b23d size: 3652 The push refers to a repository [localhost:32771/kubevirt/virt-handler] 3823713358a2: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/virt-launcher 3823713358a2: Pushed devel: digest: sha256:dddce8a7b6331b479f46fbb8d2f8cecd8a73e3998cfa35fd24161970f49539b5 size: 740 The push refers to a repository [localhost:32771/kubevirt/virt-api] 28302b9958d2: Preparing fd1b1dc4959a: Preparing f1b07f50d664: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/virt-handler f1b07f50d664: Pushed fd1b1dc4959a: Pushed 28302b9958d2: Pushed devel: digest: sha256:96222037f554207947caf2360589821c2b44e29291b4ff25163fac954a44f37c size: 1159 The push refers to a repository [localhost:32771/kubevirt/iscsi-demo-target-tgtd] 49dd1f44d9e8: Preparing 56747852d8ee: Preparing 32ca1a8fc1e5: Preparing 1824f08143b0: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/virt-api 32ca1a8fc1e5: Pushed 49dd1f44d9e8: Pushed 56747852d8ee: Pushed 1824f08143b0: Pushed devel: digest: sha256:4af14ba2ab614dbfd698c9b0218a837c3d3ae608713ba93c36b4421cc8806c2d size: 1368 The push refers to a repository [localhost:32771/kubevirt/vm-killer] dbb456e9a766: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/iscsi-demo-target-tgtd dbb456e9a766: Pushed devel: digest: sha256:88781b2c10977b62353c43e3b31630c407b68fd12dd6a6f723c3fff853b6075f size: 740 The push refers to a repository [localhost:32771/kubevirt/registry-disk-v1alpha] 6326a937b16c: Preparing 57f3c2b869b9: Preparing 6709b2da72b8: Preparing 6326a937b16c: Pushed 57f3c2b869b9: Pushed 6709b2da72b8: Pushed devel: digest: sha256:66e23eb9bb99747e903a9d927b87a23648c661541f6f2794fa1786355406eecf size: 948 The push refers to a repository [localhost:32771/kubevirt/cirros-registry-disk-demo] 2493f0030550: Preparing 6326a937b16c: Preparing 57f3c2b869b9: Preparing 6709b2da72b8: Preparing 6326a937b16c: Mounted from kubevirt/registry-disk-v1alpha 6709b2da72b8: Mounted from kubevirt/registry-disk-v1alpha 57f3c2b869b9: Mounted from kubevirt/registry-disk-v1alpha 2493f0030550: Pushed devel: digest: sha256:9878ecee291ed2dad15bf70839ac127aeb988088dfb5797eaead67357def635e size: 1160 The push refers to a repository [localhost:32771/kubevirt/fedora-cloud-registry-disk-demo] 80861469fea4: Preparing 6326a937b16c: Preparing 57f3c2b869b9: Preparing 6709b2da72b8: Preparing 6709b2da72b8: Mounted from kubevirt/cirros-registry-disk-demo 6326a937b16c: Mounted from kubevirt/cirros-registry-disk-demo 57f3c2b869b9: Mounted from kubevirt/cirros-registry-disk-demo 80861469fea4: Pushed devel: digest: sha256:8feaf237983ceb5a36b6c789a83fb7b2c9da05a72a05143262dc985ede27641d size: 1161 The push refers to a repository [localhost:32771/kubevirt/alpine-registry-disk-demo] da22f2ddf48a: Preparing 6326a937b16c: Preparing 57f3c2b869b9: Preparing 6709b2da72b8: Preparing 6709b2da72b8: Mounted from kubevirt/fedora-cloud-registry-disk-demo 6326a937b16c: Mounted from kubevirt/fedora-cloud-registry-disk-demo 57f3c2b869b9: Mounted from kubevirt/fedora-cloud-registry-disk-demo da22f2ddf48a: Pushed devel: digest: sha256:c5776ac663e53d1e226e5fcc98a9734a5e3248ea8db3ef9bc4d94cc8e1ebe54f size: 1160 The push refers to a repository [localhost:32771/kubevirt/subresource-access-test] 83318c7f1b1b: Preparing b28bfdbf9e30: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/vm-killer b28bfdbf9e30: Pushed 83318c7f1b1b: Pushed devel: digest: sha256:0c669f329c7ff2f1a845b7fc53e3e04f2faea9ce394b3757af70407539d46a6d size: 948 make[1]: Leaving directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt' 2018/04/06 13:20:53 Waiting for host: 192.168.66.101:22 2018/04/06 13:20:53 Connected to tcp://192.168.66.101:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer df1f49e34671: Pulling fs layer f42cfcd5a7ed: Pulling fs layer df1f49e34671: Verifying Checksum df1f49e34671: Download complete f42cfcd5a7ed: Verifying Checksum f42cfcd5a7ed: Download complete 2176639d844b: Verifying Checksum 2176639d844b: Download complete 2176639d844b: Pull complete df1f49e34671: Pull complete f42cfcd5a7ed: Pull complete Digest: sha256:139173556e88b8cac3c038be5916c21304b0774248ea6fb5f3d06537eeb927d4 Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer 823d32e618f3: Pulling fs layer 2986076a9449: Pulling fs layer 0a026277e748: Pulling fs layer c4cf3bc48a62: Pulling fs layer a167430290eb: Pulling fs layer 6bd066788556: Pulling fs layer e0cbc8d86655: Pulling fs layer 8c631f6eb69e: Pulling fs layer 289d25bf0192: Pulling fs layer a1e80189bea5: Waiting 6cc174edcebf: Waiting 823d32e618f3: Waiting 2986076a9449: Waiting 0a026277e748: Waiting c4cf3bc48a62: Waiting e0cbc8d86655: Waiting 8c631f6eb69e: Waiting a167430290eb: Waiting 289d25bf0192: Waiting 6bd066788556: Waiting f2ef945504a7: Verifying Checksum f2ef945504a7: Download complete a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete 2986076a9449: Verifying Checksum 2986076a9449: Download complete 0a026277e748: Verifying Checksum 0a026277e748: Download complete c4cf3bc48a62: Verifying Checksum c4cf3bc48a62: Download complete 823d32e618f3: Verifying Checksum 823d32e618f3: Download complete a167430290eb: Verifying Checksum a167430290eb: Download complete 6bd066788556: Verifying Checksum 6bd066788556: Download complete e0cbc8d86655: Verifying Checksum e0cbc8d86655: Download complete 8c631f6eb69e: Verifying Checksum 8c631f6eb69e: Download complete 289d25bf0192: Verifying Checksum 289d25bf0192: Download complete d7240bccd145: Verifying Checksum d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete 823d32e618f3: Pull complete 2986076a9449: Pull complete 0a026277e748: Pull complete c4cf3bc48a62: Pull complete a167430290eb: Pull complete 6bd066788556: Pull complete e0cbc8d86655: Pull complete 8c631f6eb69e: Pull complete 289d25bf0192: Pull complete Digest: sha256:ab5b5262079ea48863b9878741a208d0d470aebf6486bab78176347a41c3b23d Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists 9e4bb6e629cd: Pulling fs layer 9e4bb6e629cd: Verifying Checksum 9e4bb6e629cd: Download complete 9e4bb6e629cd: Pull complete Digest: sha256:dddce8a7b6331b479f46fbb8d2f8cecd8a73e3998cfa35fd24161970f49539b5 Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists bfc4816ba8a8: Pulling fs layer 5d519c008f6c: Pulling fs layer 786d373b315e: Pulling fs layer bfc4816ba8a8: Verifying Checksum bfc4816ba8a8: Download complete 5d519c008f6c: Verifying Checksum 5d519c008f6c: Download complete 786d373b315e: Verifying Checksum 786d373b315e: Download complete bfc4816ba8a8: Pull complete 5d519c008f6c: Pull complete 786d373b315e: Pull complete Digest: sha256:96222037f554207947caf2360589821c2b44e29291b4ff25163fac954a44f37c Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists cd017ee9a697: Pulling fs layer 976bd1a00d38: Pulling fs layer 3e3efc0a0202: Pulling fs layer 305dfa101de9: Pulling fs layer 305dfa101de9: Waiting 976bd1a00d38: Verifying Checksum 976bd1a00d38: Download complete 305dfa101de9: Verifying Checksum 305dfa101de9: Download complete 3e3efc0a0202: Verifying Checksum 3e3efc0a0202: Download complete cd017ee9a697: Verifying Checksum cd017ee9a697: Download complete cd017ee9a697: Pull complete 976bd1a00d38: Pull complete 3e3efc0a0202: Pull complete 305dfa101de9: Pull complete Digest: sha256:4af14ba2ab614dbfd698c9b0218a837c3d3ae608713ba93c36b4421cc8806c2d Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists b39e4e15af68: Pulling fs layer b39e4e15af68: Verifying Checksum b39e4e15af68: Download complete b39e4e15af68: Pull complete Digest: sha256:88781b2c10977b62353c43e3b31630c407b68fd12dd6a6f723c3fff853b6075f Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer d31dd609fdb4: Pulling fs layer e6ebe1dfd863: Pulling fs layer e6ebe1dfd863: Verifying Checksum e6ebe1dfd863: Download complete d31dd609fdb4: Verifying Checksum d31dd609fdb4: Download complete 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 2115d46e7396: Pull complete d31dd609fdb4: Pull complete e6ebe1dfd863: Pull complete Digest: sha256:66e23eb9bb99747e903a9d927b87a23648c661541f6f2794fa1786355406eecf Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists d31dd609fdb4: Already exists e6ebe1dfd863: Already exists 9d951451e44b: Pulling fs layer 9d951451e44b: Verifying Checksum 9d951451e44b: Download complete 9d951451e44b: Pull complete Digest: sha256:9878ecee291ed2dad15bf70839ac127aeb988088dfb5797eaead67357def635e Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists d31dd609fdb4: Already exists e6ebe1dfd863: Already exists 5d4bfaeea8d4: Pulling fs layer 5d4bfaeea8d4: Verifying Checksum 5d4bfaeea8d4: Download complete 5d4bfaeea8d4: Pull complete Digest: sha256:8feaf237983ceb5a36b6c789a83fb7b2c9da05a72a05143262dc985ede27641d Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists d31dd609fdb4: Already exists e6ebe1dfd863: Already exists de2a01f1990b: Pulling fs layer de2a01f1990b: Verifying Checksum de2a01f1990b: Download complete de2a01f1990b: Pull complete Digest: sha256:c5776ac663e53d1e226e5fcc98a9734a5e3248ea8db3ef9bc4d94cc8e1ebe54f Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists b1facae8a0a3: Pulling fs layer fcc8172a42a4: Pulling fs layer b1facae8a0a3: Verifying Checksum b1facae8a0a3: Download complete fcc8172a42a4: Verifying Checksum fcc8172a42a4: Download complete b1facae8a0a3: Pull complete fcc8172a42a4: Pull complete Digest: sha256:0c669f329c7ff2f1a845b7fc53e3e04f2faea9ce394b3757af70407539d46a6d 2018/04/06 13:23:45 Waiting for host: 192.168.66.101:22 2018/04/06 13:23:45 Connected to tcp://192.168.66.101:22 2018/04/06 13:23:48 Waiting for host: 192.168.66.102:22 2018/04/06 13:23:48 Connected to tcp://192.168.66.102:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer df1f49e34671: Pulling fs layer f42cfcd5a7ed: Pulling fs layer df1f49e34671: Verifying Checksum df1f49e34671: Download complete f42cfcd5a7ed: Verifying Checksum f42cfcd5a7ed: Download complete 2176639d844b: Verifying Checksum 2176639d844b: Download complete 2176639d844b: Pull complete df1f49e34671: Pull complete f42cfcd5a7ed: Pull complete Digest: sha256:139173556e88b8cac3c038be5916c21304b0774248ea6fb5f3d06537eeb927d4 Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer 823d32e618f3: Pulling fs layer 2986076a9449: Pulling fs layer 0a026277e748: Pulling fs layer c4cf3bc48a62: Pulling fs layer a167430290eb: Pulling fs layer 6bd066788556: Pulling fs layer e0cbc8d86655: Pulling fs layer 8c631f6eb69e: Pulling fs layer 289d25bf0192: Pulling fs layer c4cf3bc48a62: Waiting a1e80189bea5: Waiting 6cc174edcebf: Waiting a167430290eb: Waiting 823d32e618f3: Waiting 6bd066788556: Waiting 2986076a9449: Waiting e0cbc8d86655: Waiting 8c631f6eb69e: Waiting 0a026277e748: Waiting 289d25bf0192: Waiting a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete f2ef945504a7: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete 2986076a9449: Verifying Checksum 2986076a9449: Download complete 0a026277e748: Verifying Checksum 0a026277e748: Download complete 823d32e618f3: Verifying Checksum 823d32e618f3: Download complete c4cf3bc48a62: Verifying Checksum c4cf3bc48a62: Download complete a167430290eb: Verifying Checksum a167430290eb: Download complete 6bd066788556: Verifying Checksum 6bd066788556: Download complete e0cbc8d86655: Verifying Checksum e0cbc8d86655: Download complete 8c631f6eb69e: Verifying Checksum 8c631f6eb69e: Download complete 289d25bf0192: Verifying Checksum 289d25bf0192: Download complete d7240bccd145: Verifying Checksum d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete 823d32e618f3: Pull complete 2986076a9449: Pull complete 0a026277e748: Pull complete c4cf3bc48a62: Pull complete a167430290eb: Pull complete 6bd066788556: Pull complete e0cbc8d86655: Pull complete 8c631f6eb69e: Pull complete 289d25bf0192: Pull complete Digest: sha256:ab5b5262079ea48863b9878741a208d0d470aebf6486bab78176347a41c3b23d Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists 9e4bb6e629cd: Pulling fs layer 9e4bb6e629cd: Verifying Checksum 9e4bb6e629cd: Download complete 9e4bb6e629cd: Pull complete Digest: sha256:dddce8a7b6331b479f46fbb8d2f8cecd8a73e3998cfa35fd24161970f49539b5 Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists bfc4816ba8a8: Pulling fs layer 5d519c008f6c: Pulling fs layer 786d373b315e: Pulling fs layer 5d519c008f6c: Verifying Checksum 5d519c008f6c: Download complete bfc4816ba8a8: Verifying Checksum bfc4816ba8a8: Download complete 786d373b315e: Download complete bfc4816ba8a8: Pull complete 5d519c008f6c: Pull complete 786d373b315e: Pull complete Digest: sha256:96222037f554207947caf2360589821c2b44e29291b4ff25163fac954a44f37c Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists cd017ee9a697: Pulling fs layer 976bd1a00d38: Pulling fs layer 3e3efc0a0202: Pulling fs layer 305dfa101de9: Pulling fs layer 305dfa101de9: Waiting 976bd1a00d38: Verifying Checksum 976bd1a00d38: Download complete 305dfa101de9: Verifying Checksum 305dfa101de9: Download complete 3e3efc0a0202: Verifying Checksum 3e3efc0a0202: Download complete cd017ee9a697: Verifying Checksum cd017ee9a697: Download complete cd017ee9a697: Pull complete 976bd1a00d38: Pull complete 3e3efc0a0202: Pull complete 305dfa101de9: Pull complete Digest: sha256:4af14ba2ab614dbfd698c9b0218a837c3d3ae608713ba93c36b4421cc8806c2d Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists b39e4e15af68: Pulling fs layer b39e4e15af68: Verifying Checksum b39e4e15af68: Download complete b39e4e15af68: Pull complete Digest: sha256:88781b2c10977b62353c43e3b31630c407b68fd12dd6a6f723c3fff853b6075f Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer d31dd609fdb4: Pulling fs layer e6ebe1dfd863: Pulling fs layer e6ebe1dfd863: Verifying Checksum e6ebe1dfd863: Download complete d31dd609fdb4: Verifying Checksum d31dd609fdb4: Download complete 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 2115d46e7396: Pull complete d31dd609fdb4: Pull complete e6ebe1dfd863: Pull complete Digest: sha256:66e23eb9bb99747e903a9d927b87a23648c661541f6f2794fa1786355406eecf Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists d31dd609fdb4: Already exists e6ebe1dfd863: Already exists 9d951451e44b: Pulling fs layer 9d951451e44b: Verifying Checksum 9d951451e44b: Download complete 9d951451e44b: Pull complete Digest: sha256:9878ecee291ed2dad15bf70839ac127aeb988088dfb5797eaead67357def635e Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists d31dd609fdb4: Already exists e6ebe1dfd863: Already exists 5d4bfaeea8d4: Pulling fs layer 5d4bfaeea8d4: Download complete 5d4bfaeea8d4: Pull complete Digest: sha256:8feaf237983ceb5a36b6c789a83fb7b2c9da05a72a05143262dc985ede27641d Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists d31dd609fdb4: Already exists e6ebe1dfd863: Already exists de2a01f1990b: Pulling fs layer de2a01f1990b: Verifying Checksum de2a01f1990b: Download complete de2a01f1990b: Pull complete Digest: sha256:c5776ac663e53d1e226e5fcc98a9734a5e3248ea8db3ef9bc4d94cc8e1ebe54f Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists b1facae8a0a3: Pulling fs layer fcc8172a42a4: Pulling fs layer b1facae8a0a3: Verifying Checksum b1facae8a0a3: Download complete fcc8172a42a4: Download complete b1facae8a0a3: Pull complete fcc8172a42a4: Pull complete Digest: sha256:0c669f329c7ff2f1a845b7fc53e3e04f2faea9ce394b3757af70407539d46a6d 2018/04/06 13:26:26 Waiting for host: 192.168.66.102:22 2018/04/06 13:26:26 Connected to tcp://192.168.66.102:22 Done ./cluster/clean.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0-alpha.4 ++ provider_prefix=kubevirt-functional-tests-openshift-release0 ++ job_prefix=kubevirt-functional-tests-openshift-release0 + source cluster/os-3.9.0-alpha.4/provider.sh ++ set -e ++ image=os-3.9@sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8 ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0-alpha.4 ++ source hack/config-default.sh source hack/config-os-3.9.0-alpha.4.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0-alpha.4.sh ++ source hack/config-provider-os-3.9.0-alpha.4.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-alpha.4/.kubeconfig +++ docker_prefix=localhost:32771/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Cleaning up ...' Cleaning up ... + _kubectl delete ds -l kubevirt.io -n kube-system --cascade=false --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=libvirt --force --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=virt-handler --force --grace-period 0 No resources found + namespaces=(default ${namespace}) + for i in '${namespaces[@]}' + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete deployment -l kubevirt.io No resources found + _kubectl -n default delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete rs -l kubevirt.io No resources found + _kubectl -n default delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete services -l kubevirt.io No resources found + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete secrets -l kubevirt.io No resources found + _kubectl -n default delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete pv -l kubevirt.io No resources found + _kubectl -n default delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete pvc -l kubevirt.io No resources found + _kubectl -n default delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete ds -l kubevirt.io No resources found + _kubectl -n default delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n default delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete pods -l kubevirt.io No resources found + _kubectl -n default delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n default delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete rolebinding -l kubevirt.io No resources found + _kubectl -n default delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete roles -l kubevirt.io No resources found + _kubectl -n default delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete clusterroles -l kubevirt.io No resources found + _kubectl -n default delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n default get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ cluster/os-3.9.0-alpha.4/.kubectl -n default get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + for i in '${namespaces[@]}' + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete deployment -l kubevirt.io No resources found + _kubectl -n kube-system delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete rs -l kubevirt.io No resources found + _kubectl -n kube-system delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete services -l kubevirt.io No resources found + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete secrets -l kubevirt.io No resources found + _kubectl -n kube-system delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete pv -l kubevirt.io No resources found + _kubectl -n kube-system delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete pvc -l kubevirt.io No resources found + _kubectl -n kube-system delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete ds -l kubevirt.io No resources found + _kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n kube-system delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete pods -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete rolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete roles -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete clusterroles -l kubevirt.io No resources found + _kubectl -n kube-system delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ cluster/os-3.9.0-alpha.4/.kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + sleep 2 + echo Done Done ./cluster/deploy.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0-alpha.4 ++ provider_prefix=kubevirt-functional-tests-openshift-release0 ++ job_prefix=kubevirt-functional-tests-openshift-release0 + source cluster/os-3.9.0-alpha.4/provider.sh ++ set -e ++ image=os-3.9@sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8 ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0-alpha.4 ++ source hack/config-default.sh source hack/config-os-3.9.0-alpha.4.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0-alpha.4.sh ++ source hack/config-provider-os-3.9.0-alpha.4.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-alpha.4/.kubeconfig +++ docker_prefix=localhost:32771/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Deploying ...' Deploying ... + [[ -z openshift-release ]] + [[ openshift-release =~ .*-dev ]] + [[ openshift-release =~ .*-release ]] + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/demo-content.yaml =~ .*demo.* ]] + continue + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml =~ .*demo.* ]] + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml serviceaccount "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver-auth-delegator" created rolebinding "kubevirt-apiserver" created role "kubevirt-apiserver" created clusterrole "kubevirt-apiserver" created clusterrole "kubevirt-controller" created serviceaccount "kubevirt-controller" created serviceaccount "kubevirt-privileged" created clusterrolebinding "kubevirt-controller" created clusterrolebinding "kubevirt-controller-cluster-admin" created clusterrolebinding "kubevirt-privileged-cluster-admin" created customresourcedefinition "virtualmachines.kubevirt.io" created customresourcedefinition "virtualmachinereplicasets.kubevirt.io" created service "virt-api" created deployment "virt-api" created deployment "virt-controller" created daemonset "virt-handler" created customresourcedefinition "virtualmachinepresets.kubevirt.io" created customresourcedefinition "offlinevirtualmachines.kubevirt.io" created + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R persistentvolumeclaim "disk-alpine" created persistentvolume "iscsi-disk-alpine" created daemonset "iscsi-demo-target-tgtd" created serviceaccount "kubevirt-testing" created clusterrolebinding "kubevirt-testing-cluster-admin" created + '[' os-3.9.0-alpha.4 = vagrant-openshift ']' + '[' os-3.9.0-alpha.4 = os-3.9.0-alpha.4 ']' + _kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-controller"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-testing"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-privileged"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-apiserver"] + _kubectl adm policy add-scc-to-user privileged admin + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged admin scc "privileged" added to: ["admin"] + echo Done Done ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n 'iscsi-demo-target-tgtd-bn5wp 0/1 ContainerCreating 0 5s iscsi-demo-target-tgtd-p6dvs 0/1 ContainerCreating 0 5s virt-api-fd96f94b5-997lp 0/1 ContainerCreating 0 7s virt-controller-5f7c946cc4-7dnt2 0/1 ContainerCreating 0 7s virt-controller-5f7c946cc4-rf2rt 0/1 ContainerCreating 0 8s virt-handler-s9xv4 0/1 ContainerCreating 0 2s virt-handler-vhw9t 0/1 ContainerCreating 0 1s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + cluster/kubectl.sh get pods -n kube-system --no-headers + grep -v Running iscsi-demo-target-tgtd-bn5wp 0/1 ContainerCreating 0 6s iscsi-demo-target-tgtd-p6dvs 0/1 ContainerCreating 0 6s virt-api-fd96f94b5-997lp 0/1 ContainerCreating 0 8s virt-controller-5f7c946cc4-7dnt2 0/1 ContainerCreating 0 8s virt-controller-5f7c946cc4-rf2rt 0/1 ContainerCreating 0 9s virt-handler-s9xv4 0/1 ContainerCreating 0 3s virt-handler-vhw9t 0/1 ContainerCreating 0 2s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ cluster/kubectl.sh get pods -n kube-system --no-headers ++ grep -v Running + '[' -n '' ']' ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-bn5wp false iscsi-demo-target-tgtd-p6dvs + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n '' ']' ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '/virt-controller/ && /true/' ++ wc -l + '[' 1 -lt 1 ']' + kubectl get pods -n kube-system + cluster/kubectl.sh get pods -n kube-system NAME READY STATUS RESTARTS AGE iscsi-demo-target-tgtd-bn5wp 1/1 Running 1 1m iscsi-demo-target-tgtd-p6dvs 1/1 Running 1 1m virt-api-fd96f94b5-997lp 1/1 Running 0 1m virt-api-fd96f94b5-jxb8v 1/1 Running 0 1m virt-controller-5f7c946cc4-7dnt2 0/1 Running 0 1m virt-controller-5f7c946cc4-rf2rt 1/1 Running 0 1m virt-handler-s9xv4 1/1 Running 0 1m virt-handler-vhw9t 1/1 Running 0 1m + kubectl version + cluster/kubectl.sh version oc v3.9.0-alpha.4+9ab7a71 kubernetes v1.9.1+a0ce1bc657 features: Basic-Auth GSSAPI Kerberos SPNEGO Server https://127.0.0.1:32768 openshift v3.9.0-alpha.4+9ab7a71 kubernetes v1.9.1+a0ce1bc657 + ginko_params=--ginkgo.noColor + [[ -d /home/nfs/images/windows2016 ]] + FUNC_TEST_ARGS=--ginkgo.noColor + make functest hack/dockerized "hack/build-func-tests.sh" sha256:6a9ee4540f213c441209d0e9c25ce1699434ef857140b3f1b80f0a2741809637 go version go1.9.2 linux/amd64 skipping directory . go version go1.9.2 linux/amd64 Compiling tests... compiled tests.test b43163a2c613627a2c52edbd597e9d15c878801a61f4044aca139ab95f5d6424 b43163a2c613627a2c52edbd597e9d15c878801a61f4044aca139ab95f5d6424 hack/functests.sh Running Suite: Tests Suite ========================== Random Seed: 1523021343 Will run 67 of 67 specs • [SLOW TEST:21.228 seconds] VNC /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:46 with VNC connection /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:47 should allow accessing the VNC device /root/go/src/kubevirt.io/kubevirt/tests/vnc_test.go:48 ------------------------------ • [SLOW TEST:81.939 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:42 Starting and stopping the same VM /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:91 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:92 should success multiple times /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:93 ------------------------------ • [SLOW TEST:20.387 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:42 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:112 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:113 should not modify the spec on status update /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:114 ------------------------------ • [SLOW TEST:46.955 seconds] RegistryDisk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:42 Starting multiple VMs /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:130 with ephemeral registry disk /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:131 should success /root/go/src/kubevirt.io/kubevirt/tests/registry_disk_test.go:132 ------------------------------ • [SLOW TEST:42.685 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:64 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a cirros image /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:66 should return that we are running cirros /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:67 ------------------------------ • [SLOW TEST:46.223 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:64 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 with a fedora image /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:76 should return that we are running fedora /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:77 ------------------------------ • [SLOW TEST:42.935 seconds] Console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:35 A new VM /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:64 with a serial console /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:65 should be able to reconnect to console multiple times /root/go/src/kubevirt.io/kubevirt/tests/console_test.go:86 ------------------------------ • [SLOW TEST:49.557 seconds] Health Monitoring /root/go/src/kubevirt.io/kubevirt/tests/vm_monitoring_test.go:37 A VM with a watchdog device /root/go/src/kubevirt.io/kubevirt/tests/vm_monitoring_test.go:56 should be shut down when the watchdog expires /root/go/src/kubevirt.io/kubevirt/tests/vm_monitoring_test.go:57 ------------------------------ •STEP: Getting the ISCSI pod logs STEP: Checking that ISCSI is ready STEP: Starting a VM STEP: Waiting until the VM will start level=info timestamp=2018-04-06T13:35:21.711245Z pos=utils.go:203 component=tests msg="VM defined." level=info timestamp=2018-04-06T13:35:21.759481Z pos=utils.go:203 component=tests msg="VM started." STEP: Checking that the VM console has expected output ------------------------------ • Failure [224.121 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:41 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:121 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:122 should be successfully started [It] /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:123 Expected : 200000000000 to be nil /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:139 ------------------------------ • [SLOW TEST:173.096 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:41 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:121 with Alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:122 should be successfully started and stopped multiple times /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:144 ------------------------------ • [SLOW TEST:43.642 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:41 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:121 With an emptyDisk defined /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:177 should create a writeable emptyDisk with the right capacity /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:179 ------------------------------ STEP: Getting the ISCSI pod logs STEP: Checking that ISCSI is ready STEP: Starting a VM STEP: Waiting until the VM will start level=info timestamp=2018-04-06T13:42:40.570163Z pos=utils.go:203 component=tests msg="VM defined." level=error timestamp=2018-04-06T13:42:40.579141Z pos=utils.go:201 component=tests reason="unexpected warning event recieved" msg="unknown error encountered sending command Launcher.Sync: read unix @->/var/run/kubevirt/sockets/kubevirt-test-default_testvmv2hhd_sock: read: connection reset by peer" • Failure [241.004 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:41 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:121 With ephemeral alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:229 should be successfully started [It] /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:231 Unexpected Warning event recieved. Expected : Warning not to equal : Warning /root/go/src/kubevirt.io/kubevirt/tests/utils.go:205 ------------------------------ STEP: Getting the ISCSI pod logs STEP: Checking that ISCSI is ready STEP: Starting the VM STEP: Starting a VM STEP: Waiting until the VM will start level=info timestamp=2018-04-06T13:47:11.946812Z pos=utils.go:203 component=tests msg="VM defined." level=info timestamp=2018-04-06T13:47:11.960063Z pos=utils.go:203 component=tests msg="VM started." STEP: Writing an arbitrary file to it's EFI partition STEP: Killing a VM level=info timestamp=2018-04-06T13:47:27.127643Z pos=utils.go:214 component=tests msg="VM defined." level=info timestamp=2018-04-06T13:47:27.127771Z pos=utils.go:214 component=tests msg="VM started." level=info timestamp=2018-04-06T13:47:27.313566Z pos=utils.go:214 component=tests msg="VM stopping" STEP: Starting the VM again STEP: Starting a VM STEP: Waiting until the VM will start level=info timestamp=2018-04-06T13:47:27.621067Z pos=utils.go:203 component=tests msg="VM defined." level=info timestamp=2018-04-06T13:47:27.621252Z pos=utils.go:203 component=tests msg="VM started." • Failure [158.223 seconds] Storage /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:41 Starting a VM /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:121 With ephemeral alpine PVC /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:229 should not persist data [It] /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:252 Timed out after 90.119s. Expected : false to equal : true /root/go/src/kubevirt.io/kubevirt/tests/utils.go:809 ------------------------------ • [SLOW TEST:52.009 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:46 VirtualMachine with nodeNetwork definition given /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:108 should be able to reach /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 the Inbound VM /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • ------------------------------ • [SLOW TEST:7.384 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:46 VirtualMachine with nodeNetwork definition given /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:108 should be reachable via the propagated IP from a Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 on the same node from Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ • [SLOW TEST:6.160 seconds] Networking /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:46 VirtualMachine with nodeNetwork definition given /root/go/src/kubevirt.io/kubevirt/tests/vm_networking_test.go:108 should be reachable via the propagated IP from a Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 on a different node from Pod /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 ------------------------------ •• ------------------------------ • [SLOW TEST:8.440 seconds] Subresource Api /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:37 Rbac Authorization /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:48 with correct permissions /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:49 should be allowed to access subresource endpoint /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:50 ------------------------------ • [SLOW TEST:6.970 seconds] Subresource Api /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:37 Rbac Authorization /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:48 Without permissions /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:54 should not be able to access subresource endpoint /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:55 ------------------------------ •••••••• ------------------------------ • [SLOW TEST:19.988 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 should update OfflineVirtualMachine once VMs are up /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:144 ------------------------------ •• ------------------------------ • [SLOW TEST:8.270 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 shloud recreate VM if it gets deleted /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:194 ------------------------------ • [SLOW TEST:23.381 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 should stop VM if running set to false /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:210 ------------------------------ • [SLOW TEST:242.123 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 should start and stop VM multiple times /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:218 ------------------------------ • [SLOW TEST:75.154 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 should not update the VM spec if Running /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:231 ------------------------------ STEP: Creating new OVM, not running STEP: Starting the VM STEP: OVM has the running condition STEP: Getting the running VM STEP: Obtaining the serial console STEP: Guest shutdown STEP: Testing the VM is not running • Failure [283.548 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 should survive guest shutdown, multiple times [It] /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:272 Timed out after 240.000s. Expected : false to be true /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:304 ------------------------------ • [SLOW TEST:21.961 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 Using virtctl interface /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:310 should start a VM once /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:311 ------------------------------ • [SLOW TEST:28.295 seconds] OfflineVirtualMachine /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:44 A valid OfflineVirtualMachine given /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:55 Using virtctl interface /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:310 should stop a VM once /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:342 ------------------------------ • [SLOW TEST:46.659 seconds] LeaderElection /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:43 Start a VM /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:53 when the controller pod is not running /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:54 should success /root/go/src/kubevirt.io/kubevirt/tests/controller_leader_election_test.go:55 ------------------------------ •• ------------------------------ • [SLOW TEST:26.399 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:41 should update readyReplicas once VMs are up /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:102 ------------------------------ • ------------------------------ • [SLOW TEST:5.486 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:41 should remove owner references on the VM if it is orphan deleted /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:130 ------------------------------ • [SLOW TEST:5.181 seconds] VirtualMachineReplicaSet /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:41 should not scale when paused and scale when resume /root/go/src/kubevirt.io/kubevirt/tests/replicaset_test.go:168 ------------------------------ volumedisk0 compute • [SLOW TEST:47.418 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:39 VM definition /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:50 with 3 CPU cores /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:51 should report 3 cpu cores under guest OS /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:57 ------------------------------ • [SLOW TEST:44.748 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:39 New VM with all supported drives /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:109 should have all the device nodes /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:132 ------------------------------ • ------------------------------ • [SLOW TEST:20.151 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 should start it /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:61 ------------------------------ • [SLOW TEST:24.820 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 should attach virt-launcher to it /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:69 ------------------------------ • ------------------------------ • [SLOW TEST:24.599 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:99 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:100 should retry starting the VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:101 ------------------------------ • [SLOW TEST:20.820 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 with user-data /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:99 without k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:100 should log warning and proceed once the secret is there /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:132 ------------------------------ • [SLOW TEST:54.080 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 when virt-launcher crashes /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:180 should be stopped and have Failed phase /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:181 ------------------------------ • [SLOW TEST:48.683 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 when virt-handler crashes /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:210 should recover and continue management /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:211 ------------------------------ S [SKIPPING] [1.262 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:247 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-default [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Skip log query tests for JENKINS ci test environment /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:252 ------------------------------ S [SKIPPING] [0.828 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Creating a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:55 with non default namespace /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:247 should log libvirt start and stop lifecycle events of the domain /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table.go:92 kubevirt-test-alternative [It] /root/go/src/kubevirt.io/kubevirt/vendor/github.com/onsi/ginkgo/extensions/table/table_entry.go:46 Skip log query tests for JENKINS ci test environment /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:252 ------------------------------ • ------------------------------ • [SLOW TEST:25.898 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Delete a VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:321 with grace period greater than 0 /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:322 should run graceful shutdown /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:323 ------------------------------ • [SLOW TEST:36.088 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Killed VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:376 should be in Failed phase /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:377 ------------------------------ • [SLOW TEST:35.228 seconds] Vmlifecycle /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:41 Killed VM /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:376 should be left alone by virt-handler /root/go/src/kubevirt.io/kubevirt/tests/vmlifecycle_test.go:405 ------------------------------ • [SLOW TEST:40.776 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 should have cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:81 ------------------------------ • [SLOW TEST:93.360 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 with injected ssh-key /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:92 should have ssh-key under authorized keys /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:93 ------------------------------ • [SLOW TEST:44.610 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 with cloudInitNoCloud userData source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:116 should process provided cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:117 ------------------------------ • [SLOW TEST:42.763 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 should take user-data from k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:160 ------------------------------ Waiting for namespace kubevirt-test-default to be removed, this can take a while ... Waiting for namespace kubevirt-test-alternative to be removed, this can take a while ... Summarizing 4 Failures: [Fail] Storage Starting a VM with Alpine PVC [It] should be successfully started /root/go/src/kubevirt.io/kubevirt/tests/storage_test.go:139 [Fail] Storage Starting a VM With ephemeral alpine PVC [It] should be successfully started /root/go/src/kubevirt.io/kubevirt/tests/utils.go:205 [Fail] Storage Starting a VM With ephemeral alpine PVC [It] should not persist data /root/go/src/kubevirt.io/kubevirt/tests/utils.go:809 [Fail] OfflineVirtualMachine A valid OfflineVirtualMachine given [It] should survive guest shutdown, multiple times /root/go/src/kubevirt.io/kubevirt/tests/ovm_test.go:304 Ran 65 of 67 Specs in 2743.786 seconds FAIL! -- 61 Passed | 4 Failed | 0 Pending | 2 Skipped --- FAIL: TestTests (2743.79s) FAIL make: *** [functest] Error 1 + make cluster-down ./cluster/down.sh 0ec2d0ff4426 1267c2ff45d5 2107587bd2cd b0ed0a66e441 0ec2d0ff4426 1267c2ff45d5 2107587bd2cd b0ed0a66e441 kubevirt-functional-tests-openshift-release0-node01 kubevirt-functional-tests-openshift-release0-node02