+ export WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release + WORKSPACE=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release + [[ openshift-release =~ openshift-.* ]] + export PROVIDER=os-3.9.0-alpha.4 + PROVIDER=os-3.9.0-alpha.4 + export VAGRANT_NUM_NODES=1 + VAGRANT_NUM_NODES=1 + export NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + NFS_WINDOWS_DIR=/home/nfs/images/windows2016 + export NAMESPACE=kube-system + NAMESPACE=kube-system + trap '{ make cluster-down; }' EXIT + make cluster-down ./cluster/down.sh + make cluster-up ./cluster/up.sh WARNING: You're not using the default seccomp profile kubevirt-functional-tests-openshift-release0-node02 2018/04/09 15:19:32 Waiting for host: 192.168.66.102:22 2018/04/09 15:19:35 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:19:43 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:19:51 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:19:59 Problem with dial: dial tcp 192.168.66.102:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:20:07 Connected to tcp://192.168.66.102:22 Removed symlink /etc/systemd/system/docker.service.wants/origin-master-api.service. Removed symlink /etc/systemd/system/origin-node.service.wants/origin-master-api.service. Removed symlink /etc/systemd/system/docker.service.wants/origin-master-controllers.service. kubevirt-functional-tests-openshift-release0-node01 2018/04/09 15:20:13 Waiting for host: 192.168.66.101:22 2018/04/09 15:20:16 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:20:24 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:20:32 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:20:40 Problem with dial: dial tcp 192.168.66.101:22: getsockopt: no route to host. Sleeping 5s 2018/04/09 15:20:45 Connected to tcp://192.168.66.101:22 The connection to the server node01:8443 was refused - did you specify the right host or port? NAME STATUS ROLES AGE VERSION node01 Ready master 4d v1.9.1+a0ce1bc657 PING node02 (192.168.66.102) 56(84) bytes of data. 64 bytes from node02 (192.168.66.102): icmp_seq=1 ttl=64 time=0.904 ms --- node02 ping statistics --- 1 packets transmitted, 1 received, 0% packet loss, time 0ms rtt min/avg/max/mdev = 0.904/0.904/0.904/0.000 ms Found node02. Adding it to the inventory. ping: node03: Name or service not known PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) PLAY [Ensure there are new_nodes] ********************************************** TASK [fail] ******************************************************************** skipping: [localhost] TASK [fail] ******************************************************************** skipping: [localhost] PLAY [Initialization Checkpoint Start] ***************************************** TASK [Set install initialization 'In Progress'] ******************************** ok: [node01] PLAY [Populate config host groups] ********************************************* TASK [Load group name mapping variables] *************************************** ok: [localhost] TASK [Evaluate groups - g_etcd_hosts or g_new_etcd_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_master_hosts or g_new_master_hosts required] ********* skipping: [localhost] TASK [Evaluate groups - g_node_hosts or g_new_node_hosts required] ************* skipping: [localhost] TASK [Evaluate groups - g_lb_hosts required] *********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts required] ********************************** skipping: [localhost] TASK [Evaluate groups - g_nfs_hosts is single host] **************************** skipping: [localhost] TASK [Evaluate groups - g_glusterfs_hosts required] **************************** skipping: [localhost] TASK [Evaluate groups - Fail if no etcd hosts group is defined] **************** skipping: [localhost] TASK [Evaluate oo_all_hosts] *************************************************** ok: [localhost] => (item=node01) ok: [localhost] => (item=node02) TASK [Evaluate oo_masters] ***************************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_master] ************************************************ ok: [localhost] TASK [Evaluate oo_new_etcd_to_config] ****************************************** TASK [Evaluate oo_masters_to_config] ******************************************* ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_to_config] ********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_first_etcd] ************************************************** ok: [localhost] TASK [Evaluate oo_etcd_hosts_to_upgrade] *************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_etcd_hosts_to_backup] **************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_nodes_to_config] ********************************************* ok: [localhost] => (item=node02) TASK [Add master to oo_nodes_to_config] **************************************** skipping: [localhost] => (item=node01) TASK [Evaluate oo_lb_to_config] ************************************************ TASK [Evaluate oo_nfs_to_config] *********************************************** ok: [localhost] => (item=node01) TASK [Evaluate oo_glusterfs_to_config] ***************************************** TASK [Evaluate oo_etcd_to_migrate] ********************************************* ok: [localhost] => (item=node01) [WARNING]: Could not match supplied host pattern, ignoring: oo_lb_to_config PLAY [Ensure that all non-node hosts are accessible] *************************** TASK [Gathering Facts] ********************************************************* ok: [node01] PLAY [Initialize basic host facts] ********************************************* TASK [Gathering Facts] ********************************************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/deprecations.yml for node01, node02 TASK [openshift_sanitize_inventory : Check for usage of deprecated variables] *** skipping: [node01] => (item=openshift_hosted_logging_deploy) skipping: [node02] => (item=openshift_hosted_logging_deploy) skipping: [node01] => (item=openshift_hosted_logging_hostname) skipping: [node02] => (item=openshift_hosted_logging_hostname) skipping: [node01] => (item=openshift_hosted_logging_ops_hostname) skipping: [node01] => (item=openshift_hosted_logging_master_public_url) skipping: [node02] => (item=openshift_hosted_logging_ops_hostname) skipping: [node02] => (item=openshift_hosted_logging_master_public_url) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_cluster_size) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_cluster_size) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_cluster_size) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_cluster_size) skipping: [node01] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node02] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node01] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node02] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node02] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node01] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node02] => (item=openshift_hosted_logging_journal_source) skipping: [node02] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node01] => (item=openshift_hosted_logging_journal_source) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node01] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node02] => (item=openshift_hosted_logging_storage_labels) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_pvc_dynamic) skipping: [node01] => (item=openshift_hosted_logging_storage_labels) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_pvc_size) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_pvc_dynamic) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_pvc_size) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_pvc_prefix) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_pvc_prefix) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node02] => (item=openshift_hosted_loggingops_storage_labels) skipping: [node01] => (item=openshift_hosted_loggingops_storage_labels) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_dynamic) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_dynamic) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_size) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_size) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_prefix) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_pvc_prefix) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_storage_group) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_storage_kind) skipping: [node01] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node02] => (item=openshift_hosted_loggingops_storage_kind) skipping: [node01] => (item=openshift_hosted_logging_storage_kind) skipping: [node02] => (item=openshift_hosted_logging_storage_host) skipping: [node01] => (item=openshift_hosted_loggingops_storage_kind) skipping: [node02] => (item=openshift_hosted_loggingops_storage_host) skipping: [node01] => (item=openshift_hosted_logging_storage_host) skipping: [node02] => (item=openshift_hosted_logging_storage_nfs_directory) skipping: [node01] => (item=openshift_hosted_loggingops_storage_host) skipping: [node02] => (item=openshift_hosted_loggingops_storage_nfs_directory) skipping: [node01] => (item=openshift_hosted_logging_storage_nfs_directory) skipping: [node02] => (item=openshift_hosted_logging_storage_volume_name) skipping: [node01] => (item=openshift_hosted_loggingops_storage_nfs_directory) skipping: [node02] => (item=openshift_hosted_loggingops_storage_volume_name) skipping: [node01] => (item=openshift_hosted_logging_storage_volume_name) skipping: [node02] => (item=openshift_hosted_logging_storage_volume_size) skipping: [node02] => (item=openshift_hosted_loggingops_storage_volume_size) skipping: [node01] => (item=openshift_hosted_loggingops_storage_volume_name) skipping: [node02] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node01] => (item=openshift_hosted_logging_storage_volume_size) skipping: [node02] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node02] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node01] => (item=openshift_hosted_loggingops_storage_volume_size) skipping: [node02] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_enable_ops_cluster) skipping: [node02] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_image_pull_secret) skipping: [node02] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_curator_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_ops_hostname) skipping: [node02] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node01] => (item=openshift_hosted_logging_curator_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_journal_source) skipping: [node01] => (item=openshift_hosted_logging_kibana_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node01] => (item=openshift_hosted_logging_kibana_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_ops_hostname) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node01] => (item=openshift_hosted_logging_fluentd_nodeselector_label) skipping: [node02] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node01] => (item=openshift_hosted_logging_journal_source) skipping: [node02] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node01] => (item=openshift_hosted_logging_journal_read_from_head) skipping: [node02] => (item=openshift_hosted_logging_deployer_prefix) skipping: [node02] => (item=openshift_hosted_logging_deployer_version) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_instance_ram) skipping: [node02] => (item=openshift_hosted_metrics_deploy) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_nodeselector) skipping: [node02] => (item=openshift_hosted_metrics_storage_kind) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_instance_ram) skipping: [node02] => (item=openshift_hosted_metrics_storage_access_modes) skipping: [node02] => (item=openshift_hosted_metrics_storage_host) skipping: [node01] => (item=openshift_hosted_logging_elasticsearch_ops_nodeselector) skipping: [node02] => (item=openshift_hosted_metrics_storage_nfs_directory) skipping: [node01] => (item=openshift_hosted_logging_storage_access_modes) skipping: [node02] => (item=openshift_hosted_metrics_storage_volume_name) skipping: [node02] => (item=openshift_hosted_metrics_storage_volume_size) skipping: [node01] => (item=openshift_hosted_logging_deployer_prefix) skipping: [node02] => (item=openshift_hosted_metrics_storage_labels) skipping: [node01] => (item=openshift_hosted_logging_deployer_version) skipping: [node02] => (item=openshift_hosted_metrics_deployer_prefix) skipping: [node02] => (item=openshift_hosted_metrics_deployer_version) skipping: [node01] => (item=openshift_hosted_metrics_deploy) skipping: [node01] => (item=openshift_hosted_metrics_storage_kind) skipping: [node01] => (item=openshift_hosted_metrics_storage_access_modes) skipping: [node01] => (item=openshift_hosted_metrics_storage_host) skipping: [node01] => (item=openshift_hosted_metrics_storage_nfs_directory) skipping: [node01] => (item=openshift_hosted_metrics_storage_volume_name) skipping: [node01] => (item=openshift_hosted_metrics_storage_volume_size) skipping: [node01] => (item=openshift_hosted_metrics_storage_labels) skipping: [node01] => (item=openshift_hosted_metrics_deployer_prefix) skipping: [node01] => (item=openshift_hosted_metrics_deployer_version) TASK [openshift_sanitize_inventory : debug] ************************************ skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_stats] ******************************** skipping: [node01] TASK [openshift_sanitize_inventory : Assign deprecated variables to correct counterparts] *** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_logging.yml for node01, node02 included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/__deprecations_metrics.yml for node01, node02 TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : conditional_set_fact] ********************* ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : Standardize on latest variable names] ***** ok: [node01] ok: [node02] TASK [openshift_sanitize_inventory : Normalize openshift_release] ************** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Abort when openshift_release is invalid] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : include_tasks] **************************** included: /root/openshift-ansible/roles/openshift_sanitize_inventory/tasks/unsupported.yml for node01, node02 TASK [openshift_sanitize_inventory : Ensure that openshift_use_dnsmasq is true] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that openshift_node_dnsmasq_install_network_manager_hook is true] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : set_fact] ********************************* skipping: [node01] => (item=openshift_hosted_etcd_storage_kind) skipping: [node02] => (item=openshift_hosted_etcd_storage_kind) TASK [openshift_sanitize_inventory : Ensure that dynamic provisioning is set if using dynamic storage] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure the hosted registry's GlusterFS storage is configured correctly] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure clusterid is set along with the cloudprovider] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure ansible_service_broker_remove and ansible_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure template_service_broker_remove and template_service_broker_install are mutually exclusive] *** skipping: [node01] skipping: [node02] TASK [openshift_sanitize_inventory : Ensure that all requires vsphere configuration variables are set] *** skipping: [node01] skipping: [node02] TASK [Detecting Operating System from ostree_booted] *************************** ok: [node01] ok: [node02] TASK [set openshift_deployment_type if unset] ********************************** skipping: [node01] skipping: [node02] TASK [initialize_facts set fact openshift_is_atomic and openshift_is_containerized] *** ok: [node02] ok: [node01] TASK [Determine Atomic Host Docker Version] ************************************ skipping: [node01] skipping: [node02] TASK [assert atomic host docker version is 1.12 or later] ********************** skipping: [node02] skipping: [node01] PLAY [Initialize special first-master variables] ******************************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [set_fact] **************************************************************** ok: [node01] PLAY [Disable web console if required] ***************************************** TASK [set_fact] **************************************************************** skipping: [node01] PLAY [Install packages necessary for installer] ******************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Ensure openshift-ansible installer package deps are installed] *********** ok: [node02] => (item=iproute) ok: [node02] => (item=dbus-python) ok: [node02] => (item=PyYAML) ok: [node02] => (item=python-ipaddress) ok: [node02] => (item=yum-utils) TASK [Ensure various deps for running system containers are installed] ********* skipping: [node02] => (item=atomic) skipping: [node02] => (item=ostree) skipping: [node02] => (item=runc) PLAY [Initialize cluster facts] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node01] ok: [node02] TASK [Gather Cluster facts] **************************************************** changed: [node02] ok: [node01] TASK [Set fact of no_proxy_internal_hostnames] ********************************* skipping: [node01] skipping: [node02] TASK [Initialize openshift.node.sdn_mtu] *************************************** ok: [node01] ok: [node02] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* skipping: [node01] TASK [include_role] ************************************************************ skipping: [node01] TASK [debug] ******************************************************************* skipping: [node01] PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* skipping: [node02] TASK [set_fact] **************************************************************** skipping: [node02] PLAY [Ensure the requested version packages are available.] ******************** TASK [Gathering Facts] ********************************************************* skipping: [node02] TASK [include_role] ************************************************************ skipping: [node02] PLAY [Verify Requirements] ***************************************************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [Run variable sanity checks] ********************************************** ok: [node01] PLAY [Initialization Checkpoint End] ******************************************* TASK [Set install initialization 'Complete'] *********************************** ok: [node01] PLAY [Validate node hostnames] ************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Query DNS for IP address of node02] ************************************** ok: [node02] TASK [Validate openshift_hostname when defined] ******************************** skipping: [node02] TASK [Validate openshift_ip exists on node when defined] *********************** skipping: [node02] PLAY [Setup yum repositories for all hosts] ************************************ TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Install Red Hat Subscription manager] ******************* skipping: [node02] TASK [rhel_subscribe : Is host already registered?] **************************** skipping: [node02] TASK [rhel_subscribe : Register host] ****************************************** skipping: [node02] TASK [rhel_subscribe : fail] *************************************************** skipping: [node02] TASK [rhel_subscribe : Determine if OpenShift Pool Already Attached] *********** skipping: [node02] TASK [rhel_subscribe : Attach to OpenShift Pool] ******************************* skipping: [node02] TASK [rhel_subscribe : include_tasks] ****************************************** skipping: [node02] TASK [openshift_repos : openshift_repos detect ostree] ************************* ok: [node02] TASK [openshift_repos : Ensure libselinux-python is installed] ***************** ok: [node02] TASK [openshift_repos : Remove openshift_additional.repo file] ***************** ok: [node02] TASK [openshift_repos : Create any additional repos that are defined] ********** TASK [openshift_repos : include_tasks] ***************************************** skipping: [node02] TASK [openshift_repos : include_tasks] ***************************************** included: /root/openshift-ansible/roles/openshift_repos/tasks/centos_repos.yml for node02 TASK [openshift_repos : Configure origin gpg keys] ***************************** ok: [node02] TASK [openshift_repos : Configure correct origin release repository] *********** ok: [node02] => (item=/root/openshift-ansible/roles/openshift_repos/templates/CentOS-OpenShift-Origin.repo.j2) TASK [openshift_repos : Ensure clean repo cache in the event repos have been changed manually] *** changed: [node02] => { "msg": "First run of openshift_repos" } TASK [openshift_repos : Record that openshift_repos already ran] *************** ok: [node02] RUNNING HANDLER [openshift_repos : refresh cache] ****************************** changed: [node02] PLAY [Configure os_firewall] *************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [os_firewall : Detecting Atomic Host Operating System] ******************** ok: [node02] TASK [os_firewall : Set fact r_os_firewall_is_atomic] ************************** ok: [node02] TASK [os_firewall : include_tasks] ********************************************* skipping: [node02] TASK [os_firewall : include_tasks] ********************************************* included: /root/openshift-ansible/roles/os_firewall/tasks/iptables.yml for node02 TASK [os_firewall : Ensure firewalld service is not enabled] ******************* ok: [node02] TASK [os_firewall : Wait 10 seconds after disabling firewalld] ***************** skipping: [node02] TASK [os_firewall : Install iptables packages] ********************************* ok: [node02] => (item=iptables) ok: [node02] => (item=iptables-services) TASK [os_firewall : Start and enable iptables service] ************************* ok: [node02 -> node02] => (item=node02) TASK [os_firewall : need to pause here, otherwise the iptables service starting can sometimes cause ssh to fail] *** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [container_runtime : Setup the docker-storage for overlay] **************** skipping: [node02] PLAY [create oo_hosts_containerized_managed_true host group] ******************* TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [group_by] **************************************************************** ok: [node01] PLAY [oo_nodes_to_config] ****************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_excluder : Install excluders] ********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* ok: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/common/pre.yml for node02 TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Add enterprise registry, if necessary] *************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Get current installed Docker version] **************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/docker_sanity.yml for node02 TASK [container_runtime : Error out if Docker pre-installed but too old] ******* skipping: [node02] TASK [container_runtime : Error out if requested Docker is too old] ************ skipping: [node02] TASK [container_runtime : Fail if Docker version requested but downgrade is required] *** skipping: [node02] TASK [container_runtime : Error out if attempting to upgrade Docker across the 1.10 boundary] *** skipping: [node02] TASK [container_runtime : Install Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure docker.service.d directory exists] ************ ok: [node02] TASK [container_runtime : Configure Docker service unit file] ****************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Set registry params] ********************************* skipping: [node02] => (item={u'reg_conf_var': u'ADD_REGISTRY', u'reg_flag': u'--add-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'BLOCK_REGISTRY', u'reg_flag': u'--block-registry', u'reg_fact_val': []}) skipping: [node02] => (item={u'reg_conf_var': u'INSECURE_REGISTRY', u'reg_flag': u'--insecure-registry', u'reg_fact_val': []}) TASK [container_runtime : Place additional/blocked/insecure registries in /etc/containers/registries.conf] *** skipping: [node02] TASK [container_runtime : Set Proxy Settings] ********************************** skipping: [node02] => (item={u'reg_conf_var': u'HTTP_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'HTTPS_PROXY', u'reg_fact_val': u''}) skipping: [node02] => (item={u'reg_conf_var': u'NO_PROXY', u'reg_fact_val': u''}) TASK [container_runtime : Set various Docker options] ************************** ok: [node02] TASK [container_runtime : stat] ************************************************ ok: [node02] TASK [container_runtime : Configure Docker Network OPTIONS] ******************** ok: [node02] TASK [container_runtime : Detect if docker is already started] ***************** ok: [node02] TASK [container_runtime : Start the Docker service] **************************** ok: [node02] TASK [container_runtime : set_fact] ******************************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/common/post.yml for node02 TASK [container_runtime : Ensure /var/lib/containers exists] ******************* ok: [node02] TASK [container_runtime : Fix SELinux Permissions on /var/lib/containers] ****** ok: [node02] TASK [container_runtime : include_tasks] *************************************** included: /root/openshift-ansible/roles/container_runtime/tasks/registry_auth.yml for node02 TASK [container_runtime : Check for credentials file for registry auth] ******** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth] ***** skipping: [node02] TASK [container_runtime : Create credentials for docker cli registry auth (alternative)] *** skipping: [node02] TASK [container_runtime : stat the docker data dir] **************************** ok: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Fail quickly if openshift_docker_options are set] **** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Install Docker so we can use the client] ************* skipping: [node02] TASK [container_runtime : Disable Docker] ************************************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** skipping: [node02] TASK [container_runtime : debug] *********************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Pre-pull Container Engine System Container image] **** skipping: [node02] TASK [container_runtime : Ensure container-engine.service.d directory exists] *** skipping: [node02] TASK [container_runtime : Ensure /etc/docker directory exists] ***************** skipping: [node02] TASK [container_runtime : Install Container Engine System Container] *********** skipping: [node02] TASK [container_runtime : Configure Container Engine Service File] ************* skipping: [node02] TASK [container_runtime : Configure Container Engine] ************************** skipping: [node02] TASK [container_runtime : Start the Container Engine service] ****************** skipping: [node02] TASK [container_runtime : set_fact] ******************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Check we are not using node as a Docker container with CRI-O] *** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] TASK [container_runtime : Check that overlay is in the kernel] ***************** skipping: [node02] TASK [container_runtime : Add overlay to modprobe.d] *************************** skipping: [node02] TASK [container_runtime : Manually modprobe overlay into the kernel] *********** skipping: [node02] TASK [container_runtime : Enable and start systemd-modules-load] *************** skipping: [node02] TASK [container_runtime : Ensure proxies are in the atomic.conf] *************** skipping: [node02] TASK [container_runtime : debug] *********************************************** skipping: [node02] TASK [container_runtime : Pre-pull CRI-O System Container image] *************** skipping: [node02] TASK [container_runtime : Install CRI-O System Container] ********************** skipping: [node02] TASK [container_runtime : Remove CRI-O default configuration files] ************ skipping: [node02] => (item=/etc/cni/net.d/200-loopback.conf) skipping: [node02] => (item=/etc/cni/net.d/100-crio-bridge.conf) TASK [container_runtime : Create the CRI-O configuration] ********************** skipping: [node02] TASK [container_runtime : Ensure CNI configuration directory exists] *********** skipping: [node02] TASK [container_runtime : Add iptables allow rules] **************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove iptables rules] ******************************* TASK [container_runtime : Add firewalld allow rules] *************************** skipping: [node02] => (item={u'port': u'10010/tcp', u'service': u'crio'}) TASK [container_runtime : Remove firewalld allow rules] ************************ TASK [container_runtime : Configure the CNI network] *************************** skipping: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-storage] ****************** skipping: [node02] TASK [container_runtime : Create /etc/sysconfig/crio-network] ****************** skipping: [node02] TASK [container_runtime : Start the CRI-O service] ***************************** skipping: [node02] TASK [container_runtime : include_tasks] *************************************** skipping: [node02] PLAY [Determine openshift_version to configure on first master] **************** TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [include_role] ************************************************************ TASK [openshift_version : Use openshift.common.version fact as version to configure if already installed] *** ok: [node01] TASK [openshift_version : include_tasks] *************************************** included: /root/openshift-ansible/roles/openshift_version/tasks/first_master_containerized_version.yml for node01 TASK [openshift_version : Set containerized version to configure if openshift_image_tag specified] *** skipping: [node01] TASK [openshift_version : Set containerized version to configure if openshift_release specified] *** skipping: [node01] TASK [openshift_version : Lookup latest containerized version if no version specified] *** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : Set precise containerized version to configure if openshift_release specified] *** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** ok: [node01] => { "msg": "openshift_pkg_version was not defined. Falling back to -3.9.0" } TASK [openshift_version : set_fact] ******************************************** ok: [node01] TASK [openshift_version : debug] *********************************************** skipping: [node01] TASK [openshift_version : set_fact] ******************************************** skipping: [node01] TASK [debug] ******************************************************************* ok: [node01] => { "msg": "openshift_pkg_version set to -3.9.0" } PLAY [Set openshift_version for etcd, node, and master hosts] ****************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [set_fact] **************************************************************** ok: [node02] PLAY [Ensure the requested version packages are available.] ******************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [include_role] ************************************************************ TASK [openshift_version : Check openshift_version for rpm installation] ******** included: /root/openshift-ansible/roles/openshift_version/tasks/check_available_rpms.yml for node02 TASK [openshift_version : Get available origin version] ************************ ok: [node02] TASK [openshift_version : fail] ************************************************ skipping: [node02] TASK [openshift_version : Fail if rpm version and docker image version are different] *** skipping: [node02] TASK [openshift_version : For an RPM install, abort when the release requested does not match the available version.] *** skipping: [node02] TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_release": "VARIABLE IS NOT DEFINED!" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_image_tag": "v3.9.0-alpha.4" } TASK [openshift_version : debug] *********************************************** ok: [node02] => { "openshift_pkg_version": "-3.9.0" } PLAY [Node Install Checkpoint Start] ******************************************* TASK [Set Node install 'In Progress'] ****************************************** ok: [node01] PLAY [Create OpenShift certificates for node hosts] **************************** TASK [openshift_node_certificates : Ensure CA certificate exists on openshift_ca_host] *** ok: [node02 -> node01] TASK [openshift_node_certificates : fail] ************************************** skipping: [node02] TASK [openshift_node_certificates : Check status of node certificates] ********* ok: [node02] => (item=system:node:node02.crt) ok: [node02] => (item=system:node:node02.key) ok: [node02] => (item=system:node:node02.kubeconfig) ok: [node02] => (item=ca.crt) ok: [node02] => (item=server.key) ok: [node02] => (item=server.crt) TASK [openshift_node_certificates : set_fact] ********************************** ok: [node02] TASK [openshift_node_certificates : Create openshift_generated_configs_dir if it does not exist] *** ok: [node02 -> node01] TASK [openshift_node_certificates : find] ************************************** ok: [node02 -> node01] TASK [openshift_node_certificates : Generate the node client config] *********** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Generate the node server certificate] ****** changed: [node02 -> node01] => (item=node02) TASK [openshift_node_certificates : Create a tarball of the node config directories] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Retrieve the node config tarballs from the master] *** changed: [node02 -> node01] TASK [openshift_node_certificates : Ensure certificate directory exists] ******* ok: [node02] TASK [openshift_node_certificates : Unarchive the tarball on the node] ********* changed: [node02] TASK [openshift_node_certificates : Delete local temp directory] *************** ok: [node02 -> localhost] TASK [openshift_node_certificates : Copy OpenShift CA to system CA trust] ****** ok: [node02] => (item={u'cert': u'/etc/origin/node/ca.crt', u'id': u'openshift'}) PLAY [Disable excluders] ******************************************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": "false" } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": "false" } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/disable.yml for node02 TASK [openshift_excluder : Include verify_upgrade.yml when upgrading] ********** skipping: [node02] TASK [openshift_excluder : Disable excluders before the upgrade to remove older excluding expressions] *** included: /root/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* skipping: [node02] TASK [openshift_excluder : Include install.yml] ******************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Include exclude.yml] ******************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] TASK [openshift_excluder : Include unexclude.yml] ****************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/unexclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : disable docker excluder] **************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : disable openshift excluder] ************************* skipping: [node02] PLAY [Evaluate node groups] **************************************************** TASK [Gathering Facts] ********************************************************* ok: [localhost] TASK [Evaluate oo_containerized_master_nodes] ********************************** skipping: [localhost] => (item=node02) [WARNING]: Could not match supplied host pattern, ignoring: oo_containerized_master_nodes PLAY [Configure containerized nodes] ******************************************* skipping: no hosts matched PLAY [Configure nodes] ********************************************************* TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_clock : Determine if chrony is installed] ********************** changed: [node02] [WARNING]: Consider using yum, dnf or zypper module rather than running rpm TASK [openshift_clock : Install ntp package] *********************************** skipping: [node02] TASK [openshift_clock : Start and enable ntpd/chronyd] ************************* changed: [node02] TASK [openshift_cloud_provider : Set cloud provider facts] ********************* ok: [node02] TASK [openshift_cloud_provider : Create cloudprovider config dir] ************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_cloud_provider : include_tasks] ******************************** skipping: [node02] TASK [openshift_node : fail] *************************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/dnsmasq_install.yml for node02 TASK [openshift_node : Check for NetworkManager service] *********************** ok: [node02] TASK [openshift_node : Set fact using_network_manager] ************************* ok: [node02] TASK [openshift_node : Install dnsmasq] **************************************** ok: [node02] TASK [openshift_node : ensure origin/node directory exists] ******************** ok: [node02] => (item=/etc/origin) changed: [node02] => (item=/etc/origin/node) TASK [openshift_node : Install node-dnsmasq.conf] ****************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** skipping: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/dnsmasq.yml for node02 TASK [openshift_node : Install dnsmasq configuration] ************************** ok: [node02] TASK [openshift_node : Deploy additional dnsmasq.conf] ************************* skipping: [node02] TASK [openshift_node : Enable dnsmasq] ***************************************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/dnsmasq/network-manager.yml for node02 TASK [openshift_node : Install network manager dispatch script] **************** ok: [node02] TASK [openshift_node : Add iptables allow rules] ******************************* ok: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) ok: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) ok: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) ok: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove iptables rules] ********************************** TASK [openshift_node : Add firewalld allow rules] ****************************** skipping: [node02] => (item={u'port': u'10250/tcp', u'service': u'Kubernetes kubelet'}) skipping: [node02] => (item={u'port': u'80/tcp', u'service': u'http'}) skipping: [node02] => (item={u'port': u'443/tcp', u'service': u'https'}) skipping: [node02] => (item={u'cond': u'openshift_use_openshift_sdn | bool', u'port': u'4789/udp', u'service': u'OpenShift OVS sdn'}) skipping: [node02] => (item={u'cond': False, u'port': u'179/tcp', u'service': u'Calico BGP Port'}) skipping: [node02] => (item={u'cond': False, u'port': u'/tcp', u'service': u'Kubernetes service NodePort TCP'}) skipping: [node02] => (item={u'cond': False, u'port': u'/udp', u'service': u'Kubernetes service NodePort UDP'}) TASK [openshift_node : Remove firewalld allow rules] *************************** TASK [openshift_node : Disable swap] ******************************************* ok: [node02] TASK [openshift_node : include node installer] ********************************* included: /root/openshift-ansible/roles/openshift_node/tasks/install.yml for node02 TASK [openshift_node : Install Node package, sdn-ovs, conntrack packages] ****** skipping: [node02] => (item={u'name': u'origin-node-3.9.0'}) skipping: [node02] => (item={u'name': u'origin-sdn-ovs-3.9.0', u'install': True}) skipping: [node02] => (item={u'name': u'conntrack-tools'}) TASK [openshift_node : Pre-pull node image when containerized] ***************** ok: [node02] TASK [openshift_node : Restart cri-o] ****************************************** skipping: [node02] TASK [openshift_node : restart NetworkManager to ensure resolv.conf is present] *** skipping: [node02] TASK [openshift_node : sysctl] ************************************************* ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/registry_auth.yml for node02 TASK [openshift_node : Check for credentials file for registry auth] *********** skipping: [node02] TASK [openshift_node : Create credentials for registry auth] ******************* skipping: [node02] TASK [openshift_node : Create credentials for registry auth (alternative)] ***** skipping: [node02] TASK [openshift_node : Setup ro mount of /root/.docker for containerized hosts] *** skipping: [node02] TASK [openshift_node : include standard node config] *************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config.yml for node02 TASK [openshift_node : Install the systemd units] ****************************** included: /root/openshift-ansible/roles/openshift_node/tasks/systemd_units.yml for node02 TASK [openshift_node : Install Node service file] ****************************** ok: [node02] TASK [openshift_node : include node deps docker service file] ****************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/install-node-deps-docker-service-file.yml for node02 TASK [openshift_node : Install Node dependencies docker service file] ********** ok: [node02] TASK [openshift_node : include ovs service environment file] ******************* included: /root/openshift-ansible/roles/openshift_node/tasks/config/install-ovs-service-env-file.yml for node02 TASK [openshift_node : Create the openvswitch service env file] **************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/install-ovs-docker-service-file.yml for node02 TASK [openshift_node : Install OpenvSwitch docker service file] **************** ok: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/configure-node-settings.yml for node02 TASK [openshift_node : Configure Node settings] ******************************** ok: [node02] => (item={u'regex': u'^OPTIONS=', u'line': u'OPTIONS=--loglevel=2'}) ok: [node02] => (item={u'regex': u'^CONFIG_FILE=', u'line': u'CONFIG_FILE=/etc/origin/node/node-config.yaml'}) ok: [node02] => (item={u'regex': u'^IMAGE_VERSION=', u'line': u'IMAGE_VERSION=v3.9.0-alpha.4'}) TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/configure-proxy-settings.yml for node02 TASK [openshift_node : Configure Proxy Settings] ******************************* skipping: [node02] => (item={u'regex': u'^HTTP_PROXY=', u'line': u'HTTP_PROXY='}) skipping: [node02] => (item={u'regex': u'^HTTPS_PROXY=', u'line': u'HTTPS_PROXY='}) skipping: [node02] => (item={u'regex': u'^NO_PROXY=', u'line': u'NO_PROXY=[],172.30.0.0/16,10.128.0.0/14'}) TASK [openshift_node : Pull container images] ********************************** included: /root/openshift-ansible/roles/openshift_node/tasks/container_images.yml for node02 TASK [openshift_node : Install Node system container] ************************** skipping: [node02] TASK [openshift_node : Install OpenvSwitch system containers] ****************** skipping: [node02] TASK [openshift_node : Pre-pull openvswitch image] ***************************** ok: [node02] TASK [openshift_node : Start and enable openvswitch service] ******************* ok: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : file] *************************************************** skipping: [node02] TASK [openshift_node : Create the Node config] ********************************* changed: [node02] TASK [openshift_node : Configure Node Environment Variables] ******************* TASK [openshift_node : Configure AWS Cloud Provider Settings] ****************** skipping: [node02] => (item=None) skipping: [node02] => (item=None) TASK [openshift_node : Wait for master API to become available before proceeding] *** ok: [node02] TASK [openshift_node : Start and enable node dep] ****************************** changed: [node02] TASK [openshift_node : Start and enable node] ********************************** ok: [node02] TASK [openshift_node : Dump logs from node service if it failed] *************** skipping: [node02] TASK [openshift_node : Abort if node failed to start] ************************** skipping: [node02] TASK [openshift_node : set_fact] *********************************************** ok: [node02] TASK [openshift_node : NFS storage plugin configuration] *********************** included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/nfs.yml for node02 TASK [openshift_node : Install NFS storage plugin dependencies] **************** ok: [node02] TASK [openshift_node : Check for existence of nfs sebooleans] ****************** ok: [node02] => (item=virt_use_nfs) ok: [node02] => (item=virt_sandbox_use_nfs) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:34.063683', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.008203', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:34.055480', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:34.825549', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.007233', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:34.818316', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow nfs storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:34.063683', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_use_nfs'], u'rc': 0, 'item': u'virt_use_nfs', u'delta': u'0:00:00.008203', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:34.055480', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:34.825549', '_ansible_no_log': False, u'stdout': u'virt_use_nfs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_nfs'], u'rc': 0, 'item': u'virt_sandbox_use_nfs', u'delta': u'0:00:00.007233', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_nfs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_nfs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:34.818316', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : GlusterFS storage plugin configuration] ***************** included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/glusterfs.yml for node02 TASK [openshift_node : Install GlusterFS storage plugin dependencies] ********** ok: [node02] TASK [openshift_node : Check for existence of fusefs sebooleans] *************** ok: [node02] => (item=virt_use_fusefs) ok: [node02] => (item=virt_sandbox_use_fusefs) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers] *** ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:41.036258', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.006870', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:41.029388', '_ansible_ignore_errors': None, 'failed': False}) ok: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:42.197151', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.021293', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:42.175858', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Set seboolean to allow gluster storage plugin access from containers (python 3)] *** skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:41.036258', '_ansible_no_log': False, u'stdout': u'virt_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_use_fusefs'], u'rc': 0, 'item': u'virt_use_fusefs', u'delta': u'0:00:00.006870', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:41.029388', '_ansible_ignore_errors': None, 'failed': False}) skipping: [node02] => (item={'_ansible_parsed': True, 'stderr_lines': [], '_ansible_item_result': True, u'end': u'2018-04-09 15:26:42.197151', '_ansible_no_log': False, u'stdout': u'virt_sandbox_use_fusefs --> on', u'cmd': [u'getsebool', u'virt_sandbox_use_fusefs'], u'rc': 0, 'item': u'virt_sandbox_use_fusefs', u'delta': u'0:00:00.021293', u'stderr': u'', u'changed': False, u'invocation': {u'module_args': {u'creates': None, u'executable': None, u'_uses_shell': False, u'_raw_params': u'getsebool virt_sandbox_use_fusefs', u'removes': None, u'warn': True, u'chdir': None, u'stdin': None}}, 'stdout_lines': [u'virt_sandbox_use_fusefs --> on'], 'failed_when_result': False, u'start': u'2018-04-09 15:26:42.175858', '_ansible_ignore_errors': None, 'failed': False}) TASK [openshift_node : Ceph storage plugin configuration] ********************** included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/ceph.yml for node02 TASK [openshift_node : Install Ceph storage plugin dependencies] *************** ok: [node02] TASK [openshift_node : iSCSI storage plugin configuration] ********************* included: /root/openshift-ansible/roles/openshift_node/tasks/storage_plugins/iscsi.yml for node02 TASK [openshift_node : Install iSCSI storage plugin dependencies] ************** ok: [node02] => (item=iscsi-initiator-utils) ok: [node02] => (item=device-mapper-multipath) TASK [openshift_node : restart services] *************************************** ok: [node02] => (item=multipathd) ok: [node02] => (item=rpcbind) TASK [openshift_node : Template multipath configuration] *********************** changed: [node02] TASK [openshift_node : Enable multipath] *************************************** changed: [node02] TASK [openshift_node : include_tasks] ****************************************** included: /root/openshift-ansible/roles/openshift_node/tasks/config/workaround-bz1331590-ovs-oom-fix.yml for node02 TASK [openshift_node : Create OpenvSwitch service.d directory] ***************** ok: [node02] TASK [openshift_node : Install OpenvSwitch service OOM fix] ******************** ok: [node02] TASK [tuned : Check for tuned package] ***************************************** ok: [node02] TASK [tuned : Set tuned OpenShift variables] *********************************** ok: [node02] TASK [tuned : Ensure directory structure exists] ******************************* ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) TASK [tuned : Ensure files are populated from templates] *********************** skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) skipping: [node02] => (item={'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'state': 'directory', 'ctime': 1522867116.424326, 'serole': 'object_r', 'gid': 0, 'mode': '0755', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift', 'size': 24, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/recommend.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'recommend.conf', 'size': 268, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-control-plane/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-control-plane/tuned.conf', 'size': 744, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift-node/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift-node/tuned.conf', 'size': 135, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) ok: [node02] => (item={'src': u'/root/openshift-ansible/roles/tuned/templates/openshift/tuned.conf', 'group': u'root', 'uid': 0, 'selevel': 's0', 'seuser': 'unconfined_u', 'serole': 'object_r', 'ctime': 1522867116.424326, 'state': 'file', 'gid': 0, 'mode': '0644', 'mtime': 1522867116.424326, 'owner': 'root', 'path': u'openshift/tuned.conf', 'size': 593, 'root': u'/root/openshift-ansible/roles/tuned/templates', 'setype': 'admin_home_t'}) TASK [tuned : Make tuned use the recommended tuned profile on restart] ********* changed: [node02] => (item=/etc/tuned/active_profile) ok: [node02] => (item=/etc/tuned/profile_mode) TASK [tuned : Restart tuned service] ******************************************* changed: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Install logrotate] ******* ok: [node02] TASK [nickhammond.logrotate : nickhammond.logrotate | Setup logrotate.d scripts] *** RUNNING HANDLER [openshift_node : restart node] ******************************** changed: [node02] PLAY [create additional node network plugin groups] **************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] TASK [group_by] **************************************************************** ok: [node02] [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_flannel [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_calico [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_contiv [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_kuryr PLAY [etcd_client node config] ************************************************* skipping: no hosts matched PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** skipping: no hosts matched [WARNING]: Could not match supplied host pattern, ignoring: oo_nodes_use_nuage PLAY [Additional node config] ************************************************** skipping: no hosts matched PLAY [Configure Contiv masters] ************************************************ TASK [Gathering Facts] ********************************************************* ok: [node01] TASK [contiv_facts : Determine if CoreOS] ************************************** skipping: [node01] TASK [contiv_facts : Init the contiv_is_coreos fact] *************************** skipping: [node01] TASK [contiv_facts : Set the contiv_is_coreos fact] **************************** skipping: [node01] TASK [contiv_facts : Set the bin directory path for CoreOS] ******************** skipping: [node01] TASK [contiv_facts : Create the directory used to store binaries] ************** skipping: [node01] TASK [contiv_facts : Create Ansible temp directory] **************************** skipping: [node01] TASK [contiv_facts : Determine if has rpm] ************************************* skipping: [node01] TASK [contiv_facts : Init the contiv_has_rpm fact] ***************************** skipping: [node01] TASK [contiv_facts : Set the contiv_has_rpm fact] ****************************** skipping: [node01] TASK [contiv_facts : Init the contiv_has_firewalld fact] *********************** skipping: [node01] TASK [contiv_facts : Init the contiv_has_iptables fact] ************************ skipping: [node01] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] TASK [contiv : Ensure contiv_bin_dir exists] *********************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node01] PLAY [Configure rest of Contiv nodes] ****************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] ok: [node01] TASK [contiv_facts : Determine if CoreOS] ************************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_is_coreos fact] *************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Set the contiv_is_coreos fact] **************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Set the bin directory path for CoreOS] ******************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Create the directory used to store binaries] ************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Create Ansible temp directory] **************************** skipping: [node02] skipping: [node01] TASK [contiv_facts : Determine if has rpm] ************************************* skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_has_rpm fact] ***************************** skipping: [node02] skipping: [node01] TASK [contiv_facts : Set the contiv_has_rpm fact] ****************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_has_firewalld fact] *********************** skipping: [node01] skipping: [node02] TASK [contiv_facts : Init the contiv_has_iptables fact] ************************ skipping: [node02] skipping: [node01] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] skipping: [node02] TASK [contiv_facts : include_tasks] ******************************************** skipping: [node01] skipping: [node02] TASK [contiv : include_tasks] ************************************************** skipping: [node01] skipping: [node02] TASK [contiv : Ensure contiv_bin_dir exists] *********************************** skipping: [node02] skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node02] skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node02] skipping: [node01] TASK [contiv : include_tasks] ************************************************** skipping: [node02] skipping: [node01] PLAY [Configure Kuryr node] **************************************************** skipping: no hosts matched PLAY [Additional node config] ************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [openshift_manage_node : Wait for master API to become available before proceeding] *** ok: [node02 -> node01] TASK [openshift_manage_node : Wait for Node Registration] ********************** ok: [node02 -> node01] TASK [openshift_manage_node : include_tasks] *********************************** included: /root/openshift-ansible/roles/openshift_manage_node/tasks/config.yml for node02 TASK [openshift_manage_node : Set node schedulability] ************************* ok: [node02 -> node01] TASK [openshift_manage_node : Label nodes] ************************************* ok: [node02 -> node01] TASK [Create group for deployment type] **************************************** ok: [node02] PLAY [Re-enable excluder if it was previously enabled] ************************* TASK [openshift_excluder : Detecting Atomic Host Operating System] ************* ok: [node02] TASK [openshift_excluder : Debug r_openshift_excluder_enable_docker_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_docker_excluder": "false" } TASK [openshift_excluder : Debug r_openshift_excluder_enable_openshift_excluder] *** ok: [node02] => { "r_openshift_excluder_enable_openshift_excluder": "false" } TASK [openshift_excluder : Fail if invalid openshift_excluder_action provided] *** skipping: [node02] TASK [openshift_excluder : Fail if r_openshift_excluder_upgrade_target is not defined] *** skipping: [node02] TASK [openshift_excluder : Include main action task file] ********************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/enable.yml for node02 TASK [openshift_excluder : Install excluders] ********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/install.yml for node02 TASK [openshift_excluder : Install docker excluder - yum] ********************** skipping: [node02] TASK [openshift_excluder : Install docker excluder - dnf] ********************** skipping: [node02] TASK [openshift_excluder : Install openshift excluder - yum] ******************* skipping: [node02] TASK [openshift_excluder : Install openshift excluder - dnf] ******************* skipping: [node02] TASK [openshift_excluder : set_fact] ******************************************* skipping: [node02] TASK [openshift_excluder : Enable excluders] *********************************** included: /root/openshift-ansible/roles/openshift_excluder/tasks/exclude.yml for node02 TASK [openshift_excluder : Check for docker-excluder] ************************** ok: [node02] TASK [openshift_excluder : Enable docker excluder] ***************************** skipping: [node02] TASK [openshift_excluder : Check for openshift excluder] *********************** ok: [node02] TASK [openshift_excluder : Enable openshift excluder] ************************** skipping: [node02] PLAY [Node Install Checkpoint End] ********************************************* TASK [Set Node install 'Complete'] ********************************************* ok: [node01] PLAY RECAP ********************************************************************* localhost : ok=25 changed=0 unreachable=0 failed=0 node01 : ok=36 changed=0 unreachable=0 failed=0 node02 : ok=183 changed=18 unreachable=0 failed=0 INSTALLER STATUS *************************************************************** Initialization : Complete (0:00:58) Node Install : Complete (0:03:46) PLAY [new_nodes] *************************************************************** TASK [Gathering Facts] ********************************************************* ok: [node02] TASK [Restart openvswitch service] ********************************************* changed: [node02] PLAY RECAP ********************************************************************* node02 : ok=2 changed=1 unreachable=0 failed=0 2018/04/09 15:28:57 Waiting for host: 192.168.66.101:22 2018/04/09 15:28:57 Connected to tcp://192.168.66.101:22 2018/04/09 15:29:14 Waiting for host: 192.168.66.101:22 2018/04/09 15:29:14 Connected to tcp://192.168.66.101:22 Warning: Permanently added '[127.0.0.1]:32881' (ECDSA) to the list of known hosts. Warning: Permanently added '[127.0.0.1]:32881' (ECDSA) to the list of known hosts. Cluster "node01:8443" set. Cluster "node01:8443" set. ++ kubectl get nodes --no-headers ++ cluster/kubectl.sh get nodes --no-headers ++ grep -v Ready + '[' -n '' ']' + echo 'Nodes are ready:' Nodes are ready: + kubectl get nodes + cluster/kubectl.sh get nodes NAME STATUS ROLES AGE VERSION node01 Ready master 4d v1.9.1+a0ce1bc657 node02 Ready 2m v1.9.1+a0ce1bc657 + make cluster-sync ./cluster/build.sh Building ... sha256:b5b4d597a272b0a23ca6310ef9df2eff2af3e39eb9e825fc156c0ce4582d5ef1 go version go1.9.2 linux/amd64 rsync: read error: Connection reset by peer (104) rsync error: error in rsync protocol data stream (code 12) at io.c(764) [sender=3.0.9] Waiting for rsyncd to be ready skipping directory . go version go1.9.2 linux/amd64 1c00591d2736530a107829f6bad06a625b10d74ba9211d703a88e33a19bccc0a 1c00591d2736530a107829f6bad06a625b10d74ba9211d703a88e33a19bccc0a make[1]: Entering directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt' hack/dockerized "./hack/check.sh && ./hack/build-go.sh install " sha256:b5b4d597a272b0a23ca6310ef9df2eff2af3e39eb9e825fc156c0ce4582d5ef1 go version go1.9.2 linux/amd64 skipping directory . go version go1.9.2 linux/amd64 Compiling tests... compiled tests.test 11365f796f37336bdce1bfa07bcdf65f77ba9dfb4c379b5f2f3300a3144afa42 11365f796f37336bdce1bfa07bcdf65f77ba9dfb4c379b5f2f3300a3144afa42 hack/build-docker.sh build sending incremental file list ./ Dockerfile kubernetes.repo sent 854 bytes received 53 bytes 1814.00 bytes/sec total size is 1167 speedup is 1.29 Sending build context to Docker daemon 35.7 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71c3d482487e Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virt-controller ---> Using cache ---> 31e2c695509f Step 4/8 : WORKDIR /home/virt-controller ---> Using cache ---> 24d4616566ef Step 5/8 : USER 1001 ---> Using cache ---> 93387777f457 Step 6/8 : COPY virt-controller /virt-controller ---> Using cache ---> 1dd61ac6b82f Step 7/8 : ENTRYPOINT /virt-controller ---> Using cache ---> 3562c3a4700f Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-controller" '' ---> Running in 2d5093634154 ---> 34250040b9db Removing intermediate container 2d5093634154 Successfully built 34250040b9db sending incremental file list ./ Dockerfile entrypoint.sh kubevirt-sudo libvirtd.sh sh.sh sock-connector sent 3286 bytes received 129 bytes 6830.00 bytes/sec total size is 5469 speedup is 1.60 Sending build context to Docker daemon 37.44 MB Step 1/14 : FROM kubevirt/libvirt:3.7.0 ---> 60c80c8f7523 Step 2/14 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 4a94b3474ba7 Step 3/14 : RUN dnf -y install socat genisoimage util-linux libcgroup-tools ethtool sudo && dnf -y clean all && test $(id -u qemu) = 107 # make sure that the qemu user really is 107 ---> Using cache ---> 1060bf73289c Step 4/14 : COPY sock-connector /sock-connector ---> Using cache ---> 0e21145d5129 Step 5/14 : COPY sh.sh /sh.sh ---> Using cache ---> 6f40988349d4 Step 6/14 : COPY virt-launcher /virt-launcher ---> Using cache ---> 78c010573d20 Step 7/14 : COPY kubevirt-sudo /etc/sudoers.d/kubevirt ---> Using cache ---> d7a6f086c3f6 Step 8/14 : RUN chmod 0640 /etc/sudoers.d/kubevirt ---> Using cache ---> 97a3ded7f77b Step 9/14 : RUN rm -f /libvirtd.sh ---> Using cache ---> d3f98d28adf6 Step 10/14 : COPY libvirtd.sh /libvirtd.sh ---> Using cache ---> 50c57170c923 Step 11/14 : RUN chmod a+x /libvirtd.sh ---> Using cache ---> 201bf5ef4125 Step 12/14 : COPY entrypoint.sh /entrypoint.sh ---> Using cache ---> 7dde03594385 Step 13/14 : ENTRYPOINT /entrypoint.sh ---> Using cache ---> ff3592644524 Step 14/14 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-launcher" '' ---> Running in 0462fff0c4e9 ---> 2d20926a50c0 Removing intermediate container 0462fff0c4e9 Successfully built 2d20926a50c0 sending incremental file list ./ Dockerfile sent 585 bytes received 34 bytes 1238.00 bytes/sec total size is 775 speedup is 1.25 Sending build context to Docker daemon 36.37 MB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71c3d482487e Step 3/5 : COPY virt-handler /virt-handler ---> Using cache ---> 28a2921c8dbf Step 4/5 : ENTRYPOINT /virt-handler ---> Using cache ---> 3da3bc08969e Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-handler" '' ---> Running in e35370b19e8d ---> e724c8b3728f Removing intermediate container e35370b19e8d Successfully built e724c8b3728f sending incremental file list ./ Dockerfile sent 864 bytes received 34 bytes 1796.00 bytes/sec total size is 1377 speedup is 1.53 Sending build context to Docker daemon 36.12 MB Step 1/9 : FROM fedora:27 ---> 9110ae7f579f Step 2/9 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71c3d482487e Step 3/9 : RUN useradd -u 1001 --create-home -s /bin/bash virt-api ---> Using cache ---> 52824ea59532 Step 4/9 : WORKDIR /home/virt-api ---> Using cache ---> aad9b1e251be Step 5/9 : USER 1001 ---> Using cache ---> b47160c51405 Step 6/9 : RUN curl -OL https://github.com/swagger-api/swagger-ui/tarball/38f74164a7062edb5dc80ef2fdddda24f3f6eb85/swagger-ui.tar.gz && mkdir swagger-ui && tar xf swagger-ui.tar.gz -C swagger-ui --strip-components 1 && mkdir third_party && mv swagger-ui/dist third_party/swagger-ui && rm -rf swagger-ui && sed -e 's@"http://petstore.swagger.io/v2/swagger.json"@"/swaggerapi/"@' -i third_party/swagger-ui/index.html && rm swagger-ui.tar.gz && rm -rf swagger-ui ---> Using cache ---> 4dd3f32334f9 Step 7/9 : COPY virt-api /virt-api ---> Using cache ---> 7e530bfad3a1 Step 8/9 : ENTRYPOINT /virt-api ---> Using cache ---> 4407673a0c8f Step 9/9 : LABEL "kubevirt-functional-tests-openshift-release0" '' "virt-api" '' ---> Running in d25be1b91ee6 ---> a2a57b104461 Removing intermediate container d25be1b91ee6 Successfully built a2a57b104461 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/iscsi-demo-target-tgtd ./ Dockerfile run-tgt.sh sent 2185 bytes received 53 bytes 4476.00 bytes/sec total size is 3992 speedup is 1.78 Sending build context to Docker daemon 6.656 kB Step 1/10 : FROM fedora:27 ---> 9110ae7f579f Step 2/10 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71c3d482487e Step 3/10 : ENV container docker ---> Using cache ---> 453fb17b7f2a Step 4/10 : RUN dnf -y install scsi-target-utils bzip2 e2fsprogs ---> Using cache ---> 257f70388ed1 Step 5/10 : RUN mkdir -p /images ---> Using cache ---> 8c74dfc48702 Step 6/10 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /images/1-alpine.img ---> Using cache ---> 8b9a52ef2456 Step 7/10 : ADD run-tgt.sh / ---> Using cache ---> cfa30ab2f553 Step 8/10 : EXPOSE 3260 ---> Using cache ---> eb3a3602eb3c Step 9/10 : CMD /run-tgt.sh ---> Using cache ---> 56b1e43742ed Step 10/10 : LABEL "iscsi-demo-target-tgtd" '' "kubevirt-functional-tests-openshift-release0" '' ---> Using cache ---> 00c1cf7185d3 Successfully built 00c1cf7185d3 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/vm-killer ./ Dockerfile sent 602 bytes received 34 bytes 1272.00 bytes/sec total size is 787 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/5 : FROM fedora:27 ---> 9110ae7f579f Step 2/5 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71c3d482487e Step 3/5 : ENV container docker ---> Using cache ---> 453fb17b7f2a Step 4/5 : RUN dnf -y install procps-ng && dnf -y clean all ---> Using cache ---> 41a521e2b7e1 Step 5/5 : LABEL "kubevirt-functional-tests-openshift-release0" '' "vm-killer" '' ---> Using cache ---> 1cb96de563f7 Successfully built 1cb96de563f7 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/registry-disk-v1alpha ./ Dockerfile entry-point.sh sent 1529 bytes received 53 bytes 3164.00 bytes/sec total size is 2482 speedup is 1.57 Sending build context to Docker daemon 5.12 kB Step 1/7 : FROM debian:sid ---> bcec0ae8107e Step 2/7 : MAINTAINER "David Vossel" \ ---> Using cache ---> 1ca40cafe086 Step 3/7 : ENV container docker ---> Using cache ---> 453271b8b5a3 Step 4/7 : RUN apt-get update && apt-get install -y bash curl bzip2 qemu-utils && mkdir -p /disk && rm -rf /var/lib/apt/lists/* ---> Using cache ---> 4ab88b363377 Step 5/7 : ADD entry-point.sh / ---> Using cache ---> eaa7729dd9dd Step 6/7 : CMD /entry-point.sh ---> Using cache ---> 23b1b7a48aee Step 7/7 : LABEL "kubevirt-functional-tests-openshift-release0" '' "registry-disk-v1alpha" '' ---> Using cache ---> 95eaca3989a1 Successfully built 95eaca3989a1 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/cirros-registry-disk-demo ./ Dockerfile sent 630 bytes received 34 bytes 1328.00 bytes/sec total size is 825 speedup is 1.24 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:32880/kubevirt/registry-disk-v1alpha:devel ---> 95eaca3989a1 Step 2/4 : MAINTAINER "David Vossel" \ ---> Using cache ---> a6b2ad633b37 Step 3/4 : RUN curl https://download.cirros-cloud.net/0.4.0/cirros-0.4.0-x86_64-disk.img > /disk/cirros.img ---> Using cache ---> 598a6e6cc83b Step 4/4 : LABEL "cirros-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release0" '' ---> Using cache ---> a09f1622bf9a Successfully built a09f1622bf9a sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/fedora-cloud-registry-disk-demo ./ Dockerfile sent 677 bytes received 34 bytes 1422.00 bytes/sec total size is 926 speedup is 1.30 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:32880/kubevirt/registry-disk-v1alpha:devel ---> 95eaca3989a1 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 160a2436d0d4 Step 3/4 : RUN curl -g -L https://download.fedoraproject.org/pub/fedora/linux/releases/27/CloudImages/x86_64/images/Fedora-Cloud-Base-27-1.6.x86_64.qcow2 > /disk/fedora.qcow2 ---> Using cache ---> 8040c94e272a Step 4/4 : LABEL "fedora-cloud-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release0" '' ---> Using cache ---> 44429ba0701a Successfully built 44429ba0701a sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/alpine-registry-disk-demo ./ Dockerfile sent 639 bytes received 34 bytes 1346.00 bytes/sec total size is 866 speedup is 1.29 Sending build context to Docker daemon 2.56 kB Step 1/4 : FROM localhost:32880/kubevirt/registry-disk-v1alpha:devel ---> 95eaca3989a1 Step 2/4 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 160a2436d0d4 Step 3/4 : RUN curl http://dl-cdn.alpinelinux.org/alpine/v3.7/releases/x86_64/alpine-virt-3.7.0-x86_64.iso > /disk/alpine.iso ---> Using cache ---> c655975234ad Step 4/4 : LABEL "alpine-registry-disk-demo" '' "kubevirt-functional-tests-openshift-release0" '' ---> Using cache ---> 4d8ed8509be5 Successfully built 4d8ed8509be5 sending incremental file list ./ Dockerfile sent 660 bytes received 34 bytes 1388.00 bytes/sec total size is 918 speedup is 1.32 Sending build context to Docker daemon 33.59 MB Step 1/8 : FROM fedora:27 ---> 9110ae7f579f Step 2/8 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71c3d482487e Step 3/8 : RUN useradd -u 1001 --create-home -s /bin/bash virtctl ---> Using cache ---> 934d38e9f0b5 Step 4/8 : WORKDIR /home/virtctl ---> Using cache ---> 14635c1cb3c5 Step 5/8 : USER 1001 ---> Using cache ---> 868158465567 Step 6/8 : COPY subresource-access-test /subresource-access-test ---> Using cache ---> 3c92ca92fcde Step 7/8 : ENTRYPOINT /subresource-access-test ---> Using cache ---> 7e286160e75d Step 8/8 : LABEL "kubevirt-functional-tests-openshift-release0" '' "subresource-access-test" '' ---> Running in a99b70ff86c0 ---> f6a94a6cb962 Removing intermediate container a99b70ff86c0 Successfully built f6a94a6cb962 sending incremental file list created directory /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd/winrmcli ./ Dockerfile sent 773 bytes received 34 bytes 1614.00 bytes/sec total size is 1098 speedup is 1.36 Sending build context to Docker daemon 3.072 kB Step 1/9 : FROM fedora:27 ---> 9110ae7f579f Step 2/9 : MAINTAINER "The KubeVirt Project" ---> Using cache ---> 71c3d482487e Step 3/9 : ENV container docker ---> Using cache ---> 453fb17b7f2a Step 4/9 : RUN dnf -y install make git gcc && dnf -y clean all ---> Running in 2ad3f9b9a4b6  Fedora 27 - x86_64 - Updates 650 kB/s | 22 MB 00:34 Fedora 27 - x86_64 1.5 MB/s | 58 MB 00:39 Last metadata expiration check: 0:00:34 ago on Mon Apr 9 15:36:38 2018. Dependencies resolved. ================================================================================ Package Arch Version Repository Size ================================================================================ Installing: gcc x86_64 7.3.1-5.fc27 updates 21 M git x86_64 2.14.3-3.fc27 updates 1.1 M make x86_64 1:4.2.1-4.fc27 fedora 494 k Upgrading: glibc x86_64 2.26-27.fc27 updates 3.4 M glibc-common x86_64 2.26-27.fc27 updates 789 k glibc-langpack-en x86_64 2.26-27.fc27 updates 278 k libcrypt-nss x86_64 2.26-27.fc27 updates 40 k Installing dependencies: binutils x86_64 2.29-13.fc27 updates 5.9 M cpp x86_64 7.3.1-5.fc27 updates 9.4 M fipscheck x86_64 1.5.0-3.fc27 fedora 25 k fipscheck-lib x86_64 1.5.0-3.fc27 fedora 14 k gc x86_64 7.6.0-7.fc27 fedora 110 k git-core x86_64 2.14.3-3.fc27 updates 4.1 M git-core-doc x86_64 2.14.3-3.fc27 updates 2.2 M glibc-devel x86_64 2.26-27.fc27 updates 989 k glibc-headers x86_64 2.26-27.fc27 updates 504 k groff-base x86_64 1.22.3-14.fc27 updates 1.0 M guile x86_64 5:2.0.14-3.fc27 fedora 3.5 M isl x86_64 0.16.1-3.fc27 fedora 835 k kernel-headers x86_64 4.15.14-300.fc27 updates 1.2 M less x86_64 487-5.fc27 fedora 159 k libatomic_ops x86_64 7.4.6-3.fc27 fedora 33 k libedit x86_64 3.1-20.20170329cvs.fc27 fedora 99 k libgomp x86_64 7.3.1-5.fc27 updates 211 k libmpc x86_64 1.0.2-8.fc27 fedora 56 k libstdc++ x86_64 7.3.1-5.fc27 updates 482 k libtool-ltdl x86_64 2.4.6-20.fc27 fedora 55 k openssh x86_64 7.6p1-5.fc27 updates 501 k openssh-clients x86_64 7.6p1-5.fc27 updates 671 k perl-Carp noarch 1.42-394.fc27 fedora 28 k perl-Encode x86_64 4:2.94-16.fc27 updates 1.5 M perl-Errno x86_64 1.28-403.fc27 updates 72 k perl-Error noarch 1:0.17025-1.fc27 fedora 44 k perl-Exporter noarch 5.72-395.fc27 fedora 32 k perl-File-Path noarch 2.15-1.fc27 fedora 37 k perl-File-Temp noarch 0.230.400-394.fc27 fedora 61 k perl-Getopt-Long noarch 1:2.50-3.fc27 fedora 61 k perl-Git noarch 2.14.3-3.fc27 updates 68 k perl-HTTP-Tiny noarch 0.070-394.fc27 fedora 56 k perl-IO x86_64 1.38-403.fc27 updates 138 k perl-MIME-Base64 x86_64 3.15-395.fc27 fedora 29 k perl-PathTools x86_64 3.74-1.fc27 updates 88 k perl-Pod-Escapes noarch 1:1.07-394.fc27 fedora 19 k perl-Pod-Perldoc noarch 3.28-395.fc27 fedora 87 k perl-Pod-Simple noarch 1:3.35-394.fc27 fedora 211 k perl-Pod-Usage noarch 4:1.69-394.fc27 fedora 33 k perl-Scalar-List-Utils x86_64 3:1.48-1.fc27 fedora 65 k perl-Socket x86_64 4:2.027-1.fc27 updates 57 k perl-Storable x86_64 1:2.62-395.fc27 fedora 84 k perl-Term-ANSIColor noarch 4.06-395.fc27 fedora 44 k perl-Term-Cap noarch 1.17-394.fc27 fedora 21 k perl-TermReadKey x86_64 2.37-5.fc27 fedora 38 k perl-Text-ParseWords noarch 3.30-394.fc27 fedora 16 k perl-Text-Tabs+Wrap noarch 2013.0523-394.fc27 fedora 23 k perl-Time-Local noarch 1:1.250-394.fc27 fedora 30 k perl-Unicode-Normalize x86_64 1.25-395.fc27 fedora 80 k perl-constant noarch 1.33-395.fc27 fedora 24 k perl-interpreter x86_64 4:5.26.1-403.fc27 updates 6.2 M perl-libs x86_64 4:5.26.1-403.fc27 updates 1.5 M perl-macros x86_64 4:5.26.1-403.fc27 updates 68 k perl-parent noarch 1:0.236-394.fc27 fedora 18 k perl-podlators noarch 4.09-395.fc27 updates 115 k perl-threads x86_64 1:2.21-1.fc27 updates 60 k perl-threads-shared x86_64 1.58-1.fc27 updates 46 k Installing weak dependencies: perl-IO-Socket-IP noarch 0.39-4.fc27 fedora 45 k perl-Mozilla-CA noarch 20160104-6.fc27 fedora 14 k Transaction Summary ================================================================================ Install 62 Packages Upgrade 4 Packages Total download size: 71 M Downloading Packages: (1/66): gc-7.6.0-7.fc27.x86_64.rpm 134 kB/s | 110 kB 00:00 (2/66): libatomic_ops-7.4.6-3.fc27.x86_64.rpm 202 kB/s | 33 kB 00:00 (3/66): libtool-ltdl-2.4.6-20.fc27.x86_64.rpm 303 kB/s | 55 kB 00:00 (4/66): make-4.2.1-4.fc27.x86_64.rpm 360 kB/s | 494 kB 00:01 (5/66): git-2.14.3-3.fc27.x86_64.rpm 936 kB/s | 1.1 MB 00:01 (6/66): guile-2.0.14-3.fc27.x86_64.rpm 940 kB/s | 3.5 MB 00:03 (7/66): perl-Git-2.14.3-3.fc27.noarch.rpm 120 kB/s | 68 kB 00:00 (8/66): perl-Error-0.17025-1.fc27.noarch.rpm 146 kB/s | 44 kB 00:00 (9/66): perl-Getopt-Long-2.50-3.fc27.noarch.rpm 179 kB/s | 61 kB 00:00 (10/66): git-core-2.14.3-3.fc27.x86_64.rpm 1.1 MB/s | 4.1 MB 00:03 (11/66): perl-TermReadKey-2.37-5.fc27.x86_64.rp 139 kB/s | 38 kB 00:00 (12/66): perl-Carp-1.42-394.fc27.noarch.rpm 106 kB/s | 28 kB 00:00 (13/66): less-487-5.fc27.x86_64.rpm 347 kB/s | 159 kB 00:00 (14/66): perl-Exporter-5.72-395.fc27.noarch.rpm 172 kB/s | 32 kB 00:00 (15/66): perl-Time-Local-1.250-394.fc27.noarch. 313 kB/s | 30 kB 00:00 (16/66): perl-Pod-Usage-1.69-394.fc27.noarch.rp 186 kB/s | 33 kB 00:00 (17/66): perl-Scalar-List-Utils-1.48-1.fc27.x86 149 kB/s | 65 kB 00:00 (18/66): perl-Text-ParseWords-3.30-394.fc27.noa 63 kB/s | 16 kB 00:00 (19/66): perl-constant-1.33-395.fc27.noarch.rpm 211 kB/s | 24 kB 00:00 (20/66): perl-parent-0.236-394.fc27.noarch.rpm 152 kB/s | 18 kB 00:00 (21/66): perl-Pod-Perldoc-3.28-395.fc27.noarch. 284 kB/s | 87 kB 00:00 (22/66): perl-File-Temp-0.230.400-394.fc27.noar 294 kB/s | 61 kB 00:00 (23/66): perl-HTTP-Tiny-0.070-394.fc27.noarch.r 270 kB/s | 56 kB 00:00 (24/66): perl-File-Path-2.15-1.fc27.noarch.rpm 209 kB/s | 37 kB 00:00 (25/66): perl-Pod-Simple-3.35-394.fc27.noarch.r 440 kB/s | 211 kB 00:00 (26/66): perl-MIME-Base64-3.15-395.fc27.x86_64. 148 kB/s | 29 kB 00:00 (27/66): perl-Pod-Escapes-1.07-394.fc27.noarch. 134 kB/s | 19 kB 00:00 (28/66): perl-Text-Tabs+Wrap-2013.0523-394.fc27 125 kB/s | 23 kB 00:00 (29/66): cpp-7.3.1-5.fc27.x86_64.rpm 1.7 MB/s | 9.4 MB 00:05 (30/66): isl-0.16.1-3.fc27.x86_64.rpm 571 kB/s | 835 kB 00:01 (31/66): libmpc-1.0.2-8.fc27.x86_64.rpm 484 kB/s | 56 kB 00:00 (32/66): glibc-devel-2.26-27.fc27.x86_64.rpm 866 kB/s | 989 kB 00:01 (33/66): glibc-headers-2.26-27.fc27.x86_64.rpm 345 kB/s | 504 kB 00:01 (34/66): perl-interpreter-5.26.1-403.fc27.x86_6 1.1 MB/s | 6.2 MB 00:05 (35/66): gcc-7.3.1-5.fc27.x86_64.rpm 1.2 MB/s | 21 MB 00:18 (36/66): perl-Unicode-Normalize-1.25-395.fc27.x 108 kB/s | 80 kB 00:00 (37/66): perl-Errno-1.28-403.fc27.x86_64.rpm 125 kB/s | 72 kB 00:00 (38/66): perl-PathTools-3.74-1.fc27.x86_64.rpm 247 kB/s | 88 kB 00:00 (39/66): perl-podlators-4.09-395.fc27.noarch.rp 268 kB/s | 115 kB 00:00 (40/66): perl-Term-ANSIColor-4.06-395.fc27.noar 114 kB/s | 44 kB 00:00 (41/66): perl-Term-Cap-1.17-394.fc27.noarch.rpm 82 kB/s | 21 kB 00:00 (42/66): perl-Encode-2.94-16.fc27.x86_64.rpm 1.1 MB/s | 1.5 MB 00:01 (43/66): perl-Storable-2.62-395.fc27.x86_64.rpm 157 kB/s | 84 kB 00:00 (44/66): groff-base-1.22.3-14.fc27.x86_64.rpm 1.0 MB/s | 1.0 MB 00:00 (45/66): perl-libs-5.26.1-403.fc27.x86_64.rpm 181 kB/s | 1.5 MB 00:08 (46/66): perl-Socket-2.027-1.fc27.x86_64.rpm 197 kB/s | 57 kB 00:00 (47/66): perl-IO-1.38-403.fc27.x86_64.rpm 288 kB/s | 138 kB 00:00 (48/66): perl-macros-5.26.1-403.fc27.x86_64.rpm 332 kB/s | 68 kB 00:00 (49/66): perl-threads-2.21-1.fc27.x86_64.rpm 231 kB/s | 60 kB 00:00 (50/66): perl-threads-shared-1.58-1.fc27.x86_64 177 kB/s | 46 kB 00:00 (51/66): libstdc++-7.3.1-5.fc27.x86_64.rpm 363 kB/s | 482 kB 00:01 (52/66): openssh-clients-7.6p1-5.fc27.x86_64.rp 367 kB/s | 671 kB 00:01 (53/66): openssh-7.6p1-5.fc27.x86_64.rpm 385 kB/s | 501 kB 00:01 (54/66): fipscheck-lib-1.5.0-3.fc27.x86_64.rpm 41 kB/s | 14 kB 00:00 (55/66): libedit-3.1-20.20170329cvs.fc27.x86_64 164 kB/s | 99 kB 00:00 (56/66): fipscheck-1.5.0-3.fc27.x86_64.rpm 120 kB/s | 25 kB 00:00 (57/66): git-core-doc-2.14.3-3.fc27.x86_64.rpm 61 kB/s | 2.2 MB 00:37 (58/66): libgomp-7.3.1-5.fc27.x86_64.rpm 252 kB/s | 211 kB 00:00 (59/66): perl-Mozilla-CA-20160104-6.fc27.noarch 65 kB/s | 14 kB 00:00 (60/66): perl-IO-Socket-IP-0.39-4.fc27.noarch.r 148 kB/s | 45 kB 00:00 (61/66): binutils-2.29-13.fc27.x86_64.rpm 1.2 MB/s | 5.9 MB 00:04 (62/66): glibc-common-2.26-27.fc27.x86_64.rpm 445 kB/s | 789 kB 00:01 (63/66): glibc-langpack-en-2.26-27.fc27.x86_64. 330 kB/s | 278 kB 00:00 (64/66): libcrypt-nss-2.26-27.fc27.x86_64.rpm 216 kB/s | 40 kB 00:00 (65/66): kernel-headers-4.15.14-300.fc27.x86_64 94 kB/s | 1.2 MB 00:13 (66/66): glibc-2.26-27.fc27.x86_64.rpm 234 kB/s | 3.4 MB 00:14 -------------------------------------------------------------------------------- Total 1.2 MB/s | 71 MB 00:58 Running transaction check Transaction check succeeded. Running transaction test Transaction test succeeded. Running transaction Preparing : 1/1 Upgrading : glibc-common-2.26-27.fc27.x86_64 1/70 Upgrading : glibc-langpack-en-2.26-27.fc27.x86_64 2/70 Running scriptlet: glibc-2.26-27.fc27.x86_64 3/70 Upgrading : glibc-2.26-27.fc27.x86_64 3/70 Running scriptlet: glibc-2.26-27.fc27.x86_64 3/70 Upgrading : libcrypt-nss-2.26-27.fc27.x86_64 4/70 Running scriptlet: libcrypt-nss-2.26-27.fc27.x86_64 4/70 Installing : perl-Exporter-5.72-395.fc27.noarch 5/70 Installing : perl-libs-4:5.26.1-403.fc27.x86_64 6/70 Running scriptlet: perl-libs-4:5.26.1-403.fc27.x86_64 6/70 Installing : perl-Carp-1.42-394.fc27.noarch 7/70 Installing : perl-Scalar-List-Utils-3:1.48-1.fc27.x86_64 8/70 Installing : fipscheck-1.5.0-3.fc27.x86_64 9/70 Installing : fipscheck-lib-1.5.0-3.fc27.x86_64 10/70 Running scriptlet: fipscheck-lib-1.5.0-3.fc27.x86_64 10/70 Installing : perl-Text-ParseWords-3.30-394.fc27.noarch 11/70 Installing : libmpc-1.0.2-8.fc27.x86_64 12/70 Running scriptlet: libmpc-1.0.2-8.fc27.x86_64 12/70 Installing : libstdc++-7.3.1-5.fc27.x86_64 13/70 Running scriptlet: libstdc++-7.3.1-5.fc27.x86_64 13/70 Installing : groff-base-1.22.3-14.fc27.x86_64 14/70 Installing : cpp-7.3.1-5.fc27.x86_64 15/70 Running scriptlet: cpp-7.3.1-5.fc27.x86_64 15/70 Running scriptlet: openssh-7.6p1-5.fc27.x86_64 16/70 Installing : openssh-7.6p1-5.fc27.x86_64 16/70 Installing : perl-Term-ANSIColor-4.06-395.fc27.noarch 17/70 Installing : perl-macros-4:5.26.1-403.fc27.x86_64 18/70 Installing : perl-constant-1.33-395.fc27.noarch 19/70 Installing : perl-parent-1:0.236-394.fc27.noarch 20/70 Installing : perl-Text-Tabs+Wrap-2013.0523-394.fc27.noarch 21/70 Installing : perl-Unicode-Normalize-1.25-395.fc27.x86_64 22/70 Installing : perl-File-Path-2.15-1.fc27.noarch 23/70 Installing : perl-PathTools-3.74-1.fc27.x86_64 24/70 Installing : perl-Errno-1.28-403.fc27.x86_64 25/70 Installing : perl-IO-1.38-403.fc27.x86_64 26/70 Installing : perl-Socket-4:2.027-1.fc27.x86_64 27/70 Installing : perl-threads-1:2.21-1.fc27.x86_64 28/70 Installing : perl-threads-shared-1.58-1.fc27.x86_64 29/70 Installing : perl-interpreter-4:5.26.1-403.fc27.x86_64 30/70 Installing : perl-Error-1:0.17025-1.fc27.noarch 31/70 Installing : perl-MIME-Base64-3.15-395.fc27.x86_64 32/70 Installing : perl-Time-Local-1:1.250-394.fc27.noarch 33/70 Installing : perl-HTTP-Tiny-0.070-394.fc27.noarch 34/70 Installing : perl-TermReadKey-2.37-5.fc27.x86_64 35/70 Installing : perl-File-Temp-0.230.400-394.fc27.noarch 36/70 Installing : perl-Pod-Escapes-1:1.07-394.fc27.noarch 37/70 Installing : perl-Term-Cap-1.17-394.fc27.noarch 38/70 Installing : perl-Storable-1:2.62-395.fc27.x86_64 39/70 Installing : perl-Pod-Simple-1:3.35-394.fc27.noarch 40/70 Installing : perl-Pod-Usage-4:1.69-394.fc27.noarch 41/70 Installing : perl-Getopt-Long-1:2.50-3.fc27.noarch 42/70 Installing : perl-Encode-4:2.94-16.fc27.x86_64 43/70 Installing : perl-podlators-4.09-395.fc27.noarch 44/70 Installing : perl-Pod-Perldoc-3.28-395.fc27.noarch 45/70 Installing : libatomic_ops-7.4.6-3.fc27.x86_64 46/70 Running scriptlet: libatomic_ops-7.4.6-3.fc27.x86_64 46/70 Installing : gc-7.6.0-7.fc27.x86_64 47/70 Running scriptlet: gc-7.6.0-7.fc27.x86_64 47/70 Installing : libtool-ltdl-2.4.6-20.fc27.x86_64 48/70 Running scriptlet: libtool-ltdl-2.4.6-20.fc27.x86_64 48/70 Installing : guile-5:2.0.14-3.fc27.x86_64 49/70 Running scriptlet: guile-5:2.0.14-3.fc27.x86_64 49/70 Installing : less-487-5.fc27.x86_64 50/70 Installing : isl-0.16.1-3.fc27.x86_64 51/70 Running scriptlet: isl-0.16.1-3.fc27.x86_64 51/70 Installing : libedit-3.1-20.20170329cvs.fc27.x86_64 52/70 Running scriptlet: libedit-3.1-20.20170329cvs.fc27.x86_64 52/70 Installing : openssh-clients-7.6p1-5.fc27.x86_64 53/70 Installing : git-core-2.14.3-3.fc27.x86_64 54/70 Installing : git-core-doc-2.14.3-3.fc27.x86_64 55/70 Installing : perl-Git-2.14.3-3.fc27.noarch 56/70 Installing : git-2.14.3-3.fc27.x86_64 57/70 Installing : binutils-2.29-13.fc27.x86_64 58/70 Running scriptlet: binutils-2.29-13.fc27.x86_64 58/70 install-info: No such file or directory for /usr/share/info/as.info.gz install-info: No such file or directory for /usr/share/info/binutils.info.gz install-info: No such file or directory for /usr/share/info/gprof.info.gz install-info: No such file or directory for /usr/share/info/ld.info.gz Installing : libgomp-7.3.1-5.fc27.x86_64 59/70 Running scriptlet: libgomp-7.3.1-5.fc27.x86_64 59/70 Installing : kernel-headers-4.15.14-300.fc27.x86_64 60/70 Running scriptlet: glibc-headers-2.26-27.fc27.x86_64 61/70 Installing : glibc-headers-2.26-27.fc27.x86_64 61/70 Installing : glibc-devel-2.26-27.fc27.x86_64 62/70 Running scriptlet: glibc-devel-2.26-27.fc27.x86_64 62/70 Installing : gcc-7.3.1-5.fc27.x86_64 63/70 Running scriptlet: gcc-7.3.1-5.fc27.x86_64 63/70 Installing : make-1:4.2.1-4.fc27.x86_64 64/70 Running scriptlet: make-1:4.2.1-4.fc27.x86_64 64/70 Installing : perl-IO-Socket-IP-0.39-4.fc27.noarch 65/70 Installing : perl-Mozilla-CA-20160104-6.fc27.noarch 66/70 Cleanup : libcrypt-nss-2.26-26.fc27.x86_64 67/70 Running scriptlet: libcrypt-nss-2.26-26.fc27.x86_64 67/70 Cleanup : glibc-2.26-26.fc27.x86_64 68/70 Running scriptlet: glibc-2.26-26.fc27.x86_64 68/70 Cleanup : glibc-langpack-en-2.26-26.fc27.x86_64 69/70 Cleanup : glibc-common-2.26-26.fc27.x86_64 70/70 Running scriptlet: guile-5:2.0.14-3.fc27.x86_64 70/70 Verifying : make-1:4.2.1-4.fc27.x86_64 1/70 Verifying : gc-7.6.0-7.fc27.x86_64 2/70 Verifying : guile-5:2.0.14-3.fc27.x86_64 3/70 Verifying : libatomic_ops-7.4.6-3.fc27.x86_64 4/70 Verifying : libtool-ltdl-2.4.6-20.fc27.x86_64 5/70 Verifying : git-2.14.3-3.fc27.x86_64 6/70 Verifying : git-core-2.14.3-3.fc27.x86_64 7/70 Verifying : git-core-doc-2.14.3-3.fc27.x86_64 8/70 Verifying : perl-Git-2.14.3-3.fc27.noarch 9/70 Verifying : perl-Error-1:0.17025-1.fc27.noarch 10/70 Verifying : perl-Getopt-Long-1:2.50-3.fc27.noarch 11/70 Verifying : perl-TermReadKey-2.37-5.fc27.x86_64 12/70 Verifying : less-487-5.fc27.x86_64 13/70 Verifying : perl-Carp-1.42-394.fc27.noarch 14/70 Verifying : perl-Exporter-5.72-395.fc27.noarch 15/70 Verifying : perl-Time-Local-1:1.250-394.fc27.noarch 16/70 Verifying : perl-Scalar-List-Utils-3:1.48-1.fc27.x86_64 17/70 Verifying : perl-Pod-Usage-4:1.69-394.fc27.noarch 18/70 Verifying : perl-Text-ParseWords-3.30-394.fc27.noarch 19/70 Verifying : perl-constant-1.33-395.fc27.noarch 20/70 Verifying : perl-parent-1:0.236-394.fc27.noarch 21/70 Verifying : perl-Pod-Perldoc-3.28-395.fc27.noarch 22/70 Verifying : perl-File-Temp-0.230.400-394.fc27.noarch 23/70 Verifying : perl-HTTP-Tiny-0.070-394.fc27.noarch 24/70 Verifying : perl-Pod-Simple-1:3.35-394.fc27.noarch 25/70 Verifying : perl-File-Path-2.15-1.fc27.noarch 26/70 Verifying : perl-MIME-Base64-3.15-395.fc27.x86_64 27/70 Verifying : perl-Pod-Escapes-1:1.07-394.fc27.noarch 28/70 Verifying : perl-Text-Tabs+Wrap-2013.0523-394.fc27.noarch 29/70 Verifying : gcc-7.3.1-5.fc27.x86_64 30/70 Verifying : cpp-7.3.1-5.fc27.x86_64 31/70 Verifying : isl-0.16.1-3.fc27.x86_64 32/70 Verifying : libmpc-1.0.2-8.fc27.x86_64 33/70 Verifying : glibc-devel-2.26-27.fc27.x86_64 34/70 Verifying : glibc-headers-2.26-27.fc27.x86_64 35/70 Verifying : perl-interpreter-4:5.26.1-403.fc27.x86_64 36/70 Verifying : perl-libs-4:5.26.1-403.fc27.x86_64 37/70 Verifying : perl-Unicode-Normalize-1.25-395.fc27.x86_64 38/70 Verifying : perl-Errno-1.28-403.fc27.x86_64 39/70 Verifying : perl-PathTools-3.74-1.fc27.x86_64 40/70 Verifying : perl-podlators-4.09-395.fc27.noarch 41/70 Verifying : perl-Term-ANSIColor-4.06-395.fc27.noarch 42/70 Verifying : perl-Term-Cap-1.17-394.fc27.noarch 43/70 Verifying : perl-Encode-4:2.94-16.fc27.x86_64 44/70 Verifying : perl-Storable-1:2.62-395.fc27.x86_64 45/70 Verifying : groff-base-1.22.3-14.fc27.x86_64 46/70 Verifying : perl-IO-1.38-403.fc27.x86_64 47/70 Verifying : perl-Socket-4:2.027-1.fc27.x86_64 48/70 Verifying : libstdc++-7.3.1-5.fc27.x86_64 49/70 Verifying : perl-macros-4:5.26.1-403.fc27.x86_64 50/70 Verifying : perl-threads-1:2.21-1.fc27.x86_64 51/70 Verifying : perl-threads-shared-1.58-1.fc27.x86_64 52/70 Verifying : kernel-headers-4.15.14-300.fc27.x86_64 53/70 Verifying : openssh-clients-7.6p1-5.fc27.x86_64 54/70 Verifying : openssh-7.6p1-5.fc27.x86_64 55/70 Verifying : fipscheck-lib-1.5.0-3.fc27.x86_64 56/70 Verifying : libedit-3.1-20.20170329cvs.fc27.x86_64 57/70 Verifying : fipscheck-1.5.0-3.fc27.x86_64 58/70 Verifying : binutils-2.29-13.fc27.x86_64 59/70 Verifying : libgomp-7.3.1-5.fc27.x86_64 60/70 Verifying : perl-Mozilla-CA-20160104-6.fc27.noarch 61/70 Verifying : perl-IO-Socket-IP-0.39-4.fc27.noarch 62/70 Verifying : glibc-2.26-27.fc27.x86_64 63/70 Verifying : glibc-common-2.26-27.fc27.x86_64 64/70 Verifying : glibc-langpack-en-2.26-27.fc27.x86_64 65/70 Verifying : libcrypt-nss-2.26-27.fc27.x86_64 66/70 Verifying : glibc-2.26-26.fc27.x86_64 67/70 Verifying : glibc-common-2.26-26.fc27.x86_64 68/70 Verifying : glibc-langpack-en-2.26-26.fc27.x86_64 69/70 Verifying : libcrypt-nss-2.26-26.fc27.x86_64 70/70 Installed: gcc.x86_64 7.3.1-5.fc27 git.x86_64 2.14.3-3.fc27 make.x86_64 1:4.2.1-4.fc27 perl-IO-Socket-IP.noarch 0.39-4.fc27 perl-Mozilla-CA.noarch 20160104-6.fc27 binutils.x86_64 2.29-13.fc27 cpp.x86_64 7.3.1-5.fc27 fipscheck.x86_64 1.5.0-3.fc27 fipscheck-lib.x86_64 1.5.0-3.fc27 gc.x86_64 7.6.0-7.fc27 git-core.x86_64 2.14.3-3.fc27 git-core-doc.x86_64 2.14.3-3.fc27 glibc-devel.x86_64 2.26-27.fc27 glibc-headers.x86_64 2.26-27.fc27 groff-base.x86_64 1.22.3-14.fc27 guile.x86_64 5:2.0.14-3.fc27 isl.x86_64 0.16.1-3.fc27 kernel-headers.x86_64 4.15.14-300.fc27 less.x86_64 487-5.fc27 libatomic_ops.x86_64 7.4.6-3.fc27 libedit.x86_64 3.1-20.20170329cvs.fc27 libgomp.x86_64 7.3.1-5.fc27 libmpc.x86_64 1.0.2-8.fc27 libstdc++.x86_64 7.3.1-5.fc27 libtool-ltdl.x86_64 2.4.6-20.fc27 openssh.x86_64 7.6p1-5.fc27 openssh-clients.x86_64 7.6p1-5.fc27 perl-Carp.noarch 1.42-394.fc27 perl-Encode.x86_64 4:2.94-16.fc27 perl-Errno.x86_64 1.28-403.fc27 perl-Error.noarch 1:0.17025-1.fc27 perl-Exporter.noarch 5.72-395.fc27 perl-File-Path.noarch 2.15-1.fc27 perl-File-Temp.noarch 0.230.400-394.fc27 perl-Getopt-Long.noarch 1:2.50-3.fc27 perl-Git.noarch 2.14.3-3.fc27 perl-HTTP-Tiny.noarch 0.070-394.fc27 perl-IO.x86_64 1.38-403.fc27 perl-MIME-Base64.x86_64 3.15-395.fc27 perl-PathTools.x86_64 3.74-1.fc27 perl-Pod-Escapes.noarch 1:1.07-394.fc27 perl-Pod-Perldoc.noarch 3.28-395.fc27 perl-Pod-Simple.noarch 1:3.35-394.fc27 perl-Pod-Usage.noarch 4:1.69-394.fc27 perl-Scalar-List-Utils.x86_64 3:1.48-1.fc27 perl-Socket.x86_64 4:2.027-1.fc27 perl-Storable.x86_64 1:2.62-395.fc27 perl-Term-ANSIColor.noarch 4.06-395.fc27 perl-Term-Cap.noarch 1.17-394.fc27 perl-TermReadKey.x86_64 2.37-5.fc27 perl-Text-ParseWords.noarch 3.30-394.fc27 perl-Text-Tabs+Wrap.noarch 2013.0523-394.fc27 perl-Time-Local.noarch 1:1.250-394.fc27 perl-Unicode-Normalize.x86_64 1.25-395.fc27 perl-constant.noarch 1.33-395.fc27 perl-interpreter.x86_64 4:5.26.1-403.fc27 perl-libs.x86_64 4:5.26.1-403.fc27 perl-macros.x86_64 4:5.26.1-403.fc27 perl-parent.noarch 1:0.236-394.fc27 perl-podlators.noarch 4.09-395.fc27 perl-threads.x86_64 1:2.21-1.fc27 perl-threads-shared.x86_64 1.58-1.fc27 Upgraded: glibc.x86_64 2.26-27.fc27 glibc-common.x86_64 2.26-27.fc27 glibc-langpack-en.x86_64 2.26-27.fc27 libcrypt-nss.x86_64 2.26-27.fc27 Complete! 18 files removed ---> 3b80b8a7f524 Removing intermediate container 2ad3f9b9a4b6 Step 5/9 : ENV GIMME_GO_VERSION 1.9.2 ---> Running in bf3be44eeb16 ---> 2f0614c385bb Removing intermediate container bf3be44eeb16 Step 6/9 : RUN mkdir -p /gimme && curl -sL https://raw.githubusercontent.com/travis-ci/gimme/master/gimme | HOME=/gimme bash >> /etc/profile.d/gimme.sh ---> Running in 66587c7a94ea  ---> 4c3d4be00446 Removing intermediate container 66587c7a94ea Step 7/9 : ENV GOPATH "/go" GOBIN "/usr/bin" ---> Running in 4db5dd8fe449 ---> c263f5b895a3 Removing intermediate container 4db5dd8fe449 Step 8/9 : RUN mkdir -p /go && source /etc/profile.d/gimme.sh && go get github.com/masterzen/winrm-cli ---> Running in 812ea0d15c5c go version go1.9.2 linux/amd64  ---> adb2923af76f Removing intermediate container 812ea0d15c5c Step 9/9 : LABEL "kubevirt-functional-tests-openshift-release0" '' "winrmcli" '' ---> Running in 9b1785af93a4 ---> 92501309b691 Removing intermediate container 9b1785af93a4 Successfully built 92501309b691 hack/build-docker.sh push The push refers to a repository [localhost:32880/kubevirt/virt-controller] 76b1d9773953: Preparing ebdaa8997db7: Preparing 39bae602f753: Preparing ebdaa8997db7: Layer already exists 39bae602f753: Layer already exists 76b1d9773953: Pushed devel: digest: sha256:c2280c5478133cf20f7192a0182ac69fe08a517d198caf3f4dd165c980537afa size: 948 The push refers to a repository [localhost:32880/kubevirt/virt-launcher] 9025582f266c: Preparing ca8a0129689f: Preparing ca8a0129689f: Preparing e4f675c6012c: Preparing ff14c1ecb896: Preparing eee73cffd252: Preparing 686e1987fc42: Preparing 22b0b3033053: Preparing 5a3c34a960cd: Preparing 746161d2a6d5: Preparing 530cc55618cd: Preparing eee73cffd252: Waiting 34fa414dfdf6: Preparing a1359dc556dd: Preparing 22b0b3033053: Waiting 490c7c373332: Preparing 4b440db36f72: Preparing 5a3c34a960cd: Waiting 39bae602f753: Preparing 746161d2a6d5: Waiting 34fa414dfdf6: Waiting 4b440db36f72: Waiting 530cc55618cd: Waiting 490c7c373332: Waiting ff14c1ecb896: Pushed e4f675c6012c: Pushed ca8a0129689f: Pushed 9025582f266c: Pushed 22b0b3033053: Layer already exists 5a3c34a960cd: Layer already exists 746161d2a6d5: Layer already exists 530cc55618cd: Layer already exists 34fa414dfdf6: Layer already exists eee73cffd252: Pushed a1359dc556dd: Layer already exists 490c7c373332: Layer already exists 4b440db36f72: Layer already exists 39bae602f753: Layer already exists 686e1987fc42: Pushed devel: digest: sha256:3bb03c0981b761699bb142a1db3c8fee4adec4c1fc2c05c2f5598840ec5ae623 size: 3652 The push refers to a repository [localhost:32880/kubevirt/virt-handler] 131f068e9745: Preparing 39bae602f753: Preparing 39bae602f753: Layer already exists 131f068e9745: Pushed devel: digest: sha256:3638a8a743d0f540c0f7d7cb81a18eadda5fdd23f41dfd30cc4d2b897f60cf25 size: 740 The push refers to a repository [localhost:32880/kubevirt/virt-api] e8bbd6155c32: Preparing d57b3fcdcebf: Preparing 325865597484: Preparing 39bae602f753: Preparing 325865597484: Layer already exists d57b3fcdcebf: Layer already exists 39bae602f753: Layer already exists e8bbd6155c32: Pushed devel: digest: sha256:580305656a47348a2f8251377bab50a64b65b460c5a62f65d57ea117a96c31af size: 1159 The push refers to a repository [localhost:32880/kubevirt/iscsi-demo-target-tgtd] d7a94c2260cf: Preparing 51f335cf165a: Preparing 34892f148b26: Preparing 27e33247129d: Preparing 39bae602f753: Preparing 39bae602f753: Waiting 27e33247129d: Layer already exists d7a94c2260cf: Layer already exists 34892f148b26: Layer already exists 51f335cf165a: Layer already exists 39bae602f753: Layer already exists devel: digest: sha256:43943f06f85bc53a88be9b260520ed3cfcbe9a9fe45454f2ac8a25d1cdfcdf6e size: 1368 The push refers to a repository [localhost:32880/kubevirt/vm-killer] 7074bc3e2ecc: Preparing 39bae602f753: Preparing 7074bc3e2ecc: Layer already exists 39bae602f753: Layer already exists devel: digest: sha256:00ebfda087225e6a3d9fcde7830704c4f9861c6c623f9adb2161914d1da33003 size: 740 The push refers to a repository [localhost:32880/kubevirt/registry-disk-v1alpha] ba2358758a96: Preparing dd28ef9bfc97: Preparing 6709b2da72b8: Preparing dd28ef9bfc97: Layer already exists ba2358758a96: Layer already exists 6709b2da72b8: Layer already exists devel: digest: sha256:744c493fd0ab6448294021129ed7957cbf5f84eaeef228e0b7a26b47ca2cc1bc size: 948 The push refers to a repository [localhost:32880/kubevirt/cirros-registry-disk-demo] 6f1563ff3965: Preparing ba2358758a96: Preparing dd28ef9bfc97: Preparing 6709b2da72b8: Preparing 6f1563ff3965: Layer already exists dd28ef9bfc97: Layer already exists ba2358758a96: Layer already exists 6709b2da72b8: Layer already exists devel: digest: sha256:41e64f05daca4af9bb1a590f3556a3eff97ab73fff6b7cc235e372c3d5718be8 size: 1160 The push refers to a repository [localhost:32880/kubevirt/fedora-cloud-registry-disk-demo] c5b9f395fe4f: Preparing ba2358758a96: Preparing dd28ef9bfc97: Preparing 6709b2da72b8: Preparing c5b9f395fe4f: Layer already exists 6709b2da72b8: Layer already exists dd28ef9bfc97: Layer already exists ba2358758a96: Layer already exists devel: digest: sha256:ccc35b36a163a8712e70d7980de479c0d5916989fb4562cb6978a518cff82dc4 size: 1161 The push refers to a repository [localhost:32880/kubevirt/alpine-registry-disk-demo] 1e24700d96a5: Preparing ba2358758a96: Preparing dd28ef9bfc97: Preparing 6709b2da72b8: Preparing 1e24700d96a5: Layer already exists ba2358758a96: Layer already exists 6709b2da72b8: Layer already exists dd28ef9bfc97: Layer already exists devel: digest: sha256:8c47587932661803fac596f5b13f283cf2e0c7a841e9c10f57955ddcdd9602f5 size: 1160 The push refers to a repository [localhost:32880/kubevirt/subresource-access-test] 61b304e20ff0: Preparing 601899d28c74: Preparing 39bae602f753: Preparing 601899d28c74: Layer already exists 39bae602f753: Layer already exists 61b304e20ff0: Pushed devel: digest: sha256:fe4b7dad9dc54c835e6d5d67fa2cfb4bfbb55dc4de426d943b4116bb8c0be23c size: 948 The push refers to a repository [localhost:32880/kubevirt/winrmcli] 4c1b632856f3: Preparing e3fbfdd545e5: Preparing f0bec087f513: Preparing 39bae602f753: Preparing 39bae602f753: Mounted from kubevirt/subresource-access-test 4c1b632856f3: Pushed f0bec087f513: Pushed e3fbfdd545e5: Pushed devel: digest: sha256:db2b8707d3ad61bf0935b493c2a1342531deb3d902ccf91b9dd77789693be68d size: 1165 make[1]: Leaving directory `/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt' 2018/04/09 15:43:58 Waiting for host: 192.168.66.101:22 2018/04/09 15:43:58 Connected to tcp://192.168.66.101:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer 0f9c31467808: Pulling fs layer 63c6bc564da2: Pulling fs layer 0f9c31467808: Verifying Checksum 0f9c31467808: Download complete 63c6bc564da2: Verifying Checksum 63c6bc564da2: Download complete 2176639d844b: Download complete 2176639d844b: Pull complete 0f9c31467808: Pull complete 63c6bc564da2: Pull complete Digest: sha256:c2280c5478133cf20f7192a0182ac69fe08a517d198caf3f4dd165c980537afa Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer 622fd4d469ec: Pulling fs layer 424529d59559: Pulling fs layer 88a7cd7d71a5: Pulling fs layer 9d41784165ed: Pulling fs layer 0bb1eb26cd61: Pulling fs layer 67c3d70f7d71: Pulling fs layer 9820c45639c6: Pulling fs layer 2cd0bdf70106: Pulling fs layer 47d1c609f225: Pulling fs layer 9d41784165ed: Waiting 0bb1eb26cd61: Waiting 67c3d70f7d71: Waiting 9820c45639c6: Waiting 2cd0bdf70106: Waiting 47d1c609f225: Waiting a1e80189bea5: Waiting 6cc174edcebf: Waiting 622fd4d469ec: Waiting 424529d59559: Waiting 88a7cd7d71a5: Waiting f2ef945504a7: Verifying Checksum f2ef945504a7: Download complete a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete 424529d59559: Verifying Checksum 424529d59559: Download complete 88a7cd7d71a5: Verifying Checksum 88a7cd7d71a5: Download complete 622fd4d469ec: Verifying Checksum 622fd4d469ec: Download complete 0bb1eb26cd61: Verifying Checksum 0bb1eb26cd61: Download complete 67c3d70f7d71: Verifying Checksum 67c3d70f7d71: Download complete 9d41784165ed: Verifying Checksum 9d41784165ed: Download complete 9820c45639c6: Verifying Checksum 2cd0bdf70106: Verifying Checksum 2cd0bdf70106: Download complete 9820c45639c6: Download complete 47d1c609f225: Verifying Checksum 47d1c609f225: Download complete d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete 622fd4d469ec: Pull complete 424529d59559: Pull complete 88a7cd7d71a5: Pull complete 9d41784165ed: Pull complete 0bb1eb26cd61: Pull complete 67c3d70f7d71: Pull complete 9820c45639c6: Pull complete 2cd0bdf70106: Pull complete 47d1c609f225: Pull complete Digest: sha256:3bb03c0981b761699bb142a1db3c8fee4adec4c1fc2c05c2f5598840ec5ae623 Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists 5442ec147ce1: Pulling fs layer 5442ec147ce1: Verifying Checksum 5442ec147ce1: Download complete 5442ec147ce1: Pull complete Digest: sha256:3638a8a743d0f540c0f7d7cb81a18eadda5fdd23f41dfd30cc4d2b897f60cf25 Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists 37e23c3b1f9e: Pulling fs layer ff02d6919641: Pulling fs layer e5c7956144fe: Pulling fs layer 37e23c3b1f9e: Verifying Checksum 37e23c3b1f9e: Download complete ff02d6919641: Verifying Checksum ff02d6919641: Download complete e5c7956144fe: Verifying Checksum e5c7956144fe: Download complete 37e23c3b1f9e: Pull complete ff02d6919641: Pull complete e5c7956144fe: Pull complete Digest: sha256:580305656a47348a2f8251377bab50a64b65b460c5a62f65d57ea117a96c31af Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists e929cf197109: Pulling fs layer 26e3b3165052: Pulling fs layer aa37eee00b09: Pulling fs layer 6c4271069635: Pulling fs layer 6c4271069635: Waiting 26e3b3165052: Download complete 6c4271069635: Verifying Checksum 6c4271069635: Download complete aa37eee00b09: Verifying Checksum aa37eee00b09: Download complete e929cf197109: Verifying Checksum e929cf197109: Download complete e929cf197109: Pull complete 26e3b3165052: Pull complete aa37eee00b09: Pull complete 6c4271069635: Pull complete Digest: sha256:43943f06f85bc53a88be9b260520ed3cfcbe9a9fe45454f2ac8a25d1cdfcdf6e Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists 3726544f0c71: Pulling fs layer 3726544f0c71: Verifying Checksum 3726544f0c71: Download complete 3726544f0c71: Pull complete Digest: sha256:00ebfda087225e6a3d9fcde7830704c4f9861c6c623f9adb2161914d1da33003 Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer 07ea0109ea1b: Pulling fs layer c4272834948a: Pulling fs layer c4272834948a: Verifying Checksum c4272834948a: Download complete 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 07ea0109ea1b: Verifying Checksum 07ea0109ea1b: Download complete 2115d46e7396: Pull complete 07ea0109ea1b: Pull complete c4272834948a: Pull complete Digest: sha256:744c493fd0ab6448294021129ed7957cbf5f84eaeef228e0b7a26b47ca2cc1bc Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists 07ea0109ea1b: Already exists c4272834948a: Already exists 6584c8e6fa33: Pulling fs layer 6584c8e6fa33: Verifying Checksum 6584c8e6fa33: Download complete 6584c8e6fa33: Pull complete Digest: sha256:41e64f05daca4af9bb1a590f3556a3eff97ab73fff6b7cc235e372c3d5718be8 Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists 07ea0109ea1b: Already exists c4272834948a: Already exists 349669550f49: Pulling fs layer 349669550f49: Download complete 349669550f49: Pull complete Digest: sha256:ccc35b36a163a8712e70d7980de479c0d5916989fb4562cb6978a518cff82dc4 Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists 07ea0109ea1b: Already exists c4272834948a: Already exists 8bc4b617e33b: Pulling fs layer 8bc4b617e33b: Verifying Checksum 8bc4b617e33b: Download complete 8bc4b617e33b: Pull complete Digest: sha256:8c47587932661803fac596f5b13f283cf2e0c7a841e9c10f57955ddcdd9602f5 Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists 555998d3afb2: Pulling fs layer aa4c4c4a31ba: Pulling fs layer 555998d3afb2: Verifying Checksum 555998d3afb2: Download complete aa4c4c4a31ba: Verifying Checksum aa4c4c4a31ba: Download complete 555998d3afb2: Pull complete aa4c4c4a31ba: Pull complete Digest: sha256:fe4b7dad9dc54c835e6d5d67fa2cfb4bfbb55dc4de426d943b4116bb8c0be23c Trying to pull repository registry:5000/kubevirt/winrmcli ... devel: Pulling from registry:5000/kubevirt/winrmcli 2176639d844b: Already exists 494ac3d2e337: Pulling fs layer fa95949b860a: Pulling fs layer 97a795c22b24: Pulling fs layer 97a795c22b24: Verifying Checksum 97a795c22b24: Download complete 494ac3d2e337: Verifying Checksum 494ac3d2e337: Download complete fa95949b860a: Verifying Checksum fa95949b860a: Download complete 494ac3d2e337: Pull complete fa95949b860a: Pull complete 97a795c22b24: Pull complete Digest: sha256:db2b8707d3ad61bf0935b493c2a1342531deb3d902ccf91b9dd77789693be68d 2018/04/09 15:47:52 Waiting for host: 192.168.66.101:22 2018/04/09 15:47:52 Connected to tcp://192.168.66.101:22 2018/04/09 15:47:55 Waiting for host: 192.168.66.102:22 2018/04/09 15:47:55 Connected to tcp://192.168.66.102:22 Trying to pull repository registry:5000/kubevirt/virt-controller ... devel: Pulling from registry:5000/kubevirt/virt-controller 2176639d844b: Pulling fs layer 0f9c31467808: Pulling fs layer 63c6bc564da2: Pulling fs layer 0f9c31467808: Verifying Checksum 0f9c31467808: Download complete 63c6bc564da2: Verifying Checksum 63c6bc564da2: Download complete 2176639d844b: Download complete 2176639d844b: Pull complete 0f9c31467808: Pull complete 63c6bc564da2: Pull complete Digest: sha256:c2280c5478133cf20f7192a0182ac69fe08a517d198caf3f4dd165c980537afa Trying to pull repository registry:5000/kubevirt/virt-launcher ... devel: Pulling from registry:5000/kubevirt/virt-launcher 2176639d844b: Already exists d7240bccd145: Pulling fs layer f2ef945504a7: Pulling fs layer a4b9e9eb807b: Pulling fs layer a1e80189bea5: Pulling fs layer 6cc174edcebf: Pulling fs layer 622fd4d469ec: Pulling fs layer 424529d59559: Pulling fs layer 88a7cd7d71a5: Pulling fs layer 9d41784165ed: Pulling fs layer 0bb1eb26cd61: Pulling fs layer 67c3d70f7d71: Pulling fs layer 9820c45639c6: Pulling fs layer 2cd0bdf70106: Pulling fs layer 47d1c609f225: Pulling fs layer a1e80189bea5: Waiting 6cc174edcebf: Waiting 622fd4d469ec: Waiting 424529d59559: Waiting 88a7cd7d71a5: Waiting 9d41784165ed: Waiting 0bb1eb26cd61: Waiting 67c3d70f7d71: Waiting 9820c45639c6: Waiting 2cd0bdf70106: Waiting 47d1c609f225: Waiting a4b9e9eb807b: Verifying Checksum a4b9e9eb807b: Download complete f2ef945504a7: Verifying Checksum f2ef945504a7: Download complete a1e80189bea5: Verifying Checksum a1e80189bea5: Download complete 6cc174edcebf: Verifying Checksum 6cc174edcebf: Download complete 424529d59559: Verifying Checksum 424529d59559: Download complete 88a7cd7d71a5: Verifying Checksum 88a7cd7d71a5: Download complete 622fd4d469ec: Verifying Checksum 622fd4d469ec: Download complete 9d41784165ed: Verifying Checksum 9d41784165ed: Download complete 0bb1eb26cd61: Verifying Checksum 0bb1eb26cd61: Download complete 67c3d70f7d71: Verifying Checksum 67c3d70f7d71: Download complete 2cd0bdf70106: Verifying Checksum 2cd0bdf70106: Download complete 9820c45639c6: Verifying Checksum 9820c45639c6: Download complete 47d1c609f225: Verifying Checksum 47d1c609f225: Download complete d7240bccd145: Verifying Checksum d7240bccd145: Download complete d7240bccd145: Pull complete f2ef945504a7: Pull complete a4b9e9eb807b: Pull complete a1e80189bea5: Pull complete 6cc174edcebf: Pull complete 622fd4d469ec: Pull complete 424529d59559: Pull complete 88a7cd7d71a5: Pull complete 9d41784165ed: Pull complete 0bb1eb26cd61: Pull complete 67c3d70f7d71: Pull complete 9820c45639c6: Pull complete 2cd0bdf70106: Pull complete 47d1c609f225: Pull complete Digest: sha256:3bb03c0981b761699bb142a1db3c8fee4adec4c1fc2c05c2f5598840ec5ae623 Trying to pull repository registry:5000/kubevirt/virt-handler ... devel: Pulling from registry:5000/kubevirt/virt-handler 2176639d844b: Already exists 5442ec147ce1: Pulling fs layer 5442ec147ce1: Verifying Checksum 5442ec147ce1: Download complete 5442ec147ce1: Pull complete Digest: sha256:3638a8a743d0f540c0f7d7cb81a18eadda5fdd23f41dfd30cc4d2b897f60cf25 Trying to pull repository registry:5000/kubevirt/virt-api ... devel: Pulling from registry:5000/kubevirt/virt-api 2176639d844b: Already exists 37e23c3b1f9e: Pulling fs layer ff02d6919641: Pulling fs layer e5c7956144fe: Pulling fs layer 37e23c3b1f9e: Verifying Checksum 37e23c3b1f9e: Download complete ff02d6919641: Verifying Checksum ff02d6919641: Download complete 37e23c3b1f9e: Pull complete e5c7956144fe: Verifying Checksum e5c7956144fe: Download complete ff02d6919641: Pull complete e5c7956144fe: Pull complete Digest: sha256:580305656a47348a2f8251377bab50a64b65b460c5a62f65d57ea117a96c31af Trying to pull repository registry:5000/kubevirt/iscsi-demo-target-tgtd ... devel: Pulling from registry:5000/kubevirt/iscsi-demo-target-tgtd 2176639d844b: Already exists e929cf197109: Pulling fs layer 26e3b3165052: Pulling fs layer aa37eee00b09: Pulling fs layer 6c4271069635: Pulling fs layer 6c4271069635: Waiting 26e3b3165052: Verifying Checksum 26e3b3165052: Download complete 6c4271069635: Verifying Checksum 6c4271069635: Download complete aa37eee00b09: Verifying Checksum aa37eee00b09: Download complete e929cf197109: Verifying Checksum e929cf197109: Download complete e929cf197109: Pull complete 26e3b3165052: Pull complete aa37eee00b09: Pull complete 6c4271069635: Pull complete Digest: sha256:43943f06f85bc53a88be9b260520ed3cfcbe9a9fe45454f2ac8a25d1cdfcdf6e Trying to pull repository registry:5000/kubevirt/vm-killer ... devel: Pulling from registry:5000/kubevirt/vm-killer 2176639d844b: Already exists 3726544f0c71: Pulling fs layer 3726544f0c71: Verifying Checksum 3726544f0c71: Download complete 3726544f0c71: Pull complete Digest: sha256:00ebfda087225e6a3d9fcde7830704c4f9861c6c623f9adb2161914d1da33003 Trying to pull repository registry:5000/kubevirt/registry-disk-v1alpha ... devel: Pulling from registry:5000/kubevirt/registry-disk-v1alpha 2115d46e7396: Pulling fs layer 07ea0109ea1b: Pulling fs layer c4272834948a: Pulling fs layer c4272834948a: Verifying Checksum 07ea0109ea1b: Verifying Checksum 07ea0109ea1b: Download complete 2115d46e7396: Verifying Checksum 2115d46e7396: Download complete 2115d46e7396: Pull complete 07ea0109ea1b: Pull complete c4272834948a: Pull complete Digest: sha256:744c493fd0ab6448294021129ed7957cbf5f84eaeef228e0b7a26b47ca2cc1bc Trying to pull repository registry:5000/kubevirt/cirros-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/cirros-registry-disk-demo 2115d46e7396: Already exists 07ea0109ea1b: Already exists c4272834948a: Already exists 6584c8e6fa33: Pulling fs layer 6584c8e6fa33: Verifying Checksum 6584c8e6fa33: Download complete 6584c8e6fa33: Pull complete Digest: sha256:41e64f05daca4af9bb1a590f3556a3eff97ab73fff6b7cc235e372c3d5718be8 Trying to pull repository registry:5000/kubevirt/fedora-cloud-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/fedora-cloud-registry-disk-demo 2115d46e7396: Already exists 07ea0109ea1b: Already exists c4272834948a: Already exists 349669550f49: Pulling fs layer 349669550f49: Verifying Checksum 349669550f49: Download complete 349669550f49: Pull complete Digest: sha256:ccc35b36a163a8712e70d7980de479c0d5916989fb4562cb6978a518cff82dc4 Trying to pull repository registry:5000/kubevirt/alpine-registry-disk-demo ... devel: Pulling from registry:5000/kubevirt/alpine-registry-disk-demo 2115d46e7396: Already exists 07ea0109ea1b: Already exists c4272834948a: Already exists 8bc4b617e33b: Pulling fs layer 8bc4b617e33b: Verifying Checksum 8bc4b617e33b: Download complete 8bc4b617e33b: Pull complete Digest: sha256:8c47587932661803fac596f5b13f283cf2e0c7a841e9c10f57955ddcdd9602f5 Trying to pull repository registry:5000/kubevirt/subresource-access-test ... devel: Pulling from registry:5000/kubevirt/subresource-access-test 2176639d844b: Already exists 555998d3afb2: Pulling fs layer aa4c4c4a31ba: Pulling fs layer 555998d3afb2: Verifying Checksum 555998d3afb2: Download complete aa4c4c4a31ba: Verifying Checksum aa4c4c4a31ba: Download complete 555998d3afb2: Pull complete aa4c4c4a31ba: Pull complete Digest: sha256:fe4b7dad9dc54c835e6d5d67fa2cfb4bfbb55dc4de426d943b4116bb8c0be23c Trying to pull repository registry:5000/kubevirt/winrmcli ... devel: Pulling from registry:5000/kubevirt/winrmcli 2176639d844b: Already exists 494ac3d2e337: Pulling fs layer fa95949b860a: Pulling fs layer 97a795c22b24: Pulling fs layer 97a795c22b24: Verifying Checksum 97a795c22b24: Download complete 494ac3d2e337: Verifying Checksum 494ac3d2e337: Download complete fa95949b860a: Verifying Checksum fa95949b860a: Download complete 494ac3d2e337: Pull complete fa95949b860a: Pull complete 97a795c22b24: Pull complete Digest: sha256:db2b8707d3ad61bf0935b493c2a1342531deb3d902ccf91b9dd77789693be68d 2018/04/09 15:53:09 Waiting for host: 192.168.66.102:22 2018/04/09 15:53:09 Connected to tcp://192.168.66.102:22 Done ./cluster/clean.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0-alpha.4 ++ provider_prefix=kubevirt-functional-tests-openshift-release0 ++ job_prefix=kubevirt-functional-tests-openshift-release0 + source cluster/os-3.9.0-alpha.4/provider.sh ++ set -e ++ image=os-3.9@sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8 ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0-alpha.4 ++ source hack/config-default.sh source hack/config-os-3.9.0-alpha.4.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0-alpha.4.sh ++ source hack/config-provider-os-3.9.0-alpha.4.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-alpha.4/.kubeconfig +++ docker_prefix=localhost:32880/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Cleaning up ...' Cleaning up ... + _kubectl delete ds -l kubevirt.io -n kube-system --cascade=false --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=libvirt --force --grace-period 0 No resources found + _kubectl delete pods -n kube-system -l=kubevirt.io=virt-handler --force --grace-period 0 No resources found + namespaces=(default ${namespace}) + for i in '${namespaces[@]}' + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete deployment -l kubevirt.io No resources found + _kubectl -n default delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete rs -l kubevirt.io No resources found + _kubectl -n default delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete services -l kubevirt.io No resources found + _kubectl -n default delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete apiservices -l kubevirt.io No resources found + _kubectl -n default delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete secrets -l kubevirt.io No resources found + _kubectl -n default delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete pv -l kubevirt.io No resources found + _kubectl -n default delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete pvc -l kubevirt.io No resources found + _kubectl -n default delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete ds -l kubevirt.io No resources found + _kubectl -n default delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n default delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete pods -l kubevirt.io No resources found + _kubectl -n default delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n default delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete rolebinding -l kubevirt.io No resources found + _kubectl -n default delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete roles -l kubevirt.io No resources found + _kubectl -n default delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete clusterroles -l kubevirt.io No resources found + _kubectl -n default delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n default delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n default get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ cluster/os-3.9.0-alpha.4/.kubectl -n default get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + for i in '${namespaces[@]}' + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete deployment -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete deployment -l kubevirt.io No resources found + _kubectl -n kube-system delete rs -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete rs -l kubevirt.io No resources found + _kubectl -n kube-system delete services -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete services -l kubevirt.io No resources found + _kubectl -n kube-system delete apiservices -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete apiservices -l kubevirt.io No resources found + _kubectl -n kube-system delete secrets -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete secrets -l kubevirt.io No resources found + _kubectl -n kube-system delete pv -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete pv -l kubevirt.io No resources found + _kubectl -n kube-system delete pvc -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete pvc -l kubevirt.io No resources found + _kubectl -n kube-system delete ds -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete ds -l kubevirt.io No resources found + _kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete customresourcedefinitions -l kubevirt.io No resources found + _kubectl -n kube-system delete pods -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete pods -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterrolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete clusterrolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete rolebinding -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete rolebinding -l kubevirt.io No resources found + _kubectl -n kube-system delete roles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete roles -l kubevirt.io No resources found + _kubectl -n kube-system delete clusterroles -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete clusterroles -l kubevirt.io No resources found + _kubectl -n kube-system delete serviceaccounts -l kubevirt.io + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl -n kube-system delete serviceaccounts -l kubevirt.io No resources found ++ _kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io ++ wc -l ++ export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig ++ cluster/os-3.9.0-alpha.4/.kubectl -n kube-system get crd offlinevirtualmachines.kubevirt.io Error from server (NotFound): customresourcedefinitions.apiextensions.k8s.io "offlinevirtualmachines.kubevirt.io" not found + '[' 0 -gt 0 ']' + sleep 2 + echo Done Done ./cluster/deploy.sh + source hack/common.sh ++++ dirname 'hack/common.sh[0]' +++ cd hack/../ +++ pwd ++ KUBEVIRT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt ++ OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out ++ VENDOR_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/vendor ++ CMD_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/cmd ++ TESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/tests ++ APIDOCS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/apidocs ++ MANIFESTS_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests ++ PYTHON_CLIENT_OUT_DIR=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/client-python ++ PROVIDER=os-3.9.0-alpha.4 ++ provider_prefix=kubevirt-functional-tests-openshift-release0 ++ job_prefix=kubevirt-functional-tests-openshift-release0 + source cluster/os-3.9.0-alpha.4/provider.sh ++ set -e ++ image=os-3.9@sha256:6d5e570e78981f7faf2deb6df2bb2f887a6ed2dd458b4178897ebe877316f8f8 ++ source cluster/ephemeral-provider-common.sh +++ set -e +++ _cli='docker run --privileged --rm -v /var/run/docker.sock:/var/run/docker.sock kubevirtci/cli@sha256:b0023d1863338ef04fa0b8a8ee5956ae08616200d89ffd2e230668ea3deeaff4' + source hack/config.sh ++ unset binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig manifest_docker_prefix namespace ++ PROVIDER=os-3.9.0-alpha.4 ++ source hack/config-default.sh source hack/config-os-3.9.0-alpha.4.sh +++ binaries='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virtctl cmd/fake-qemu-process cmd/virt-api cmd/subresource-access-test' +++ docker_images='cmd/virt-controller cmd/virt-launcher cmd/virt-handler cmd/virt-api images/iscsi-demo-target-tgtd images/vm-killer cmd/registry-disk-v1alpha images/cirros-registry-disk-demo images/fedora-cloud-registry-disk-demo images/alpine-registry-disk-demo cmd/subresource-access-test images/winrmcli' +++ docker_prefix=kubevirt +++ docker_tag=latest +++ master_ip=192.168.200.2 +++ network_provider=flannel +++ kubeconfig=cluster/vagrant/.kubeconfig +++ namespace=kube-system ++ test -f hack/config-provider-os-3.9.0-alpha.4.sh ++ source hack/config-provider-os-3.9.0-alpha.4.sh +++ master_ip=127.0.0.1 +++ docker_tag=devel +++ kubeconfig=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/cluster/os-3.9.0-alpha.4/.kubeconfig +++ docker_prefix=localhost:32880/kubevirt +++ manifest_docker_prefix=registry:5000/kubevirt ++ test -f hack/config-local.sh ++ export binaries docker_images docker_prefix docker_tag manifest_templates master_ip network_provider kubeconfig namespace + echo 'Deploying ...' Deploying ... + [[ -z openshift-release ]] + [[ openshift-release =~ .*-dev ]] + [[ openshift-release =~ .*-release ]] + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/demo-content.yaml =~ .*demo.* ]] + continue + for manifest in '${MANIFESTS_OUT_DIR}/release/*' + [[ /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml =~ .*demo.* ]] + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/release/kubevirt.yaml serviceaccount "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver" created clusterrolebinding "kubevirt-apiserver-auth-delegator" created rolebinding "kubevirt-apiserver" created role "kubevirt-apiserver" created clusterrole "kubevirt-apiserver" created clusterrole "kubevirt-controller" created serviceaccount "kubevirt-controller" created serviceaccount "kubevirt-privileged" created clusterrolebinding "kubevirt-controller" created clusterrolebinding "kubevirt-controller-cluster-admin" created clusterrolebinding "kubevirt-privileged-cluster-admin" created service "virt-api" created deployment "virt-api" created deployment "virt-controller" created daemonset "virt-handler" created customresourcedefinition "virtualmachines.kubevirt.io" created customresourcedefinition "virtualmachinereplicasets.kubevirt.io" created customresourcedefinition "virtualmachinepresets.kubevirt.io" created customresourcedefinition "offlinevirtualmachines.kubevirt.io" created + _kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl create -f /var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/go/src/kubevirt.io/kubevirt/_out/manifests/testing -R persistentvolumeclaim "disk-alpine" created persistentvolume "iscsi-disk-alpine" created daemonset "iscsi-demo-target-tgtd" created serviceaccount "kubevirt-testing" created clusterrolebinding "kubevirt-testing-cluster-admin" created + '[' os-3.9.0-alpha.4 = vagrant-openshift ']' + '[' os-3.9.0-alpha.4 = os-3.9.0-alpha.4 ']' + _kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-controller -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-controller"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-testing -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-testing"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-privileged -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-privileged"] + _kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged -z kubevirt-apiserver -n kube-system scc "privileged" added to: ["system:serviceaccount:kube-system:kubevirt-apiserver"] + _kubectl adm policy add-scc-to-user privileged admin + export KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + KUBECONFIG=cluster/os-3.9.0-alpha.4/.kubeconfig + cluster/os-3.9.0-alpha.4/.kubectl adm policy add-scc-to-user privileged admin scc "privileged" added to: ["admin"] + echo Done Done ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n 'iscsi-demo-target-tgtd-5dfbz 0/1 ContainerCreating 0 4s iscsi-demo-target-tgtd-n5f4c 0/1 ContainerCreating 0 4s virt-api-fd96f94b5-2pn9v 0/1 ContainerCreating 0 7s virt-api-fd96f94b5-qcqm5 0/1 ContainerCreating 0 7s virt-controller-5f7c946cc4-qd8wt 0/1 ContainerCreating 0 7s virt-controller-5f7c946cc4-rjfqz 0/1 ContainerCreating 0 7s virt-handler-jq2mx 0/1 Pending 0 1s virt-handler-xtnkb 0/1 Pending 0 2s' ']' + echo 'Waiting for kubevirt pods to enter the Running state ...' Waiting for kubevirt pods to enter the Running state ... + kubectl get pods -n kube-system --no-headers + grep -v Running + cluster/kubectl.sh get pods -n kube-system --no-headers iscsi-demo-target-tgtd-5dfbz 0/1 ContainerCreating 0 6s virt-api-fd96f94b5-qcqm5 0/1 ContainerCreating 0 9s virt-controller-5f7c946cc4-qd8wt 0/1 ContainerCreating 0 9s virt-controller-5f7c946cc4-rjfqz 0/1 ContainerCreating 0 9s virt-handler-jq2mx 0/1 ContainerCreating 0 3s virt-handler-xtnkb 0/1 ContainerCreating 0 4s + sleep 10 ++ kubectl get pods -n kube-system --no-headers ++ grep -v Running ++ cluster/kubectl.sh get pods -n kube-system --no-headers + '[' -n '' ']' ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' + '[' -n 'false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers false iscsi-demo-target-tgtd-5dfbz false iscsi-demo-target-tgtd-n5f4c + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n 'false iscsi-demo-target-tgtd-5dfbz' ']' + echo 'Waiting for KubeVirt containers to become ready ...' Waiting for KubeVirt containers to become ready ... + kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + awk '!/virt-controller/ && /false/' + cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + sleep 10 ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '!/virt-controller/ && /false/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers + '[' -n '' ']' ++ kubectl get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ awk '/virt-controller/ && /true/' ++ cluster/kubectl.sh get pods -n kube-system '-ocustom-columns=status:status.containerStatuses[*].ready,metadata:metadata.name' --no-headers ++ wc -l + '[' 1 -lt 1 ']' + kubectl get pods -n kube-system + cluster/kubectl.sh get pods -n kube-system NAME READY STATUS RESTARTS AGE iscsi-demo-target-tgtd-5dfbz 1/1 Running 1 1m iscsi-demo-target-tgtd-n5f4c 1/1 Running 1 1m virt-api-fd96f94b5-2pn9v 1/1 Running 0 1m virt-api-fd96f94b5-qcqm5 1/1 Running 0 1m virt-controller-5f7c946cc4-qd8wt 1/1 Running 0 1m virt-controller-5f7c946cc4-rjfqz 0/1 Running 0 1m virt-handler-jq2mx 1/1 Running 0 1m virt-handler-xtnkb 1/1 Running 0 1m + kubectl version + cluster/kubectl.sh version oc v3.9.0-alpha.4+9ab7a71 kubernetes v1.9.1+a0ce1bc657 features: Basic-Auth GSSAPI Kerberos SPNEGO Server https://127.0.0.1:32877 openshift v3.9.0-alpha.4+9ab7a71 kubernetes v1.9.1+a0ce1bc657 + ginko_params='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/junit.xml' + [[ -d /home/nfs/images/windows2016 ]] + FUNC_TEST_ARGS='--ginkgo.noColor --junit-output=/var/lib/swarm/workspace/kubevirt-functional-tests-openshift-release/junit.xml' + make functest hack/dockerized "hack/build-func-tests.sh" sha256:b5b4d597a272b0a23ca6310ef9df2eff2af3e39eb9e825fc156c0ce4582d5ef1 go version go1.9.2 linux/amd64 rsync: read error: Connection reset by peer (104) rsync error: error in rsync protocol data stream (code 12) at io.c(764) [sender=3.0.9] Waiting for rsyncd to be ready skipping directory . go version go1.9.2 linux/amd64 Compiling tests... compiled tests.test 74f6380cee8f1d049577e12b3f119e3b35820ab491d67f155a68004c2c0f7f05 74f6380cee8f1d049577e12b3f119e3b35820ab491d67f155a68004c2c0f7f05 hack/functests.sh Running Suite: Tests Suite ========================== Random Seed: 1523289564 Will run 76 of 76 specs volumedisk0 compute • [SLOW TEST:45.634 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:39 VM definition /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:50 with 3 CPU cores /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:51 should report 3 cpu cores under guest OS /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:57 ------------------------------ • [SLOW TEST:45.356 seconds] Configurations /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:39 New VM with all supported drives /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:109 should have all the device nodes /root/go/src/kubevirt.io/kubevirt/tests/vm_configuration_test.go:132 ------------------------------ • [SLOW TEST:7.493 seconds] Subresource Api /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:37 Rbac Authorization /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:48 with correct permissions /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:49 should be allowed to access subresource endpoint /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:50 ------------------------------ • [SLOW TEST:8.011 seconds] Subresource Api /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:37 Rbac Authorization /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:48 Without permissions /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:54 should not be able to access subresource endpoint /root/go/src/kubevirt.io/kubevirt/tests/subresource_api_test.go:55 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.008 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:54 should success to start a vm [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:128 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1046 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.010 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:54 should success to stop a running vm [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:134 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1046 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.009 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:54 with winrm connection [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:145 should have correct UUID /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:187 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1046 ------------------------------ S [SKIPPING] in Spec Setup (BeforeEach) [0.011 seconds] Windows VM /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:54 with winrm connection [BeforeEach] /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:145 should have pod IP /root/go/src/kubevirt.io/kubevirt/tests/windows_test.go:203 Skip Windows tests that requires PVC disk-windows /root/go/src/kubevirt.io/kubevirt/tests/utils.go:1046 ------------------------------ • [SLOW TEST:39.751 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 should have cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:81 ------------------------------ • [SLOW TEST:95.753 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 with cloudInitNoCloud userDataBase64 source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:80 with injected ssh-key /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:92 should have ssh-key under authorized keys /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:93 ------------------------------ • [SLOW TEST:45.700 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 with cloudInitNoCloud userData source /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:116 should process provided cloud-init data /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:117 ------------------------------ • [SLOW TEST:43.887 seconds] CloudInit UserData /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:46 A new VM /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:79 should take user-data from k8s secret /root/go/src/kubevirt.io/kubevirt/tests/vm_userdata_test.go:160 ------------------------------ Received interrupt. Emitting contents of GinkgoWriter... --------------------------------------------------------- STEP: Starting a VM --------------------------------------------------------- Received interrupt. Running AfterSuite... ^C again to terminate immediately ++ make cluster-down