[Rdo-list] OverCloud deploy fails with error "No valid host was found"
Sasha Chuzhoy
sasha at redhat.com
Fri Oct 16 15:44:49 UTC 2015
Hi Esra,
if the undercloud nodes are UP - you can login with: ssh heat-admin@<IP>
You can see the IP of the nodes with: "nova list".
BTW,
What do you see if you run "sudo systemctl|grep ironic" on the undercloud?
Best regards,
Sasha Chuzhoy.
----- Original Message -----
> From: "Esra Celik" <celik.esra at tubitak.gov.tr>
> To: "Sasha Chuzhoy" <sasha at redhat.com>
> Cc: "Marius Cornea" <mcornea at redhat.com>, rdo-list at redhat.com
> Sent: Friday, October 16, 2015 1:40:16 AM
> Subject: Re: [Rdo-list] OverCloud deploy fails with error "No valid host was found"
>
> Hi Sasha,
>
> I have 3 nodes, 1 Undercloud, 1 Overcloud-Controller, 1 Overcloud-Compute
>
> This is my undercloud.conf file:
>
> image_path = .
> local_ip = 192.0.2.1/24
> local_interface = em2
> masquerade_network = 192.0.2.0/24
> dhcp_start = 192.0.2.5
> dhcp_end = 192.0.2.24
> network_cidr = 192.0.2.0/24
> network_gateway = 192.0.2.1
> inspection_interface = br-ctlplane
> inspection_iprange = 192.0.2.100,192.0.2.120
> inspection_runbench = false
> undercloud_debug = true
> enable_tuskar = false
> enable_tempest = false
>
> IP configuration for the Undercloud is as follows:
>
> stack at undercloud ~]$ ip addr
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
> inet 127.0.0.1/8 scope host lo
> valid_lft forever preferred_lft forever
> inet6 ::1/128 scope host
> valid_lft forever preferred_lft forever
> 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen
> 1000
> link/ether 08:9e:01:50:8a:21 brd ff:ff:ff:ff:ff:ff
> inet 10.1.34.81/24 brd 10.1.34.255 scope global em1
> valid_lft forever preferred_lft forever
> inet6 fe80::a9e:1ff:fe50:8a21/64 scope link
> valid_lft forever preferred_lft forever
> 3: em2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master ovs-system
> state UP qlen 1000
> link/ether 08:9e:01:50:8a:22 brd ff:ff:ff:ff:ff:ff
> 4: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> link/ether 9a:a8:0f:ec:42:15 brd ff:ff:ff:ff:ff:ff
> 5: br-ctlplane: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
> state UNKNOWN
> link/ether 08:9e:01:50:8a:22 brd ff:ff:ff:ff:ff:ff
> inet 192.0.2.1/24 brd 192.0.2.255 scope global br-ctlplane
> valid_lft forever preferred_lft forever
> inet6 fe80::a9e:1ff:fe50:8a22/64 scope link
> valid_lft forever preferred_lft forever
> 6: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN
> link/ether fa:85:ac:92:f5:41 brd ff:ff:ff:ff:ff:ff
>
> And I attached two screenshots showing the boot stage for overcloud nodes
>
> Is there a way to login the overcloud nodes to see their IP configuration?
>
> Thanks
>
> Esra ÇELİK
> TÜBİTAK BİLGEM
> www.bilgem.tubitak.gov.tr
> celik.esra at tubitak.gov.tr
>
> ----- Orijinal Mesaj -----
>
> > Kimden: "Sasha Chuzhoy" <sasha at redhat.com>
> > Kime: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > Kk: "Marius Cornea" <mcornea at redhat.com>, rdo-list at redhat.com
> > Gönderilenler: 15 Ekim Perşembe 2015 16:58:41
> > Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid host was
> > found"
>
> > Just my 2 cents.
> > Did you make sure that all the registered nodes are configured to boot off
> > the right NIC first?
> > Can you watch the console and see what happens on the problematic nodes
> > upon
> > boot?
>
> > Best regards,
> > Sasha Chuzhoy.
>
> > ----- Original Message -----
> > > From: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > > To: "Marius Cornea" <mcornea at redhat.com>
> > > Cc: rdo-list at redhat.com
> > > Sent: Thursday, October 15, 2015 4:40:46 AM
> > > Subject: Re: [Rdo-list] OverCloud deploy fails with error "No valid host
> > > was found"
> > >
> > >
> > > Sorry for the late reply
> > >
> > > ironic node-show results are below. I have my nodes power on after
> > > introspection bulk start. And I get the following warning
> > > Introspection didn't finish for nodes
> > > 5b28998f-4dc8-42aa-8a51-521e20b1e5ed,6f35ac24-135d-4b99-8a24-fa2b731bd218
> > >
> > > Doesn't seem to be the same issue with
> > > https://www.redhat.com/archives/rdo-list/2015-October/msg00174.html
> > >
> > >
> > >
> > >
> > > [stack at undercloud ~]$ ironic node-list
> > > +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > | UUID | Name | Instance UUID | Power State | Provisioning State |
> > > | Maintenance |
> > > +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > | 5b28998f-4dc8-42aa-8a51-521e20b1e5ed | None | None | power on |
> > > | available
> > > | |
> > > | False |
> > > | 6f35ac24-135d-4b99-8a24-fa2b731bd218 | None | None | power on |
> > > | available
> > > | |
> > > | False |
> > > +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > >
> > >
> > > [stack at undercloud ~]$ ironic node-show
> > > 5b28998f-4dc8-42aa-8a51-521e20b1e5ed
> > > +------------------------+-------------------------------------------------------------------------+
> > > | Property | Value |
> > > +------------------------+-------------------------------------------------------------------------+
> > > | target_power_state | None |
> > > | extra | {} |
> > > | last_error | None |
> > > | updated_at | 2015-10-15T08:26:42+00:00 |
> > > | maintenance_reason | None |
> > > | provision_state | available |
> > > | clean_step | {} |
> > > | uuid | 5b28998f-4dc8-42aa-8a51-521e20b1e5ed |
> > > | console_enabled | False |
> > > | target_provision_state | None |
> > > | provision_updated_at | 2015-10-15T08:26:42+00:00 |
> > > | maintenance | False |
> > > | inspection_started_at | None |
> > > | inspection_finished_at | None |
> > > | power_state | power on |
> > > | driver | pxe_ipmitool |
> > > | reservation | None |
> > > | properties | {u'memory_mb': u'8192', u'cpu_arch': u'x86_64',
> > > | u'local_gb':
> > > | u'10', |
> > > | | u'cpus': u'4', u'capabilities': u'boot_option:local'} |
> > > | instance_uuid | None |
> > > | name | None |
> > > | driver_info | {u'ipmi_password': u'******', u'ipmi_address':
> > > | u'192.168.0.18', |
> > > | | u'ipmi_username': u'root', u'deploy_kernel':
> > > | | u'49a2c8d4-a283-4bdf-8d6f-
> > > | | |
> > > | | e83ae28da047', u'deploy_ramdisk': u'3db3dbed- |
> > > | | 0d88-4632-af98-8defb05ca6e2'} |
> > > | created_at | 2015-10-15T07:49:08+00:00 |
> > > | driver_internal_info | {u'clean_steps': None} |
> > > | chassis_uuid | |
> > > | instance_info | {} |
> > > +------------------------+-------------------------------------------------------------------------+
> > >
> > >
> > > [stack at undercloud ~]$ ironic node-show
> > > 6f35ac24-135d-4b99-8a24-fa2b731bd218
> > > +------------------------+-------------------------------------------------------------------------+
> > > | Property | Value |
> > > +------------------------+-------------------------------------------------------------------------+
> > > | target_power_state | None |
> > > | extra | {} |
> > > | last_error | None |
> > > | updated_at | 2015-10-15T08:26:42+00:00 |
> > > | maintenance_reason | None |
> > > | provision_state | available |
> > > | clean_step | {} |
> > > | uuid | 6f35ac24-135d-4b99-8a24-fa2b731bd218 |
> > > | console_enabled | False |
> > > | target_provision_state | None |
> > > | provision_updated_at | 2015-10-15T08:26:42+00:00 |
> > > | maintenance | False |
> > > | inspection_started_at | None |
> > > | inspection_finished_at | None |
> > > | power_state | power on |
> > > | driver | pxe_ipmitool |
> > > | reservation | None |
> > > | properties | {u'memory_mb': u'8192', u'cpu_arch': u'x86_64',
> > > | u'local_gb':
> > > | u'100', |
> > > | | u'cpus': u'4', u'capabilities': u'boot_option:local'} |
> > > | instance_uuid | None |
> > > | name | None |
> > > | driver_info | {u'ipmi_password': u'******', u'ipmi_address':
> > > | u'192.168.0.19', |
> > > | | u'ipmi_username': u'root', u'deploy_kernel':
> > > | | u'49a2c8d4-a283-4bdf-8d6f-
> > > | | |
> > > | | e83ae28da047', u'deploy_ramdisk': u'3db3dbed- |
> > > | | 0d88-4632-af98-8defb05ca6e2'} |
> > > | created_at | 2015-10-15T07:49:08+00:00 |
> > > | driver_internal_info | {u'clean_steps': None} |
> > > | chassis_uuid | |
> > > | instance_info | {} |
> > > +------------------------+-------------------------------------------------------------------------+
> > > [stack at undercloud ~]$
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > And below I added my history for the stack user. I don't think I am doing
> > > something other than
> > > https://repos.fedorapeople.org/repos/openstack-m/rdo-manager-docs/liberty
> > > doc
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > 1 vi instackenv.json
> > > 2 sudo yum -y install epel-release
> > > 3 sudo curl -o /etc/yum.repos.d/delorean.repo
> > > http://trunk.rdoproject.org/centos7-liberty/current-passed-ci/delorean.repo
> > > 4 sudo curl -o /etc/yum.repos.d/delorean-current.repo
> > > http://trunk.rdoproject.org/centos7-liberty/current/delorean.repo
> > > 5 sudo sed -i 's/\[delorean\]/\[delorean-current\]/'
> > > /etc/yum.repos.d/delorean-current.repo
> > > 6 sudo /bin/bash -c "cat <<EOF>>/etc/yum.repos.d/delorean-current.repo
> > >
> > > includepkgs=diskimage-builder,openstack-heat,instack,instack-undercloud,openstack-ironic,openstack-ironic-inspector,os-cloud-config,os-net-config,python-ironic-inspector-client,python-tripleoclient,tripleo-common,openstack-tripleo-heat-templates,openstack-tripleo-image-elements,openstack-tuskar-ui-extras,openstack-puppet-modules
> > > EOF"
> > > 7 sudo curl -o /etc/yum.repos.d/delorean-deps.repo
> > > http://trunk.rdoproject.org/centos7-liberty/delorean-deps.repo
> > > 8 sudo yum -y install yum-plugin-priorities
> > > 9 sudo yum install -y python-tripleoclient
> > > 10 cp /usr/share/instack-undercloud/undercloud.conf.sample
> > > ~/undercloud.conf
> > > 11 vi undercloud.conf
> > > 12 export DIB_INSTALLTYPE_puppet_modules=source
> > > 13 openstack undercloud install
> > > 14 source stackrc
> > > 15 export NODE_DIST=centos7
> > > 16 export DIB_YUM_REPO_CONF="/etc/yum.repos.d/delorean.repo
> > > /etc/yum.repos.d/delorean-deps.repo"
> > > 17 export DIB_INSTALLTYPE_puppet_modules=source
> > > 18 openstack overcloud image build --all
> > > 19 ls
> > > 20 openstack overcloud image upload
> > > 21 openstack baremetal import --json instackenv.json
> > > 22 openstack baremetal configure boot
> > > 23 ironic node-list
> > > 24 openstack baremetal introspection bulk start
> > > 25 ironic node-list
> > > 26 ironic node-show 5b28998f-4dc8-42aa-8a51-521e20b1e5ed
> > > 27 ironic node-show 6f35ac24-135d-4b99-8a24-fa2b731bd218
> > > 28 history
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > Thanks
> > >
> > >
> > >
> > > Esra ÇELİK
> > > TÜBİTAK BİLGEM
> > > www.bilgem.tubitak.gov.tr
> > > celik.esra at tubitak.gov.tr
> > >
> > >
> > > Kimden: "Marius Cornea" <mcornea at redhat.com>
> > > Kime: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > > Kk: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com
> > > Gönderilenler: 14 Ekim Çarşamba 2015 19:40:07
> > > Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid host was
> > > found"
> > >
> > > Can you do ironic node-show for your ironic nodes and post the results?
> > > Also
> > > check the following suggestion if you're experiencing the same issue:
> > > https://www.redhat.com/archives/rdo-list/2015-October/msg00174.html
> > >
> > > ----- Original Message -----
> > > > From: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > > > To: "Marius Cornea" <mcornea at redhat.com>
> > > > Cc: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com
> > > > Sent: Wednesday, October 14, 2015 3:22:20 PM
> > > > Subject: Re: [Rdo-list] OverCloud deploy fails with error "No valid
> > > > host
> > > > was found"
> > > >
> > > >
> > > >
> > > > Well in the early stage of the introspection I can see Client IP of
> > > > nodes
> > > > (screenshot attached). But then I see continuous ironic-python-agent
> > > > errors
> > > > (screenshot-2 attached). Errors repeat after time out.. And the nodes
> > > > are
> > > > not powered off.
> > > >
> > > > Seems like I am stuck in introspection stage..
> > > >
> > > > I can use ipmitool command to successfully power on/off the nodes
> > > >
> > > >
> > > >
> > > > [stack at undercloud ~]$ ipmitool -I lanplus -H 192.168.0.19 -L
> > > > ADMINISTRATOR
> > > > -U
> > > > root -R 3 -N 5 -P <password> power status
> > > > Chassis Power is on
> > > >
> > > >
> > > > [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P
> > > > <password> chassis power status
> > > > Chassis Power is on
> > > > [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P
> > > > <password> chassis power off
> > > > Chassis Power Control: Down/Off
> > > > [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P
> > > > <password> chassis power status
> > > > Chassis Power is off
> > > > [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P
> > > > <password> chassis power on
> > > > Chassis Power Control: Up/On
> > > > [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P
> > > > <password> chassis power status
> > > > Chassis Power is on
> > > >
> > > >
> > > > Esra ÇELİK
> > > > TÜBİTAK BİLGEM
> > > > www.bilgem.tubitak.gov.tr
> > > > celik.esra at tubitak.gov.tr
> > > >
> > > >
> > > > ----- Orijinal Mesaj -----
> > > >
> > > > Kimden: "Marius Cornea" <mcornea at redhat.com>
> > > > Kime: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > > > Kk: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com
> > > > Gönderilenler: 14 Ekim Çarşamba 2015 14:59:30
> > > > Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid host
> > > > was
> > > > found"
> > > >
> > > >
> > > > ----- Original Message -----
> > > > > From: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > > > > To: "Marius Cornea" <mcornea at redhat.com>
> > > > > Cc: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com
> > > > > Sent: Wednesday, October 14, 2015 10:49:01 AM
> > > > > Subject: Re: [Rdo-list] OverCloud deploy fails with error "No valid
> > > > > host
> > > > > was found"
> > > > >
> > > > >
> > > > > Well today I started with re-installing the OS and nothing seems
> > > > > wrong
> > > > > with
> > > > > undercloud installation, then;
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > I see an error during image build
> > > > >
> > > > >
> > > > > [stack at undercloud ~]$ openstack overcloud image build --all
> > > > > ...
> > > > > a lot of log
> > > > > ...
> > > > > ++ cat /etc/dib_dracut_drivers
> > > > > + dracut -N --install ' curl partprobe lsblk targetcli tail head awk
> > > > > ifconfig
> > > > > cut expr route ping nc wget tftp grep' --kernel-cmdline 'rd.shell
> > > > > rd.debug
> > > > > rd.neednet=1 rd.driver.pre=ahci' --include
> > > > > /var/tmp/image.YVhwuArQ/mnt/
> > > > > /
> > > > > --kver 3.10.0-229.14.1.el7.x86_64 --add-drivers ' virtio virtio_net
> > > > > virtio_blk target_core_mod iscsi_target_mod target_core_iblock
> > > > > target_core_file target_core_pscsi configfs' -o 'dash plymouth'
> > > > > /tmp/ramdisk
> > > > > cat: write error: Broken pipe
> > > > > + cp /boot/vmlinuz-3.10.0-229.14.1.el7.x86_64 /tmp/kernel
> > > > > + chmod o+r /tmp/kernel
> > > > > + trap EXIT
> > > > > + target_tag=99-build-dracut-ramdisk
> > > > > + date +%s.%N
> > > > > + output '99-build-dracut-ramdisk completed'
> > > > > ...
> > > > > a lot of log
> > > > > ...
> > > >
> > > > You can ignore that afaik, if you end up having all the required images
> > > > it
> > > > should be ok.
> > > >
> > > > >
> > > > > Then, during introspection stage I see ironic-python-agent errors on
> > > > > nodes
> > > > > (screenshot attached) and the following warnings
> > > > >
> > > >
> > > > That looks odd. Is it showing up in the early stage of the
> > > > introspection?
> > > > At
> > > > some point it should receive an address by DHCP and the Network is
> > > > unreachable error should disappear. Does the introspection complete and
> > > > the
> > > > nodes are turned off?
> > > >
> > > > >
> > > > >
> > > > > [root at localhost ~]# journalctl -fl -u
> > > > > openstack-ironic-conductor.service
> > > > > |
> > > > > grep -i "warning\|error"
> > > > > Oct 14 10:30:12 undercloud.rdo ironic-conductor[619]: 2015-10-14
> > > > > 10:30:12.119
> > > > > 619 WARNING oslo_config.cfg [req-eccf8cb5-6e93-4d8f-9a05-0e8c8d2aab7b
> > > > > ]
> > > > > Option "http_url" from group "pxe" is deprecated. Use option
> > > > > "http_url"
> > > > > from
> > > > > group "deploy".
> > > > > Oct 14 10:30:12 undercloud.rdo ironic-conductor[619]: 2015-10-14
> > > > > 10:30:12.119
> > > > > 619 WARNING oslo_config.cfg [req-eccf8cb5-6e93-4d8f-9a05-0e8c8d2aab7b
> > > > > ]
> > > > > Option "http_root" from group "pxe" is deprecated. Use option
> > > > > "http_root"
> > > > > from group "deploy".
> > > > >
> > > > >
> > > > > Before deployment ironic node-list:
> > > > >
> > > >
> > > > This is odd too as I'm expecting the nodes to be powered off before
> > > > running
> > > > deployment.
> > > >
> > > > >
> > > > >
> > > > > [stack at undercloud ~]$ ironic node-list
> > > > > +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > | UUID | Name | Instance UUID | Power State | Provisioning State |
> > > > > | Maintenance |
> > > > > +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > | acfc1bb4-469d-479a-af70-c0bdd669b32d | None | None | power on |
> > > > > | available
> > > > > | |
> > > > > | False |
> > > > > | b5811c06-d5d1-41f1-87b3-2fd55ae63553 | None | None | power on |
> > > > > | available
> > > > > | |
> > > > > | False |
> > > > > +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > >
> > > > > During deployment I get following errors
> > > > >
> > > > > [root at localhost ~]# journalctl -fl -u
> > > > > openstack-ironic-conductor.service
> > > > > |
> > > > > grep -i "warning\|error"
> > > > > Oct 14 11:29:01 undercloud.rdo ironic-conductor[619]: 2015-10-14
> > > > > 11:29:01.739
> > > > > 619 ERROR ironic.drivers.modules.ipmitool [-] IPMI Error while
> > > > > attempting
> > > > > "ipmitool -I lanplus -H 192.168.0.19 -L ADMINISTRATOR -U root -R 3 -N
> > > > > 5
> > > > > -f
> > > > > /tmp/tmpSCKHIv power status"for node
> > > > > b5811c06-d5d1-41f1-87b3-2fd55ae63553.
> > > > > Error: Unexpected error while running command.
> > > > > Oct 14 11:29:01 undercloud.rdo ironic-conductor[619]: 2015-10-14
> > > > > 11:29:01.739
> > > > > 619 WARNING ironic.drivers.modules.ipmitool [-] IPMI power status
> > > > > failed
> > > > > for
> > > > > node b5811c06-d5d1-41f1-87b3-2fd55ae63553 with error: Unexpected
> > > > > error
> > > > > while
> > > > > running command.
> > > > > Oct 14 11:29:01 undercloud.rdo ironic-conductor[619]: 2015-10-14
> > > > > 11:29:01.740
> > > > > 619 WARNING ironic.conductor.manager [-] During sync_power_state,
> > > > > could
> > > > > not
> > > > > get power state for node b5811c06-d5d1-41f1-87b3-2fd55ae63553,
> > > > > attempt
> > > > > 1
> > > > > of
> > > > > 3. Error: IPMI call failed: power status..
> > > > >
> > > >
> > > > This looks like an ipmi error, can you try to manually run commands
> > > > using
> > > > the
> > > > ipmitool and see if you get any success? It's also worth filing a bug
> > > > with
> > > > details such as the ipmitool version, server model, drac firmware
> > > > version.
> > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > Thanks a lot
> > > > >
> > > > >
> > > > >
> > > > > ----- Orijinal Mesaj -----
> > > > >
> > > > > Kimden: "Marius Cornea" <mcornea at redhat.com>
> > > > > Kime: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > > > > Kk: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com
> > > > > Gönderilenler: 13 Ekim Salı 2015 21:16:14
> > > > > Konu: Re: Yan: Re: [Rdo-list] OverCloud deploy fails with error "No
> > > > > valid
> > > > > host was found"
> > > > >
> > > > >
> > > > > ----- Original Message -----
> > > > > > From: "Esra Celik" <celik.esra at tubitak.gov.tr>
> > > > > > To: "Marius Cornea" <mcornea at redhat.com>
> > > > > > Cc: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com
> > > > > > Sent: Tuesday, October 13, 2015 5:02:09 PM
> > > > > > Subject: Yan: Re: [Rdo-list] OverCloud deploy fails with error "No
> > > > > > valid
> > > > > > host was found"
> > > > > >
> > > > > > During deployment they are powering on and deploying the images. I
> > > > > > see
> > > > > > lot
> > > > > > of
> > > > > > connection error messages about ironic-python-agent but ignore them
> > > > > > as
> > > > > > mentioned here
> > > > > > (https://www.redhat.com/archives/rdo-list/2015-October/msg00052.html)
> > > > >
> > > > > That was referring to the introspection stage. From what I can tell
> > > > > you
> > > > > are
> > > > > experiencing issues during deployment as it fails to provision the
> > > > > nova
> > > > > instances, can you check if during that stage the nodes get powered
> > > > > on?
> > > > >
> > > > > Make sure that before overcloud deploy the ironic nodes are available
> > > > > for
> > > > > provisioning (ironic node-list and check the provisioning state
> > > > > column).
> > > > > Also check that you didn't miss any step in the docs in regards to
> > > > > kernel
> > > > > and ramdisk assignment, introspection, flavor creation(so it matches
> > > > > the
> > > > > nodes resources)
> > > > > https://repos.fedorapeople.org/repos/openstack-m/rdo-manager-docs/liberty/basic_deployment/basic_deployment_cli.html
> > > > >
> > > > >
> > > > > > In instackenv.json file I do not need to add the undercloud node,
> > > > > > or
> > > > > > do
> > > > > > I?
> > > > >
> > > > > No, the nodes details should be enough.
> > > > >
> > > > > > And which log files should I watch during deployment?
> > > > >
> > > > > You can check the openstack-ironic-conductor logs(journalctl -fl -u
> > > > > openstack-ironic-conductor.service) and the logs in /var/log/nova.
> > > > >
> > > > > > Thanks
> > > > > > Esra
> > > > > >
> > > > > >
> > > > > > ----- Orijinal Mesaj -----Kimden: Marius Cornea
> > > > > > <mcornea at redhat.com>Kime:
> > > > > > Esra Celik <celik.esra at tubitak.gov.tr>Kk: Ignacio Bravo
> > > > > > <ibravo at ltgfederal.com>, rdo-list at redhat.comGönderilenler:
> > > > > > Tue,
> > > > > > 13
> > > > > > Oct
> > > > > > 2015 17:25:00 +0300 (EEST)Konu: Re: [Rdo-list] OverCloud deploy
> > > > > > fails
> > > > > > with
> > > > > > error "No valid host was found"
> > > > > >
> > > > > > ----- Original Message -----> From: "Esra Celik"
> > > > > > <celik.esra at tubitak.gov.tr>>
> > > > > > To: "Ignacio Bravo" <ibravo at ltgfederal.com>> Cc:
> > > > > > rdo-list at redhat.com>
> > > > > > Sent:
> > > > > > Tuesday, October 13, 2015 3:47:57 PM> Subject: Re: [Rdo-list]
> > > > > > OverCloud
> > > > > > deploy fails with error "No valid host was found"> > > > Actually I
> > > > > > re-installed the OS for Undercloud before deploying. However I did>
> > > > > > not
> > > > > > re-install the OS in Compute and Controller nodes.. I will
> > > > > > reinstall>
> > > > > > basic
> > > > > > OS for them too, and retry..
> > > > > >
> > > > > > You don't need to reinstall the OS on the controller and compute,
> > > > > > they
> > > > > > will
> > > > > > get the image served by the undercloud. I'd recommend that during
> > > > > > deployment
> > > > > > you watch the servers console and make sure they get powered on,
> > > > > > pxe
> > > > > > boot,
> > > > > > and actually get the image deployed.
> > > > > >
> > > > > > Thanks
> > > > > >
> > > > > > > Thanks> > > > Esra ÇELİK> TÜBİTAK BİLGEM>
> > > > > > > www.bilgem.tubitak.gov.tr> celik.esra at tubitak.gov.tr> > > Kimden:
> > > > > > > "Ignacio
> > > > > > > Bravo" <ibravo at ltgfederal.com>> Kime: "Esra Celik"
> > > > > > > <celik.esra at tubitak.gov.tr>> Kk: rdo-list at redhat.com>
> > > > > > > Gönderilenler:
> > > > > > > 13 Ekim Salı 2015 16:36:06> Konu: Re: [Rdo-list] OverCloud deploy
> > > > > > > fails
> > > > > > > with error "No valid host was> found"> > Esra,> > I encountered
> > > > > > > the
> > > > > > > same
> > > > > > > problem after deleting the stack and re-deploying.> > It turns
> > > > > > > out
> > > > > > > that
> > > > > > > 'heat stack-delete overcloud’ does remove the nodes from>
> > > > > > > ‘nova list’ and one would assume that the baremetal
> > > > > > > servers
> > > > > > > are now ready to> be used for the next stack, but when
> > > > > > > redeploying,
> > > > > > > I
> > > > > > > get
> > > > > > > the same message of> not enough hosts available.> > You can look
> > > > > > > into
> > > > > > > the
> > > > > > > nova logs and it mentions something about ‘node xxx is>
> > > > > > > already
> > > > > > > associated with UUID yyyy’ and ‘I tried 3 times and
> > > > > > > I’m
> > > > > > > giving up’.> The issue is that the UUID yyyy belonged to a
> > > > > > > prior
> > > > > > > unsuccessful deployment.> > I’m now redeploying the basic
> > > > > > > OS
> > > > > > > to
> > > > > > > start from scratch again.> > IB> > __> Ignacio Bravo> LTG
> > > > > > > Federal,
> > > > > > > Inc>
> > > > > > > www.ltgfederal.com> Office: (703) 951-7760> > > > On Oct 13,
> > > > > > > 2015,
> > > > > > > at
> > > > > > > 9:25
> > > > > > > AM, Esra Celik < celik.esra at tubitak.gov.tr > wrote:> > > Hi all,>
> > > > > > > >
> > > > > > > OverCloud deploy fails with error "No valid host was found"> >
> > > > > > > [stack at undercloud ~]$ openstack overcloud deploy --templates>
> > > > > > > Deploying
> > > > > > > templates in the directory>
> > > > > > > /usr/share/openstack-tripleo-heat-templates>
> > > > > > > Stack failed with status: Resource CREATE failed:
> > > > > > > resources.Compute:>
> > > > > > > ResourceInError: resources[0].resources.NovaCompute: Went to
> > > > > > > status
> > > > > > > ERROR>
> > > > > > > due to "Message: No valid host was found. There are not enough
> > > > > > > hosts>
> > > > > > > available., Code: 500"> Heat Stack create failed.> > Here are
> > > > > > > some
> > > > > > > logs:>
> > > > > > > > Every 2.0s: heat resource-list -n 5 overcloud | grep -v
> > > > > > > > COMPLETE
> > > > > > > > Tue
> > > > > > > > Oct
> > > > > > > 13> 16:18:17 2015> >
> > > > > > > +-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+>
> > > > > > > | resource_name | physical_resource_id | resource_type |
> > > > > > > | resource_status
> > > > > > > |> | updated_time | stack_name |>
> > > > > > > +-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+>
> > > > > > > | Compute | e33b6b1e-8740-4ded-ad7f-720617a03393 |
> > > > > > > | OS::Heat::ResourceGroup
> > > > > > > |> | CREATE_FAILED | 2015-10-13T10:20:36 | overcloud |> |
> > > > > > > |> | Controller
> > > > > > > |> | |
> > > > > > > 116c57ff-debb-4c12-92e1-e4163b67dc17 | OS::Heat::ResourceGroup> |
> > > > > > > |
> > > > > > > CREATE_FAILED | 2015-10-13T10:20:36 | overcloud |> | 0 |
> > > > > > > 342a9023-de8f-4b5b-b3ec-498d99b56dc4 | OS::TripleO::Controller |>
> > > > > > > |
> > > > > > > CREATE_IN_PROGRESS | 2015-10-13T10:20:52 |> |
> > > > > > > overcloud-Controller-45bbw24xxhxs |> | 0 |
> > > > > > > e420a7bd-86f8-4cc1-b6a0-5ba8d1412453 | OS::TripleO::Compute |> |
> > > > > > > CREATE_FAILED | 2015-10-13T10:20:54 |
> > > > > > > overcloud-Compute-vqk632ysg64r
> > > > > > > |>
> > > > > > > |
> > > > > > > Controller | 2e9ac712-0566-49b5-958f-c3e151bb24d7 |
> > > > > > > OS::Nova::Server
> > > > > > > |>
> > > > > > > |
> > > > > > > CREATE_IN_PROGRESS | 2015-10-13T10:20:54 |> |
> > > > > > > overcloud-Controller-45bbw24xxhxs-0-3vyhjiak2rsk |> | NovaCompute
> > > > > > > |
> > > > > > > 96efee56-81cb-46af-beef-84f4a3af761a | OS::Nova::Server |> |
> > > > > > > CREATE_FAILED
> > > > > > > | 2015-10-13T10:20:56 |> |
> > > > > > > | overcloud-Compute-vqk632ysg64r-0-32nalzkofmef
> > > > > > > |>
> > > > > > > +-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+>
> > > > > > > > > [stack at undercloud ~]$ heat resource-show overcloud Compute>
> > > > > > > +------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+>
> > > > > > > | Property | Value |>
> > > > > > > +------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+>
> > > > > > > | attributes | { |> | | "attributes": null, |> | | "refs": null
> > > > > > > | |>
> > > > > > > | |
> > > > > > > | |
> > > > > > > | }
> > > > > > > |> | creation_time | 2015-10-13T10:20:36 |> | description | |> |
> > > > > > > |> | links
> > > > > > > |> | |>
> > > > > > > |
> > > > > > > http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overcloud/620ada83-fb7d-4769-b1e8-431cfbb37d70/resources/Compute>
> > > > > > > | (self) |> | |
> > > > > > > http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overcloud/620ada83-fb7d-4769-b1e8-431cfbb37d70>
> > > > > > > | | (stack) |> | |
> > > > > > > http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overcloud-Compute-vqk632ysg64r/e33b6b1e-8740-4ded-ad7f-720617a03393>
> > > > > > > | | (nested) |> | logical_resource_id | Compute |> |
> > > > > > > | | physical_resource_id
> > > > > > > | e33b6b1e-8740-4ded-ad7f-720617a03393 |> | required_by |
> > > > > > > ComputeAllNodesDeployment |> | | ComputeNodesPostDeployment |> |
> > > > > > > |
> > > > > > > ComputeCephDeployment |> | | ComputeAllNodesValidationDeployment
> > > > > > > |>
> > > > > > > |
> > > > > > > |
> > > > > > > AllNodesExtraConfig |> | | allNodesConfig |> | resource_name |
> > > > > > > Compute
> > > > > > > |>
> > > > > > > | resource_status | CREATE_FAILED |> | resource_status_reason |
> > > > > > > resources.Compute: ResourceInError:> |
> > > > > > > resources[0].resources.NovaCompute:
> > > > > > > Went to status ERROR due to "Message:> | No valid host was found.
> > > > > > > There
> > > > > > > are not enough hosts available., Code: 500"> | |> | resource_type
> > > > > > > |
> > > > > > > OS::Heat::ResourceGroup |> | updated_time | 2015-10-13T10:20:36
> > > > > > > |>
> > > > > > > +------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+>
> > > > > > > > > > This is my instackenv.json for 1 compute and 1 control node
> > > > > > > > > > to
> > > > > > > > > > be
> > > > > > > deployed.> > {> "nodes": [> {> "pm_type":"pxe_ipmitool",>
> > > > > > > "mac":[>
> > > > > > > "08:9E:01:58:CC:A1"> ],> "cpu":"4",> "memory":"8192",>
> > > > > > > "disk":"10",>
> > > > > > > "arch":"x86_64",> "pm_user":"root",> "pm_password":"calvin",>
> > > > > > > "pm_addr":"192.168.0.18"> },> {> "pm_type":"pxe_ipmitool",>
> > > > > > > "mac":[>
> > > > > > > "08:9E:01:58:D0:3D"> ],> "cpu":"4",> "memory":"8192",>
> > > > > > > "disk":"100",>
> > > > > > > "arch":"x86_64",> "pm_user":"root",> "pm_password":"calvin",>
> > > > > > > "pm_addr":"192.168.0.19"> }> ]> }> > > Any ideas? Thanks in
> > > > > > > advance>
> > > > > > > >
> > > > > > > >
> > > > > > > Esra ÇELİK> TÜBİTAK BİLGEM>
> > > > > > > www.bilgem.tubitak.gov.tr>
> > > > > > > celik.esra at tubitak.gov.tr> >
> > > > > > > _______________________________________________> Rdo-list mailing
> > > > > > > list>
> > > > > > > Rdo-list at redhat.com>
> > > > > > > https://www.redhat.com/mailman/listinfo/rdo-list>
> > > > > > > >
> > > > > > > To unsubscribe: rdo-list-unsubscribe at redhat.com> > > >
> > > > > > > _______________________________________________> Rdo-list mailing
> > > > > > > list>
> > > > > > > Rdo-list at redhat.com>
> > > > > > > https://www.redhat.com/mailman/listinfo/rdo-list>
> > > > > > > >
> > > > > > > To unsubscribe: rdo-list-unsubscribe at redhat.com
> > > > > >
> > > > >
> > > > >
> > > >
> > > >
> > >
> > >
> > > _______________________________________________
> > > Rdo-list mailing list
> > > Rdo-list at redhat.com
> > > https://www.redhat.com/mailman/listinfo/rdo-list
> > >
> > > To unsubscribe: rdo-list-unsubscribe at redhat.com
>
More information about the dev
mailing list