Hi Sasha
This is my instackenv.json. MAC addresses are, em2 interface’s MAC address of
the nodes
{
"nodes": [
{
"pm_type":"pxe_ipmitool",
"mac":[
"08:9E:01:58:CC:A1"
],
"cpu":"4",
"memory":"8192",
"disk":"10",
"arch":"x86_64",
"pm_user":"root",
"pm_password”:”<pwd>”,
"pm_addr":"192.168.0.18"
},
{
"pm_type":"pxe_ipmitool",
"mac":[
"08:9E:01:58:D0:3D"
],
"cpu":"4",
"memory":"8192",
"disk":"100",
"arch":"x86_64",
"pm_user":"root",
"pm_password”:”<pwd>”,
"pm_addr":"192.168.0.19"
}
]
}
This is my undercloud.conf file:
image_path = .
local_ip = 192.0.2.1/24
local_interface = em2
masquerade_network = 192.0.2.0/24
dhcp_start = 192.0.2.5
dhcp_end = 192.0.2.24
network_cidr = 192.0.2.0/24
network_gateway = 192.0.2.1
inspection_interface = br-ctlplane
inspection_iprange = 192.0.2.100,192.0.2.120
inspection_runbench = false
undercloud_debug = true
enable_tuskar = false
enable_tempest = false
I have previously sent the screenshot of the consoles during introspection stage. Now I am
attaching them again.
I cannot login to consoles because introspection stage is not completed successfully and I
don't know the IP addresses. (nova list is empty)
(I don't know if I can login with the IP addresses that I was previously set by
myself. I am not able to reach the nodes now, from home.)
I ran the flavor-create command after the introspection stage. But introspection was not
completed successfully,
I just ran deploy command to see if nova list fills during deployment.
Esra ÇELİK
TÜBİTAK BİLGEM
www.bilgem.tubitak.gov.tr
celik.esra(a)tubitak.gov.tr
----- Sasha Chuzhoy <sasha(a)redhat.com> şöyle yazıyor:> Esra, Is it
possible to check the console of the nodes being introspected and/or deployed? I wonder if
the instackenv.json file is accurate. Also, what's the output from 'nova
flavor-list'? Thanks. Best regards, Sasha Chuzhoy. ----- Original Message ----- >
From: "Esra Celik" <celik.esra(a)tubitak.gov.tr> > To: "Marius
Cornea" <mcornea(a)redhat.com> > Cc: "Sasha Chuzhoy"
<sasha(a)redhat.com>, rdo-list(a)redhat.com > Sent: Monday, October 19, 2015 9:51:51
AM > Subject: Re: [Rdo-list] OverCloud deploy fails with error "No valid host was
found" > > All 3 baremetal nodes (1 undercloud, 2 overcloud) have 2 nics. >
> the undercloud machine's ip config is as follows: > > [stack@undercloud ~]$
ip addr > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > inet 127.0.0.1/8 scope
host lo > valid_lft forever preferred_lft forever > inet6 ::1/128 scope host >
valid_lft forever preferred_lft forever > 2: em1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen > 1000 >
link/ether 08:9e:01:50:8a:21 brd ff:ff:ff:ff:ff:ff > inet 10.1.34.81/24 brd 10.1.34.255
scope global em1 > valid_lft forever preferred_lft forever > inet6
fe80::a9e:1ff:fe50:8a21/64 scope link > valid_lft forever preferred_lft forever > 3:
em2: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master ovs-system >
state UP qlen 1000 > link/ether 08:9e:01:50:8a:22 brd ff:ff:ff:ff:ff:ff > 4:
ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > link/ether
9a:a8:0f:ec:42:15 brd ff:ff:ff:ff:ff:ff > 5: br-ctlplane:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue > state UNKNOWN >
link/ether 08:9e:01:50:8a:22 brd ff:ff:ff:ff:ff:ff > inet 192.0.2.1/24 brd 192.0.2.255
scope global br-ctlplane > valid_lft forever preferred_lft forever > inet6
fe80::a9e:1ff:fe50:8a22/64 scope link > valid_lft forever preferred_lft forever > 6:
br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > link/ether
fa:85:ac:92:f5:41 brd ff:ff:ff:ff:ff:ff > > I am using em2 for pxe boot on the other
machines.. So I configured > instackenv.json to have em2's MAC address > For
overcloud nodes, em1 was configured to have 10.1.34.x ip, but after image > deploy I am
not sure what happened for that nic. > > Thanks > > Esra ÇELİK >
TÜBİTAK BİLGEM >
www.bilgem.tubitak.gov.tr > celik.esra(a)tubitak.gov.tr >
----- Orijinal Mesaj ----- > > > Kimden: "Marius Cornea"
<mcornea(a)redhat.com> > > Kime: "Esra Celik"
<celik.esra(a)tubitak.gov.tr> > > Kk: "Sasha Chuzhoy"
<sasha(a)redhat.com>, rdo-list(a)redhat.com > > Gönderilenler: 19 Ekim
Pazartesi 2015 15:36:58 > > Konu: Re: [Rdo-list] OverCloud deploy fails with error
"No valid host was > > found" > > > Hi, > > > I believe
the nodes were stuck in introspection so they were not ready for > > deployment thus
the not enough hosts message. Can you describe the > > networking setup (how many
nics the nodes have and to what networks they're > > connected)? > > >
Thanks, > > Marius > > > ----- Original Message ----- > > > From:
"Esra Celik" <celik.esra(a)tubitak.gov.tr> > > > To: "Sasha
Chuzhoy" <sasha(a)redhat.com> > > > Cc: "Marius Cornea"
<mcornea(a)redhat.com>, rdo-list(a)redhat.com > > > Sent: Monday, October 19,
2015 12:34:32 PM > > > Subject: Re: [Rdo-list] OverCloud deploy fails with error
"No valid host > > > was found" > > > > > > Hi again,
> > > > > > "nova list" was empty after introspection stage
which was not completed > > > successfully. So I cloud not ssh the nodes.. Is
there another way to > > > obtain > > > the IP addresses? > > >
> > > [stack@undercloud ~]$ sudo systemctl|grep ironic > > >
openstack-ironic-api.service loaded active running OpenStack Ironic API > > >
service > > > openstack-ironic-conductor.service loaded active running OpenStack
Ironic > > > Conductor service > > >
openstack-ironic-inspector-dnsmasq.service loaded active running PXE boot > > >
dnsmasq service for Ironic Inspector > > > openstack-ironic-inspector.service
loaded active running Hardware > > > introspection service for OpenStack Ironic
> > > > > > If I start deployment anyway I get 2 nodes in ERROR state
> > > > > > [stack@undercloud ~]$ openstack overcloud deploy --templates
> > > Deploying templates in the directory > > >
/usr/share/openstack-tripleo-heat-templates > > > Stack failed with status:
resources.Controller: resources[0]: > > > ResourceInError: resources.Controller:
Went to status ERROR due to > > > "Message: > > > No valid host was
found. There are not enough hosts available., Code: > > > 500" > >
> > > > [stack@undercloud ~]$ nova list > > >
+--------------------------------------+-------------------------+--------+------------+-------------+----------+
> > > | ID | Name | Status | Task State | Power State | Networks | > > >
+--------------------------------------+-------------------------+--------+------------+-------------+----------+
> > > | 3a8e1fe4-d189-4fce-9912-dcf49fefb000 | overcloud-controller-0 | ERROR |
> > > | - > > > | | > > > | NOSTATE | | > > > |
616b45c6-2749-418f-8aa4-fe2bfe164782 | overcloud-novacompute-0 | ERROR > > > | |
> > > | - > > > | | NOSTATE | | > > >
+--------------------------------------+-------------------------+--------+------------+-------------+----------+
> > > > > > Did the repositories update during weekend? Should I better
restart the > > > overall Undercloud and Overcloud installation from the
beginning? > > > > > > Thanks. > > > > > > Esra
ÇELİK > > > Uzman Araştırmacı > > > Bilişim Teknolojileri
Enstitüsü > > > TÜBİTAK BİLGEM > > > 41470
GEBZE - KOCAELİ > > > T +90 262 675 3140 > > > F +90 262 646 3187 >
> >
www.bilgem.tubitak.gov.tr > > > celik.esra(a)tubitak.gov.tr > >
> ................................................................ > > > >
> > Sorumluluk Reddi > > > > > > ----- Orijinal Mesaj ----- >
> > > > > > Kimden: "Sasha Chuzhoy" <sasha(a)redhat.com>
> > > > Kime: "Esra Celik" <celik.esra(a)tubitak.gov.tr> >
> > > Kk: "Marius Cornea" <mcornea(a)redhat.com>,
rdo-list(a)redhat.com > > > > Gönderilenler: 16 Ekim Cuma 2015 18:44:49
> > > > Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid
host > > > > was > > > > found" > > > > > >
> Hi Esra, > > > > if the undercloud nodes are UP - you can login with: ssh
> > > > heat-admin@<IP> > > > > You can see the IP of the
nodes with: "nova list". > > > > BTW, > > > > What do
you see if you run "sudo systemctl|grep ironic" on the > > > >
undercloud? > > > > > > > Best regards, > > > > Sasha
Chuzhoy. > > > > > > > ----- Original Message ----- > > >
> > From: "Esra Celik" <celik.esra(a)tubitak.gov.tr> > > >
> > To: "Sasha Chuzhoy" <sasha(a)redhat.com> > > > > >
Cc: "Marius Cornea" <mcornea(a)redhat.com>, rdo-list(a)redhat.com > >
> > > Sent: Friday, October 16, 2015 1:40:16 AM > > > > > Subject:
Re: [Rdo-list] OverCloud deploy fails with error "No valid > > > > >
host > > > > > was found" > > > > > > > > >
> Hi Sasha, > > > > > > > > > > I have 3 nodes, 1
Undercloud, 1 Overcloud-Controller, 1 > > > > > Overcloud-Compute > >
> > > > > > > > This is my undercloud.conf file: > > >
> > > > > > > image_path = . > > > > > local_ip =
192.0.2.1/24 > > > > > local_interface = em2 > > > > >
masquerade_network = 192.0.2.0/24 > > > > > dhcp_start = 192.0.2.5 >
> > > > dhcp_end = 192.0.2.24 > > > > > network_cidr =
192.0.2.0/24 > > > > > network_gateway = 192.0.2.1 > > > > >
inspection_interface = br-ctlplane > > > > > inspection_iprange =
192.0.2.100,192.0.2.120 > > > > > inspection_runbench = false > >
> > > undercloud_debug = true > > > > > enable_tuskar = false >
> > > > enable_tempest = false > > > > > > > > >
> IP configuration for the Undercloud is as follows: > > > > > > >
> > > stack@undercloud ~]$ ip addr > > > > > 1: lo:
<LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN > > > >
> link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 > > > > > inet
127.0.0.1/8 scope host lo > > > > > valid_lft forever preferred_lft forever
> > > > > inet6 ::1/128 scope host > > > > > valid_lft
forever preferred_lft forever > > > > > 2: em1:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP > > > >
> qlen > > > > > 1000 > > > > > link/ether
08:9e:01:50:8a:21 brd ff:ff:ff:ff:ff:ff > > > > > inet 10.1.34.81/24 brd
10.1.34.255 scope global em1 > > > > > valid_lft forever preferred_lft
forever > > > > > inet6 fe80::a9e:1ff:fe50:8a21/64 scope link > >
> > > valid_lft forever preferred_lft forever > > > > > 3: em2:
<BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq master > > > > >
ovs-system > > > > > state UP qlen 1000 > > > > > link/ether
08:9e:01:50:8a:22 brd ff:ff:ff:ff:ff:ff > > > > > 4: ovs-system:
<BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > > > > >
link/ether 9a:a8:0f:ec:42:15 brd ff:ff:ff:ff:ff:ff > > > > > 5:
br-ctlplane: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc > > > >
> noqueue > > > > > state UNKNOWN > > > > > link/ether
08:9e:01:50:8a:22 brd ff:ff:ff:ff:ff:ff > > > > > inet 192.0.2.1/24 brd
192.0.2.255 scope global br-ctlplane > > > > > valid_lft forever
preferred_lft forever > > > > > inet6 fe80::a9e:1ff:fe50:8a22/64 scope link
> > > > > valid_lft forever preferred_lft forever > > > > >
6: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN > > > >
> link/ether fa:85:ac:92:f5:41 brd ff:ff:ff:ff:ff:ff > > > > > > >
> > > And I attached two screenshots showing the boot stage for overcloud >
> > > > nodes > > > > > > > > > > Is there a way
to login the overcloud nodes to see their IP > > > > > configuration? >
> > > > > > > > > Thanks > > > > > > >
> > > Esra ÇELİK > > > > > TÜBİTAK BİLGEM >
> > > >
www.bilgem.tubitak.gov.tr > > > > >
celik.esra(a)tubitak.gov.tr > > > > > > > > > > ----- Orijinal
Mesaj ----- > > > > > > > > > > > Kimden: "Sasha
Chuzhoy" <sasha(a)redhat.com> > > > > > > Kime: "Esra
Celik" <celik.esra(a)tubitak.gov.tr> > > > > > > Kk:
"Marius Cornea" <mcornea(a)redhat.com>, rdo-list(a)redhat.com > > >
> > > Gönderilenler: 15 Ekim Perşembe 2015 16:58:41 > > > >
> > Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid > >
> > > > host > > > > > > was > > > > > >
found" > > > > > > > > > > > Just my 2 cents. >
> > > > > Did you make sure that all the registered nodes are configured to
> > > > > > boot > > > > > > off > > > >
> > the right NIC first? > > > > > > Can you watch the console and
see what happens on the problematic > > > > > > nodes > > >
> > > upon > > > > > > boot? > > > > > > >
> > > > Best regards, > > > > > > Sasha Chuzhoy. > >
> > > > > > > > > ----- Original Message ----- > > >
> > > > From: "Esra Celik" <celik.esra(a)tubitak.gov.tr> >
> > > > > > To: "Marius Cornea" <mcornea(a)redhat.com>
> > > > > > > Cc: rdo-list(a)redhat.com > > > > > >
> Sent: Thursday, October 15, 2015 4:40:46 AM > > > > > > >
Subject: Re: [Rdo-list] OverCloud deploy fails with error "No > > > >
> > > valid > > > > > > > host > > > > > >
> was found" > > > > > > > > > > > > > >
> > > > > > > Sorry for the late reply > > > > > >
> > > > > > > > ironic node-show results are below. I have my
nodes power on > > > > > > > after > > > > > > >
introspection bulk start. And I get the following warning > > > > > >
> Introspection didn't finish for nodes > > > > > > >
5b28998f-4dc8-42aa-8a51-521e20b1e5ed,6f35ac24-135d-4b99-8a24-fa2b731bd218 > > >
> > > > > > > > > > > Doesn't seem to be the same
issue with > > > > > > >
https://www.redhat.com/archives/rdo-list/2015-October/msg00174.html > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > [stack@undercloud ~]$
ironic node-list > > > > > > >
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > > > | UUID | Name | Instance UUID | Power State |
Provisioning State > > > > > > > | | > > > > > >
> | Maintenance | > > > > > > >
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > > > | 5b28998f-4dc8-42aa-8a51-521e20b1e5ed | None | None |
power on | > > > > > > > | available > > > > > >
> | | > > > > > > > | False | > > > > > > > |
6f35ac24-135d-4b99-8a24-fa2b731bd218 | None | None | power on | > > > > >
> > | available > > > > > > > | | > > > > > >
> | False | > > > > > > >
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > > > > > > > > > > > > > >
> > > [stack@undercloud ~]$ ironic node-show > > > > > > >
5b28998f-4dc8-42aa-8a51-521e20b1e5ed > > > > > > >
+------------------------+-------------------------------------------------------------------------+
> > > > > > > | Property | Value | > > > > > > >
+------------------------+-------------------------------------------------------------------------+
> > > > > > > | target_power_state | None | > > > > >
> > | extra | {} | > > > > > > > | last_error | None | >
> > > > > > | updated_at | 2015-10-15T08:26:42+00:00 | > > >
> > > > | maintenance_reason | None | > > > > > > > |
provision_state | available | > > > > > > > | clean_step | {} | >
> > > > > > | uuid | 5b28998f-4dc8-42aa-8a51-521e20b1e5ed | > >
> > > > > | console_enabled | False | > > > > > > > |
target_provision_state | None | > > > > > > > | provision_updated_at
| 2015-10-15T08:26:42+00:00 | > > > > > > > | maintenance | False |
> > > > > > > | inspection_started_at | None | > > > >
> > > | inspection_finished_at | None | > > > > > > > |
power_state | power on | > > > > > > > | driver | pxe_ipmitool | >
> > > > > > | reservation | None | > > > > > > > |
properties | {u'memory_mb': u'8192', u'cpu_arch':
u'x86_64', > > > > > > > | u'local_gb': > >
> > > > > | u'10', | > > > > > > > | |
u'cpus': u'4', u'capabilities': u'boot_option:local'} |
> > > > > > > | instance_uuid | None | > > > > > >
> | name | None | > > > > > > > | driver_info |
{u'ipmi_password': u'******', u'ipmi_address': > > > >
> > > | u'192.168.0.18', | > > > > > > > | |
u'ipmi_username': u'root', u'deploy_kernel': > > > >
> > > | | u'49a2c8d4-a283-4bdf-8d6f- > > > > > > > | | |
> > > > > > > | | e83ae28da047', u'deploy_ramdisk':
u'3db3dbed- | > > > > > > > | | 0d88-4632-af98-8defb05ca6e2'}
| > > > > > > > | created_at | 2015-10-15T07:49:08+00:00 | > >
> > > > > | driver_internal_info | {u'clean_steps': None} | >
> > > > > > | chassis_uuid | | > > > > > > > |
instance_info | {} | > > > > > > >
+------------------------+-------------------------------------------------------------------------+
> > > > > > > > > > > > > > > > > >
> > > [stack@undercloud ~]$ ironic node-show > > > > > > >
6f35ac24-135d-4b99-8a24-fa2b731bd218 > > > > > > >
+------------------------+-------------------------------------------------------------------------+
> > > > > > > | Property | Value | > > > > > > >
+------------------------+-------------------------------------------------------------------------+
> > > > > > > | target_power_state | None | > > > > >
> > | extra | {} | > > > > > > > | last_error | None | >
> > > > > > | updated_at | 2015-10-15T08:26:42+00:00 | > > >
> > > > | maintenance_reason | None | > > > > > > > |
provision_state | available | > > > > > > > | clean_step | {} | >
> > > > > > | uuid | 6f35ac24-135d-4b99-8a24-fa2b731bd218 | > >
> > > > > | console_enabled | False | > > > > > > > |
target_provision_state | None | > > > > > > > | provision_updated_at
| 2015-10-15T08:26:42+00:00 | > > > > > > > | maintenance | False |
> > > > > > > | inspection_started_at | None | > > > >
> > > | inspection_finished_at | None | > > > > > > > |
power_state | power on | > > > > > > > | driver | pxe_ipmitool | >
> > > > > > | reservation | None | > > > > > > > |
properties | {u'memory_mb': u'8192', u'cpu_arch':
u'x86_64', > > > > > > > | u'local_gb': > >
> > > > > | u'100', | > > > > > > > | |
u'cpus': u'4', u'capabilities': u'boot_option:local'} |
> > > > > > > | instance_uuid | None | > > > > > >
> | name | None | > > > > > > > | driver_info |
{u'ipmi_password': u'******', u'ipmi_address': > > > >
> > > | u'192.168.0.19', | > > > > > > > | |
u'ipmi_username': u'root', u'deploy_kernel': > > > >
> > > | | u'49a2c8d4-a283-4bdf-8d6f- > > > > > > > | | |
> > > > > > > | | e83ae28da047', u'deploy_ramdisk':
u'3db3dbed- | > > > > > > > | | 0d88-4632-af98-8defb05ca6e2'}
| > > > > > > > | created_at | 2015-10-15T07:49:08+00:00 | > >
> > > > > | driver_internal_info | {u'clean_steps': None} | >
> > > > > > | chassis_uuid | | > > > > > > > |
instance_info | {} | > > > > > > >
+------------------------+-------------------------------------------------------------------------+
> > > > > > > [stack@undercloud ~]$ > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > And below I added my history for the
stack user. I don't think I > > > > > > > am > > > >
> > > doing > > > > > > > something other than > >
> > > > >
https://repos.fedorapeople.org/repos/openstack-m/rdo-manager-docs/liberty > > >
> > > > doc > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > 1 vi instackenv.json > > > > > > >
2 sudo yum -y install epel-release > > > > > > > 3 sudo curl -o
/etc/yum.repos.d/delorean.repo > > > > > > >
http://trunk.rdoproject.org/centos7-liberty/current-passed-ci/delorean.repo > > >
> > > > 4 sudo curl -o /etc/yum.repos.d/delorean-current.repo > > >
> > > >
http://trunk.rdoproject.org/centos7-liberty/current/delorean.repo >
> > > > > > 5 sudo sed -i 's/\[delorean\]/\[delorean-current\]/'
> > > > > > > /etc/yum.repos.d/delorean-current.repo > > >
> > > > 6 sudo /bin/bash -c "cat > > > > > > >
<<EOF>>/etc/yum.repos.d/delorean-current.repo > > > > > >
> > > > > > > >
includepkgs=diskimage-builder,openstack-heat,instack,instack-undercloud,openstack-ironic,openstack-ironic-inspector,os-cloud-config,os-net-config,python-ironic-inspector-client,python-tripleoclient,tripleo-common,openstack-tripleo-heat-templates,openstack-tripleo-image-elements,openstack-tuskar-ui-extras,openstack-puppet-modules
> > > > > > > EOF" > > > > > > > 7 sudo
curl -o /etc/yum.repos.d/delorean-deps.repo > > > > > > >
http://trunk.rdoproject.org/centos7-liberty/delorean-deps.repo > > > > >
> > 8 sudo yum -y install yum-plugin-priorities > > > > > > > 9
sudo yum install -y python-tripleoclient > > > > > > > 10 cp
/usr/share/instack-undercloud/undercloud.conf.sample > > > > > > >
~/undercloud.conf > > > > > > > 11 vi undercloud.conf > > >
> > > > 12 export DIB_INSTALLTYPE_puppet_modules=source > > > >
> > > 13 openstack undercloud install > > > > > > > 14
source stackrc > > > > > > > 15 export NODE_DIST=centos7 > >
> > > > > 16 export DIB_YUM_REPO_CONF="/etc/yum.repos.d/delorean.repo
> > > > > > > /etc/yum.repos.d/delorean-deps.repo" > >
> > > > > 17 export DIB_INSTALLTYPE_puppet_modules=source > > >
> > > > 18 openstack overcloud image build --all > > > > > >
> 19 ls > > > > > > > 20 openstack overcloud image upload >
> > > > > > 21 openstack baremetal import --json instackenv.json >
> > > > > > 22 openstack baremetal configure boot > > > >
> > > 23 ironic node-list > > > > > > > 24 openstack
baremetal introspection bulk start > > > > > > > 25 ironic node-list
> > > > > > > 26 ironic node-show
5b28998f-4dc8-42aa-8a51-521e20b1e5ed > > > > > > > 27 ironic
node-show 6f35ac24-135d-4b99-8a24-fa2b731bd218 > > > > > > > 28
history > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > Thanks > > > > > > > > > > > >
> > > > > > > > > > > > > > > > Esra
ÇELİK > > > > > > > TÜBİTAK BİLGEM > > >
> > > >
www.bilgem.tubitak.gov.tr > > > > > > >
celik.esra(a)tubitak.gov.tr > > > > > > > > > > > > >
> > > > > > > > Kimden: "Marius Cornea"
<mcornea(a)redhat.com> > > > > > > > Kime: "Esra Celik"
<celik.esra(a)tubitak.gov.tr> > > > > > > > Kk: "Ignacio
Bravo" <ibravo(a)ltgfederal.com>, rdo-list(a)redhat.com > > > > >
> > Gönderilenler: 14 Ekim Çarşamba 2015 19:40:07 > > >
> > > > Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid
> > > > > > > host > > > > > > > was > >
> > > > > found" > > > > > > > > > >
> > > > Can you do ironic node-show for your ironic nodes and post the >
> > > > > > results? > > > > > > > Also > >
> > > > > check the following suggestion if you're experiencing the
same > > > > > > > issue: > > > > > > >
https://www.redhat.com/archives/rdo-list/2015-October/msg00174.html > > > >
> > > > > > > > > > ----- Original Message ----- > >
> > > > > > From: "Esra Celik"
<celik.esra(a)tubitak.gov.tr> > > > > > > > > To: "Marius
Cornea" <mcornea(a)redhat.com> > > > > > > > > Cc:
"Ignacio Bravo" <ibravo(a)ltgfederal.com>, > > > > > >
> > rdo-list(a)redhat.com > > > > > > > > Sent: Wednesday,
October 14, 2015 3:22:20 PM > > > > > > > > Subject: Re:
[Rdo-list] OverCloud deploy fails with error "No > > > > > > >
> valid > > > > > > > > host > > > > > > >
> was found" > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
Well in the early stage of the introspection I can see Client > > > > >
> > > IP > > > > > > > > of > > > > > >
> > nodes > > > > > > > > (screenshot attached). But then I
see continuous > > > > > > > > ironic-python-agent > > >
> > > > > errors > > > > > > > > (screenshot-2
attached). Errors repeat after time out.. And the > > > > > > > >
nodes > > > > > > > > are > > > > > > > >
not powered off. > > > > > > > > > > > > > >
> > Seems like I am stuck in introspection stage.. > > > > > >
> > > > > > > > > > I can use ipmitool command to
successfully power on/off the > > > > > > > > nodes > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > [stack@undercloud ~]$ ipmitool -I
lanplus -H 192.168.0.19 -L > > > > > > > > ADMINISTRATOR > >
> > > > > > -U > > > > > > > > root -R 3 -N 5 -P
<password> power status > > > > > > > > Chassis Power is on
> > > > > > > > > > > > > > > > > >
> > > > > > [stack@undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U
> > > > > > > > root > > > > > > > > -P
> > > > > > > > <password> chassis power status > >
> > > > > > Chassis Power is on > > > > > > > >
[stack@undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U > > > > > >
> > root > > > > > > > > -P > > > > > >
> > <password> chassis power off > > > > > > > >
Chassis Power Control: Down/Off > > > > > > > > [stack@undercloud
~]$ ipmitool -H 192.168.0.18 -I lanplus -U > > > > > > > > root
> > > > > > > > -P > > > > > > > >
<password> chassis power status > > > > > > > > Chassis
Power is off > > > > > > > > [stack@undercloud ~]$ ipmitool -H
192.168.0.18 -I lanplus -U > > > > > > > > root > > >
> > > > > -P > > > > > > > > <password>
chassis power on > > > > > > > > Chassis Power Control: Up/On >
> > > > > > > [stack@undercloud ~]$ ipmitool -H 192.168.0.18 -I
lanplus -U > > > > > > > > root > > > > > > >
> -P > > > > > > > > <password> chassis power status >
> > > > > > > Chassis Power is on > > > > > > >
> > > > > > > > > > > > > > > > > Esra
ÇELİK > > > > > > > > TÜBİTAK BİLGEM > >
> > > > > >
www.bilgem.tubitak.gov.tr > > > > > > >
> celik.esra(a)tubitak.gov.tr > > > > > > > > > > > >
> > > > > > > > > > > > ----- Orijinal Mesaj -----
> > > > > > > > > > > > > > > > Kimden:
"Marius Cornea" <mcornea(a)redhat.com> > > > > > > >
> Kime: "Esra Celik" <celik.esra(a)tubitak.gov.tr> > > > >
> > > > Kk: "Ignacio Bravo" <ibravo(a)ltgfederal.com>, > >
> > > > > > rdo-list(a)redhat.com > > > > > > > >
Gönderilenler: 14 Ekim Çarşamba 2015 14:59:30 > > > > >
> > > Konu: Re: [Rdo-list] OverCloud deploy fails with error "No > >
> > > > > > valid > > > > > > > > host > >
> > > > > > was > > > > > > > > found" >
> > > > > > > > > > > > > > > > > >
> > > > > ----- Original Message ----- > > > > > > >
> > From: "Esra Celik" <celik.esra(a)tubitak.gov.tr> > > >
> > > > > > To: "Marius Cornea" <mcornea(a)redhat.com>
> > > > > > > > > Cc: "Ignacio Bravo"
<ibravo(a)ltgfederal.com>, > > > > > > > > >
rdo-list(a)redhat.com > > > > > > > > > Sent: Wednesday, October
14, 2015 10:49:01 AM > > > > > > > > > Subject: Re: [Rdo-list]
OverCloud deploy fails with error "No > > > > > > > > >
valid > > > > > > > > > host > > > > > > >
> > was found" > > > > > > > > > > > > >
> > > > > > > > > > > > > > Well today I started
with re-installing the OS and nothing > > > > > > > > > seems
> > > > > > > > > wrong > > > > > > > >
> with > > > > > > > > > undercloud installation, then; >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > I see an error during image build > > >
> > > > > > > > > > > > > > > > > >
> > > > > > [stack@undercloud ~]$ openstack overcloud image build --all
> > > > > > > > > ... > > > > > > > >
> a lot of log > > > > > > > > > ... > > > >
> > > > > ++ cat /etc/dib_dracut_drivers > > > > > > >
> > + dracut -N --install ' curl partprobe lsblk targetcli tail > > >
> > > > > > head > > > > > > > > > awk >
> > > > > > > > ifconfig > > > > > > > >
> cut expr route ping nc wget tftp grep' --kernel-cmdline > > > > >
> > > > 'rd.shell > > > > > > > > > rd.debug
> > > > > > > > > rd.neednet=1 rd.driver.pre=ahci'
--include > > > > > > > > > /var/tmp/image.YVhwuArQ/mnt/ >
> > > > > > > > / > > > > > > > > >
--kver 3.10.0-229.14.1.el7.x86_64 --add-drivers ' virtio > > > > > >
> > > virtio_net > > > > > > > > > virtio_blk
target_core_mod iscsi_target_mod > > > > > > > > >
target_core_iblock > > > > > > > > > target_core_file
target_core_pscsi configfs' -o 'dash > > > > > > > > >
plymouth' > > > > > > > > > /tmp/ramdisk > > >
> > > > > > cat: write error: Broken pipe > > > > > >
> > > + cp /boot/vmlinuz-3.10.0-229.14.1.el7.x86_64 /tmp/kernel > > >
> > > > > > + chmod o+r /tmp/kernel > > > > > > >
> > + trap EXIT > > > > > > > > > +
target_tag=99-build-dracut-ramdisk > > > > > > > > > + date
+%s.%N > > > > > > > > > + output '99-build-dracut-ramdisk
completed' > > > > > > > > > ... > > > > >
> > > > a lot of log > > > > > > > > > ... >
> > > > > > > > > > > > > > > You can ignore
that afaik, if you end up having all the > > > > > > > > required
> > > > > > > > images > > > > > > > > it
> > > > > > > > should be ok. > > > > > > >
> > > > > > > > > > > > > > > > > >
> Then, during introspection stage I see ironic-python-agent > > > > >
> > > > errors > > > > > > > > > on > > >
> > > > > > nodes > > > > > > > > >
(screenshot attached) and the following warnings > > > > > > > >
> > > > > > > > > > > > > > > > > That
looks odd. Is it showing up in the early stage of the > > > > > > >
> introspection? > > > > > > > > At > > > > >
> > > some point it should receive an address by DHCP and the Network > >
> > > > > > is > > > > > > > > unreachable error
should disappear. Does the introspection > > > > > > > > complete
> > > > > > > > and > > > > > > > > the
> > > > > > > > nodes are turned off? > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > [root@localhost ~]# journalctl
-fl -u > > > > > > > > > openstack-ironic-conductor.service
> > > > > > > > > | > > > > > > > >
> grep -i "warning\|error" > > > > > > > > > Oct
14 10:30:12 undercloud.rdo ironic-conductor[619]: > > > > > > > >
> 2015-10-14 > > > > > > > > > 10:30:12.119 > > >
> > > > > > 619 WARNING oslo_config.cfg > > > > > >
> > > [req-eccf8cb5-6e93-4d8f-9a05-0e8c8d2aab7b > > > > > >
> > > ] > > > > > > > > > Option "http_url"
from group "pxe" is deprecated. Use option > > > > > > >
> > "http_url" > > > > > > > > > from > >
> > > > > > > group "deploy". > > > > > >
> > > Oct 14 10:30:12 undercloud.rdo ironic-conductor[619]: > > > >
> > > > > 2015-10-14 > > > > > > > > >
10:30:12.119 > > > > > > > > > 619 WARNING oslo_config.cfg >
> > > > > > > > [req-eccf8cb5-6e93-4d8f-9a05-0e8c8d2aab7b >
> > > > > > > > ] > > > > > > > > >
Option "http_root" from group "pxe" is deprecated. Use option >
> > > > > > > > "http_root" > > > > >
> > > > from group "deploy". > > > > > > > >
> > > > > > > > > > > > > > > > > >
> Before deployment ironic node-list: > > > > > > > > > >
> > > > > > > > > > > > > > > This is odd too
as I'm expecting the nodes to be powered off > > > > > > > >
before > > > > > > > > running > > > > > > >
> deployment. > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > [stack@undercloud ~]$ ironic node-list > > > > > > >
> >
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > > > > > | UUID | Name | Instance UUID | Power State |
Provisioning > > > > > > > > > | State > > > > >
> > > > | | > > > > > > > > > | Maintenance | >
> > > > > > > >
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > > > > > | acfc1bb4-469d-479a-af70-c0bdd669b32d | None
| None | power > > > > > > > > > | on > > > > >
> > > > | | > > > > > > > > > | available > >
> > > > > > > | | > > > > > > > > > |
False | > > > > > > > > > |
b5811c06-d5d1-41f1-87b3-2fd55ae63553 | None | None | power > > > > > >
> > > | on > > > > > > > > > | | > > > >
> > > > > | available > > > > > > > > > | | >
> > > > > > > > | False | > > > > > > > >
>
+--------------------------------------+------+---------------+-------------+--------------------+-------------+
> > > > > > > > > > > > > > > > > >
During deployment I get following errors > > > > > > > > > >
> > > > > > > > [root@localhost ~]# journalctl -fl -u > >
> > > > > > > openstack-ironic-conductor.service > > > >
> > > > > | > > > > > > > > > grep -i
"warning\|error" > > > > > > > > > Oct 14 11:29:01
undercloud.rdo ironic-conductor[619]: > > > > > > > > >
2015-10-14 > > > > > > > > > 11:29:01.739 > > > >
> > > > > 619 ERROR ironic.drivers.modules.ipmitool [-] IPMI Error >
> > > > > > > > while > > > > > > > > >
attempting > > > > > > > > > "ipmitool -I lanplus -H
192.168.0.19 -L ADMINISTRATOR -U root > > > > > > > > > -R >
> > > > > > > > 3 > > > > > > > > > -N
> > > > > > > > > 5 > > > > > > > >
> -f > > > > > > > > > /tmp/tmpSCKHIv power status"for
node > > > > > > > > > b5811c06-d5d1-41f1-87b3-2fd55ae63553.
> > > > > > > > > Error: Unexpected error while running
command. > > > > > > > > > Oct 14 11:29:01 undercloud.rdo
ironic-conductor[619]: > > > > > > > > > 2015-10-14 > >
> > > > > > > 11:29:01.739 > > > > > > > >
> 619 WARNING ironic.drivers.modules.ipmitool [-] IPMI power > > > > >
> > > > status > > > > > > > > > failed > >
> > > > > > > for > > > > > > > > > node
b5811c06-d5d1-41f1-87b3-2fd55ae63553 with error: > > > > > > > >
> Unexpected > > > > > > > > > error > > > >
> > > > > while > > > > > > > > > running
command. > > > > > > > > > Oct 14 11:29:01 undercloud.rdo
ironic-conductor[619]: > > > > > > > > > 2015-10-14 > >
> > > > > > > 11:29:01.740 > > > > > > > >
> 619 WARNING ironic.conductor.manager [-] During > > > > > > >
> > sync_power_state, > > > > > > > > > could > >
> > > > > > > not > > > > > > > > > get
power state for node > > > > > > > > >
b5811c06-d5d1-41f1-87b3-2fd55ae63553, > > > > > > > > > attempt
> > > > > > > > > 1 > > > > > > > >
> of > > > > > > > > > 3. Error: IPMI call failed: power
status.. > > > > > > > > > > > > > > > >
> > > > > > > > > This looks like an ipmi error, can you try to
manually run > > > > > > > > commands > > > > >
> > > using > > > > > > > > the > > > > >
> > > ipmitool and see if you get any success? It's also worth filing >
> > > > > > > a > > > > > > > > bug > >
> > > > > > with > > > > > > > > details such as
the ipmitool version, server model, drac > > > > > > > > firmware
> > > > > > > > version. > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
Thanks a lot > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > ----- Orijinal Mesaj ----- > > > > > > > > >
> > > > > > > > > Kimden: "Marius Cornea"
<mcornea(a)redhat.com> > > > > > > > > > Kime: "Esra
Celik" <celik.esra(a)tubitak.gov.tr> > > > > > > > > >
Kk: "Ignacio Bravo" <ibravo(a)ltgfederal.com>, > > > > > >
> > > rdo-list(a)redhat.com > > > > > > > > >
Gönderilenler: 13 Ekim Salı 2015 21:16:14 > > > > > > > >
> Konu: Re: Yan: Re: [Rdo-list] OverCloud deploy fails with > > > > >
> > > > error > > > > > > > > > "No > >
> > > > > > > valid > > > > > > > > > host
was found" > > > > > > > > > > > > > > >
> > > > > > > > > > > > ----- Original Message -----
> > > > > > > > > > From: "Esra Celik"
<celik.esra(a)tubitak.gov.tr> > > > > > > > > > > To:
"Marius Cornea" <mcornea(a)redhat.com> > > > > > > >
> > > Cc: "Ignacio Bravo" <ibravo(a)ltgfederal.com>, > > >
> > > > > > > rdo-list(a)redhat.com > > > > > > >
> > > Sent: Tuesday, October 13, 2015 5:02:09 PM > > > > > >
> > > > Subject: Yan: Re: [Rdo-list] OverCloud deploy fails with > >
> > > > > > > > error > > > > > > > > >
> "No > > > > > > > > > > valid > > > >
> > > > > > host was found" > > > > > > > >
> > > > > > > > > > > > During deployment they are
powering on and deploying the > > > > > > > > > > images.
> > > > > > > > > > I > > > > > > >
> > > see > > > > > > > > > > lot > > >
> > > > > > > of > > > > > > > > > >
connection error messages about ironic-python-agent but > > > > > > >
> > > ignore > > > > > > > > > > them > >
> > > > > > > > as > > > > > > > > >
> mentioned here > > > > > > > > > >
(
https://www.redhat.com/archives/rdo-list/2015-October/msg00052.html) > > > >
> > > > > > > > > > > > > > That was referring
to the introspection stage. From what I > > > > > > > > > can
> > > > > > > > > tell > > > > > > > >
> you > > > > > > > > > are > > > > > >
> > > experiencing issues during deployment as it fails to > > > >
> > > > > provision > > > > > > > > > the >
> > > > > > > > nova > > > > > > > > >
instances, can you check if during that stage the nodes get > > > > > >
> > > powered > > > > > > > > > on? > > >
> > > > > > > > > > > > > > > Make sure that
before overcloud deploy the ironic nodes are > > > > > > > > >
available > > > > > > > > > for > > > > > >
> > > provisioning (ironic node-list and check the provisioning > > >
> > > > > > state > > > > > > > > > column).
> > > > > > > > > Also check that you didn't miss any step
in the docs in > > > > > > > > > regards > > > >
> > > > > to > > > > > > > > > kernel > >
> > > > > > > and ramdisk assignment, introspection, flavor
creation(so it > > > > > > > > > matches > > > >
> > > > > the > > > > > > > > > nodes resources)
> > > > > > > > >
https://repos.fedorapeople.org/repos/openstack-m/rdo-manager-docs/liberty...
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > In instackenv.json file I do not need to
add the undercloud > > > > > > > > > > node, > > >
> > > > > > > or > > > > > > > > > > do
> > > > > > > > > > I? > > > > > > >
> > > > > > > > > > > No, the nodes details should be
enough. > > > > > > > > > > > > > > > >
> > > And which log files should I watch during deployment? > > > >
> > > > > > > > > > > > > > You can check the
openstack-ironic-conductor logs(journalctl > > > > > > > > >
-fl > > > > > > > > > -u > > > > > > >
> > openstack-ironic-conductor.service) and the logs in > > > > >
> > > > /var/log/nova. > > > > > > > > > > >
> > > > > > > > Thanks > > > > > > > >
> > Esra > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > ----- Orijinal
Mesaj -----Kimden: Marius Cornea > > > > > > > > > >
<mcornea(a)redhat.com>Kime: > > > > > > > > > > Esra
Celik <celik.esra(a)tubitak.gov.tr>Kk: Ignacio Bravo > > > > > >
> > > > <ibravo(a)ltgfederal.com>, > > > > > > > >
> > rdo-list(a)redhat.comG&ouml;nderilenler: > > > > > > >
> > > Tue, > > > > > > > > > > 13 > > >
> > > > > > > Oct > > > > > > > > > >
2015 17:25:00 +0300 (EEST)Konu: Re: [Rdo-list] OverCloud > > > > > >
> > > > deploy > > > > > > > > > > fails >
> > > > > > > > > with > > > > > > > >
> > error "No valid host was found" > > > > > > >
> > > > > > > > > > > > > ----- Original Message
-----> From: "Esra Celik" > > > > > > > > > >
<celik.esra(a)tubitak.gov.tr>> > > > > > > > > > >
To: "Ignacio Bravo" <ibravo(a)ltgfederal.com>> Cc: > > > >
> > > > > > rdo-list(a)redhat.com> > > > > > > >
> > > Sent: > > > > > > > > > > Tuesday, October
13, 2015 3:47:57 PM> Subject: Re: > > > > > > > > > >
[Rdo-list] > > > > > > > > > > OverCloud > > > >
> > > > > > deploy fails with error "No valid host was
found"> > > > > > > > > > > > > > Actually
> > > > > > > > > > I > > > > > > >
> > > re-installed the OS for Undercloud before deploying. > > > >
> > > > > > However > > > > > > > > > > I
> > > > > > > > > > did> > > > > > >
> > > > not > > > > > > > > > > re-install the
OS in Compute and Controller nodes.. I will > > > > > > > > >
> reinstall> > > > > > > > > > > basic > > >
> > > > > > > OS for them too, and retry.. > > > > >
> > > > > > > > > > > > > > > You don't
need to reinstall the OS on the controller and > > > > > > > >
> > compute, > > > > > > > > > > they > > >
> > > > > > > will > > > > > > > > > >
get the image served by the undercloud. I'd recommend that > > > > >
> > > > > during > > > > > > > > > >
deployment > > > > > > > > > > you watch the servers console
and make sure they get > > > > > > > > > > powered > >
> > > > > > > > on, > > > > > > > > >
> pxe > > > > > > > > > > boot, > > > > >
> > > > > and actually get the image deployed. > > > > >
> > > > > > > > > > > > > > > Thanks >
> > > > > > > > > > > > > > > > > >
> > Thanks> > > > Esra ÇELİK> TÜBİTAK BİLGEM>
> > > > > > > > > > >
www.bilgem.tubitak.gov.tr>
celik.esra(a)tubitak.gov.tr> > > > > > > > > > > > >
> Kimden: > > > > > > > > > > > "Ignacio >
> > > > > > > > > > Bravo"
<ibravo(a)ltgfederal.com>> Kime: "Esra Celik" > > > > >
> > > > > > <celik.esra(a)tubitak.gov.tr>> Kk:
rdo-list(a)redhat.com> > > > > > > > > > > >
Gönderilenler: > > > > > > > > > > > 13 Ekim
Salı 2015 16:36:06> Konu: Re: [Rdo-list] > > > > > > > > >
> > OverCloud > > > > > > > > > > > deploy >
> > > > > > > > > > fails > > > > > > >
> > > > with error "No valid host was> found"> > Esra,>
> I > > > > > > > > > > > encountered > > >
> > > > > > > > the > > > > > > > > >
> > same > > > > > > > > > > > problem after
deleting the stack and re-deploying.> > It > > > > > > > >
> > > turns > > > > > > > > > > > out > >
> > > > > > > > > that > > > > > > > >
> > > 'heat stack-delete overcloud’ does remove the nodes > >
> > > > > > > > > from> > > > > > > >
> > > > ‘nova list’ and one would assume that the >
> > > > > > > > > > baremetal > > > > > >
> > > > > servers > > > > > > > > > > >
are now ready to> be used for the next stack, but when > > > > > >
> > > > > redeploying, > > > > > > > > > >
> I > > > > > > > > > > > get > > > > >
> > > > > > the same message of> not enough hosts available.> >
You > > > > > > > > > > > can > > > > >
> > > > > > look > > > > > > > > > > >
into > > > > > > > > > > > the > > > > >
> > > > > > nova logs and it mentions something about ‘node
xxx > > > > > > > > > > > is> > > > > >
> > > > > > already > > > > > > > > > >
> associated with UUID yyyy’ and ‘I tried 3 > > > >
> > > > > > > times > > > > > > > > > >
> and > > > > > > > > > > > I’m > >
> > > > > > > > > giving up’.> The issue is that
the UUID yyyy > > > > > > > > > > > belonged > >
> > > > > > > > > to > > > > > > > >
> > > a > > > > > > > > > > > prior > >
> > > > > > > > > unsuccessful deployment.> >
I’m now redeploying the > > > > > > > > > > >
basic > > > > > > > > > > > OS > > > > >
> > > > > > to > > > > > > > > > > >
start from scratch again.> > IB> > __> Ignacio Bravo> LTG > > >
> > > > > > > > Federal, > > > > > > > >
> > > Inc> > > > > > > > > > > >
www.ltgfederal.com> Office: (703) 951-7760> > > > On Oct > > >
> > > > > > > > 13, > > > > > > > > >
> > 2015, > > > > > > > > > > > at > > >
> > > > > > > > 9:25 > > > > > > > > >
> > AM, Esra Celik < celik.esra(a)tubitak.gov.tr > wrote:> > > >
> > > > > > > > > > Hi > > > > > > >
> > > > all,> > > > > > > > > > > > >
> > > > > > > > > > > OverCloud deploy fails with error
"No valid host was > > > > > > > > > > >
found"> > > > > > > > > > > > > > > >
> > > > > > > > [stack@undercloud ~]$ openstack overcloud deploy
> > > > > > > > > > > --templates> > > > >
> > > > > > > Deploying > > > > > > > > >
> > templates in the directory> > > > > > > > > > >
> /usr/share/openstack-tripleo-heat-templates> > > > > > > >
> > > > Stack failed with status: Resource CREATE failed: > > > >
> > > > > > > resources.Compute:> > > > > > >
> > > > > ResourceInError: resources[0].resources.NovaCompute: Went >
> > > > > > > > > > to > > > > > > >
> > > > status > > > > > > > > > > >
ERROR> > > > > > > > > > > > due to "Message: No
valid host was found. There are not > > > > > > > > > > >
enough > > > > > > > > > > > hosts> > > >
> > > > > > > > available., Code: 500"> Heat Stack create
failed.> > Here > > > > > > > > > > > are > >
> > > > > > > > > some > > > > > > > >
> > > logs:> > > > > > > > > > > > > Every
2.0s: heat resource-list -n 5 overcloud | grep -v > > > > > > > >
> > > > COMPLETE > > > > > > > > > > > >
Tue > > > > > > > > > > > > Oct > > > >
> > > > > > > 13> 16:18:17 2015> > > > > > >
> > > > > >
+-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+>
> > > > > > > > > > > | resource_name |
physical_resource_id | resource_type | > > > > > > > > > >
> | resource_status > > > > > > > > > > > |> |
updated_time | stack_name |> > > > > > > > > > > >
+-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+>
> > > > > > > > > > > | Compute |
e33b6b1e-8740-4ded-ad7f-720617a03393 | > > > > > > > > > >
> | OS::Heat::ResourceGroup > > > > > > > > > > >
|> | CREATE_FAILED | 2015-10-13T10:20:36 | overcloud |> | > > > > >
> > > > > > |> | Controller > > > > > > > >
> > > |> | | > > > > > > > > > > >
116c57ff-debb-4c12-92e1-e4163b67dc17 | > > > > > > > > > >
> OS::Heat::ResourceGroup> > > > > > > > > > > > |
> > > > > > > > > > > | > > > > > >
> > > > > CREATE_FAILED | 2015-10-13T10:20:36 | overcloud |> | 0 | >
> > > > > > > > > > 342a9023-de8f-4b5b-b3ec-498d99b56dc4 |
> > > > > > > > > > > OS::TripleO::Controller > >
> > > > > > > > > |> > > > > > > > >
> > > | > > > > > > > > > > > CREATE_IN_PROGRESS
| 2015-10-13T10:20:52 |> | > > > > > > > > > > >
overcloud-Controller-45bbw24xxhxs |> | 0 | > > > > > > > > >
> > e420a7bd-86f8-4cc1-b6a0-5ba8d1412453 | > > > > > > > >
> > > OS::TripleO::Compute > > > > > > > > > > >
|> > > > > > > > > > > > | > > > > >
> > > > > > CREATE_FAILED | 2015-10-13T10:20:54 | > > > >
> > > > > > > overcloud-Compute-vqk632ysg64r > > > > >
> > > > > > |> > > > > > > > > > > >
| > > > > > > > > > > > Controller |
2e9ac712-0566-49b5-958f-c3e151bb24d7 | > > > > > > > > > >
> OS::Nova::Server > > > > > > > > > > > |> >
> > > > > > > > > > | > > > > > > >
> > > > CREATE_IN_PROGRESS | 2015-10-13T10:20:54 |> | > > > >
> > > > > > > overcloud-Controller-45bbw24xxhxs-0-3vyhjiak2rsk |>
| > > > > > > > > > > > NovaCompute > > > >
> > > > > > > | > > > > > > > > > >
> 96efee56-81cb-46af-beef-84f4a3af761a | OS::Nova::Server > > > > > >
> > > > > |> > > > > > > > > > > > |
> > > > > > > > > > > CREATE_FAILED > > > >
> > > > > > > | 2015-10-13T10:20:56 |> | > > > > >
> > > > > > | overcloud-Compute-vqk632ysg64r-0-32nalzkofmef > >
> > > > > > > > > |> > > > > > > > >
> > >
+-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+>
> > > > > > > > > > > > > [stack@undercloud ~]$
heat resource-show overcloud > > > > > > > > > > > >
> Compute> > > > > > > > > > > >
+------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+>
> > > > > > > > > > > | Property | Value |> > >
> > > > > > > > >
+------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+>
> > > > > > > > > > > | attributes | { |> | |
"attributes": null, |> | | > > > > > > > > > >
> | "refs": > > > > > > > > > > > | null
> > > > > > > > > > > | |> > > > > >
> > > > > > | | > > > > > > > > > > > |
| > > > > > > > > > > > | } > > > > > >
> > > > > |> | creation_time | 2015-10-13T10:20:36 |> | description
> > > > > > > > > > > |> | | > > > > >
> > > > > > |> | |> > > > > > > > > >
> > |> | | > > > > > > > > > > > |> | links
> > > > > > > > > > > |> | |> > > > >
> > > > > > > | > > > > > > > > > >
>
http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overclou...
> > > > > > > > > > > | (self) |> | | > > >
> > > > > > > >
http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overclou...
> > > > > > > > > > > | | (stack) |> | | > >
> > > > > > > > >
http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overclou...
> > > > > > > > > > > | | (nested) |> |
logical_resource_id | Compute |> | > > > > > > > > > >
> | | physical_resource_id > > > > > > > > > > > |
e33b6b1e-8740-4ded-ad7f-720617a03393 |> | required_by | > > > > > >
> > > > > ComputeAllNodesDeployment |> | | > > > > > >
> > > > > ComputeNodesPostDeployment > > > > > > >
> > > > |> > > > > > > > > > > > | >
> > > > > > > > > > | > > > > > > >
> > > > ComputeCephDeployment |> | | > > > > > > >
> > > > ComputeAllNodesValidationDeployment > > > > > > >
> > > > |> > > > > > > > > > > > | >
> > > > > > > > > > | > > > > > > >
> > > > AllNodesExtraConfig |> | | allNodesConfig |> | > > >
> > > > > > > > resource_name > > > > > > >
> > > > | > > > > > > > > > > > Compute >
> > > > > > > > > > |> > > > > > > >
> > > > | resource_status | CREATE_FAILED |> | > > > > >
> > > > > > | resource_status_reason > > > > > > >
> > > > | | > > > > > > > > > > >
resources.Compute: ResourceInError:> | > > > > > > > > >
> > resources[0].resources.NovaCompute: > > > > > > > > >
> > Went to status ERROR due to "Message:> | No valid host > > >
> > > > > > > > was > > > > > > > > >
> > found. > > > > > > > > > > > There > >
> > > > > > > > > are not enough hosts available., Code:
500"> | |> | > > > > > > > > > > >
resource_type > > > > > > > > > > > | > > > >
> > > > > > > OS::Heat::ResourceGroup |> | updated_time | >
> > > > > > > > > > 2015-10-13T10:20:36 > > > >
> > > > > > > |> > > > > > > > > > >
>
+------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+>
> > > > > > > > > > > > > > This is my
instackenv.json for 1 compute and 1 > > > > > > > > > > >
> > > control > > > > > > > > > > > > >
> node > > > > > > > > > > > > > > to >
> > > > > > > > > > > > > be > > > >
> > > > > > > deployed.> > {> "nodes": [> {>
"pm_type":"pxe_ipmitool",> > > > > > > > >
> > > "mac":[> > > > > > > > > > >
> "08:9E:01:58:CC:A1"> ],> "cpu":"4",>
"memory":"8192",> > > > > > > > > > >
> "disk":"10",> > > > > > > > > >
> > "arch":"x86_64",>
"pm_user":"root",> > > > > > > > > >
> > "pm_password":"calvin",> > > > > > >
> > > > > "pm_addr":"192.168.0.18"> },> {>
> > > > > > > > > > >
"pm_type":"pxe_ipmitool",> > > > > > > > >
> > > "mac":[> > > > > > > > > > >
> "08:9E:01:58:D0:3D"> ],> "cpu":"4",>
"memory":"8192",> > > > > > > > > > >
> "disk":"100",> > > > > > > > > >
> > "arch":"x86_64",>
"pm_user":"root",> > > > > > > > > >
> > "pm_password":"calvin",> > > > > > >
> > > > > "pm_addr":"192.168.0.19"> }> ]>
}> > > Any ideas? Thanks > > > > > > > > > > >
in > > > > > > > > > > > advance> > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > Esra ÇELİK>
TÜBİTAK BİLGEM> > > > > > > > > > > >
www.bilgem.tubitak.gov.tr> > > > > > > > > > > >
celik.esra(a)tubitak.gov.tr> > > > > > > > > > > > >
_______________________________________________> Rdo-list > > > > > >
> > > > > mailing > > > > > > > > > > >
list> > > > > > > > > > > > Rdo-list(a)redhat.com>
> > > > > > > > > > >
https://www.redhat.com/mailman/listinfo/rdo-list> > > > > > > >
> > > > > > > > > > > > > > > > To
unsubscribe: rdo-list-unsubscribe(a)redhat.com> > > > > > > > >
> > > > > > _______________________________________________> Rdo-list
> > > > > > > > > > > mailing > > > > >
> > > > > > list> > > > > > > > > > >
> Rdo-list(a)redhat.com> > > > > > > > > > > >
https://www.redhat.com/mailman/listinfo/rdo-list> > > > > > > >
> > > > > > > > > > > > > > > > To
unsubscribe: rdo-list-unsubscribe(a)redhat.com > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > > > > > > > > > > > > > > > > >
> > _______________________________________________ > > > > > >
> Rdo-list mailing list > > > > > > > Rdo-list(a)redhat.com >
> > > > > >
https://www.redhat.com/mailman/listinfo/rdo-list > >
> > > > > > > > > > > > To unsubscribe:
rdo-list-unsubscribe(a)redhat.com > > > > > > > > >