[Rdo-list] OverCloud deploy fails with error "No valid host was found"

Esra Celik celik.esra at tubitak.gov.tr
Thu Oct 15 08:40:46 UTC 2015


Sorry for the late reply 

ironic node-show results are below. I have my nodes power on after introspection bulk start. And I get the following warning 
Introspection didn't finish for nodes 5b28998f-4dc8-42aa-8a51-521e20b1e5ed,6f35ac24-135d-4b99-8a24-fa2b731bd218 

Doesn't seem to be the same issue with https://www.redhat.com/archives/rdo-list/2015-October/msg00174.html 




[stack at undercloud ~]$ ironic node-list 
+--------------------------------------+------+---------------+-------------+--------------------+-------------+ 
| UUID | Name | Instance UUID | Power State | Provisioning State | Maintenance | 
+--------------------------------------+------+---------------+-------------+--------------------+-------------+ 
| 5b28998f-4dc8-42aa-8a51-521e20b1e5ed | None | None | power on | available | False | 
| 6f35ac24-135d-4b99-8a24-fa2b731bd218 | None | None | power on | available | False | 
+--------------------------------------+------+---------------+-------------+--------------------+-------------+ 


[stack at undercloud ~]$ ironic node-show 5b28998f-4dc8-42aa-8a51-521e20b1e5ed 
+------------------------+-------------------------------------------------------------------------+ 
| Property | Value | 
+------------------------+-------------------------------------------------------------------------+ 
| target_power_state | None | 
| extra | {} | 
| last_error | None | 
| updated_at | 2015-10-15T08:26:42+00:00 | 
| maintenance_reason | None | 
| provision_state | available | 
| clean_step | {} | 
| uuid | 5b28998f-4dc8-42aa-8a51-521e20b1e5ed | 
| console_enabled | False | 
| target_provision_state | None | 
| provision_updated_at | 2015-10-15T08:26:42+00:00 | 
| maintenance | False | 
| inspection_started_at | None | 
| inspection_finished_at | None | 
| power_state | power on | 
| driver | pxe_ipmitool | 
| reservation | None | 
| properties | {u'memory_mb': u'8192', u'cpu_arch': u'x86_64', u'local_gb': u'10', | 
| | u'cpus': u'4', u'capabilities': u'boot_option:local'} | 
| instance_uuid | None | 
| name | None | 
| driver_info | {u'ipmi_password': u'******', u'ipmi_address': u'192.168.0.18', | 
| | u'ipmi_username': u'root', u'deploy_kernel': u'49a2c8d4-a283-4bdf-8d6f- | 
| | e83ae28da047', u'deploy_ramdisk': u'3db3dbed- | 
| | 0d88-4632-af98-8defb05ca6e2'} | 
| created_at | 2015-10-15T07:49:08+00:00 | 
| driver_internal_info | {u'clean_steps': None} | 
| chassis_uuid | | 
| instance_info | {} | 
+------------------------+-------------------------------------------------------------------------+ 


[stack at undercloud ~]$ ironic node-show 6f35ac24-135d-4b99-8a24-fa2b731bd218 
+------------------------+-------------------------------------------------------------------------+ 
| Property | Value | 
+------------------------+-------------------------------------------------------------------------+ 
| target_power_state | None | 
| extra | {} | 
| last_error | None | 
| updated_at | 2015-10-15T08:26:42+00:00 | 
| maintenance_reason | None | 
| provision_state | available | 
| clean_step | {} | 
| uuid | 6f35ac24-135d-4b99-8a24-fa2b731bd218 | 
| console_enabled | False | 
| target_provision_state | None | 
| provision_updated_at | 2015-10-15T08:26:42+00:00 | 
| maintenance | False | 
| inspection_started_at | None | 
| inspection_finished_at | None | 
| power_state | power on | 
| driver | pxe_ipmitool | 
| reservation | None | 
| properties | {u'memory_mb': u'8192', u'cpu_arch': u'x86_64', u'local_gb': u'100', | 
| | u'cpus': u'4', u'capabilities': u'boot_option:local'} | 
| instance_uuid | None | 
| name | None | 
| driver_info | {u'ipmi_password': u'******', u'ipmi_address': u'192.168.0.19', | 
| | u'ipmi_username': u'root', u'deploy_kernel': u'49a2c8d4-a283-4bdf-8d6f- | 
| | e83ae28da047', u'deploy_ramdisk': u'3db3dbed- | 
| | 0d88-4632-af98-8defb05ca6e2'} | 
| created_at | 2015-10-15T07:49:08+00:00 | 
| driver_internal_info | {u'clean_steps': None} | 
| chassis_uuid | | 
| instance_info | {} | 
+------------------------+-------------------------------------------------------------------------+ 
[stack at undercloud ~]$ 









And below I added my history for the stack user. I don't think I am doing something other than https://repos.fedorapeople.org/repos/openstack-m/rdo-manager-docs/liberty doc 







1 vi instackenv.json 
2 sudo yum -y install epel-release 
3 sudo curl -o /etc/yum.repos.d/delorean.repo http://trunk.rdoproject.org/centos7-liberty/current-passed-ci/delorean.repo 
4 sudo curl -o /etc/yum.repos.d/delorean-current.repo http://trunk.rdoproject.org/centos7-liberty/current/delorean.repo 
5 sudo sed -i 's/\[delorean\]/\[delorean-current\]/' /etc/yum.repos.d/delorean-current.repo 
6 sudo /bin/bash -c "cat <<EOF>>/etc/yum.repos.d/delorean-current.repo 

includepkgs=diskimage-builder,openstack-heat,instack,instack-undercloud,openstack-ironic,openstack-ironic-inspector,os-cloud-config,os-net-config,python-ironic-inspector-client,python-tripleoclient,tripleo-common,openstack-tripleo-heat-templates,openstack-tripleo-image-elements,openstack-tuskar-ui-extras,openstack-puppet-modules 
EOF" 
7 sudo curl -o /etc/yum.repos.d/delorean-deps.repo http://trunk.rdoproject.org/centos7-liberty/delorean-deps.repo 
8 sudo yum -y install yum-plugin-priorities 
9 sudo yum install -y python-tripleoclient 
10 cp /usr/share/instack-undercloud/undercloud.conf.sample ~/undercloud.conf 
11 vi undercloud.conf 
12 export DIB_INSTALLTYPE_puppet_modules=source 
13 openstack undercloud install 
14 source stackrc 
15 export NODE_DIST=centos7 
16 export DIB_YUM_REPO_CONF="/etc/yum.repos.d/delorean.repo /etc/yum.repos.d/delorean-deps.repo" 
17 export DIB_INSTALLTYPE_puppet_modules=source 
18 openstack overcloud image build --all 
19 ls 
20 openstack overcloud image upload 
21 openstack baremetal import --json instackenv.json 
22 openstack baremetal configure boot 
23 ironic node-list 
24 openstack baremetal introspection bulk start 
25 ironic node-list 
26 ironic node-show 5b28998f-4dc8-42aa-8a51-521e20b1e5ed 
27 ironic node-show 6f35ac24-135d-4b99-8a24-fa2b731bd218 
28 history 







Thanks 



Esra ÇELİK 
TÜBİTAK BİLGEM 
www.bilgem.tubitak.gov.tr 
celik.esra at tubitak.gov.tr 

----- Orijinal Mesaj -----

Kimden: "Marius Cornea" <mcornea at redhat.com> 
Kime: "Esra Celik" <celik.esra at tubitak.gov.tr> 
Kk: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com 
Gönderilenler: 14 Ekim Çarşamba 2015 19:40:07 
Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid host was found" 

Can you do ironic node-show for your ironic nodes and post the results? Also check the following suggestion if you're experiencing the same issue: 
https://www.redhat.com/archives/rdo-list/2015-October/msg00174.html 

----- Original Message ----- 
> From: "Esra Celik" <celik.esra at tubitak.gov.tr> 
> To: "Marius Cornea" <mcornea at redhat.com> 
> Cc: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com 
> Sent: Wednesday, October 14, 2015 3:22:20 PM 
> Subject: Re: [Rdo-list] OverCloud deploy fails with error "No valid host was found" 
> 
> 
> 
> Well in the early stage of the introspection I can see Client IP of nodes 
> (screenshot attached). But then I see continuous ironic-python-agent errors 
> (screenshot-2 attached). Errors repeat after time out.. And the nodes are 
> not powered off. 
> 
> Seems like I am stuck in introspection stage.. 
> 
> I can use ipmitool command to successfully power on/off the nodes 
> 
> 
> 
> [stack at undercloud ~]$ ipmitool -I lanplus -H 192.168.0.19 -L ADMINISTRATOR -U 
> root -R 3 -N 5 -P <password> power status 
> Chassis Power is on 
> 
> 
> [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P 
> <password> chassis power status 
> Chassis Power is on 
> [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P 
> <password> chassis power off 
> Chassis Power Control: Down/Off 
> [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P 
> <password> chassis power status 
> Chassis Power is off 
> [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P 
> <password> chassis power on 
> Chassis Power Control: Up/On 
> [stack at undercloud ~]$ ipmitool -H 192.168.0.18 -I lanplus -U root -P 
> <password> chassis power status 
> Chassis Power is on 
> 
> 
> Esra ÇELİK 
> TÜBİTAK BİLGEM 
> www.bilgem.tubitak.gov.tr 
> celik.esra at tubitak.gov.tr 
> 
> 
> ----- Orijinal Mesaj ----- 
> 
> Kimden: "Marius Cornea" <mcornea at redhat.com> 
> Kime: "Esra Celik" <celik.esra at tubitak.gov.tr> 
> Kk: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com 
> Gönderilenler: 14 Ekim Çarşamba 2015 14:59:30 
> Konu: Re: [Rdo-list] OverCloud deploy fails with error "No valid host was 
> found" 
> 
> 
> ----- Original Message ----- 
> > From: "Esra Celik" <celik.esra at tubitak.gov.tr> 
> > To: "Marius Cornea" <mcornea at redhat.com> 
> > Cc: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com 
> > Sent: Wednesday, October 14, 2015 10:49:01 AM 
> > Subject: Re: [Rdo-list] OverCloud deploy fails with error "No valid host 
> > was found" 
> > 
> > 
> > Well today I started with re-installing the OS and nothing seems wrong with 
> > undercloud installation, then; 
> > 
> > 
> > 
> > 
> > 
> > 
> > I see an error during image build 
> > 
> > 
> > [stack at undercloud ~]$ openstack overcloud image build --all 
> > ... 
> > a lot of log 
> > ... 
> > ++ cat /etc/dib_dracut_drivers 
> > + dracut -N --install ' curl partprobe lsblk targetcli tail head awk 
> > ifconfig 
> > cut expr route ping nc wget tftp grep' --kernel-cmdline 'rd.shell rd.debug 
> > rd.neednet=1 rd.driver.pre=ahci' --include /var/tmp/image.YVhwuArQ/mnt/ / 
> > --kver 3.10.0-229.14.1.el7.x86_64 --add-drivers ' virtio virtio_net 
> > virtio_blk target_core_mod iscsi_target_mod target_core_iblock 
> > target_core_file target_core_pscsi configfs' -o 'dash plymouth' 
> > /tmp/ramdisk 
> > cat: write error: Broken pipe 
> > + cp /boot/vmlinuz-3.10.0-229.14.1.el7.x86_64 /tmp/kernel 
> > + chmod o+r /tmp/kernel 
> > + trap EXIT 
> > + target_tag=99-build-dracut-ramdisk 
> > + date +%s.%N 
> > + output '99-build-dracut-ramdisk completed' 
> > ... 
> > a lot of log 
> > ... 
> 
> You can ignore that afaik, if you end up having all the required images it 
> should be ok. 
> 
> > 
> > Then, during introspection stage I see ironic-python-agent errors on nodes 
> > (screenshot attached) and the following warnings 
> > 
> 
> That looks odd. Is it showing up in the early stage of the introspection? At 
> some point it should receive an address by DHCP and the Network is 
> unreachable error should disappear. Does the introspection complete and the 
> nodes are turned off? 
> 
> > 
> > 
> > [root at localhost ~]# journalctl -fl -u openstack-ironic-conductor.service | 
> > grep -i "warning\|error" 
> > Oct 14 10:30:12 undercloud.rdo ironic-conductor[619]: 2015-10-14 
> > 10:30:12.119 
> > 619 WARNING oslo_config.cfg [req-eccf8cb5-6e93-4d8f-9a05-0e8c8d2aab7b ] 
> > Option "http_url" from group "pxe" is deprecated. Use option "http_url" 
> > from 
> > group "deploy". 
> > Oct 14 10:30:12 undercloud.rdo ironic-conductor[619]: 2015-10-14 
> > 10:30:12.119 
> > 619 WARNING oslo_config.cfg [req-eccf8cb5-6e93-4d8f-9a05-0e8c8d2aab7b ] 
> > Option "http_root" from group "pxe" is deprecated. Use option "http_root" 
> > from group "deploy". 
> > 
> > 
> > Before deployment ironic node-list: 
> > 
> 
> This is odd too as I'm expecting the nodes to be powered off before running 
> deployment. 
> 
> > 
> > 
> > [stack at undercloud ~]$ ironic node-list 
> > +--------------------------------------+------+---------------+-------------+--------------------+-------------+ 
> > | UUID | Name | Instance UUID | Power State | Provisioning State | 
> > | Maintenance | 
> > +--------------------------------------+------+---------------+-------------+--------------------+-------------+ 
> > | acfc1bb4-469d-479a-af70-c0bdd669b32d | None | None | power on | available 
> > | | 
> > | False | 
> > | b5811c06-d5d1-41f1-87b3-2fd55ae63553 | None | None | power on | available 
> > | | 
> > | False | 
> > +--------------------------------------+------+---------------+-------------+--------------------+-------------+ 
> > 
> > During deployment I get following errors 
> > 
> > [root at localhost ~]# journalctl -fl -u openstack-ironic-conductor.service | 
> > grep -i "warning\|error" 
> > Oct 14 11:29:01 undercloud.rdo ironic-conductor[619]: 2015-10-14 
> > 11:29:01.739 
> > 619 ERROR ironic.drivers.modules.ipmitool [-] IPMI Error while attempting 
> > "ipmitool -I lanplus -H 192.168.0.19 -L ADMINISTRATOR -U root -R 3 -N 5 -f 
> > /tmp/tmpSCKHIv power status"for node b5811c06-d5d1-41f1-87b3-2fd55ae63553. 
> > Error: Unexpected error while running command. 
> > Oct 14 11:29:01 undercloud.rdo ironic-conductor[619]: 2015-10-14 
> > 11:29:01.739 
> > 619 WARNING ironic.drivers.modules.ipmitool [-] IPMI power status failed 
> > for 
> > node b5811c06-d5d1-41f1-87b3-2fd55ae63553 with error: Unexpected error 
> > while 
> > running command. 
> > Oct 14 11:29:01 undercloud.rdo ironic-conductor[619]: 2015-10-14 
> > 11:29:01.740 
> > 619 WARNING ironic.conductor.manager [-] During sync_power_state, could not 
> > get power state for node b5811c06-d5d1-41f1-87b3-2fd55ae63553, attempt 1 of 
> > 3. Error: IPMI call failed: power status.. 
> > 
> 
> This looks like an ipmi error, can you try to manually run commands using the 
> ipmitool and see if you get any success? It's also worth filing a bug with 
> details such as the ipmitool version, server model, drac firmware version. 
> 
> > 
> > 
> > 
> > 
> > 
> > Thanks a lot 
> > 
> > 
> > 
> > ----- Orijinal Mesaj ----- 
> > 
> > Kimden: "Marius Cornea" <mcornea at redhat.com> 
> > Kime: "Esra Celik" <celik.esra at tubitak.gov.tr> 
> > Kk: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com 
> > Gönderilenler: 13 Ekim Salı 2015 21:16:14 
> > Konu: Re: Yan: Re: [Rdo-list] OverCloud deploy fails with error "No valid 
> > host was found" 
> > 
> > 
> > ----- Original Message ----- 
> > > From: "Esra Celik" <celik.esra at tubitak.gov.tr> 
> > > To: "Marius Cornea" <mcornea at redhat.com> 
> > > Cc: "Ignacio Bravo" <ibravo at ltgfederal.com>, rdo-list at redhat.com 
> > > Sent: Tuesday, October 13, 2015 5:02:09 PM 
> > > Subject: Yan: Re: [Rdo-list] OverCloud deploy fails with error "No valid 
> > > host was found" 
> > > 
> > > During deployment they are powering on and deploying the images. I see 
> > > lot 
> > > of 
> > > connection error messages about ironic-python-agent but ignore them as 
> > > mentioned here 
> > > (https://www.redhat.com/archives/rdo-list/2015-October/msg00052.html) 
> > 
> > That was referring to the introspection stage. From what I can tell you are 
> > experiencing issues during deployment as it fails to provision the nova 
> > instances, can you check if during that stage the nodes get powered on? 
> > 
> > Make sure that before overcloud deploy the ironic nodes are available for 
> > provisioning (ironic node-list and check the provisioning state column). 
> > Also check that you didn't miss any step in the docs in regards to kernel 
> > and ramdisk assignment, introspection, flavor creation(so it matches the 
> > nodes resources) 
> > https://repos.fedorapeople.org/repos/openstack-m/rdo-manager-docs/liberty/basic_deployment/basic_deployment_cli.html 
> > 
> > 
> > > In instackenv.json file I do not need to add the undercloud node, or do 
> > > I? 
> > 
> > No, the nodes details should be enough. 
> > 
> > > And which log files should I watch during deployment? 
> > 
> > You can check the openstack-ironic-conductor logs(journalctl -fl -u 
> > openstack-ironic-conductor.service) and the logs in /var/log/nova. 
> > 
> > > Thanks 
> > > Esra 
> > > 
> > > 
> > > ----- Orijinal Mesaj -----Kimden: Marius Cornea <mcornea at redhat.com>Kime: 
> > > Esra Celik <celik.esra at tubitak.gov.tr>Kk: Ignacio Bravo 
> > > <ibravo at ltgfederal.com>, rdo-list at redhat.comGönderilenler: Tue, 13 
> > > Oct 
> > > 2015 17:25:00 +0300 (EEST)Konu: Re: [Rdo-list] OverCloud deploy fails 
> > > with 
> > > error "No valid host was found" 
> > > 
> > > ----- Original Message -----> From: "Esra Celik" 
> > > <celik.esra at tubitak.gov.tr>> 
> > > To: "Ignacio Bravo" <ibravo at ltgfederal.com>> Cc: rdo-list at redhat.com> 
> > > Sent: 
> > > Tuesday, October 13, 2015 3:47:57 PM> Subject: Re: [Rdo-list] OverCloud 
> > > deploy fails with error "No valid host was found"> > > > Actually I 
> > > re-installed the OS for Undercloud before deploying. However I did> not 
> > > re-install the OS in Compute and Controller nodes.. I will reinstall> 
> > > basic 
> > > OS for them too, and retry.. 
> > > 
> > > You don't need to reinstall the OS on the controller and compute, they 
> > > will 
> > > get the image served by the undercloud. I'd recommend that during 
> > > deployment 
> > > you watch the servers console and make sure they get powered on, pxe 
> > > boot, 
> > > and actually get the image deployed. 
> > > 
> > > Thanks 
> > > 
> > > > Thanks> > > > Esra ÇELİK> TÜBİTAK BİLGEM> 
> > > > www.bilgem.tubitak.gov.tr> celik.esra at tubitak.gov.tr> > > Kimden: 
> > > > "Ignacio 
> > > > Bravo" <ibravo at ltgfederal.com>> Kime: "Esra Celik" 
> > > > <celik.esra at tubitak.gov.tr>> Kk: rdo-list at redhat.com> 
> > > > Gönderilenler: 
> > > > 13 Ekim Salı 2015 16:36:06> Konu: Re: [Rdo-list] OverCloud deploy fails 
> > > > with error "No valid host was> found"> > Esra,> > I encountered the 
> > > > same 
> > > > problem after deleting the stack and re-deploying.> > It turns out that 
> > > > 'heat stack-delete overcloud’ does remove the nodes from> 
> > > > ‘nova list’ and one would assume that the baremetal servers 
> > > > are now ready to> be used for the next stack, but when redeploying, I 
> > > > get 
> > > > the same message of> not enough hosts available.> > You can look into 
> > > > the 
> > > > nova logs and it mentions something about ‘node xxx is> already 
> > > > associated with UUID yyyy’ and ‘I tried 3 times and 
> > > > I’m 
> > > > giving up’.> The issue is that the UUID yyyy belonged to a prior 
> > > > unsuccessful deployment.> > I’m now redeploying the basic OS to 
> > > > start from scratch again.> > IB> > __> Ignacio Bravo> LTG Federal, Inc> 
> > > > www.ltgfederal.com> Office: (703) 951-7760> > > > On Oct 13, 2015, at 
> > > > 9:25 
> > > > AM, Esra Celik < celik.esra at tubitak.gov.tr > wrote:> > > Hi all,> > 
> > > > OverCloud deploy fails with error "No valid host was found"> > 
> > > > [stack at undercloud ~]$ openstack overcloud deploy --templates> Deploying 
> > > > templates in the directory> 
> > > > /usr/share/openstack-tripleo-heat-templates> 
> > > > Stack failed with status: Resource CREATE failed: resources.Compute:> 
> > > > ResourceInError: resources[0].resources.NovaCompute: Went to status 
> > > > ERROR> 
> > > > due to "Message: No valid host was found. There are not enough hosts> 
> > > > available., Code: 500"> Heat Stack create failed.> > Here are some 
> > > > logs:> 
> > > > > Every 2.0s: heat resource-list -n 5 overcloud | grep -v COMPLETE Tue 
> > > > > Oct 
> > > > 13> 16:18:17 2015> > 
> > > > +-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+> 
> > > > | resource_name | physical_resource_id | resource_type | 
> > > > | resource_status 
> > > > |> | updated_time | stack_name |> 
> > > > +-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+> 
> > > > | Compute | e33b6b1e-8740-4ded-ad7f-720617a03393 | 
> > > > | OS::Heat::ResourceGroup 
> > > > |> | CREATE_FAILED | 2015-10-13T10:20:36 | overcloud |> | Controller | 
> > > > 116c57ff-debb-4c12-92e1-e4163b67dc17 | OS::Heat::ResourceGroup> | | 
> > > > CREATE_FAILED | 2015-10-13T10:20:36 | overcloud |> | 0 | 
> > > > 342a9023-de8f-4b5b-b3ec-498d99b56dc4 | OS::TripleO::Controller |> | 
> > > > CREATE_IN_PROGRESS | 2015-10-13T10:20:52 |> | 
> > > > overcloud-Controller-45bbw24xxhxs |> | 0 | 
> > > > e420a7bd-86f8-4cc1-b6a0-5ba8d1412453 | OS::TripleO::Compute |> | 
> > > > CREATE_FAILED | 2015-10-13T10:20:54 | overcloud-Compute-vqk632ysg64r |> 
> > > > | 
> > > > Controller | 2e9ac712-0566-49b5-958f-c3e151bb24d7 | OS::Nova::Server |> 
> > > > | 
> > > > CREATE_IN_PROGRESS | 2015-10-13T10:20:54 |> | 
> > > > overcloud-Controller-45bbw24xxhxs-0-3vyhjiak2rsk |> | NovaCompute | 
> > > > 96efee56-81cb-46af-beef-84f4a3af761a | OS::Nova::Server |> | 
> > > > CREATE_FAILED 
> > > > | 2015-10-13T10:20:56 |> | 
> > > > | overcloud-Compute-vqk632ysg64r-0-32nalzkofmef 
> > > > |> 
> > > > +-------------------------------------------+-----------------------------------------------+---------------------------------------------------+--------------------+---------------------+---------------------------------------------------------------------------------+> 
> > > > > > [stack at undercloud ~]$ heat resource-show overcloud Compute> 
> > > > +------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+> 
> > > > | Property | Value |> 
> > > > +------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+> 
> > > > | attributes | { |> | | "attributes": null, |> | | "refs": null |> | | 
> > > > | } 
> > > > |> | creation_time | 2015-10-13T10:20:36 |> | description | |> | links 
> > > > |> | |> 
> > > > | 
> > > > http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overcloud/620ada83-fb7d-4769-b1e8-431cfbb37d70/resources/Compute> 
> > > > | (self) |> | | 
> > > > http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overcloud/620ada83-fb7d-4769-b1e8-431cfbb37d70> 
> > > > | | (stack) |> | | 
> > > > http://192.0.2.1:8004/v1/3c5942d315bc445c8f1f7b32bd445de5/stacks/overcloud-Compute-vqk632ysg64r/e33b6b1e-8740-4ded-ad7f-720617a03393> 
> > > > | | (nested) |> | logical_resource_id | Compute |> | 
> > > > | | physical_resource_id 
> > > > | e33b6b1e-8740-4ded-ad7f-720617a03393 |> | required_by | 
> > > > ComputeAllNodesDeployment |> | | ComputeNodesPostDeployment |> | | 
> > > > ComputeCephDeployment |> | | ComputeAllNodesValidationDeployment |> | | 
> > > > AllNodesExtraConfig |> | | allNodesConfig |> | resource_name | Compute 
> > > > |> 
> > > > | resource_status | CREATE_FAILED |> | resource_status_reason | 
> > > > resources.Compute: ResourceInError:> | 
> > > > resources[0].resources.NovaCompute: 
> > > > Went to status ERROR due to "Message:> | No valid host was found. There 
> > > > are not enough hosts available., Code: 500"> | |> | resource_type | 
> > > > OS::Heat::ResourceGroup |> | updated_time | 2015-10-13T10:20:36 |> 
> > > > +------------------------+------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+> 
> > > > > > > This is my instackenv.json for 1 compute and 1 control node to be 
> > > > deployed.> > {> "nodes": [> {> "pm_type":"pxe_ipmitool",> "mac":[> 
> > > > "08:9E:01:58:CC:A1"> ],> "cpu":"4",> "memory":"8192",> "disk":"10",> 
> > > > "arch":"x86_64",> "pm_user":"root",> "pm_password":"calvin",> 
> > > > "pm_addr":"192.168.0.18"> },> {> "pm_type":"pxe_ipmitool",> "mac":[> 
> > > > "08:9E:01:58:D0:3D"> ],> "cpu":"4",> "memory":"8192",> "disk":"100",> 
> > > > "arch":"x86_64",> "pm_user":"root",> "pm_password":"calvin",> 
> > > > "pm_addr":"192.168.0.19"> }> ]> }> > > Any ideas? Thanks in advance> > 
> > > > > 
> > > > Esra ÇELİK> TÜBİTAK BİLGEM> www.bilgem.tubitak.gov.tr> 
> > > > celik.esra at tubitak.gov.tr> > 
> > > > _______________________________________________> Rdo-list mailing list> 
> > > > Rdo-list at redhat.com> https://www.redhat.com/mailman/listinfo/rdo-list> 
> > > > > 
> > > > To unsubscribe: rdo-list-unsubscribe at redhat.com> > > > 
> > > > _______________________________________________> Rdo-list mailing list> 
> > > > Rdo-list at redhat.com> https://www.redhat.com/mailman/listinfo/rdo-list> 
> > > > > 
> > > > To unsubscribe: rdo-list-unsubscribe at redhat.com 
> > > 
> > 
> > 
> 
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.rdoproject.org/pipermail/dev/attachments/20151015/267c5e28/attachment.html>


More information about the dev mailing list