[rdo-list] [tripleo] [oooq] Deployment to baremetal fails; "No valid host was found"
John Trowbridge
trown at redhat.com
Mon Jun 20 14:34:55 UTC 2016
Unfortunately, "No valid host" is the most generic error in OpenStack.
Maybe someday Nova will provide a better error message for that error,
but in the meantime we need to check the scheduler logs
(/var/log/nova/nova-scheduler.log) for more clues.
I usually grep for 'returned 0 host' to find which filter is failing to
match any hosts. That narrows the search space to investigate further
why the filter fails.
On 06/20/2016 05:30 AM, Gerard Braad wrote:
> Hi,
>
>
> as mentioned in a previous email, I am deploying baremetal nodes using
> the quickstart. At the moment I can introspect nodes correctly, but am
> unable to deploy to them.
>
> I performed the checks as mentioned in
> /tripleo-docs/doc/source/troubleshooting/troubleshooting-overcloud.rst:
>
>
> The flavor list I have is unchanged:
> [stack at undercloud ~]$ openstack flavor list
> +--------------------------------------+---------------+------+------+-----------+-------+-----------+
> | ID | Name | RAM | Disk |
> Ephemeral | VCPUs | Is Public |
> +--------------------------------------+---------------+------+------+-----------+-------+-----------+
> | 2e72ffb5-c6d7-46fd-ad75-448c0ad6855f | baremetal | 4096 | 40 |
> 0 | 1 | True |
> | 6b8b37e4-618d-4841-b5e3-f556ef27fd4d | oooq_compute | 8192 | 49 |
> 0 | 1 | True |
> | 973b58c3-8730-4b1f-96b2-fda253c15dbc | oooq_control | 8192 | 49 |
> 0 | 1 | True |
> | e22dc516-f53f-4a71-9793-29c614999801 | oooq_ceph | 8192 | 49 |
> 0 | 1 | True |
> | e3dce62a-ac8d-41ba-9f97-84554b247faa | block-storage | 4096 | 40 |
> 0 | 1 | True |
> | f5fe9ba6-cf5c-4ef3-adc2-34f3b4381915 | control | 4096 | 40 |
> 0 | 1 | True |
> | fabf81d8-44cb-4c25-8ed0-2afd124425db | compute | 4096 | 40 |
> 0 | 1 | True |
> | fe512696-2294-40cb-9d20-12415f45c1a9 | ceph-storage | 4096 | 40 |
> 0 | 1 | True |
> | ffc859af-dbfd-4e27-99fb-9ab02f4afa79 | swift-storage | 4096 | 40 |
> 0 | 1 | True |
> +--------------------------------------+---------------+------+------+-----------+-------+-----------+
>
>
> In instackenv.json the nodes have been assigned as:
> [stack at undercloud ~]$ cat instackenv.json
> {
> "nodes":[
> {
> "_comment": "ooo1",
> "pm_type":"pxe_ipmitool",
> "mac": [
> "00:26:9e:9b:c3:36"
> ],
> "cpu": "16",
> "memory": "65536",
> "disk": "370",
> "arch": "x86_64",
> "pm_user":"root",
> "pm_password":"admin",
> "pm_addr":"10.0.108.126",
> "capabilities": "profile:control,boot_option:local"
> },
> {
> "_comment": "ooo2",
> "pm_type":"pxe_ipmitool",
> "mac": [
> "00:26:9e:9c:38:a6"
> ],
> "cpu": "16",
> "memory": "65536",
> "disk": "370",
> "arch": "x86_64",
> "pm_user":"root",
> "pm_password":"admin",
> "pm_addr":"10.0.108.127",
> "capabilities": "profile:compute,boot_option:local"
> }
> ]
> }
>
> [stack at undercloud ~]$ ironic node-list
> +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> | UUID | Name | Instance UUID | Power
> State | Provisioning State | Maintenance |
> +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> | 0956df36-b642-44b8-a67f-0df88270372b | None | None | power
> off | manageable | False |
> | cc311355-f373-4e5c-99be-31ba3185639d | None | None | power
> off | manageable | False |
> +--------------------------------------+------+---------------+-------------+--------------------+-------------+
>
>
> And manually I perform the introspection:
>
> [stack at undercloud ~]$ openstack baremetal introspection bulk start
> Setting nodes for introspection to manageable...
> Starting introspection of node: 0956df36-b642-44b8-a67f-0df88270372b
> Starting introspection of node: cc311355-f373-4e5c-99be-31ba3185639d
> Waiting for introspection to finish...
> Introspection for UUID 0956df36-b642-44b8-a67f-0df88270372b finished
> successfully.
> Introspection for UUID cc311355-f373-4e5c-99be-31ba3185639d finished
> successfully.
> Setting manageable nodes to available...
> Node 0956df36-b642-44b8-a67f-0df88270372b has been set to available.
> Node cc311355-f373-4e5c-99be-31ba3185639d has been set to available.
> Introspection completed.
>
>
> [stack at undercloud ~]$ ironic node-list
> +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> | UUID | Name | Instance UUID | Power
> State | Provisioning State | Maintenance |
> +--------------------------------------+------+---------------+-------------+--------------------+-------------+
> | 0956df36-b642-44b8-a67f-0df88270372b | None | None | power
> off | available | False |
> | cc311355-f373-4e5c-99be-31ba3185639d | None | None | power
> off | available | False |
> +--------------------------------------+------+---------------+-------------+--------------------+-------------+
>
>
> After this, I start the deployment. I have defined the compute and
> control flavor to be of the respective type.
>
> [stack at undercloud ~]$ ./overcloud-deploy.sh
> <snip>
> + openstack overcloud deploy --templates --timeout 60 --control-scale
> 1 --control-flavor control --compute-scale 1 --compute-flavor compute
> --ntp-server pool.ntp.org -e /tmp/deploy_env.yaml
> Deploying templates in the directory /usr/share/openstack-tripleo-heat-templates
> 2016-06-20 08:18:33 [overcloud]: CREATE_IN_PROGRESS Stack CREATE started
> 2016-06-20 08:18:33 [HorizonSecret]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:33 [RabbitCookie]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:33 [PcsdPassword]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:33 [MysqlClusterUniquePart]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:33 [MysqlRootPassword]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:33 [Networks]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:34 [VipConfig]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:34 [HeatAuthEncryptionKey]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:18:34 [overcloud-VipConfig-i4dgmk37z6hg]:
> CREATE_IN_PROGRESS Stack CREATE started
> 2016-06-20 08:18:34 [overcloud-Networks-4pb3htxq7rkd]:
> CREATE_IN_PROGRESS Stack CREATE started
> <snip>
> 2016-06-20 08:19:06 [Controller]: CREATE_FAILED ResourceInError:
> resources.Controller: Went to status ERROR due to "Message: No valid
> host was found. There are not enough hosts available., Code: 500"
> 2016-06-20 08:19:06 [Controller]: DELETE_IN_PROGRESS state changed
> 2016-06-20 08:19:06 [NovaCompute]: CREATE_FAILED ResourceInError:
> resources.NovaCompute: Went to status ERROR due to "Message: No valid
> host was found. There are not enough hosts available., Code: 500"
> 2016-06-20 08:19:06 [NovaCompute]: DELETE_IN_PROGRESS state changed
> 2016-06-20 08:19:09 [Controller]: DELETE_COMPLETE state changed
> 2016-06-20 08:19:09 [NovaCompute]: DELETE_COMPLETE state changed
> 2016-06-20 08:19:12 [Controller]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:19:12 [NovaCompute]: CREATE_IN_PROGRESS state changed
> 2016-06-20 08:19:14 [Controller]: CREATE_FAILED ResourceInError:
> resources.Controller: Went to status ERROR due to "Message: No valid
> host was found. There are not enough hosts available., Code: 500"
> 2016-06-20 08:19:14 [Controller]: DELETE_IN_PROGRESS state changed
> 2016-06-20 08:19:14 [NovaCompute]: CREATE_FAILED ResourceInError:
> resources.NovaCompute: Went to status ERROR due to "Message: No valid
> host was found. There are not enough hosts available., Code: 500"
> 2016-06-20 08:19:14 [NovaCompute]: DELETE_IN_PROGRESS state changed
>
>
> But as you can see, the deployment fails.
>
> I check the introspection information and verify that the disk, local
> memory and cpus are matching or exceeding the flavor:
>
> [stack at undercloud ~]$ ironic node-show 0956df36-b642-44b8-a67f-0df88270372b
> +------------------------+-------------------------------------------------------------------------+
> | Property | Value
> |
> +------------------------+-------------------------------------------------------------------------+
> | chassis_uuid |
> |
> | clean_step | {}
> |
> | console_enabled | False
> |
> | created_at | 2016-06-20T05:51:17+00:00
> |
> | driver | pxe_ipmitool
> |
> | driver_info | {u'ipmi_password': u'******',
> u'ipmi_address': u'10.0.108.126', |
> | | u'ipmi_username': u'root',
> u'deploy_kernel': |
> | | u'07c794a6-b427-4e75-ba58-7c555abbf2f8',
> u'deploy_ramdisk': u'67a66b7b- |
> | | 637f-4b25-bcef-ed39ae32a1f4'}
> |
> | driver_internal_info | {}
> |
> | extra | {u'hardware_swift_object':
> u'extra_hardware-0956df36-b642-44b8-a67f- |
> | | 0df88270372b'}
> |
> | inspection_finished_at | None
> |
> | inspection_started_at | None
> |
> | instance_info | {}
> |
> | instance_uuid | None
> |
> | last_error | None
> |
> | maintenance | False
> |
> | maintenance_reason | None
> |
> | name | None
> |
> | power_state | power off
> |
> | properties | {u'memory_mb': u'65536', u'cpu_arch':
> u'x86_64', u'local_gb': u'371', |
> | | u'cpus': u'16', u'capabilities':
> u'profile:control,boot_option:local'} |
> | provision_state | available
> |
> | provision_updated_at | 2016-06-20T07:32:46+00:00
> |
> | raid_config |
> |
> | reservation | None
> |
> | target_power_state | None
> |
> | target_provision_state | None
> |
> | target_raid_config |
> |
> | updated_at | 2016-06-20T07:32:46+00:00
> |
> | uuid | 0956df36-b642-44b8-a67f-0df88270372b
> |
> +------------------------+-------------------------------------------------------------------------+
>
> And also the hypervisor stats are set, but only matching the node count.
> [stack at undercloud ~]$ nova hypervisor-stats
> +----------------------+-------+
> | Property | Value |
> +----------------------+-------+
> | count | 2 |
> | current_workload | 0 |
> | disk_available_least | 0 |
> | free_disk_gb | 0 |
> | free_ram_mb | 0 |
> | local_gb | 0 |
> | local_gb_used | 0 |
> | memory_mb | 0 |
> | memory_mb_used | 0 |
> | running_vms | 0 |
> | vcpus | 0 |
> | vcpus_used | 0 |
> +----------------------+-------+
>
> Registering the nodes as profile:baremetal has the same effect.
>
> What other parameters are used in making the decision if a node can be
> deployed to? I probably miss a small detail... what can I check to
> make sure the deployment starts?
>
> regards,
>
>
> Gerard
>
>
More information about the dev
mailing list