Solved

Redeploying didn't help.
The cause was that a Job was stuck in the DELL iDRAC, deleting the job solved the problem.

Thanks
Samuel

On Tue, Mar 28, 2017 at 7:55 PM Dan Sneddon <dsneddon@redhat.com> wrote:
It may be another unknown cause. Were you blue to redeploy after multiple delete/redeploy attempts?

Dan Sneddon  |  Senior Principal OpenStack Engineer  |  dsneddon@redhat.com

On Mar 28, 2017, at 3:52 AM, Samuel Monderer <smonderer@vasonanetworks.com> wrote:

Hi Dan,

When running ironic node-list it showed that all nodes are available.

Samuel

On Mon, Mar 27, 2017 at 7:54 PM Dan Sneddon <dsneddon@redhat.com> wrote:
On 03/27/2017 07:10 AM, Samuel Monderer wrote:
> Hi,
>
> After deleting a stack and running the deployment again I get No hosts
> available for the compute node.
> /var/log/nova/nova-scheduler shows the following
> 2017-03-27 16:34:56.510 19784 DEBUG
> nova.scheduler.filters.compute_capabilities_filter
> [req-12df93c6-1226-4292-9938-e6ace00f88cc
> eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
> (edge-rhos-director.vasonanetworks.com
> <http://edge-rhos-director.vasonanetworks.com>,
> 82d68395-d33e-4332-99ff-e48daa38861a) ram: 0MB disk: 0MB io_ops: 0
> instances: 0 fails extra_spec requirements. 'compute' does not match
> 'control' _satisfies_extra_specs
> /usr/lib/python2.7/site-packages/nova/scheduler/filters/compute_capabilities_filter.py:103
> 2017-03-27 16:34:56.511 19784 DEBUG
> nova.scheduler.filters.compute_capabilities_filter
> [req-12df93c6-1226-4292-9938-e6ace00f88cc
> eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
> (edge-rhos-director.vasonanetworks.com
> <http://edge-rhos-director.vasonanetworks.com>,
> 82d68395-d33e-4332-99ff-e48daa38861a) ram: 0MB disk: 0MB io_ops: 0
> instances: 0 fails instance_type extra_specs requirements host_passes
> /usr/lib/python2.7/site-packages/nova/scheduler/filters/compute_capabilities_filter.py:112
> 2017-03-27 16:34:56.512 19784 INFO nova.filters
> [req-12df93c6-1226-4292-9938-e6ace00f88cc
> eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
> Filter ComputeCapabilitiesFilter returned 0 hosts
> 2017-03-27 16:34:56.512 19784 DEBUG nova.filters
> [req-12df93c6-1226-4292-9938-e6ace00f88cc
> eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
> Filtering removed all hosts for the request with instance ID
> '4fa2af02-8c0b-4440-aa3b-10ffe2016d48'. Filter results: [('RetryFilter',
> [(u'edge-rhos-director.vasonanetworks.com
> <http://edge-rhos-director.vasonanetworks.com>',
> u'82d68395-d33e-4332-99ff-e48daa38861a')]),
> ('TripleOCapabilitiesFilter', [(u'edge-rhos-director.vasonanetworks.com
> <http://edge-rhos-director.vasonanetworks.com>',
> u'82d68395-d33e-4332-99ff-e48daa38861a')]),
> ('ComputeCapabilitiesFilter', None)] get_filtered_objects
> /usr/lib/python2.7/site-packages/nova/filters.py:129
> 2017-03-27 16:34:56.513 19784 INFO nova.filters
> [req-12df93c6-1226-4292-9938-e6ace00f88cc
> eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
> Filtering removed all hosts for the request with instance ID
> '4fa2af02-8c0b-4440-aa3b-10ffe2016d48'. Filter results: ['RetryFilter:
> (start: 2, end: 1)', 'TripleOCapabilitiesFilter: (start: 1, end: 1)',
> 'ComputeCapabilitiesFilter: (start: 1, end: 0)']
> 2017-03-27 16:34:56.514 19784 DEBUG nova.scheduler.filter_scheduler
> [req-12df93c6-1226-4292-9938-e6ace00f88cc
> eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
> There are 0 hosts available but 1 instances requested to build.
> select_destinations
> /usr/lib/python2.7/site-packages/nova/scheduler/filter_scheduler.py:71
>
> But when check the available node for compute I get
> [stack@edge-rhos-director ~]$ ironic node-list --provision-state
> available --maintenance false --associated false
> +--------------------------------------+---------------+---------------+-------------+--------------------+-------------+
> | UUID                                 | Name          | Instance UUID |
> Power State | Provisioning State | Maintenance |
> +--------------------------------------+---------------+---------------+-------------+--------------------+-------------+
> | ded46c41-c5a5-4aa7-a1ee-3df75e6cf976 | rhos-compute0 | None          |
> power off   | available          | False       |
> +--------------------------------------+---------------+---------------+-------------+--------------------+-------------+
>
> [stack@edge-rhos-director ~]$ ironic node-show rhos-compute0
> +------------------------+-------------------------------------------------------------------------+
> | Property               | Value
>                           |
> +------------------------+-------------------------------------------------------------------------+
> | chassis_uuid           |
>                           |
> | clean_step             | {}
>                            |
> | console_enabled        | False
>                           |
> | created_at             | 2017-03-23T14:41:35+00:00
>                           |
> | driver                 | pxe_drac
>                            |
> | driver_info            | {u'deploy_kernel':
> u'95a09fe8-919a-4c15-9f0e-c5936b6bcecf',             |
> |                        | u'drac_password': u'******',
> u'drac_username': u'root',                 |
> |                        | u'deploy_ramdisk':
> u'5c5f25c4-c9bf-4859-aa33-2e4352eb5620',             |
> |                        | u'drac_host': u'192.168.61.7'}
>                            |
> | driver_internal_info   | {u'agent_url': u'http://192.0.2.9:9999',
> u'root_uuid_or_disk_id':       |
> |                        | u'c2bb0683-ebe8-4541-9c5c-a811d0326ae5',
> u'is_whole_disk_image': False, |
> |                        | u'drac_boot_device': {u'boot_device': u'pxe',
> u'persistent': True},     |
> |                        | u'agent_last_heartbeat': 1490597251}
>                            |
> | extra                  | {u'hardware_swift_object':
> u'extra_hardware-ded46c41-c5a5-4aa7-a1ee-    |
> |                        | 3df75e6cf976'}
>                            |
> | inspection_finished_at | None
>                            |
> | inspection_started_at  | None
>                            |
> | instance_info          | {}
>                            |
> | instance_uuid          | None
>                            |
> | last_error             | None
>                            |
> | maintenance            | False
>                           |
> | maintenance_reason     | None
>                            |
> | name                   | rhos-compute0
>                           |
> | network_interface      |
>                           |
> | power_state            | power off
>                           |
> | properties             | {u'memory_mb': u'131072', u'cpu_arch':
> u'x86_64', u'local_gb': u'557',  |
> |                        | u'cpus': u'48', u'capabilities':
> u'profile:compute,boot_option:local'}  |
> | provision_state        | available
>                           |
> | provision_updated_at   | 2017-03-27T13:34:51+00:00
>                           |
> | raid_config            |
>                           |
> | reservation            | None
>                            |
> | resource_class         |
>                           |
> | target_power_state     | None
>                            |
> | target_provision_state | None
>                            |
> | target_raid_config     |
>                           |
> | updated_at             | 2017-03-27T13:34:51+00:00
>                           |
> | uuid                   | ded46c41-c5a5-4aa7-a1ee-3df75e6cf976
>                            |
> +------------------------+-------------------------------------------------------------------------+
>
> and it suites the compute flavor
> [stack@edge-rhos-director ~]$ openstack flavor show compute
> +----------------------------+------------------------------------------------------------------+
> | Field                      | Value
>                        |
> +----------------------------+------------------------------------------------------------------+
> | OS-FLV-DISABLED:disabled   | False
>                        |
> | OS-FLV-EXT-DATA:ephemeral  | 0
>                        |
> | access_project_ids         | None
>                         |
> | disk                       | 40
>                         |
> | id                         | 1f44e78f-5121-4b8f-ab9d-19a64fb1d020
>                         |
> | name                       | compute
>                        |
> | os-flavor-access:is_public | True
>                         |
> | properties                 | capabilities:boot_option='local',
> capabilities:profile='compute' |
> | ram                        | 4096
>                         |
> | rxtx_factor                | 1.0
>                        |
> | swap                       |
>                        |
> | vcpus                      | 1
>                        |
> +----------------------------+------------------------------------------------------------------+
>
> Any Idea what could be the reason why it's not finding a suitable host?
>
> Samuel
>
>
> _______________________________________________
> rdo-list mailing list
> rdo-list@redhat.com
> https://www.redhat.com/mailman/listinfo/rdo-list
>
> To unsubscribe: rdo-list-unsubscribe@redhat.com
>

When I have seen this, it was because the stack didn't fully reset all
the nodes to available (assuming the stack was fully deleted). Can you
run "ironic node-list" at the undercloud command-line, and make sure
that all your nodes are set to "available"? If not, you can manually set
the provisioning mode with "ironic node-set-provision-state <UUID>
provide" for each node. Also, make sure that the stack has been fully
deleted ("heat stack-list"), although I don't think that is the issue here.

--
Dan Sneddon         |  Senior Principal Software Engineer
dsneddon@redhat.comredhat.com/openstack
dsneddon:irc        |  @dxs:twitter