Hi,
After deleting a stack and running the deployment again I get No hosts
available for the compute node.
/var/log/nova/nova-scheduler shows the following
2017-03-27 16:34:56.510 19784 DEBUG
nova.scheduler.filters.compute_capabilities_filter
[req-12df93c6-1226-4292-9938-e6ace00f88cc
eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
(
edge-rhos-director.vasonanetworks.com
<
http://edge-rhos-director.vasonanetworks.com>,
82d68395-d33e-4332-99ff-e48daa38861a) ram: 0MB disk: 0MB io_ops: 0
instances: 0 fails extra_spec requirements. 'compute' does not match
'control' _satisfies_extra_specs
/usr/lib/python2.7/site-packages/nova/scheduler/filters/compute_capabilities_filter.py:103
2017-03-27 16:34:56.511 19784 DEBUG
nova.scheduler.filters.compute_capabilities_filter
[req-12df93c6-1226-4292-9938-e6ace00f88cc
eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
(
edge-rhos-director.vasonanetworks.com
<
http://edge-rhos-director.vasonanetworks.com>,
82d68395-d33e-4332-99ff-e48daa38861a) ram: 0MB disk: 0MB io_ops: 0
instances: 0 fails instance_type extra_specs requirements host_passes
/usr/lib/python2.7/site-packages/nova/scheduler/filters/compute_capabilities_filter.py:112
2017-03-27 16:34:56.512 19784 INFO nova.filters
[req-12df93c6-1226-4292-9938-e6ace00f88cc
eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
Filter ComputeCapabilitiesFilter returned 0 hosts
2017-03-27 16:34:56.512 19784 DEBUG nova.filters
[req-12df93c6-1226-4292-9938-e6ace00f88cc
eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
Filtering removed all hosts for the request with instance ID
'4fa2af02-8c0b-4440-aa3b-10ffe2016d48'. Filter results: [('RetryFilter',
[(u'edge-rhos-director.vasonanetworks.com
<
http://edge-rhos-director.vasonanetworks.com>';,
u'82d68395-d33e-4332-99ff-e48daa38861a')]),
('TripleOCapabilitiesFilter', [(u'edge-rhos-director.vasonanetworks.com
<
http://edge-rhos-director.vasonanetworks.com>';,
u'82d68395-d33e-4332-99ff-e48daa38861a')]),
('ComputeCapabilitiesFilter', None)] get_filtered_objects
/usr/lib/python2.7/site-packages/nova/filters.py:129
2017-03-27 16:34:56.513 19784 INFO nova.filters
[req-12df93c6-1226-4292-9938-e6ace00f88cc
eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
Filtering removed all hosts for the request with instance ID
'4fa2af02-8c0b-4440-aa3b-10ffe2016d48'. Filter results: ['RetryFilter:
(start: 2, end: 1)', 'TripleOCapabilitiesFilter: (start: 1, end: 1)',
'ComputeCapabilitiesFilter: (start: 1, end: 0)']
2017-03-27 16:34:56.514 19784 DEBUG nova.scheduler.filter_scheduler
[req-12df93c6-1226-4292-9938-e6ace00f88cc
eece59c584b8439e992306a27ab78eb7 5495048e618b444288911f261b2c10e2 - - -]
There are 0 hosts available but 1 instances requested to build.
select_destinations
/usr/lib/python2.7/site-packages/nova/scheduler/filter_scheduler.py:71
But when check the available node for compute I get
[stack@edge-rhos-director ~]$ ironic node-list --provision-state
available --maintenance false --associated false
+--------------------------------------+---------------+---------------+-------------+--------------------+-------------+
| UUID | Name | Instance UUID |
Power State | Provisioning State | Maintenance |
+--------------------------------------+---------------+---------------+-------------+--------------------+-------------+
| ded46c41-c5a5-4aa7-a1ee-3df75e6cf976 | rhos-compute0 | None |
power off | available | False |
+--------------------------------------+---------------+---------------+-------------+--------------------+-------------+
[stack@edge-rhos-director ~]$ ironic node-show rhos-compute0
+------------------------+-------------------------------------------------------------------------+
| Property | Value
|
+------------------------+-------------------------------------------------------------------------+
| chassis_uuid |
|
| clean_step | {}
|
| console_enabled | False
|
| created_at | 2017-03-23T14:41:35+00:00
|
| driver | pxe_drac
|
| driver_info | {u'deploy_kernel':
u'95a09fe8-919a-4c15-9f0e-c5936b6bcecf', |
| | u'drac_password': u'******',
u'drac_username': u'root', |
| | u'deploy_ramdisk':
u'5c5f25c4-c9bf-4859-aa33-2e4352eb5620', |
| | u'drac_host': u'192.168.61.7'}
|
| driver_internal_info | {u'agent_url': u'http://192.0.2.9:9999',
u'root_uuid_or_disk_id': |
| | u'c2bb0683-ebe8-4541-9c5c-a811d0326ae5',
u'is_whole_disk_image': False, |
| | u'drac_boot_device': {u'boot_device':
u'pxe',
u'persistent': True}, |
| | u'agent_last_heartbeat': 1490597251}
|
| extra | {u'hardware_swift_object':
u'extra_hardware-ded46c41-c5a5-4aa7-a1ee- |
| | 3df75e6cf976'}
|
| inspection_finished_at | None
|
| inspection_started_at | None
|
| instance_info | {}
|
| instance_uuid | None
|
| last_error | None
|
| maintenance | False
|
| maintenance_reason | None
|
| name | rhos-compute0
|
| network_interface |
|
| power_state | power off
|
| properties | {u'memory_mb': u'131072',
u'cpu_arch':
u'x86_64', u'local_gb': u'557', |
| | u'cpus': u'48', u'capabilities':
u'profile:compute,boot_option:local'} |
| provision_state | available
|
| provision_updated_at | 2017-03-27T13:34:51+00:00
|
| raid_config |
|
| reservation | None
|
| resource_class |
|
| target_power_state | None
|
| target_provision_state | None
|
| target_raid_config |
|
| updated_at | 2017-03-27T13:34:51+00:00
|
| uuid | ded46c41-c5a5-4aa7-a1ee-3df75e6cf976
|
+------------------------+-------------------------------------------------------------------------+
and it suites the compute flavor
[stack@edge-rhos-director ~]$ openstack flavor show compute
+----------------------------+------------------------------------------------------------------+
| Field | Value
|
+----------------------------+------------------------------------------------------------------+
| OS-FLV-DISABLED:disabled | False
|
| OS-FLV-EXT-DATA:ephemeral | 0
|
| access_project_ids | None
|
| disk | 40
|
| id | 1f44e78f-5121-4b8f-ab9d-19a64fb1d020
|
| name | compute
|
| os-flavor-access:is_public | True
|
| properties | capabilities:boot_option='local',
capabilities:profile='compute' |
| ram | 4096
|
| rxtx_factor | 1.0
|
| swap |
|
| vcpus | 1
|
+----------------------------+------------------------------------------------------------------+
Any Idea what could be the reason why it's not finding a suitable host?
Samuel
_______________________________________________
rdo-list mailing list
rdo-list(a)redhat.com
https://www.redhat.com/mailman/listinfo/rdo-list
To unsubscribe: rdo-list-unsubscribe(a)redhat.com
When I have seen this, it was because the stack didn't fully reset all
the nodes to available (assuming the stack was fully deleted). Can you
run "ironic node-list" at the undercloud command-line, and make sure
that all your nodes are set to "available"? If not, you can manually set
the provisioning mode with "ironic node-set-provision-state <UUID>
provide" for each node. Also, make sure that the stack has been fully
deleted ("heat stack-list"), although I don't think that is the issue here.
--
Dan Sneddon | Senior Principal Software Engineer
dsneddon(a)redhat.com |