[Rdo-list] baremetal rdo manager overcloud deployment issues

Marius Cornea marius at remote-lab.net
Mon Jun 22 20:13:22 UTC 2015


Hi Pedro,

Networking looks the same to me before and after running cloud-init so
I don't why is the route error message there. Just to rule out the
networking part can you also check that name resolution works on the
nodes? Check /etc/resolv.conf and see if the nameservers answer the
DNS queries.

Thanks,
Marius

On Mon, Jun 22, 2015 at 11:28 AM, Pedro Sousa <pgsousa at gmail.com> wrote:
> Hi Marius,
>
> Before:
>
> [root at ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip a
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>     inet 127.0.0.1/8 scope host lo
>        valid_lft forever preferred_lft forever
>     inet6 ::1/128 scope host
>        valid_lft forever preferred_lft forever
> 2: enp0s25: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> state UP qlen 1000
>     link/ether 00:a0:d1:e3:dd:ed brd ff:ff:ff:ff:ff:ff
>     inet 192.168.1.149/24 brd 192.168.1.255 scope global dynamic enp0s25
>        valid_lft 42508sec preferred_lft 42508sec
>     inet6 fe80::2a0:d1ff:fee3:dded/64 scope link
>        valid_lft forever preferred_lft forever
> 3: enp1s0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
> UP qlen 1000
>     link/ether 00:a0:d1:e3:dd:ec brd ff:ff:ff:ff:ff:ff
>     inet 192.168.21.20/24 brd 192.168.21.255 scope global dynamic enp1s0
>        valid_lft 85708sec preferred_lft 85708sec
>     inet6 fe80::2a0:d1ff:fee3:ddec/64 scope link
>        valid_lft forever preferred_lft forever
> 4: p55p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
> state DOWN qlen 1000
>     link/ether 00:15:17:68:7b:42 brd ff:ff:ff:ff:ff:ff
> 5: p55p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
> state DOWN qlen 1000
>     link/ether 00:15:17:68:7b:43 brd ff:ff:ff:ff:ff:ff
>
> [root at ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip r
> default via 192.168.1.246 dev enp0s25
> 169.254.169.254 via 192.168.21.180 dev enp1s0  proto static
> 192.168.1.0/24 dev enp0s25  proto kernel  scope link  src 192.168.1.149
> 192.168.21.0/24 dev enp1s0  proto kernel  scope link  src 192.168.21.20
>
>
> After:
>
> [root at ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip a
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>     inet 127.0.0.1/8 scope host lo
>        valid_lft forever preferred_lft forever
>     inet6 ::1/128 scope host
>        valid_lft forever preferred_lft forever
> 2: enp0s25: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
> state UP qlen 1000
>     link/ether 00:a0:d1:e3:dd:ed brd ff:ff:ff:ff:ff:ff
>     inet 192.168.1.149/24 brd 192.168.1.255 scope global dynamic enp0s25
>        valid_lft 42468sec preferred_lft 42468sec
>     inet6 fe80::2a0:d1ff:fee3:dded/64 scope link
>        valid_lft forever preferred_lft forever
> 3: enp1s0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
> UP qlen 1000
>     link/ether 00:a0:d1:e3:dd:ec brd ff:ff:ff:ff:ff:ff
>     inet 192.168.21.20/24 brd 192.168.21.255 scope global dynamic enp1s0
>        valid_lft 85668sec preferred_lft 85668sec
>     inet6 fe80::2a0:d1ff:fee3:ddec/64 scope link
>        valid_lft forever preferred_lft forever
> 4: p55p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
> state DOWN qlen 1000
>     link/ether 00:15:17:68:7b:42 brd ff:ff:ff:ff:ff:ff
> 5: p55p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
> state DOWN qlen 1000
>     link/ether 00:15:17:68:7b:43 brd ff:ff:ff:ff:ff:ff
>
> [root at ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip r
> default via 192.168.1.246 dev enp0s25
> 169.254.169.254 via 192.168.21.180 dev enp1s0  proto static
> 192.168.1.0/24 dev enp0s25  proto kernel  scope link  src 192.168.1.149
> 192.168.21.0/24 dev enp1s0  proto kernel  scope link  src 192.168.21.20
>
>
>
> Thanks,
> Pedro Sousa
>
>
> On Fri, Jun 19, 2015 at 9:33 PM, Marius Cornea <marius at remote-lab.net>
> wrote:
>>
>> Hm..interesting. Could you post the output of 'ip a; ip r' before and
>> after restarting cloud-init please?
>>
>> Thanks
>>
>> On Fri, Jun 19, 2015 at 10:16 PM, Pedro Sousa <pgsousa at gmail.com> wrote:
>> > Hi Marius,
>> >
>> > Yes.
>> >
>> > Regards,
>> > Pedro Sousa
>> >
>> > Em 19/06/2015 21:14, "Marius Cornea" <marius at remote-lab.net> escreveu:
>> >>
>> >> Hi Pedro,
>> >>
>> >> Just to make sure I understand it correctly - you are able to SSH to
>> >> the overcloud nodes and restart cloud-init ?
>> >>
>> >> Thanks,
>> >> Marius
>> >>
>> >> On Fri, Jun 19, 2015 at 6:42 PM, Pedro Sousa <pgsousa at gmail.com> wrote:
>> >> > Hi Maruis,
>> >> >
>> >> > thank you for your reply. Yes the nodes can reach dhcp, but my
>> >> > understanding
>> >> > is that cloud init starts first than I get ip addresses.
>> >> >
>> >> > If I restart cloud-init I will see the routes tables being created
>> >> > properly.
>> >> > However, after I restart cloud-init nothing happens, I don't see the
>> >> > deployment to resume, I only see this in the logs:
>> >> >
>> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.357
>> >> > 1518
>> >> > WARNING os_collect_config.heat [-] No auth_url configured.
>> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
>> >> > 1518
>> >> > WARNING os_collect_config.request [-] No metadata_url configured.
>> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
>> >> > 1518
>> >> > WARNING os-collect-config [-] Source [request] Unavailable.
>> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
>> >> > 1518
>> >> > WARNING os_collect_config.local [-]
>> >> > /var/lib/os-collect-config/local-data
>> >> > not found. Skipping
>> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
>> >> > 1518
>> >> > WARNING os_collect_config.local [-] No local metadata found
>> >> > (['/var/lib/os-collect-config/local-data'])
>> >> >
>> >> > I also see that it gets stuck here:
>> >> >
>> >> > [stack at instack ~]$ heat resource-show overcloud Controller
>> >> >
>> >> >
>> >> > +------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------+
>> >> > | Property               | Value
>> >> > |
>> >> >
>> >> >
>> >> > +------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------+
>> >> > | attributes             | {
>> >> > |
>> >> > |                        |   "attributes": null,
>> >> > |
>> >> > |                        |   "refs": null
>> >> > |
>> >> > |                        | }
>> >> > |
>> >> > | description            |
>> >> > |
>> >> > | links                  |
>> >> >
>> >> >
>> >> > http://192.168.21.180:8004/v1/9fcf8994049b48d6af5ea6fe5323a21d/stacks/overcloud/32a0c1d4-4915-4ff9-8a8a-f8f40590fcae/resources/Controller
>> >> > (self)      |
>> >> > |                        |
>> >> >
>> >> >
>> >> > http://192.168.21.180:8004/v1/9fcf8994049b48d6af5ea6fe5323a21d/stacks/overcloud/32a0c1d4-4915-4ff9-8a8a-f8f40590fcae
>> >> > (stack)                          |
>> >> > |                        |
>> >> >
>> >> >
>> >> > http://192.168.21.180:8004/v1/9fcf8994049b48d6af5ea6fe5323a21d/stacks/overcloud-Controller-vkqyq5m4na2v/89b52bb2-de3b-45c5-8ec4-43c2545b8d04
>> >> > (nested) |
>> >> > | logical_resource_id    | Controller
>> >> > |
>> >> > | physical_resource_id   | 89b52bb2-de3b-45c5-8ec4-43c2545b8d04
>> >> > |
>> >> > | required_by            | allNodesConfig
>> >> > |
>> >> > |                        | VipDeployment
>> >> > |
>> >> > |                        | ControllerAllNodesDeployment
>> >> > |
>> >> > |                        | ControllerIpListMap
>> >> > |
>> >> > |                        | CephClusterConfig
>> >> > |
>> >> > |                        | ControllerBootstrapNodeConfig
>> >> > |
>> >> > |                        | ControllerCephDeployment
>> >> > |
>> >> > |                        | ControllerBootstrapNodeDeployment
>> >> > |
>> >> > |                        | ControllerClusterConfig
>> >> > |
>> >> > |                        | ControllerSwiftDeployment
>> >> > |
>> >> > |                        | SwiftDevicesAndProxyConfig
>> >> > |
>> >> > |                        | ControllerNodesPostDeployment
>> >> > |
>> >> > |                        | ControllerClusterDeployment
>> >> > |
>> >> > | resource_name          | Controller
>> >> > |
>> >> > | resource_status        | CREATE_IN_PROGRESS
>> >> > |
>> >> > | resource_status_reason | state changed
>> >> > |
>> >> > | resource_type          | OS::Heat::ResourceGroup
>> >> > |
>> >> > | updated_time           | 2015-06-19T17:25:45Z
>> >> > |
>> >> >
>> >> >
>> >> > +------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------+
>> >> >
>> >> >
>> >> > Regards,
>> >> > Pedro Sousa
>> >> >
>> >> >
>> >> > On Thu, Jun 18, 2015 at 9:16 PM, Marius Cornea
>> >> > <marius at remote-lab.net>
>> >> > wrote:
>> >> >>
>> >> >> Hi Pedro,
>> >> >>
>> >> >> Can you check if the nodes can reach the dhcp server on the
>> >> >> undercloud
>> >> >> node? Looks to me that the nodes can't get an IP address:
>> >> >>
>> >> >> systemctl status neutron-dhcp-agent.service # check service status
>> >> >> ip netns list # check if dhcp namespace is there
>> >> >> cat /var/lib/neutron/dhcp/<namespace_uuid>/leases # check if the
>> >> >> file
>> >> >> shows leases for your nodes nic mac addresses
>> >> >>
>> >> >> Thanks,
>> >> >> Marius
>> >> >>
>> >> >>
>> >> >> On Wed, Jun 17, 2015 at 7:17 PM, Pedro Sousa <pgsousa at gmail.com>
>> >> >> wrote:
>> >> >> > Hi all,
>> >> >> >
>> >> >> > I'm trying to deploy 2 nodes, one compute and one controller using
>> >> >> > RDO.
>> >> >> > However my heat stack times out and I don't understand why, I see
>> >> >> > that
>> >> >> > openstack doesn't get configured, checking the logs, I see this:
>> >> >> >
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
>> >> >> > +++++++++++++++++++++++Net
>> >> >> > device info+++++++++++++++++++++++
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
>> >> >> > +--------+------+-----------+-----------+-------------------+
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: | Device |  Up  |
>> >> >> > Address  |
>> >> >> > Mask   |     Hw-Address    |
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
>> >> >> > +--------+------+-----------+-----------+-------------------+
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: |  lo:   | True |
>> >> >> > 127.0.0.1 |
>> >> >> > 255.0.0.0 |         .         |
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: |  em2:  | True |
>> >> >> > .
>> >> >> > |
>> >> >> > .     | d4:ae:52:a1:cd:80 |
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: |  em1:  | True |
>> >> >> > .
>> >> >> > |
>> >> >> > .     | d4:ae:52:a1:cd:7f |
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: | p2p1:  | True |
>> >> >> > .
>> >> >> > |
>> >> >> > .     | 68:05:ca:16:db:94 |
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
>> >> >> > +--------+------+-----------+-----------+-------------------+
>> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
>> >> >> > !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!Route info
>> >> >> > failed!!!!!!!!!!!!!!!!!!!!!!!
>> >> >> >
>> >> >> >
>> >> >> > os-collect-config: 2015-06-17 15:39:22.319 1663 WARNING
>> >> >> > os_collect_config.cfn [-] 403 Client Error: AccessDenied
>> >> >> >
>> >> >> > Any hint?
>> >> >> >
>> >> >> > Thanks,
>> >> >> > Pedro Sousa
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> > _______________________________________________
>> >> >> > Rdo-list mailing list
>> >> >> > Rdo-list at redhat.com
>> >> >> > https://www.redhat.com/mailman/listinfo/rdo-list
>> >> >> >
>> >> >> > To unsubscribe: rdo-list-unsubscribe at redhat.com
>> >> >
>> >> >
>
>




More information about the dev mailing list