Hi Pedro,
Networking looks the same to me before and after running cloud-init so
I don't why is the route error message there. Just to rule out the
networking part can you also check that name resolution works on the
nodes? Check /etc/resolv.conf and see if the nameservers answer the
DNS queries.
Thanks,
Marius
On Mon, Jun 22, 2015 at 11:28 AM, Pedro Sousa <pgsousa(a)gmail.com> wrote:
Hi Marius,
Before:
[root@ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: enp0s25: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
state UP qlen 1000
link/ether 00:a0:d1:e3:dd:ed brd ff:ff:ff:ff:ff:ff
inet 192.168.1.149/24 brd 192.168.1.255 scope global dynamic enp0s25
valid_lft 42508sec preferred_lft 42508sec
inet6 fe80::2a0:d1ff:fee3:dded/64 scope link
valid_lft forever preferred_lft forever
3: enp1s0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
UP qlen 1000
link/ether 00:a0:d1:e3:dd:ec brd ff:ff:ff:ff:ff:ff
inet 192.168.21.20/24 brd 192.168.21.255 scope global dynamic enp1s0
valid_lft 85708sec preferred_lft 85708sec
inet6 fe80::2a0:d1ff:fee3:ddec/64 scope link
valid_lft forever preferred_lft forever
4: p55p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
state DOWN qlen 1000
link/ether 00:15:17:68:7b:42 brd ff:ff:ff:ff:ff:ff
5: p55p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
state DOWN qlen 1000
link/ether 00:15:17:68:7b:43 brd ff:ff:ff:ff:ff:ff
[root@ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip r
default via 192.168.1.246 dev enp0s25
169.254.169.254 via 192.168.21.180 dev enp1s0 proto static
192.168.1.0/24 dev enp0s25 proto kernel scope link src 192.168.1.149
192.168.21.0/24 dev enp1s0 proto kernel scope link src 192.168.21.20
After:
[root@ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: enp0s25: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
state UP qlen 1000
link/ether 00:a0:d1:e3:dd:ed brd ff:ff:ff:ff:ff:ff
inet 192.168.1.149/24 brd 192.168.1.255 scope global dynamic enp0s25
valid_lft 42468sec preferred_lft 42468sec
inet6 fe80::2a0:d1ff:fee3:dded/64 scope link
valid_lft forever preferred_lft forever
3: enp1s0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
UP qlen 1000
link/ether 00:a0:d1:e3:dd:ec brd ff:ff:ff:ff:ff:ff
inet 192.168.21.20/24 brd 192.168.21.255 scope global dynamic enp1s0
valid_lft 85668sec preferred_lft 85668sec
inet6 fe80::2a0:d1ff:fee3:ddec/64 scope link
valid_lft forever preferred_lft forever
4: p55p1: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
state DOWN qlen 1000
link/ether 00:15:17:68:7b:42 brd ff:ff:ff:ff:ff:ff
5: p55p2: <NO-CARRIER,BROADCAST,MULTICAST,UP> mtu 1500 qdisc pfifo_fast
state DOWN qlen 1000
link/ether 00:15:17:68:7b:43 brd ff:ff:ff:ff:ff:ff
[root@ov-iagiwqs7y3w-0-dq2lxejobfzq-controller-57rzpswhpzf6 ~]# ip r
default via 192.168.1.246 dev enp0s25
169.254.169.254 via 192.168.21.180 dev enp1s0 proto static
192.168.1.0/24 dev enp0s25 proto kernel scope link src 192.168.1.149
192.168.21.0/24 dev enp1s0 proto kernel scope link src 192.168.21.20
Thanks,
Pedro Sousa
On Fri, Jun 19, 2015 at 9:33 PM, Marius Cornea <marius(a)remote-lab.net>
wrote:
>
> Hm..interesting. Could you post the output of 'ip a; ip r' before and
> after restarting cloud-init please?
>
> Thanks
>
> On Fri, Jun 19, 2015 at 10:16 PM, Pedro Sousa <pgsousa(a)gmail.com> wrote:
> > Hi Marius,
> >
> > Yes.
> >
> > Regards,
> > Pedro Sousa
> >
> > Em 19/06/2015 21:14, "Marius Cornea" <marius(a)remote-lab.net>
escreveu:
> >>
> >> Hi Pedro,
> >>
> >> Just to make sure I understand it correctly - you are able to SSH to
> >> the overcloud nodes and restart cloud-init ?
> >>
> >> Thanks,
> >> Marius
> >>
> >> On Fri, Jun 19, 2015 at 6:42 PM, Pedro Sousa <pgsousa(a)gmail.com>
wrote:
> >> > Hi Maruis,
> >> >
> >> > thank you for your reply. Yes the nodes can reach dhcp, but my
> >> > understanding
> >> > is that cloud init starts first than I get ip addresses.
> >> >
> >> > If I restart cloud-init I will see the routes tables being created
> >> > properly.
> >> > However, after I restart cloud-init nothing happens, I don't see
the
> >> > deployment to resume, I only see this in the logs:
> >> >
> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.357
> >> > 1518
> >> > WARNING os_collect_config.heat [-] No auth_url configured.
> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
> >> > 1518
> >> > WARNING os_collect_config.request [-] No metadata_url configured.
> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
> >> > 1518
> >> > WARNING os-collect-config [-] Source [request] Unavailable.
> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
> >> > 1518
> >> > WARNING os_collect_config.local [-]
> >> > /var/lib/os-collect-config/local-data
> >> > not found. Skipping
> >> > Jun 19 16:38:19 localhost os-collect-config: 2015-06-19 16:38:19.359
> >> > 1518
> >> > WARNING os_collect_config.local [-] No local metadata found
> >> > (['/var/lib/os-collect-config/local-data'])
> >> >
> >> > I also see that it gets stuck here:
> >> >
> >> > [stack@instack ~]$ heat resource-show overcloud Controller
> >> >
> >> >
> >> >
+------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------+
> >> > | Property | Value
> >> > |
> >> >
> >> >
> >> >
+------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------+
> >> > | attributes | {
> >> > |
> >> > | | "attributes": null,
> >> > |
> >> > | | "refs": null
> >> > |
> >> > | | }
> >> > |
> >> > | description |
> >> > |
> >> > | links |
> >> >
> >> >
> >> >
http://192.168.21.180:8004/v1/9fcf8994049b48d6af5ea6fe5323a21d/stacks/ove...
> >> > (self) |
> >> > | |
> >> >
> >> >
> >> >
http://192.168.21.180:8004/v1/9fcf8994049b48d6af5ea6fe5323a21d/stacks/ove...
> >> > (stack) |
> >> > | |
> >> >
> >> >
> >> >
http://192.168.21.180:8004/v1/9fcf8994049b48d6af5ea6fe5323a21d/stacks/ove...
> >> > (nested) |
> >> > | logical_resource_id | Controller
> >> > |
> >> > | physical_resource_id | 89b52bb2-de3b-45c5-8ec4-43c2545b8d04
> >> > |
> >> > | required_by | allNodesConfig
> >> > |
> >> > | | VipDeployment
> >> > |
> >> > | | ControllerAllNodesDeployment
> >> > |
> >> > | | ControllerIpListMap
> >> > |
> >> > | | CephClusterConfig
> >> > |
> >> > | | ControllerBootstrapNodeConfig
> >> > |
> >> > | | ControllerCephDeployment
> >> > |
> >> > | | ControllerBootstrapNodeDeployment
> >> > |
> >> > | | ControllerClusterConfig
> >> > |
> >> > | | ControllerSwiftDeployment
> >> > |
> >> > | | SwiftDevicesAndProxyConfig
> >> > |
> >> > | | ControllerNodesPostDeployment
> >> > |
> >> > | | ControllerClusterDeployment
> >> > |
> >> > | resource_name | Controller
> >> > |
> >> > | resource_status | CREATE_IN_PROGRESS
> >> > |
> >> > | resource_status_reason | state changed
> >> > |
> >> > | resource_type | OS::Heat::ResourceGroup
> >> > |
> >> > | updated_time | 2015-06-19T17:25:45Z
> >> > |
> >> >
> >> >
> >> >
+------------------------+-------------------------------------------------------------------------------------------------------------------------------------------------------+
> >> >
> >> >
> >> > Regards,
> >> > Pedro Sousa
> >> >
> >> >
> >> > On Thu, Jun 18, 2015 at 9:16 PM, Marius Cornea
> >> > <marius(a)remote-lab.net>
> >> > wrote:
> >> >>
> >> >> Hi Pedro,
> >> >>
> >> >> Can you check if the nodes can reach the dhcp server on the
> >> >> undercloud
> >> >> node? Looks to me that the nodes can't get an IP address:
> >> >>
> >> >> systemctl status neutron-dhcp-agent.service # check service status
> >> >> ip netns list # check if dhcp namespace is there
> >> >> cat /var/lib/neutron/dhcp/<namespace_uuid>/leases # check if
the
> >> >> file
> >> >> shows leases for your nodes nic mac addresses
> >> >>
> >> >> Thanks,
> >> >> Marius
> >> >>
> >> >>
> >> >> On Wed, Jun 17, 2015 at 7:17 PM, Pedro Sousa
<pgsousa(a)gmail.com>
> >> >> wrote:
> >> >> > Hi all,
> >> >> >
> >> >> > I'm trying to deploy 2 nodes, one compute and one
controller using
> >> >> > RDO.
> >> >> > However my heat stack times out and I don't understand
why, I see
> >> >> > that
> >> >> > openstack doesn't get configured, checking the logs, I see
this:
> >> >> >
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
> >> >> > +++++++++++++++++++++++Net
> >> >> > device info+++++++++++++++++++++++
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
> >> >> > +--------+------+-----------+-----------+-------------------+
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: | Device | Up
|
> >> >> > Address |
> >> >> > Mask | Hw-Address |
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
> >> >> > +--------+------+-----------+-----------+-------------------+
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: | lo: | True
|
> >> >> > 127.0.0.1 |
> >> >> > 255.0.0.0 | . |
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: | em2: | True
|
> >> >> > .
> >> >> > |
> >> >> > . | d4:ae:52:a1:cd:80 |
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: | em1: | True
|
> >> >> > .
> >> >> > |
> >> >> > . | d4:ae:52:a1:cd:7f |
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info: | p2p1: | True
|
> >> >> > .
> >> >> > |
> >> >> > . | 68:05:ca:16:db:94 |
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
> >> >> > +--------+------+-----------+-----------+-------------------+
> >> >> > Jun 17 16:44:12 localhost cloud-init: ci-info:
> >> >> > !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!Route info
> >> >> > failed!!!!!!!!!!!!!!!!!!!!!!!
> >> >> >
> >> >> >
> >> >> > os-collect-config: 2015-06-17 15:39:22.319 1663 WARNING
> >> >> > os_collect_config.cfn [-] 403 Client Error: AccessDenied
> >> >> >
> >> >> > Any hint?
> >> >> >
> >> >> > Thanks,
> >> >> > Pedro Sousa
> >> >> >
> >> >> >
> >> >> >
> >> >> > _______________________________________________
> >> >> > Rdo-list mailing list
> >> >> > Rdo-list(a)redhat.com
> >> >> >
https://www.redhat.com/mailman/listinfo/rdo-list
> >> >> >
> >> >> > To unsubscribe: rdo-list-unsubscribe(a)redhat.com
> >> >
> >> >