Hi all,
I'm trying to poupulate the compute dbs on controller node, but I'm getting
this error, I can't figure out why 'cause I'm able to login :
*[root@controller nova]# su -s /bin/sh -c "nova-manage db sync" nova*
*Command failed, please check log for more info*
*[root@controller nova]# more /var/log/nova/nova-manage.log*
*2016-01-14 13:11:15.269 4286 CRITICAL nova [-] OperationalError:
(_mysql_exceptions.OperationalError) (1045, "Access denied for user
'nova'@'localhost' (using password: YES)")*
*2016-01-14 13:11:15.269 4286 ERROR nova Traceback (most recent call last):*
*2016-01-14 13:11:15.269 4286 ERROR nova File "/usr/bin/nova-manage",
line 10, in <module>*
*2016-01-14 13:11:15.269 4286 ERROR nova sys.exit(main())*
but :
*[root@controller nova]# mysql -u nova -p *
*Enter password: r00tme*
*Welcome to the MariaDB monitor. Commands end with ; or \g.*
*Your MariaDB connection id is 12*
*Server version: 5.5.44-MariaDB MariaDB Server*
*Copyright (c) 2000, 2015, Oracle, MariaDB Corporation Ab and others.*
*Type 'help;' or '\h' for help. Type '\c' to clear the current
input
statement.*
*MariaDB [(none)]> *
connection string in nova.conf file is
*connection=mysql://nova:r00tme@controller/nova*
Any suggestions ?
Cheers
Gabriele
2016-01-14 11:20 GMT+00:00 Soputhi Sea <puthi(a)live.com>:
Hi,
Openstack Juno's Live Migration, I've been trying to get live-migration to
work on this version but i keep getting the same error as below.
I wonder if anybody can point me to the right direction to where to debug
the problem. Or if anybody come across this problem before please share
some ideas.
I google around for a few days already but so far I haven't got any luck.
Note: the same nova, neutron and libvirt configuration work on Icehouse
and Liberty on a different cluster, as i tested.
Thanks
Puthi
Nova Version tested: 2014.2.3 and 2014.2.4
Nova Error Log
============
2016-01-14 17:34:08.818 6173 ERROR oslo.messaging.rpc.dispatcher
[req-54581412-a194-40d5-9208-b1bf6d04f8d8 ] Exception during message
handling: A NetworkModel is required here
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher Traceback
(most recent call last):
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/oslo/messaging/rpc/dispatcher.py", line
134, in _dispatch_and_reply
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
incoming.message))
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/oslo/messaging/rpc/dispatcher.py", line
177, in _dispatch
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
return self._do_dispatch(endpoint, method, ctxt, args)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/oslo/messaging/rpc/dispatcher.py", line
123, in _do_dispatch
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
result = getattr(endpoint, method)(ctxt, **new_args)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/exception.py", line 88, in wrapped
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
payload)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/openstack/common/excutils.py", line
82, in __exit__
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
six.reraise(self.type_, self.value, self.tb)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/exception.py", line 71, in wrapped
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
return f(self, context, *args, **kw)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 335, in
decorated_function
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
kwargs['instance'], e, sys.exc_info())
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/openstack/common/excutils.py", line
82, in __exit__
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
six.reraise(self.type_, self.value, self.tb)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 323, in
decorated_function
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
return function(self, context, *args, **kwargs)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 4978, in
live_migration
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
expected_attrs=expected)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/objects/instance.py", line 300, in
_from_db_object
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
db_inst['info_cache'])
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/objects/instance_info_cache.py",
line 45, in _from_db_object
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
info_cache[field] = db_obj[field]
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/objects/base.py", line 474, in
__setitem__
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
setattr(self, name, value)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/objects/base.py", line 75, in setter
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
field_value = field.coerce(self, name, value)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/objects/fields.py", line 189, in
coerce
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
return self._type.coerce(obj, attr, value)
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher File
"/usr/lib/python2.7/site-packages/nova/objects/fields.py", line 516, in
coerce
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher raise
ValueError(_('A NetworkModel is required here'))
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
ValueError: A NetworkModel is required here
2016-01-14 17:34:08.818 6173 TRACE oslo.messaging.rpc.dispatcher
Nova Config
===========================
[DEFAULT]
rpc_backend = qpid
qpid_hostname = management-host
auth_strategy = keystone
my_ip = 10.201.171.244
vnc_enabled = True
novncproxy_host=0.0.0.0
novncproxy_port=6080
novncproxy_base_url=http://management-host:6080/vnc_auto.html
network_api_class = nova.network.neutronv2.api.API
linuxnet_interface_driver = nova.network.linux_net.LinuxOVSInterfaceDriver
firewall_driver = nova.virt.firewall.NoopFirewallDriver
vncserver_listen=0.0.0.0
vncserver_proxyclient_address=10.201.171.244
[baremetal]
[cells]
[cinder]
[conductor]
[database]
connection = mysql://nova:novadbpassword@db-host/nova
[ephemeral_storage_encryption]
[glance]
host = glance-host
port = 9292
api_servers=$host:$port
[hyperv]
[image_file_url]
[ironic]
[keymgr]
[keystone_authtoken]
auth_uri =
http://management-host:5000/v2.0
identity_uri =
http://management-host:35357
admin_user = nova
admin_tenant_name = service
admin_password = nova2014agprod2
[libvirt]
live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER,
VIR_MIGRATE_LIVE #, VIR_MIGRATE_TUNNELLED
block_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER,
VIR_MIGRATE_NON_SHARED_INC, VIR_MIGRATE_LIVE
[matchmaker_redis]
[matchmaker_ring]
[metrics]
[neutron]
url =
http://management-host:9696
admin_username = neutron
admin_password = neutronpassword
admin_tenant_name = service
admin_auth_url =
http://management-host:35357/v2.0
auth_strategy = keystone
[osapi_v3]
[rdp]
[serial_console]
[spice]
[ssl]
[trusted_computing]
[upgrade_levels]
compute=icehouse
conductor=icehouse
[vmware]
[xenserver]
[zookeeper]
Neutron Config
============
[DEFAULT]
auth_strategy = keystone
rpc_backend = neutron.openstack.common.rpc.impl_qpid
qpid_hostname = management-host
core_plugin = ml2
service_plugins = router
dhcp_lease_duration = 604800
dhcp_agents_per_network = 3
[matchmaker_redis]
[matchmaker_ring]
[quotas]
[agent]
[keystone_authtoken]
auth_uri =
http://management-host:5000
identity_uri =
http://management-host:35357
admin_tenant_name = service
admin_user = neutron
admin_password = neutronpassword
auth_host = management-host
auth_protocol = http
auth_port = 35357
[database]
[service_providers]
service_provider=LOADBALANCER:Haproxy:neutron.services.loadbalancer.drivers.haproxy.plugin_driver.HaproxyOnHostPluginDriver:default
service_provider=VPN:openswan:neutron.services.vpn.service_drivers.ipsec.IPsecVPNDriver:default
Neutron Plugin
============
[ml2]
type_drivers = local,flat
mechanism_drivers = openvswitch
[ml2_type_flat]
flat_networks = physnet3
[ml2_type_vlan]
[ml2_type_gre]
tunnel_id_ranges = 1:1000
[ml2_type_vxlan]
[securitygroup]
firewall_driver = neutron.agent.firewall.NoopFirewallDriver
enable_security_group = False
[ovs]
enable_tunneling = False
local_ip = 10.201.171.244
network_vlan_ranges = physnet3
bridge_mappings = physnet3:br-bond0
Libvirt Config
===========
/etc/sysconfig/libvirtd
Uncomment
LIBVIRTD_ARGS="--listen"
/etc/libvirt/libvirtd.conf
listen_tls = 0
listen_tcp = 1
auth_tcp = “none”
_______________________________________________
Rdo-list mailing list
Rdo-list(a)redhat.com
https://www.redhat.com/mailman/listinfo/rdo-list
To unsubscribe: rdo-list-unsubscribe(a)redhat.com