[rdo-users] [rdo][ussuri][TripleO][nova][kvm] libvirt.libvirtError: internal error: process exited while connecting to monitor

Arkady Shtempler ashtempl at redhat.com
Tue Jun 30 14:02:16 UTC 2020


Hi all!

I was able to analyze the attached log files and I hope that the results
may help you understand what's going wrong with instance creation.
You can find *Log_Tool's unique exported Error blocks* here:
http://paste.openstack.org/show/795356/

*Some statistics and problematical messages:*
##### Statistics - Number of Errors/Warnings per Standard OSP log since:
2020-06-30 12:30:00 #####
Total_Number_Of_Errors --> 9
/home/ashtempl/Ruslanas/controller/neutron/server.log --> 1
/home/ashtempl/Ruslanas/compute/stdouts/ovn_controller.log --> 1
/home/ashtempl/Ruslanas/compute/nova/nova-compute.log --> 7

*nova-compute.log*
*default default] Error launching a defined domain with XML: <domain
type='kvm'>*
368-2020-06-30 12:30:10.815 7 *ERROR* nova.compute.manager
[req-87bef18f-ad3d-4147-a1b3-196b5b64b688 7bdb8c3bf8004f98aae1b16d938ac09b
69134106b56941698e58c61...
70dc50f] Instance *failed* to spawn: *libvirt.libvirtError*: internal
*error*: qemu unexpectedly closed the monitor: 2020-06-30T10:30:10.182675Z
qemu-kvm: *error*: failed to set MSR 0...
he monitor: 2020-06-30T10:30:10.182675Z *qemu-kvm: error: failed to set MSR
0x48e to 0xfff9fffe04006172*
_msrs: Assertion `ret == cpu->kvm_msr_buf->nmsrs' *failed*.
 [instance: 128f372c-cb2e-47d9-b1bf-ce17270dc50f] *Traceback* (most recent
call last):
375-2020-06-30 12:30:10.815 7* ERROR* nova.compute.manager [instance:
128f372c-cb2e-47d9-b1bf-ce17270dc50f]   File
"/usr/lib/python3.6/site-packages/nova/vir...

*server.log *
5821c815-d213-498d-9394-fe25c6849918', 'status': 'failed', *'code': 422}
returned with failed status*

*ovn_controller.log*
272-2020-06-30T12:30:10.126079625+02:00 stderr F
2020-06-30T10:30:10Z|00247|patch|WARN|*Bridge 'br-ex' not found for network
'datacentre'*

Thanks!

On Tue, Jun 30, 2020 at 2:13 PM Ruslanas Gžibovskis <ruslanas at lpic.lt>
wrote:

> attaching logs here. let's see if it will work.
>
> On Tue, 30 Jun 2020 at 12:55, Ruslanas Gžibovskis <ruslanas at lpic.lt>
> wrote:
>
>> hi all,
>>
>> I am back, had some issues with MTU.
>> Now looks good, at least deployment part.
>>
>> So I have installed back what I had, and still failing at same point as
>> in first message.
>>
>> I have tried to use: LogTool, how to use it? well, I launched it, but it
>> always say [0] detailed output:
>>   File "./PyTool.py", line 596, in <module>
>>     random_node=random.choice(overcloud_nodes)
>>
>> I do not get, how to make it work, should it get from stackrc ? as I see
>> in
>>     overcloud_nodes = []
>>     all_nodes = exec_command_line_command('source ' + source_rc_file_path
>> + 'stackrc;openstack server list -f json')[
>>
>> [0] http://paste.openstack.org/show/795345/
>>
>> On Wed, 24 Jun 2020 at 20:02, Arkady Shtempler <ashtempl at redhat.com>
>> wrote:
>>
>>> Hi Ruslanas!
>>>
>>> Is it possible to get all logs under /var/log/containers somehow?
>>>
>>> Thanks!
>>>
>>> On Wed, Jun 24, 2020 at 2:18 AM Ruslanas Gžibovskis <ruslanas at lpic.lt>
>>> wrote:
>>>
>>>> Hi Alfredo,
>>>>
>>>> Compute nodes are baremetal or virtualized?, I've seen similar bug
>>>>>> reports when using nested virtualization in other OSes.
>>>>>>
>>>>> baremetal. Dell R630 if to be VERY precise.
>>>>
>>>>
>>>>
>>>>>> When using podman, the recommended way to restart containers is using
>>>>> systemd:
>>>>>
>>>>>
>>>>> https://docs.openstack.org/project-deploy-guide/tripleo-docs/latest/deployment/tips_tricks.html
>>>>>
>>>>
>>>> Thank you, I will try. I also modified a file, and it looked like it
>>>> relaunched podman container once config was changed. Either way, if I
>>>> understand Linux config correctly, the default value for user and group is
>>>> root, if commented out:
>>>> #user = "root"
>>>> #group = "root"
>>>>
>>>> also in some logs, I saw, that it detected, that it is not AMD CPU :)
>>>> and it is really not AMD CPU.
>>>>
>>>>
>>>> Just for fun, it might be important, here is how my node info looks.
>>>>   ComputeS01Parameters:
>>>>     NovaReservedHostMemory: 16384
>>>>     KernelArgs: "crashkernel=no rhgb"
>>>>   ComputeS01ExtraConfig:
>>>>     nova::cpu_allocation_ratio: 4.0
>>>>     nova::compute::libvirt::rx_queue_size: 1024
>>>>     nova::compute::libvirt::tx_queue_size: 1024
>>>>     nova::compute::resume_guests_state_on_host_boot: true
>>>> _______________________________________________
>>>> users mailing list
>>>> users at lists.rdoproject.org
>>>> http://lists.rdoproject.org/mailman/listinfo/users
>>>>
>>>> To unsubscribe: users-unsubscribe at lists.rdoproject.org
>>>>
>>>
>>
>> --
>> Ruslanas Gžibovskis
>> +370 6030 7030
>>
>
>
> --
> Ruslanas Gžibovskis
> +370 6030 7030
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.rdoproject.org/pipermail/users/attachments/20200630/c53e5b92/attachment.html>


More information about the users mailing list