Hi all!

I was able to analyze the attached log files and I hope that the results may help you understand what's going wrong with instance creation.
You can find Log_Tool's unique exported Error blocks here: http://paste.openstack.org/show/795356/

Some statistics and problematical messages:
##### Statistics - Number of Errors/Warnings per Standard OSP log since: 2020-06-30 12:30:00 #####
Total_Number_Of_Errors --> 9
/home/ashtempl/Ruslanas/controller/neutron/server.log --> 1
/home/ashtempl/Ruslanas/compute/stdouts/ovn_controller.log --> 1
/home/ashtempl/Ruslanas/compute/nova/nova-compute.log --> 7


nova-compute.log
default default] Error launching a defined domain with XML: <domain type='kvm'>
368-2020-06-30 12:30:10.815 7 ERROR nova.compute.manager [req-87bef18f-ad3d-4147-a1b3-196b5b64b688 7bdb8c3bf8004f98aae1b16d938ac09b 69134106b56941698e58c61...
70dc50f] Instance failed to spawn: libvirt.libvirtError: internal error: qemu unexpectedly closed the monitor: 2020-06-30T10:30:10.182675Z qemu-kvm: error: failed to set MSR 0...
he monitor: 2020-06-30T10:30:10.182675Z qemu-kvm: error: failed to set MSR 0x48e to 0xfff9fffe04006172
_msrs: Assertion `ret == cpu->kvm_msr_buf->nmsrs' failed.
 [instance: 128f372c-cb2e-47d9-b1bf-ce17270dc50f] Traceback (most recent call last):
375-2020-06-30 12:30:10.815 7 ERROR nova.compute.manager [instance: 128f372c-cb2e-47d9-b1bf-ce17270dc50f]   File "/usr/lib/python3.6/site-packages/nova/vir...


server.log 
5821c815-d213-498d-9394-fe25c6849918', 'status': 'failed', 'code': 422} returned with failed status

ovn_controller.log
272-2020-06-30T12:30:10.126079625+02:00 stderr F 2020-06-30T10:30:10Z|00247|patch|WARN|Bridge 'br-ex' not found for network 'datacentre'

Thanks!

On Tue, Jun 30, 2020 at 2:13 PM Ruslanas Gžibovskis <ruslanas@lpic.lt> wrote:
attaching logs here. let's see if it will work.

On Tue, 30 Jun 2020 at 12:55, Ruslanas Gžibovskis <ruslanas@lpic.lt> wrote:
hi all,

I am back, had some issues with MTU.
Now looks good, at least deployment part.

So I have installed back what I had, and still failing at same point as in first message.

I have tried to use: LogTool, how to use it? well, I launched it, but it always say [0] detailed output:
  File "./PyTool.py", line 596, in <module>
    random_node=random.choice(overcloud_nodes)

I do not get, how to make it work, should it get from stackrc ? as I see in 
    overcloud_nodes = []
    all_nodes = exec_command_line_command('source ' + source_rc_file_path + 'stackrc;openstack server list -f json')[


On Wed, 24 Jun 2020 at 20:02, Arkady Shtempler <ashtempl@redhat.com> wrote:
Hi Ruslanas!

Is it possible to get all logs under /var/log/containers somehow?

Thanks!

On Wed, Jun 24, 2020 at 2:18 AM Ruslanas Gžibovskis <ruslanas@lpic.lt> wrote:
Hi Alfredo,

Compute nodes are baremetal or virtualized?, I've seen similar bug reports when using nested virtualization in other OSes.
baremetal. Dell R630 if to be VERY precise. 



When using podman, the recommended way to restart containers is using systemd:

 
Thank you, I will try. I also modified a file, and it looked like it relaunched podman container once config was changed. Either way, if I understand Linux config correctly, the default value for user and group is root, if commented out:
#user = "root"
#group = "root"

also in some logs, I saw, that it detected, that it is not AMD CPU :) and it is really not AMD CPU.


Just for fun, it might be important, here is how my node info looks.
  ComputeS01Parameters:
    NovaReservedHostMemory: 16384
    KernelArgs: "crashkernel=no rhgb"
  ComputeS01ExtraConfig:
    nova::cpu_allocation_ratio: 4.0
    nova::compute::libvirt::rx_queue_size: 1024
    nova::compute::libvirt::tx_queue_size: 1024
    nova::compute::resume_guests_state_on_host_boot: true
_______________________________________________
users mailing list
users@lists.rdoproject.org
http://lists.rdoproject.org/mailman/listinfo/users

To unsubscribe: users-unsubscribe@lists.rdoproject.org


--
Ruslanas Gžibovskis
+370 6030 7030


--
Ruslanas Gžibovskis
+370 6030 7030