[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: xen cache colors in ARM
Hi Oleg, On 16/05/2023 14:15, Oleg Nikitenko wrote: > > > > Hello, > > Thanks a lot Michal. > > Then the next question. > When I just started my experiments with xen, Stefano mentioned that each > cache's color size is 256M. > Is it possible to extend this figure ? With 16 colors (e.g. on Cortex-A53) and 4GB of memory, roughly each color is 256M (i.e. 4GB/16 = 256M). So as you can see this figure depends on the number of colors and memory size. ~Michal > > Regards, > Oleg > > пн, 15 мая 2023 г. в 11:57, Michal Orzel <michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>: > > Hi Oleg, > > On 15/05/2023 10:51, Oleg Nikitenko wrote: > > > > > > > > Hello guys, > > > > Thanks a lot. > > After a long problem list I was able to run xen with Dom0 with a cache > color. > > One more question from my side. > > I want to run a guest with color mode too. > > I inserted a string into guest config file llc-colors = "9-13" > > I got an error > > [ 457.517004] loop0: detected capacity change from 0 to 385840 > > Parsing config from /xen/red_config.cfg > > /xen/red_config.cfg:26: config parsing error near `-colors': lexical > error > > warning: Config file looks like it contains Python code. > > warning: Arbitrary Python is no longer supported. > > warning: See https://wiki.xen.org/wiki/PythonInXlConfig > <https://wiki.xen.org/wiki/PythonInXlConfig> > <https://wiki.xen.org/wiki/PythonInXlConfig > <https://wiki.xen.org/wiki/PythonInXlConfig>> > > Failed to parse config: Invalid argument > > So this is a question. > > Is it possible to assign a color mode for the DomU by config file ? > > If so, what string should I use? > Please, always refer to the relevant documentation. In this case, for > xl.cfg: > > https://github.com/Xilinx/xen/blob/xlnx_rebase_4.17/docs/man/xl.cfg.5.pod.in#L2890 > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.17/docs/man/xl.cfg.5.pod.in#L2890> > > ~Michal > > > > > Regards, > > Oleg > > > > чт, 11 мая 2023 г. в 13:32, Oleg Nikitenko <oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>: > > > > Hi Michal, > > > > Thanks. > > This compilation previously had a name CONFIG_COLORING. > > It mixed me up. > > > > Regards, > > Oleg > > > > чт, 11 мая 2023 г. в 13:15, Michal Orzel <michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>>: > > > > Hi Oleg, > > > > On 11/05/2023 12:02, Oleg Nikitenko wrote: > > > > > > > > > > > > Hello, > > > > > > Thanks Stefano. > > > Then the next question. > > > I cloned xen repo from xilinx site > https://github.com/Xilinx/xen.git <https://github.com/Xilinx/xen.git> > <https://github.com/Xilinx/xen.git <https://github.com/Xilinx/xen.git>> > <https://github.com/Xilinx/xen.git <https://github.com/Xilinx/xen.git> > <https://github.com/Xilinx/xen.git <https://github.com/Xilinx/xen.git>>> > > > I managed to build a xlnx_rebase_4.17 branch in my > environment. > > > I did it without coloring first. I did not find any color > footprints at this branch. > > > I realized coloring is not in the xlnx_rebase_4.17 branch yet. > > This is not true. Cache coloring is in xlnx_rebase_4.17. Please > see the docs: > > > https://github.com/Xilinx/xen/blob/xlnx_rebase_4.17/docs/misc/arm/cache-coloring.rst > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.17/docs/misc/arm/cache-coloring.rst> > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.17/docs/misc/arm/cache-coloring.rst > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.17/docs/misc/arm/cache-coloring.rst>> > > > > It describes the feature and documents the required properties. > > > > ~Michal > > > > > > > > > > > вт, 9 мая 2023 г. в 22:49, Stefano Stabellini > <sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>>: > > > > > > We test Xen Cache Coloring regularly on zcu102. Every > Petalinux release > > > (twice a year) is tested with cache coloring enabled. The > last Petalinux > > > release is 2023.1 and the kernel used is this: > > > > https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS > <https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS> > <https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS > <https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS>> > <https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS > <https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS> > <https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS > <https://github.com/Xilinx/linux-xlnx/tree/xlnx_rebase_v6.1_LTS>>> > > > > > > > > > On Tue, 9 May 2023, Oleg Nikitenko wrote: > > > > Hello guys, > > > > > > > > I have a couple of more questions. > > > > Have you ever run xen with the cache coloring at Zynq > UltraScale+ MPSoC zcu102 xczu15eg ? > > > > When did you run xen with the cache coloring last time ? > > > > What kernel version did you use for Dom0 when you ran > xen with the cache coloring last time ? > > > > > > > > Regards, > > > > Oleg > > > > > > > > пт, 5 мая 2023 г. в 11:48, Oleg Nikitenko > <oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>>>: > > > > Hi Michal, > > > > > > > > Thanks. > > > > > > > > Regards, > > > > Oleg > > > > > > > > пт, 5 мая 2023 г. в 11:34, Michal Orzel > <michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>>>>: > > > > Hi Oleg, > > > > > > > > Replying, so that you do not need to wait for > Stefano. > > > > > > > > On 05/05/2023 10:28, Oleg Nikitenko wrote: > > > > > > > > > > > > > > > > > > > > Hello Stefano, > > > > > > > > > > I would like to try a xen cache color property > from this repo https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git> > <https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git>> > <https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git> > <https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git>>> > > > > <https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git> > <https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git>> > <https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git> > <https://xenbits.xen.org/git-http/xen.git > <https://xenbits.xen.org/git-http/xen.git>>>> > > > > > Could you tell whot branch I should use ? > > > > Cache coloring feature is not part of the > upstream tree and it is still under review. > > > > You can only find it integrated in the Xilinx Xen > tree. > > > > > > > > ~Michal > > > > > > > > > > > > > > Regards, > > > > > Oleg > > > > > > > > > > пт, 28 апр. 2023 г. в 00:51, Stefano Stabellini > <sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>>>: > > > > > > > > > > I am familiar with the zcu102 but I don't > know how you could possibly > > > > > generate a SError. > > > > > > > > > > I suggest to try to use ImageBuilder [1] to > generate the boot > > > > > configuration as a test because that is > known to work well for zcu102. > > > > > > > > > > [1] > https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder> > <https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder>> > <https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder> > <https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder>>> > <https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder> > <https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder>> > <https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder> > <https://gitlab.com/xen-project/imagebuilder > <https://gitlab.com/xen-project/imagebuilder>>>> > > > > > > > > > > > > > > > On Thu, 27 Apr 2023, Oleg Nikitenko wrote: > > > > > > Hello Stefano, > > > > > > > > > > > > Thanks for clarification. > > > > > > We nighter use ImageBuilder nor uboot > boot script. > > > > > > A model is zcu102 compatible. > > > > > > > > > > > > Regards, > > > > > > O. > > > > > > > > > > > > вт, 25 апр. 2023 г. в 21:21, Stefano > Stabellini <sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>>>: > > > > > > This is interesting. Are you using > Xilinx hardware by any chance? If so, > > > > > > which board? > > > > > > > > > > > > Are you using ImageBuilder to > generate your boot.scr boot script? If so, > > > > > > could you please post your > ImageBuilder config file? If not, can you > > > > > > post the source of your uboot boot > script? > > > > > > > > > > > > SErrors are supposed to be related > to a hardware failure of some kind. > > > > > > You are not supposed to be able to > trigger an SError easily by > > > > > > "mistake". I have not seen SErrors > due to wrong cache coloring > > > > > > configurations on any Xilinx board > before. > > > > > > > > > > > > The differences between Xen with > and without cache coloring from a > > > > > > hardware perspective are: > > > > > > > > > > > > - With cache coloring, the SMMU is > enabled and does address translations > > > > > > even for dom0. Without cache > coloring the SMMU could be disabled, and > > > > > > if enabled, the SMMU doesn't do > any address translations for Dom0. If > > > > > > there is a hardware failure > related to SMMU address translation it > > > > > > could only trigger with cache > coloring. This would be my normal > > > > > > suggestion for you to explore, > but the failure happens too early > > > > > > before any DMA-capable device is > programmed. So I don't think this can > > > > > > be the issue. > > > > > > > > > > > > - With cache coloring, the memory > allocation is very different so you'll > > > > > > end up using different DDR > regions for Dom0. So if your DDR is > > > > > > defective, you might only see a > failure with cache coloring enabled > > > > > > because you end up using > different regions. > > > > > > > > > > > > > > > > > > On Tue, 25 Apr 2023, Oleg Nikitenko > wrote: > > > > > > > Hi Stefano, > > > > > > > > > > > > > > Thank you. > > > > > > > If I build xen without colors > support there is not this error. > > > > > > > All the domains are booted well. > > > > > > > Hense it can not be a hardware > issue. > > > > > > > This panic arrived during > unpacking the rootfs. > > > > > > > Here I attached the boot log > xen/Dom0 without color. > > > > > > > A highlighted strings printed > exactly after the place where 1-st time panic arrived. > > > > > > > > > > > > > > Xen 4.16.1-pre > > > > > > > (XEN) Xen version 4.16.1-pre > (nole2390@(none)) (aarch64-portable-linux-gcc (GCC) 11.3.0) debug=y > > > > 2023-04-21 > > > > > > > (XEN) Latest ChangeSet: Wed Apr > 19 12:56:14 2023 +0300 git:321687b231-dirty > > > > > > > (XEN) build-id: > c1847258fdb1b79562fc710dda40008f96c0fde5 > > > > > > > (XEN) Processor: > 00000000410fd034: "ARM Limited", variant: 0x0, part 0xd03,rev 0x4 > > > > > > > (XEN) 64-bit Execution: > > > > > > > (XEN) Processor Features: > 0000000000002222 0000000000000000 > > > > > > > (XEN) Exception Levels: > EL3:64+32 EL2:64+32 EL1:64+32 EL0:64+32 > > > > > > > (XEN) Extensions: > FloatingPoint AdvancedSIMD > > > > > > > (XEN) Debug Features: > 0000000010305106 0000000000000000 > > > > > > > (XEN) Auxiliary Features: > 0000000000000000 0000000000000000 > > > > > > > (XEN) Memory Model Features: > 0000000000001122 0000000000000000 > > > > > > > (XEN) ISA Features: > 0000000000011120 0000000000000000 > > > > > > > (XEN) 32-bit Execution: > > > > > > > (XEN) Processor Features: > 0000000000000131:0000000000011011 > > > > > > > (XEN) Instruction Sets: > AArch32 A32 Thumb Thumb-2 Jazelle > > > > > > > (XEN) Extensions: > GenericTimer Security > > > > > > > (XEN) Debug Features: > 0000000003010066 > > > > > > > (XEN) Auxiliary Features: > 0000000000000000 > > > > > > > (XEN) Memory Model Features: > 0000000010201105 0000000040000000 > > > > > > > (XEN) > 0000000001260000 0000000002102211 > > > > > > > (XEN) ISA Features: > 0000000002101110 0000000013112111 0000000021232042 > > > > > > > (XEN) > 0000000001112131 0000000000011142 0000000000011121 > > > > > > > (XEN) Using SMC Calling > Convention v1.2 > > > > > > > (XEN) Using PSCI v1.1 > > > > > > > (XEN) SMP: Allowing 4 CPUs > > > > > > > (XEN) Generic Timer IRQ: phys=30 > hyp=26 virt=27 Freq: 100000 KHz > > > > > > > (XEN) GICv2 initialization: > > > > > > > (XEN) > gic_dist_addr=00000000f9010000 > > > > > > > (XEN) > gic_cpu_addr=00000000f9020000 > > > > > > > (XEN) > gic_hyp_addr=00000000f9040000 > > > > > > > (XEN) > gic_vcpu_addr=00000000f9060000 > > > > > > > (XEN) > gic_maintenance_irq=25 > > > > > > > (XEN) GICv2: Adjusting CPU > interface base to 0xf902f000 > > > > > > > (XEN) GICv2: 192 lines, 4 cpus, > secure (IID 0200143b). > > > > > > > (XEN) Using scheduler: null > Scheduler (null) > > > > > > > (XEN) Initializing null scheduler > > > > > > > (XEN) WARNING: This is > experimental software in development. > > > > > > > (XEN) Use at your own risk. > > > > > > > (XEN) Allocated console ring of > 32 KiB. > > > > > > > (XEN) CPU0: Guest atomics will > try 12 times before pausing the domain > > > > > > > (XEN) Bringing up CPU1 > > > > > > > (XEN) CPU1: Guest atomics will > try 13 times before pausing the domain > > > > > > > (XEN) CPU 1 booted. > > > > > > > (XEN) Bringing up CPU2 > > > > > > > (XEN) CPU2: Guest atomics will > try 13 times before pausing the domain > > > > > > > (XEN) CPU 2 booted. > > > > > > > (XEN) Bringing up CPU3 > > > > > > > (XEN) CPU3: Guest atomics will > try 13 times before pausing the domain > > > > > > > (XEN) Brought up 4 CPUs > > > > > > > (XEN) CPU 3 booted. > > > > > > > (XEN) smmu: /axi/smmu@fd800000: > probing hardware configuration... > > > > > > > (XEN) smmu: /axi/smmu@fd800000: > SMMUv2 with: > > > > > > > (XEN) smmu: /axi/smmu@fd800000: > stage 2 translation > > > > > > > (XEN) smmu: /axi/smmu@fd800000: > stream matching with 48 register groups, mask 0x7fff<2>smmu: > > > > /axi/smmu@fd800000: 16 context > > > > > > banks (0 > > > > > > > stage-2 only) > > > > > > > (XEN) smmu: /axi/smmu@fd800000: > Stage-2: 48-bit IPA -> 48-bit PA > > > > > > > (XEN) smmu: /axi/smmu@fd800000: > registered 29 master devices > > > > > > > (XEN) I/O virtualisation enabled > > > > > > > (XEN) - Dom0 mode: Relaxed > > > > > > > (XEN) P2M: 40-bit IPA with 40-bit > PA and 8-bit VMID > > > > > > > (XEN) P2M: 3 levels with order-1 > root, VTCR 0x0000000080023558 > > > > > > > (XEN) Scheduling granularity: > cpu, 1 CPU per sched-resource > > > > > > > (XEN) alternatives: Patching with > alt table 00000000002cc5c8 -> 00000000002ccb2c > > > > > > > (XEN) *** LOADING DOMAIN 0 *** > > > > > > > (XEN) Loading d0 kernel from boot > module @ 0000000001000000 > > > > > > > (XEN) Loading ramdisk from boot > module @ 0000000002000000 > > > > > > > (XEN) Allocating 1:1 mappings > totalling 1600MB for dom0: > > > > > > > (XEN) BANK[0] > 0x00000010000000-0x00000020000000 (256MB) > > > > > > > (XEN) BANK[1] > 0x00000024000000-0x00000028000000 (64MB) > > > > > > > (XEN) BANK[2] > 0x00000030000000-0x00000080000000 (1280MB) > > > > > > > (XEN) Grant table range: > 0x00000000e00000-0x00000000e40000 > > > > > > > (XEN) smmu: /axi/smmu@fd800000: > d0: p2maddr 0x000000087bf94000 > > > > > > > (XEN) Allocating PPI 16 for event > channel interrupt > > > > > > > (XEN) Extended region 0: > 0x81200000->0xa0000000 > > > > > > > (XEN) Extended region 1: > 0xb1200000->0xc0000000 > > > > > > > (XEN) Extended region 2: > 0xc8000000->0xe0000000 > > > > > > > (XEN) Extended region 3: > 0xf0000000->0xf9000000 > > > > > > > (XEN) Extended region 4: > 0x100000000->0x600000000 > > > > > > > (XEN) Extended region 5: > 0x880000000->0x8000000000 > > > > > > > (XEN) Extended region 6: > 0x8001000000->0x10000000000 > > > > > > > (XEN) Loading zImage from > 0000000001000000 to 0000000010000000-0000000010e41008 > > > > > > > (XEN) Loading d0 initrd from > 0000000002000000 to 0x0000000013600000-0x000000001ff3a617 > > > > > > > (XEN) Loading d0 DTB to > 0x0000000013400000-0x000000001340cbdc > > > > > > > (XEN) Initial low memory virq > threshold set at 0x4000 pages. > > > > > > > (XEN) Std. Loglevel: All > > > > > > > (XEN) Guest Loglevel: All > > > > > > > (XEN) *** Serial input to DOM0 > (type 'CTRL-a' three times to switch input) > > > > > > > (XEN) null.c:353: 0 <-- d0v0 > > > > > > > (XEN) Freed 356kB init memory. > > > > > > > (XEN) d0v0 Unhandled SMC/HVC: > 0x84000050 > > > > > > > (XEN) d0v0 Unhandled SMC/HVC: > 0x8600ff01 > > > > > > > (XEN) d0v0: vGICD: unhandled word > write 0x000000ffffffff to ICACTIVER4 > > > > > > > (XEN) d0v0: vGICD: unhandled word > write 0x000000ffffffff to ICACTIVER8 > > > > > > > (XEN) d0v0: vGICD: unhandled word > write 0x000000ffffffff to ICACTIVER12 > > > > > > > (XEN) d0v0: vGICD: unhandled word > write 0x000000ffffffff to ICACTIVER16 > > > > > > > (XEN) d0v0: vGICD: unhandled word > write 0x000000ffffffff to ICACTIVER20 > > > > > > > (XEN) d0v0: vGICD: unhandled word > write 0x000000ffffffff to ICACTIVER0 > > > > > > > [ 0.000000] Booting Linux on > physical CPU 0x0000000000 [0x410fd034] > > > > > > > [ 0.000000] Linux version > 5.15.72-xilinx-v2022.1 (oe-user@oe-host) (aarch64-portable-linux-gcc (GCC) > > > > 11.3.0, GNU ld (GNU > > > > > > Binutils) > > > > > > > 2.38.20220708) #1 SMP Tue Feb 21 > 05:47:54 UTC 2023 > > > > > > > [ 0.000000] Machine model: D14 > Viper Board - White Unit > > > > > > > [ 0.000000] Xen 4.16 support > found > > > > > > > [ 0.000000] Zone ranges: > > > > > > > [ 0.000000] DMA [mem > 0x0000000010000000-0x000000007fffffff] > > > > > > > [ 0.000000] DMA32 empty > > > > > > > [ 0.000000] Normal empty > > > > > > > [ 0.000000] Movable zone start > for each node > > > > > > > [ 0.000000] Early memory node > ranges > > > > > > > [ 0.000000] node 0: [mem > 0x0000000010000000-0x000000001fffffff] > > > > > > > [ 0.000000] node 0: [mem > 0x0000000022000000-0x0000000022147fff] > > > > > > > [ 0.000000] node 0: [mem > 0x0000000022200000-0x0000000022347fff] > > > > > > > [ 0.000000] node 0: [mem > 0x0000000024000000-0x0000000027ffffff] > > > > > > > [ 0.000000] node 0: [mem > 0x0000000030000000-0x000000007fffffff] > > > > > > > [ 0.000000] Initmem setup node > 0 [mem 0x0000000010000000-0x000000007fffffff] > > > > > > > [ 0.000000] On node 0, zone > DMA: 8192 pages in unavailable ranges > > > > > > > [ 0.000000] On node 0, zone > DMA: 184 pages in unavailable ranges > > > > > > > [ 0.000000] On node 0, zone > DMA: 7352 pages in unavailable ranges > > > > > > > [ 0.000000] cma: Reserved 256 > MiB at 0x000000006e000000 > > > > > > > [ 0.000000] psci: probing for > conduit method from DT. > > > > > > > [ 0.000000] psci: PSCIv1.1 > detected in firmware. > > > > > > > [ 0.000000] psci: Using > standard PSCI v0.2 function IDs > > > > > > > [ 0.000000] psci: Trusted OS > migration not required > > > > > > > [ 0.000000] psci: SMC Calling > Convention v1.1 > > > > > > > [ 0.000000] percpu: Embedded > 16 pages/cpu s32792 r0 d32744 u65536 > > > > > > > [ 0.000000] Detected VIPT > I-cache on CPU0 > > > > > > > [ 0.000000] CPU features: > kernel page table isolation forced ON by KASLR > > > > > > > [ 0.000000] CPU features: > detected: Kernel page table isolation (KPTI) > > > > > > > [ 0.000000] Built 1 zonelists, > mobility grouping on. Total pages: 403845 > > > > > > > [ 0.000000] Kernel command > line: console=hvc0 earlycon=xen earlyprintk=xen clk_ignore_unused fips=1 > > > > root=/dev/ram0 > > > > > > maxcpus=2 > > > > > > > [ 0.000000] Unknown kernel > command line parameters "earlyprintk=xen fips=1", will be passed to user > > > > space. > > > > > > > [ 0.000000] Dentry cache hash > table entries: 262144 (order: 9, 2097152 bytes, linear) > > > > > > > [ 0.000000] Inode-cache hash > table entries: 131072 (order: 8, 1048576 bytes, linear) > > > > > > > [ 0.000000] mem auto-init: > stack:off, heap alloc:on, heap free:on > > > > > > > [ 0.000000] mem auto-init: > clearing system memory may take some time... > > > > > > > [ 0.000000] Memory: > 1121936K/1641024K available (9728K kernel code, 836K rwdata, 2396K rodata, > 1536K > > > > init, 262K bss, > > > > > > 256944K reserved, > > > > > > > 262144K cma-reserved) > > > > > > > [ 0.000000] SLUB: HWalign=64, > Order=0-3, MinObjects=0, CPUs=2, Nodes=1 > > > > > > > [ 0.000000] rcu: Hierarchical > RCU implementation. > > > > > > > [ 0.000000] rcu: RCU event > tracing is enabled. > > > > > > > [ 0.000000] rcu: RCU > restricting CPUs from NR_CPUS=8 to nr_cpu_ids=2. > > > > > > > [ 0.000000] rcu: RCU > calculated value of scheduler-enlistment delay is 25 jiffies. > > > > > > > [ 0.000000] rcu: Adjusting > geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 > > > > > > > [ 0.000000] NR_IRQS: 64, > nr_irqs: 64, preallocated irqs: 0 > > > > > > > [ 0.000000] Root IRQ handler: > gic_handle_irq > > > > > > > [ 0.000000] arch_timer: cp15 > timer(s) running at 100.00MHz (virt). > > > > > > > [ 0.000000] clocksource: > arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x171024e7e0, > > > > max_idle_ns: 440795205315 ns > > > > > > > [ 0.000000] sched_clock: 56 > bits at 100MHz, resolution 10ns, wraps every 4398046511100ns > > > > > > > [ 0.000258] Console: colour > dummy device 80x25 > > > > > > > [ 0.310231] printk: console > [hvc0] enabled > > > > > > > [ 0.314403] Calibrating delay > loop (skipped), value calculated using timer frequency.. 200.00 BogoMIPS > > > > (lpj=400000) > > > > > > > [ 0.324851] pid_max: default: > 32768 minimum: 301 > > > > > > > [ 0.329706] LSM: Security > Framework initializing > > > > > > > [ 0.334204] Yama: becoming > mindful. > > > > > > > [ 0.337865] Mount-cache hash > table entries: 4096 (order: 3, 32768 bytes, linear) > > > > > > > [ 0.345180] Mountpoint-cache > hash table entries: 4096 (order: 3, 32768 bytes, linear) > > > > > > > [ 0.354743] xen:grant_table: > Grant tables using version 1 layout > > > > > > > [ 0.359132] Grant table > initialized > > > > > > > [ 0.362664] xen:events: Using > FIFO-based ABI > > > > > > > [ 0.366993] Xen: initializing > cpu0 > > > > > > > [ 0.370515] rcu: Hierarchical > SRCU implementation. > > > > > > > [ 0.375930] smp: Bringing up > secondary CPUs ... > > > > > > > (XEN) null.c:353: 1 <-- d0v1 > > > > > > > (XEN) d0v1: vGICD: unhandled word > write 0x000000ffffffff to ICACTIVER0 > > > > > > > [ 0.382549] Detected VIPT > I-cache on CPU1 > > > > > > > [ 0.388712] Xen: initializing > cpu1 > > > > > > > [ 0.388743] CPU1: Booted > secondary processor 0x0000000001 [0x410fd034] > > > > > > > [ 0.388829] smp: Brought up 1 > node, 2 CPUs > > > > > > > [ 0.406941] SMP: Total of 2 > processors activated. > > > > > > > [ 0.411698] CPU features: > detected: 32-bit EL0 Support > > > > > > > [ 0.416888] CPU features: > detected: CRC32 instructions > > > > > > > [ 0.422121] CPU: All CPU(s) > started at EL1 > > > > > > > [ 0.426248] alternatives: > patching kernel code > > > > > > > [ 0.431424] devtmpfs: > initialized > > > > > > > [ 0.441454] KASLR enabled > > > > > > > [ 0.441602] clocksource: > jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: > > > > 7645041785100000 ns > > > > > > > [ 0.448321] futex hash table > entries: 512 (order: 3, 32768 bytes, linear) > > > > > > > [ 0.496183] NET: Registered > PF_NETLINK/PF_ROUTE protocol family > > > > > > > [ 0.498277] DMA: preallocated > 256 KiB GFP_KERNEL pool for atomic allocations > > > > > > > [ 0.503772] DMA: preallocated > 256 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations > > > > > > > [ 0.511610] DMA: preallocated > 256 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations > > > > > > > [ 0.519478] audit: > initializing netlink subsys (disabled) > > > > > > > [ 0.524985] audit: type=2000 > audit(0.336:1): state=initialized audit_enabled=0 res=1 > > > > > > > [ 0.529169] thermal_sys: > Registered thermal governor 'step_wise' > > > > > > > [ 0.533023] hw-breakpoint: > found 6 breakpoint and 4 watchpoint registers. > > > > > > > [ 0.545608] ASID allocator > initialised with 32768 entries > > > > > > > [ 0.551030] xen:swiotlb_xen: > Warning: only able to allocate 4 MB for software IO TLB > > > > > > > [ 0.559332] software IO TLB: > mapped [mem 0x0000000011800000-0x0000000011c00000] (4MB) > > > > > > > [ 0.583565] HugeTLB registered > 1.00 GiB page size, pre-allocated 0 pages > > > > > > > [ 0.584721] HugeTLB registered > 32.0 MiB page size, pre-allocated 0 pages > > > > > > > [ 0.591478] HugeTLB registered > 2.00 MiB page size, pre-allocated 0 pages > > > > > > > [ 0.598225] HugeTLB registered > 64.0 KiB page size, pre-allocated 0 pages > > > > > > > [ 0.636520] DRBG: Continuing > without Jitter RNG > > > > > > > [ 0.737187] raid6: neonx8 > gen() 2143 MB/s > > > > > > > [ 0.805294] raid6: neonx8 > xor() 1589 MB/s > > > > > > > [ 0.873406] raid6: neonx4 > gen() 2177 MB/s > > > > > > > [ 0.941499] raid6: neonx4 > xor() 1556 MB/s > > > > > > > [ 1.009612] raid6: neonx2 > gen() 2072 MB/s > > > > > > > [ 1.077715] raid6: neonx2 > xor() 1430 MB/s > > > > > > > [ 1.145834] raid6: neonx1 > gen() 1769 MB/s > > > > > > > [ 1.213935] raid6: neonx1 > xor() 1214 MB/s > > > > > > > [ 1.282046] raid6: int64x8 > gen() 1366 MB/s > > > > > > > [ 1.350132] raid6: int64x8 > xor() 773 MB/s > > > > > > > [ 1.418259] raid6: int64x4 > gen() 1602 MB/s > > > > > > > [ 1.486349] raid6: int64x4 > xor() 851 MB/s > > > > > > > [ 1.554464] raid6: int64x2 > gen() 1396 MB/s > > > > > > > [ 1.622561] raid6: int64x2 > xor() 744 MB/s > > > > > > > [ 1.690687] raid6: int64x1 > gen() 1033 MB/s > > > > > > > [ 1.758770] raid6: int64x1 > xor() 517 MB/s > > > > > > > [ 1.758809] raid6: using > algorithm neonx4 gen() 2177 MB/s > > > > > > > [ 1.762941] raid6: .... xor() > 1556 MB/s, rmw enabled > > > > > > > [ 1.767957] raid6: using neon > recovery algorithm > > > > > > > [ 1.772824] xen:balloon: > Initialising balloon driver > > > > > > > [ 1.778021] iommu: Default > domain type: Translated > > > > > > > [ 1.782584] iommu: DMA domain > TLB invalidation policy: strict mode > > > > > > > [ 1.789149] SCSI subsystem > initialized > > > > > > > [ 1.792820] usbcore: > registered new interface driver usbfs > > > > > > > [ 1.798254] usbcore: > registered new interface driver hub > > > > > > > [ 1.803626] usbcore: > registered new device driver usb > > > > > > > [ 1.808761] pps_core: LinuxPPS > API ver. 1 registered > > > > > > > [ 1.813716] pps_core: Software > ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti <giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx> <mailto:giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx>> <mailto:giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx> <mailto:giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx>>> > > > > <mailto:giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx> <mailto:giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx>> <mailto:giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx> <mailto:giometti@xxxxxxxx > <mailto:giometti@xxxxxxxx>>>>> > > > > > > > [ 1.822903] PTP clock support > registered > > > > > > > [ 1.826893] EDAC MC: Ver: 3.0.0 > > > > > > > [ 1.830375] zynqmp-ipi-mbox > mailbox@ff990400: Registered ZynqMP IPI mbox with TX/RX channels. > > > > > > > [ 1.838863] zynqmp-ipi-mbox > mailbox@ff990600: Registered ZynqMP IPI mbox with TX/RX channels. > > > > > > > [ 1.847356] zynqmp-ipi-mbox > mailbox@ff990800: Registered ZynqMP IPI mbox with TX/RX channels. > > > > > > > [ 1.855907] FPGA manager > framework > > > > > > > [ 1.859952] clocksource: > Switched to clocksource arch_sys_counter > > > > > > > [ 1.871712] NET: Registered > PF_INET protocol family > > > > > > > [ 1.871838] IP idents hash > table entries: 32768 (order: 6, 262144 bytes, linear) > > > > > > > [ 1.879392] > tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, > linear) > > > > > > > [ 1.887078] Table-perturb hash > table entries: 65536 (order: 6, 262144 bytes, linear) > > > > > > > [ 1.894846] TCP established > hash table entries: 16384 (order: 5, 131072 bytes, linear) > > > > > > > [ 1.902900] TCP bind hash > table entries: 16384 (order: 6, 262144 bytes, linear) > > > > > > > [ 1.910350] TCP: Hash tables > configured (established 16384 bind 16384) > > > > > > > [ 1.916778] UDP hash table > entries: 1024 (order: 3, 32768 bytes, linear) > > > > > > > [ 1.923509] UDP-Lite hash > table entries: 1024 (order: 3, 32768 bytes, linear) > > > > > > > [ 1.930759] NET: Registered > PF_UNIX/PF_LOCAL protocol family > > > > > > > [ 1.936834] RPC: Registered > named UNIX socket transport module. > > > > > > > [ 1.942342] RPC: Registered > udp transport module. > > > > > > > [ 1.947088] RPC: Registered > tcp transport module. > > > > > > > [ 1.951843] RPC: Registered > tcp NFSv4.1 backchannel transport module. > > > > > > > [ 1.958334] PCI: CLS 0 bytes, > default 64 > > > > > > > [ 1.962709] Trying to unpack > rootfs image as initramfs... > > > > > > > [ 1.977090] workingset: > timestamp_bits=62 max_order=19 bucket_order=0 > > > > > > > [ 1.982863] Installing knfsd > (copyright (C) 1996 okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx> > <mailto:okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx>> > <mailto:okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx> > <mailto:okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx>>> > <mailto:okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx> > <mailto:okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx>> > <mailto:okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx> > <mailto:okir@xxxxxxxxxxxx <mailto:okir@xxxxxxxxxxxx>>>>). > > > > > > > [ 2.021045] NET: Registered > PF_ALG protocol family > > > > > > > [ 2.021122] xor: measuring > software checksum speed > > > > > > > [ 2.029347] 8regs > : 2366 MB/sec > > > > > > > [ 2.033081] 32regs > : 2802 MB/sec > > > > > > > [ 2.038223] arm64_neon > : 2320 MB/sec > > > > > > > [ 2.038385] xor: using > function: 32regs (2802 MB/sec) > > > > > > > [ 2.043614] Block layer SCSI > generic (bsg) driver version 0.4 loaded (major 247) > > > > > > > [ 2.050959] io scheduler > mq-deadline registered > > > > > > > [ 2.055521] io scheduler kyber > registered > > > > > > > [ 2.068227] xen:xen_evtchn: > Event-channel device installed > > > > > > > [ 2.069281] Serial: 8250/16550 > driver, 4 ports, IRQ sharing disabled > > > > > > > [ 2.076190] cacheinfo: Unable > to detect cache hierarchy for CPU 0 > > > > > > > [ 2.085548] brd: module loaded > > > > > > > [ 2.089290] loop: module loaded > > > > > > > [ 2.089341] Invalid max_queues > (4), will use default max: 2. > > > > > > > [ 2.094565] tun: Universal > TUN/TAP device driver, 1.6 > > > > > > > [ 2.098655] xen_netfront: > Initialising Xen virtual ethernet driver > > > > > > > [ 2.104156] usbcore: > registered new interface driver rtl8150 > > > > > > > [ 2.109813] usbcore: > registered new interface driver r8152 > > > > > > > [ 2.115367] usbcore: > registered new interface driver asix > > > > > > > [ 2.120794] usbcore: > registered new interface driver ax88179_178a > > > > > > > [ 2.126934] usbcore: > registered new interface driver cdc_ether > > > > > > > [ 2.132816] usbcore: > registered new interface driver cdc_eem > > > > > > > [ 2.138527] usbcore: > registered new interface driver net1080 > > > > > > > [ 2.144256] usbcore: > registered new interface driver cdc_subset > > > > > > > [ 2.150205] usbcore: > registered new interface driver zaurus > > > > > > > [ 2.155837] usbcore: > registered new interface driver cdc_ncm > > > > > > > [ 2.161550] usbcore: > registered new interface driver r8153_ecm > > > > > > > [ 2.168240] usbcore: > registered new interface driver cdc_acm > > > > > > > [ 2.173109] cdc_acm: USB > Abstract Control Model driver for USB modems and ISDN adapters > > > > > > > [ 2.181358] usbcore: > registered new interface driver uas > > > > > > > [ 2.186547] usbcore: > registered new interface driver usb-storage > > > > > > > [ 2.192643] usbcore: > registered new interface driver ftdi_sio > > > > > > > [ 2.198384] usbserial: USB > Serial support registered for FTDI USB Serial Device > > > > > > > [ 2.206118] udc-core: couldn't > find an available UDC - added [g_mass_storage] to list of pending > > > > drivers > > > > > > > [ 2.215332] i2c_dev: i2c /dev > entries driver > > > > > > > [ 2.220467] xen_wdt xen_wdt: > initialized (timeout=60s, nowayout=0) > > > > > > > [ 2.225923] device-mapper: > uevent: version 1.0.3 > > > > > > > [ 2.230668] device-mapper: > ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx> <mailto:dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx>> <mailto:dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx> <mailto:dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx>>> > > > > <mailto:dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx> <mailto:dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx>> <mailto:dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx> <mailto:dm-devel@xxxxxxxxxx > <mailto:dm-devel@xxxxxxxxxx>>>> > > > > > > > [ 2.239315] EDAC MC0: Giving > out device to module 1 controller synps_ddr_controller: DEV synps_edac > > > > (INTERRUPT) > > > > > > > [ 2.249405] EDAC DEVICE0: > Giving out device to module zynqmp-ocm-edac controller zynqmp_ocm: DEV > > > > > > ff960000.memory-controller > (INTERRUPT) > > > > > > > [ 2.261719] sdhci: Secure > Digital Host Controller Interface driver > > > > > > > [ 2.267487] sdhci: > Copyright(c) Pierre Ossman > > > > > > > [ 2.271890] sdhci-pltfm: SDHCI > platform and OF driver helper > > > > > > > [ 2.278157] ledtrig-cpu: > registered to indicate activity on CPUs > > > > > > > [ 2.283816] > zynqmp_firmware_probe Platform Management API v1.1 > > > > > > > [ 2.289554] > zynqmp_firmware_probe Trustzone version v1.0 > > > > > > > [ 2.327875] securefw securefw: > securefw probed > > > > > > > [ 2.328324] alg: No test for > xilinx-zynqmp-aes (zynqmp-aes) > > > > > > > [ 2.332563] zynqmp_aes > firmware:zynqmp-firmware:zynqmp-aes: AES Successfully Registered > > > > > > > [ 2.341183] alg: No test for > xilinx-zynqmp-rsa (zynqmp-rsa) > > > > > > > [ 2.347667] remoteproc > remoteproc0: ff9a0000.rf5ss:r5f_0 is available > > > > > > > [ 2.353003] remoteproc > remoteproc1: ff9a0000.rf5ss:r5f_1 is available > > > > > > > [ 2.362605] fpga_manager > fpga0: Xilinx ZynqMP FPGA Manager registered > > > > > > > [ 2.366540] viper-xen-proxy > viper-xen-proxy: Viper Xen Proxy registered > > > > > > > [ 2.372525] viper-vdpp > a4000000.vdpp: Device Tree Probing > > > > > > > [ 2.377778] viper-vdpp > a4000000.vdpp: VDPP Version: 1.3.9.0 Info: 1.512.15.0 KeyLen: 32 > > > > > > > [ 2.386432] viper-vdpp > a4000000.vdpp: Unable to register tamper handler. Retrying... > > > > > > > [ 2.394094] viper-vdpp-net > a5000000.vdpp_net: Device Tree Probing > > > > > > > [ 2.399854] viper-vdpp-net > a5000000.vdpp_net: Device registered > > > > > > > [ 2.405931] viper-vdpp-stat > a8000000.vdpp_stat: Device Tree Probing > > > > > > > [ 2.412037] viper-vdpp-stat > a8000000.vdpp_stat: Build parameters: VTI Count: 512 Event Count: 32 > > > > > > > [ 2.420856] default preset > > > > > > > [ 2.423797] viper-vdpp-stat > a8000000.vdpp_stat: Device registered > > > > > > > [ 2.430054] viper-vdpp-rng > ac000000.vdpp_rng: Device Tree Probing > > > > > > > [ 2.435948] viper-vdpp-rng > ac000000.vdpp_rng: Device registered > > > > > > > [ 2.441976] vmcu driver init > > > > > > > [ 2.444922] VMCU: : (240:0) > registered > > > > > > > [ 2.444956] In K81 Updater init > > > > > > > [ 2.449003] pktgen: Packet > Generator for packet performance testing. Version: 2.75 > > > > > > > [ 2.468833] Initializing XFRM > netlink socket > > > > > > > [ 2.468902] NET: Registered > PF_PACKET protocol family > > > > > > > [ 2.472729] Bridge firewalling > registered > > > > > > > [ 2.476785] 8021q: 802.1Q VLAN > Support v1.8 > > > > > > > [ 2.481341] registered > taskstats version 1 > > > > > > > [ 2.486394] Btrfs loaded, > crc32c=crc32c-generic, zoned=no, fsverity=no > > > > > > > [ 2.503145] ff010000.serial: > ttyPS1 at MMIO 0xff010000 (irq = 36, base_baud = 6250000) is a xuartps > > > > > > > [ 2.507103] of-fpga-region > fpga-full: FPGA Region probed > > > > > > > [ 2.512986] xilinx-zynqmp-dma > fd500000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.520267] xilinx-zynqmp-dma > fd510000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.528239] xilinx-zynqmp-dma > fd520000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.536152] xilinx-zynqmp-dma > fd530000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.544153] xilinx-zynqmp-dma > fd540000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.552127] xilinx-zynqmp-dma > fd550000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.560178] xilinx-zynqmp-dma > ffa80000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.567987] xilinx-zynqmp-dma > ffa90000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.576018] xilinx-zynqmp-dma > ffaa0000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.583889] xilinx-zynqmp-dma > ffab0000.dma-controller: ZynqMP DMA driver Probe success > > > > > > > [ 2.946379] spi-nor spi0.0: > mt25qu512a (131072 Kbytes) > > > > > > > [ 2.946467] 2 fixed-partitions > partitions found on MTD device spi0.0 > > > > > > > [ 2.952393] Creating 2 MTD > partitions on "spi0.0": > > > > > > > [ 2.957231] > 0x000004000000-0x000008000000 : "bank A" > > > > > > > [ 2.963332] > 0x000000000000-0x000004000000 : "bank B" > > > > > > > [ 2.968694] macb > ff0b0000.ethernet: Not enabling partial store and forward > > > > > > > [ 2.975333] macb > ff0b0000.ethernet eth0: Cadence GEM rev 0x50070106 at 0xff0b0000 irq 25 > > > > (18:41:fe:0f:ff:02) > > > > > > > [ 2.984472] macb > ff0c0000.ethernet: Not enabling partial store and forward > > > > > > > [ 2.992144] macb > ff0c0000.ethernet eth1: Cadence GEM rev 0x50070106 at 0xff0c0000 irq 26 > > > > (18:41:fe:0f:ff:03) > > > > > > > [ 3.001043] viper_enet > viper_enet: Viper power GPIOs initialised > > > > > > > [ 3.007313] viper_enet > viper_enet vnet0 (uninitialized): Validate interface QSGMII > > > > > > > [ 3.014914] viper_enet > viper_enet vnet1 (uninitialized): Validate interface QSGMII > > > > > > > [ 3.022138] viper_enet > viper_enet vnet1 (uninitialized): Validate interface type 18 > > > > > > > [ 3.030274] viper_enet > viper_enet vnet2 (uninitialized): Validate interface QSGMII > > > > > > > [ 3.037785] viper_enet > viper_enet vnet3 (uninitialized): Validate interface QSGMII > > > > > > > [ 3.045301] viper_enet > viper_enet: Viper enet registered > > > > > > > [ 3.050958] xilinx-axipmon > ffa00000.perf-monitor: Probed Xilinx APM > > > > > > > [ 3.057135] xilinx-axipmon > fd0b0000.perf-monitor: Probed Xilinx APM > > > > > > > [ 3.063538] xilinx-axipmon > fd490000.perf-monitor: Probed Xilinx APM > > > > > > > [ 3.069920] xilinx-axipmon > ffa10000.perf-monitor: Probed Xilinx APM > > > > > > > [ 3.097729] si70xx: probe of > 2-0040 failed with error -5 > > > > > > > [ 3.098042] cdns-wdt > fd4d0000.watchdog: Xilinx Watchdog Timer with timeout 60s > > > > > > > [ 3.105111] cdns-wdt > ff150000.watchdog: Xilinx Watchdog Timer with timeout 10s > > > > > > > [ 3.112457] viper-tamper > viper-tamper: Device registered > > > > > > > [ 3.117593] active_bank > active_bank: boot bank: 1 > > > > > > > [ 3.122184] active_bank > active_bank: boot mode: (0x02) qspi32 > > > > > > > [ 3.128247] viper-vdpp > a4000000.vdpp: Device Tree Probing > > > > > > > [ 3.133439] viper-vdpp > a4000000.vdpp: VDPP Version: 1.3.9.0 Info: 1.512.15.0 KeyLen: 32 > > > > > > > [ 3.142151] viper-vdpp > a4000000.vdpp: Tamper handler registered > > > > > > > [ 3.147438] viper-vdpp > a4000000.vdpp: Device registered > > > > > > > [ 3.153007] lpc55_l2 spi1.0: > registered handler for protocol 0 > > > > > > > [ 3.158582] lpc55_user > lpc55_user: The major number for your device is 236 > > > > > > > [ 3.165976] lpc55_l2 spi1.0: > registered handler for protocol 1 > > > > > > > [ 3.181999] rtc-lpc55 > rtc_lpc55: lpc55_rtc_get_time: bad result: 1 > > > > > > > [ 3.182856] rtc-lpc55 > rtc_lpc55: registered as rtc0 > > > > > > > [ 3.188656] lpc55_l2 spi1.0: > (2) mcu still not ready? > > > > > > > [ 3.193744] lpc55_l2 spi1.0: > (3) mcu still not ready? > > > > > > > [ 3.198848] lpc55_l2 spi1.0: > (4) mcu still not ready? > > > > > > > [ 3.202932] mmc0: SDHCI > controller on ff160000.mmc [ff160000.mmc] using ADMA 64-bit > > > > > > > [ 3.210689] lpc55_l2 spi1.0: > (5) mcu still not ready? > > > > > > > [ 3.215694] lpc55_l2 spi1.0: > rx error: -110 > > > > > > > [ 3.284438] mmc0: new HS200 > MMC card at address 0001 > > > > > > > [ 3.285179] mmcblk0: mmc0:0001 > SEM16G 14.6 GiB > > > > > > > [ 3.291784] mmcblk0: p1 p2 p3 > p4 p5 p6 p7 p8 > > > > > > > [ 3.293915] mmcblk0boot0: > mmc0:0001 SEM16G 4.00 MiB > > > > > > > [ 3.299054] mmcblk0boot1: > mmc0:0001 SEM16G 4.00 MiB > > > > > > > [ 3.303905] mmcblk0rpmb: > mmc0:0001 SEM16G 4.00 MiB, chardev (244:0) > > > > > > > [ 3.582676] rtc-lpc55 > rtc_lpc55: lpc55_rtc_get_time: bad result: 1 > > > > > > > [ 3.583332] rtc-lpc55 > rtc_lpc55: hctosys: unable to read the hardware clock > > > > > > > [ 3.591252] cdns-i2c > ff020000.i2c: recovery information complete > > > > > > > [ 3.597085] at24 0-0050: > supply vcc not found, using dummy regulator > > > > > > > [ 3.603011] lpc55_l2 spi1.0: > (2) mcu still not ready? > > > > > > > [ 3.608093] at24 0-0050: 256 > byte spd EEPROM, read-only > > > > > > > [ 3.613620] lpc55_l2 spi1.0: > (3) mcu still not ready? > > > > > > > [ 3.619362] lpc55_l2 spi1.0: > (4) mcu still not ready? > > > > > > > [ 3.624224] rtc-rv3028 0-0052: > registered as rtc1 > > > > > > > [ 3.628343] lpc55_l2 spi1.0: > (5) mcu still not ready? > > > > > > > [ 3.633253] lpc55_l2 spi1.0: > rx error: -110 > > > > > > > [ 3.639104] k81_bootloader > 0-0010: probe > > > > > > > [ 3.641628] VMCU: : (235:0) > registered > > > > > > > [ 3.641635] k81_bootloader > 0-0010: probe completed > > > > > > > [ 3.668346] cdns-i2c > ff020000.i2c: 400 kHz mmio ff020000 irq 28 > > > > > > > [ 3.669154] cdns-i2c > ff030000.i2c: recovery information complete > > > > > > > [ 3.675412] lm75 1-0048: > supply vs not found, using dummy regulator > > > > > > > [ 3.682920] lm75 1-0048: > hwmon1: sensor 'tmp112' > > > > > > > [ 3.686548] i2c i2c-1: Added > multiplexed i2c bus 3 > > > > > > > [ 3.690795] i2c i2c-1: Added > multiplexed i2c bus 4 > > > > > > > [ 3.695629] i2c i2c-1: Added > multiplexed i2c bus 5 > > > > > > > [ 3.700492] i2c i2c-1: Added > multiplexed i2c bus 6 > > > > > > > [ 3.705157] pca954x 1-0070: > registered 4 multiplexed busses for I2C switch pca9546 > > > > > > > [ 3.713049] at24 1-0054: > supply vcc not found, using dummy regulator > > > > > > > [ 3.720067] at24 1-0054: 1024 > byte 24c08 EEPROM, read-only > > > > > > > [ 3.724761] cdns-i2c > ff030000.i2c: 100 kHz mmio ff030000 irq 29 > > > > > > > [ 3.731272] sfp > viper_enet:sfp-eth1: Host maximum power 2.0W > > > > > > > [ 3.737549] > sfp_register_socket: got sfp_bus > > > > > > > [ 3.740709] > sfp_register_socket: register sfp_bus > > > > > > > [ 3.745459] sfp_register_bus: > ops ok! > > > > > > > [ 3.749179] sfp_register_bus: > Try to attach > > > > > > > [ 3.753419] sfp_register_bus: > Attach succeeded > > > > > > > [ 3.757914] sfp_register_bus: > upstream ops attach > > > > > > > [ 3.762677] sfp_register_bus: > Bus registered > > > > > > > [ 3.766999] > sfp_register_socket: register sfp_bus succeeded > > > > > > > [ 3.775870] of_cfs_init > > > > > > > [ 3.776000] of_cfs_init: OK > > > > > > > [ 3.778211] clk: Not disabling > unused clocks > > > > > > > [ 11.278477] Freeing initrd > memory: 206056K > > > > > > > [ 11.279406] Freeing unused > kernel memory: 1536K > > > > > > > [ 11.314006] Checked W+X > mappings: passed, no W+X pages found > > > > > > > [ 11.314142] Run /init as init > process > > > > > > > INIT: version 3.01 booting > > > > > > > fsck (busybox 1.35.0) > > > > > > > /dev/mmcblk0p1: clean, 12/102400 > files, 238162/409600 blocks > > > > > > > /dev/mmcblk0p2: clean, 12/102400 > files, 171972/409600 blocks > > > > > > > /dev/mmcblk0p3 was not cleanly > unmounted, check forced. > > > > > > > /dev/mmcblk0p3: 20/4096 files > (0.0% non-contiguous), 663/16384 blocks > > > > > > > [ 11.553073] EXT4-fs > (mmcblk0p3): mounted filesystem without journal. Opts: (null). Quota mode: > > > > disabled. > > > > > > > Starting random number generator > daemon. > > > > > > > [ 11.580662] random: crng init > done > > > > > > > Starting udev > > > > > > > [ 11.613159] udevd[142]: > starting version 3.2.10 > > > > > > > [ 11.620385] udevd[143]: > starting eudev-3.2.10 > > > > > > > [ 11.704481] macb > ff0b0000.ethernet control_red: renamed from eth0 > > > > > > > [ 11.720264] macb > ff0c0000.ethernet control_black: renamed from eth1 > > > > > > > [ 12.063396] > ip_local_port_range: prefer different parity for start/end values. > > > > > > > [ 12.084801] rtc-lpc55 > rtc_lpc55: lpc55_rtc_get_time: bad result: 1 > > > > > > > hwclock: RTC_RD_TIME: Invalid > exchange > > > > > > > Mon Feb 27 08:40:53 UTC 2023 > > > > > > > [ 12.115309] rtc-lpc55 > rtc_lpc55: lpc55_rtc_set_time: bad result > > > > > > > hwclock: RTC_SET_TIME: Invalid > exchange > > > > > > > [ 12.131027] rtc-lpc55 > rtc_lpc55: lpc55_rtc_get_time: bad result: 1 > > > > > > > Starting mcud > > > > > > > INIT: Entering runlevel: 5 > > > > > > > Configuring network interfaces... > done. > > > > > > > resetting network interface > > > > > > > [ 12.718295] macb > ff0b0000.ethernet control_red: PHY [ff0b0000.ethernet-ffffffff:02] driver > [Xilinx > > > > PCS/PMA PHY] (irq=POLL) > > > > > > > [ 12.723919] macb > ff0b0000.ethernet control_red: configuring for phy/gmii link mode > > > > > > > [ 12.732151] pps pps0: new PPS > source ptp0 > > > > > > > [ 12.735563] macb > ff0b0000.ethernet: gem-ptp-timer ptp clock registered. > > > > > > > [ 12.745724] macb > ff0c0000.ethernet control_black: PHY [ff0c0000.ethernet-ffffffff:01] driver > [Xilinx > > > > PCS/PMA PHY] > > > > > > (irq=POLL) > > > > > > > [ 12.753469] macb > ff0c0000.ethernet control_black: configuring for phy/gmii link mode > > > > > > > [ 12.761804] pps pps1: new PPS > source ptp1 > > > > > > > [ 12.765398] macb > ff0c0000.ethernet: gem-ptp-timer ptp clock registered. > > > > > > > Auto-negotiation: off > > > > > > > Auto-negotiation: off > > > > > > > [ 16.828151] macb > ff0b0000.ethernet control_red: unable to generate target frequency: 125000000 > Hz > > > > > > > [ 16.834553] macb > ff0b0000.ethernet control_red: Link is Up - 1Gbps/Full - flow control off > > > > > > > [ 16.860552] macb > ff0c0000.ethernet control_black: unable to generate target frequency: > 125000000 Hz > > > > > > > [ 16.867052] macb > ff0c0000.ethernet control_black: Link is Up - 1Gbps/Full - flow control off > > > > > > > Starting Failsafe Secure Shell > server in port 2222: sshd > > > > > > > done. > > > > > > > Starting rpcbind daemon...done. > > > > > > > > > > > > > > [ 17.093019] rtc-lpc55 > rtc_lpc55: lpc55_rtc_get_time: bad result: 1 > > > > > > > hwclock: RTC_RD_TIME: Invalid > exchange > > > > > > > Starting State Manager Service > > > > > > > Start state-manager restarter... > > > > > > > (XEN) d0v1 Forwarding AES > operation: 3254779951 > > > > > > > Starting /usr/sbin/xenstored....[ > 17.265256] BTRFS: device fsid 80efc224-c202-4f8e-a949-4dae7f04a0aa > > > > devid 1 transid 744 > > > > > > /dev/dm-0 > > > > > > > scanned by udevd (385) > > > > > > > [ 17.349933] BTRFS info (device > dm-0): disk space caching is enabled > > > > > > > [ 17.350670] BTRFS info (device > dm-0): has skinny extents > > > > > > > [ 17.364384] BTRFS info (device > dm-0): enabling ssd optimizations > > > > > > > [ 17.830462] BTRFS: device fsid > 27ff666b-f4e5-4f90-9054-c210db5b2e2e devid 1 transid 6 > > > > /dev/mapper/client_prov scanned by > > > > > > mkfs.btrfs > > > > > > > (526) > > > > > > > [ 17.872699] BTRFS info (device > dm-1): using free space tree > > > > > > > [ 17.872771] BTRFS info (device > dm-1): has skinny extents > > > > > > > [ 17.878114] BTRFS info (device > dm-1): flagging fs with big metadata feature > > > > > > > [ 17.894289] BTRFS info (device > dm-1): enabling ssd optimizations > > > > > > > [ 17.895695] BTRFS info (device > dm-1): checking UUID tree > > > > > > > > > > > > > > Setting domain 0 name, domid and > JSON config... > > > > > > > Done setting up Dom0 > > > > > > > Starting xenconsoled... > > > > > > > Starting QEMU as disk backend for > dom0 > > > > > > > Starting domain watchdog daemon: > xenwatchdogd startup > > > > > > > > > > > > > > [ 18.408647] BTRFS: device fsid > 5e08d5e9-bc2a-46b9-af6a-44c7087b8921 devid 1 transid 6 > > > > /dev/mapper/client_config scanned by > > > > > > mkfs.btrfs > > > > > > > (574) > > > > > > > [done] > > > > > > > [ 18.465552] BTRFS info (device > dm-2): using free space tree > > > > > > > [ 18.465629] BTRFS info (device > dm-2): has skinny extents > > > > > > > [ 18.471002] BTRFS info (device > dm-2): flagging fs with big metadata feature > > > > > > > Starting crond: [ 18.482371] > BTRFS info (device dm-2): enabling ssd optimizations > > > > > > > [ 18.486659] BTRFS info (device > dm-2): checking UUID tree > > > > > > > OK > > > > > > > starting rsyslogd ... Log > partition ready after 0 poll loops > > > > > > > done > > > > > > > rsyslogd: cannot connect to > 172.18.0.1:514 <http://172.18.0.1:514> <http://172.18.0.1:514 > <http://172.18.0.1:514>> <http://172.18.0.1:514 <http://172.18.0.1:514> > <http://172.18.0.1:514 <http://172.18.0.1:514>>> <http://172.18.0.1:514 > <http://172.18.0.1:514> <http://172.18.0.1:514 <http://172.18.0.1:514>> > <http://172.18.0.1:514 <http://172.18.0.1:514> <http://172.18.0.1:514 > <http://172.18.0.1:514>>>>: Network is unreachable [v8.2208.0 try > > > > https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027> <https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027>> <https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027> <https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027>>> <https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027> <https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027>> <https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027> <https://www.rsyslog.com/e/2027 > <https://www.rsyslog.com/e/2027>>>> ] > > > > > > > [ 18.670637] BTRFS: device fsid > 39d7d9e1-967d-478e-94ae-690deb722095 devid 1 transid 608 /dev/dm-3 > > > > scanned by udevd (518) > > > > > > > > > > > > > > Please insert USB token and enter > your role in login prompt. > > > > > > > > > > > > > > login: > > > > > > > > > > > > > > Regards, > > > > > > > O. > > > > > > > > > > > > > > > > > > > > > пн, 24 апр. 2023 г. в 23:39, > Stefano Stabellini <sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>>>: > > > > > > > Hi Oleg, > > > > > > > > > > > > > > Here is the issue from your > logs: > > > > > > > > > > > > > > SError Interrupt on CPU0, > code 0xbe000000 -- SError > > > > > > > > > > > > > > SErrors are special signals > to notify software of serious hardware > > > > > > > errors. Something is going > very wrong. Defective hardware is a > > > > > > > possibility. Another > possibility if software accessing address ranges > > > > > > > that it is not supposed to, > sometimes it causes SErrors. > > > > > > > > > > > > > > Cheers, > > > > > > > > > > > > > > Stefano > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Mon, 24 Apr 2023, Oleg > Nikitenko wrote: > > > > > > > > > > > > > > > Hello, > > > > > > > > > > > > > > > > Thanks guys. > > > > > > > > I found out where the > problem was. > > > > > > > > Now dom0 booted more. But > I have a new one. > > > > > > > > This is a kernel panic > during Dom0 loading. > > > > > > > > Maybe someone is able to > suggest something ? > > > > > > > > > > > > > > > > Regards, > > > > > > > > O. > > > > > > > > > > > > > > > > [ 3.771362] > sfp_register_bus: upstream ops attach > > > > > > > > [ 3.776119] > sfp_register_bus: Bus registered > > > > > > > > [ 3.780459] > sfp_register_socket: register sfp_bus succeeded > > > > > > > > [ 3.789399] of_cfs_init > > > > > > > > [ 3.789499] > of_cfs_init: OK > > > > > > > > [ 3.791685] clk: Not > disabling unused clocks > > > > > > > > [ 11.010355] SError > Interrupt on CPU0, code 0xbe000000 -- SError > > > > > > > > [ 11.010380] CPU: 0 > PID: 9 Comm: kworker/u4:0 Not tainted 5.15.72-xilinx-v2022.1 #1 > > > > > > > > [ 11.010393] Workqueue: > events_unbound async_run_entry_fn > > > > > > > > [ 11.010414] pstate: > 60000005 (nZCv daif -PAN -UAO -TCO -DIT -SSBS BTYPE=--) > > > > > > > > [ 11.010422] pc : > simple_write_end+0xd0/0x130 > > > > > > > > [ 11.010431] lr : > generic_perform_write+0x118/0x1e0 > > > > > > > > [ 11.010438] sp : > ffffffc00809b910 > > > > > > > > [ 11.010441] x29: > ffffffc00809b910 x28: 0000000000000000 x27: ffffffef69ba88c0 > > > > > > > > [ 11.010451] x26: > 0000000000003eec x25: ffffff807515db00 x24: 0000000000000000 > > > > > > > > [ 11.010459] x23: > ffffffc00809ba90 x22: 0000000002aac000 x21: ffffff807315a260 > > > > > > > > [ 11.010472] x20: > 0000000000001000 x19: fffffffe02000000 x18: 0000000000000000 > > > > > > > > [ 11.010481] x17: > 00000000ffffffff x16: 0000000000008000 x15: 0000000000000000 > > > > > > > > [ 11.010490] x14: > 0000000000000000 x13: 0000000000000000 x12: 0000000000000000 > > > > > > > > [ 11.010498] x11: > 0000000000000000 x10: 0000000000000000 x9 : 0000000000000000 > > > > > > > > [ 11.010507] x8 : > 0000000000000000 x7 : ffffffef693ba680 x6 : 000000002d89b700 > > > > > > > > [ 11.010515] x5 : > fffffffe02000000 x4 : ffffff807315a3c8 x3 : 0000000000001000 > > > > > > > > [ 11.010524] x2 : > 0000000002aab000 x1 : 0000000000000001 x0 : 0000000000000005 > > > > > > > > [ 11.010534] Kernel > panic - not syncing: Asynchronous SError Interrupt > > > > > > > > [ 11.010539] CPU: 0 > PID: 9 Comm: kworker/u4:0 Not tainted 5.15.72-xilinx-v2022.1 #1 > > > > > > > > [ 11.010545] Hardware > name: D14 Viper Board - White Unit (DT) > > > > > > > > [ 11.010548] Workqueue: > events_unbound async_run_entry_fn > > > > > > > > [ 11.010556] Call trace: > > > > > > > > [ 11.010558] > dump_backtrace+0x0/0x1c4 > > > > > > > > [ 11.010567] > show_stack+0x18/0x2c > > > > > > > > [ 11.010574] > dump_stack_lvl+0x7c/0xa0 > > > > > > > > [ 11.010583] > dump_stack+0x18/0x34 > > > > > > > > [ 11.010588] > panic+0x14c/0x2f8 > > > > > > > > [ 11.010597] > print_tainted+0x0/0xb0 > > > > > > > > [ 11.010606] > arm64_serror_panic+0x6c/0x7c > > > > > > > > [ 11.010614] > do_serror+0x28/0x60 > > > > > > > > [ 11.010621] > el1h_64_error_handler+0x30/0x50 > > > > > > > > [ 11.010628] > el1h_64_error+0x78/0x7c > > > > > > > > [ 11.010633] > simple_write_end+0xd0/0x130 > > > > > > > > [ 11.010639] > generic_perform_write+0x118/0x1e0 > > > > > > > > [ 11.010644] > __generic_file_write_iter+0x138/0x1c4 > > > > > > > > [ 11.010650] > generic_file_write_iter+0x78/0xd0 > > > > > > > > [ 11.010656] > __kernel_write+0xfc/0x2ac > > > > > > > > [ 11.010665] > kernel_write+0x88/0x160 > > > > > > > > [ 11.010673] > xwrite+0x44/0x94 > > > > > > > > [ 11.010680] > do_copy+0xa8/0x104 > > > > > > > > [ 11.010686] > write_buffer+0x38/0x58 > > > > > > > > [ 11.010692] > flush_buffer+0x4c/0xbc > > > > > > > > [ 11.010698] > __gunzip+0x280/0x310 > > > > > > > > [ 11.010704] > gunzip+0x1c/0x28 > > > > > > > > [ 11.010709] > unpack_to_rootfs+0x170/0x2b0 > > > > > > > > [ 11.010715] > do_populate_rootfs+0x80/0x164 > > > > > > > > [ 11.010722] > async_run_entry_fn+0x48/0x164 > > > > > > > > [ 11.010728] > process_one_work+0x1e4/0x3a0 > > > > > > > > [ 11.010736] > worker_thread+0x7c/0x4c0 > > > > > > > > [ 11.010743] > kthread+0x120/0x130 > > > > > > > > [ 11.010750] > ret_from_fork+0x10/0x20 > > > > > > > > [ 11.010757] SMP: > stopping secondary CPUs > > > > > > > > [ 11.010784] Kernel > Offset: 0x2f61200000 from 0xffffffc008000000 > > > > > > > > [ 11.010788] > PHYS_OFFSET: 0x0 > > > > > > > > [ 11.010790] CPU > features: 0x00000401,00000842 > > > > > > > > [ 11.010795] Memory > Limit: none > > > > > > > > [ 11.277509] ---[ end > Kernel panic - not syncing: Asynchronous SError Interrupt ]--- > > > > > > > > > > > > > > > > пт, 21 апр. 2023 г. в > 15:52, Michal Orzel <michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>>>>>: > > > > > > > > Hi Oleg, > > > > > > > > > > > > > > > > On 21/04/2023 > 14:49, Oleg Nikitenko wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Hello Michal, > > > > > > > > > > > > > > > > > > I was not able to > enable earlyprintk in the xen for now. > > > > > > > > > I decided to > choose another way. > > > > > > > > > This is a xen's > command line that I found out completely. > > > > > > > > > > > > > > > > > > (XEN) $$$$ > console=dtuart dtuart=serial0 dom0_mem=1600M dom0_max_vcpus=2 dom0_vcpus_pin > > > > bootscrub=0 > > > > > > vwfi=native > > > > > > > sched=null > > > > > > > > timer_slop=0 > > > > > > > > Yes, adding a > printk() in Xen was also a good idea. > > > > > > > > > > > > > > > > > > > > > > > > > > So you are > absolutely right about a command line. > > > > > > > > > Now I am going to > find out why xen did not have the correct parameters from the device > > > > tree. > > > > > > > > Maybe you will find > this document helpful: > > > > > > > > > https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt> > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt>> > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt> > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt>>> > > > > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt> > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt>> > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt> > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt > > <https://github.com/Xilinx/xen/blob/xlnx_rebase_4.16/docs/misc/arm/device-tree/booting.txt>>>> > > > > > > > > > > > > > > > > ~Michal > > > > > > > > > > > > > > > > > > > > > > > > > > Regards, > > > > > > > > > Oleg > > > > > > > > > > > > > > > > > > пт, 21 апр. 2023 > г. в 11:16, Michal Orzel <michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>>> > > > > <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>>>>>: > > > > > > > > > > > > > > > > > > > > > > > > > > > On 21/04/2023 > 10:04, Oleg Nikitenko wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Hello > Michal, > > > > > > > > > > > > > > > > > > > > Yes, I use > yocto. > > > > > > > > > > > > > > > > > > > > Yesterday > all day long I tried to follow your suggestions. > > > > > > > > > > I faced a > problem. > > > > > > > > > > Manually in > the xen config build file I pasted the strings: > > > > > > > > > In the > .config file or in some Yocto file (listing additional Kconfig options) added > > > > to SRC_URI? > > > > > > > > > You shouldn't > really modify .config file but if you do, you should execute "make > > > > olddefconfig" > > > > > > afterwards. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > CONFIG_EARLY_PRINTK > > > > > > > > > > > CONFIG_EARLY_PRINTK_ZYNQMP > > > > > > > > > > > CONFIG_EARLY_UART_CHOICE_CADENCE > > > > > > > > > I hope you > added =y to them. > > > > > > > > > > > > > > > > > > Anyway, you > have at least the following solutions: > > > > > > > > > 1) Run > bitbake xen -c menuconfig to properly set early printk > > > > > > > > > 2) Find out > how you enable other Kconfig options in your project (e.g. > > > > CONFIG_COLORING=y that is not > > > > > > enabled by > > > > > > > default) > > > > > > > > > 3) Append the > following to "xen/arch/arm/configs/arm64_defconfig": > > > > > > > > > > CONFIG_EARLY_PRINTK_ZYNQMP=y > > > > > > > > > > > > > > > > > > ~Michal > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Host hangs > in build time. > > > > > > > > > > Maybe I did > not set something in the config build file ? > > > > > > > > > > > > > > > > > > > > Regards, > > > > > > > > > > Oleg > > > > > > > > > > > > > > > > > > > > чт, 20 апр. > 2023 г. в 11:57, Oleg Nikitenko <oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>> > > > > <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>>> > > > > > > > > > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>> > > > > <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>>>>>: > > > > > > > > > > > > > > > > > > > > Thanks > Michal, > > > > > > > > > > > > > > > > > > > > You > gave me an idea. > > > > > > > > > > I am > going to try it today. > > > > > > > > > > > > > > > > > > > > Regards, > > > > > > > > > > O. > > > > > > > > > > > > > > > > > > > > чт, 20 > апр. 2023 г. в 11:56, Oleg Nikitenko <oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>> > > > > <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>>> > > > > > > > > > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>> > > > > <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>>>>>: > > > > > > > > > > > > > > > > > > > > > Thanks Stefano. > > > > > > > > > > > > > > > > > > > > I > am going to do it today. > > > > > > > > > > > > > > > > > > > > > Regards, > > > > > > > > > > O. > > > > > > > > > > > > > > > > > > > > ср, > 19 апр. 2023 г. в 23:05, Stefano Stabellini <sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> > > > > <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>> > > > > > > <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>>> > > > > > > > > > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>> > > > > <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>>>>>: > > > > > > > > > > > > > > > > > > > > > On Wed, 19 Apr 2023, Oleg Nikitenko wrote: > > > > > > > > > > > > Hi Michal, > > > > > > > > > > > > > > > > > > > > > > > > I corrected xen's command line. > > > > > > > > > > > > Now it is > > > > > > > > > > > > xen,xen-bootargs = "console=dtuart dtuart=serial0 dom0_mem=1600M > > > > dom0_max_vcpus=2 > > > > > > dom0_vcpus_pin > > > > > > > > bootscrub=0 > vwfi=native sched=null > > > > > > > > > > > > timer_slop=0 way_size=65536 xen_colors=0-3 dom0_colors=4-7"; > > > > > > > > > > > > > > > > > > > > > 4 colors is way too many for xen, just do xen_colors=0-0. There is no > > > > > > > > > > > advantage in using more than 1 color for Xen. > > > > > > > > > > > > > > > > > > > > > 4 colors is too few for dom0, if you are giving 1600M of memory to > > > > Dom0. > > > > > > > > > > > Each color is 256M. For 1600M you should give at least 7 colors. Try: > > > > > > > > > > > > > > > > > > > > > xen_colors=0-0 dom0_colors=1-8 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Unfortunately the result was the same. > > > > > > > > > > > > > > > > > > > > > > > > (XEN) - Dom0 mode: Relaxed > > > > > > > > > > > > (XEN) P2M: 40-bit IPA with 40-bit PA and 8-bit VMID > > > > > > > > > > > > (XEN) P2M: 3 levels with order-1 root, VTCR 0x0000000080023558 > > > > > > > > > > > > (XEN) Scheduling granularity: cpu, 1 CPU per sched-resource > > > > > > > > > > > > (XEN) Coloring general information > > > > > > > > > > > > (XEN) Way size: 64kB > > > > > > > > > > > > (XEN) Max. number of colors available: 16 > > > > > > > > > > > > (XEN) Xen color(s): [ 0 ] > > > > > > > > > > > > (XEN) alternatives: Patching with alt table 00000000002cc690 -> > > > > 00000000002ccc0c > > > > > > > > > > > > (XEN) Color array allocation failed for dom0 > > > > > > > > > > > > (XEN) > > > > > > > > > > > > (XEN) **************************************** > > > > > > > > > > > > (XEN) Panic on CPU 0: > > > > > > > > > > > > (XEN) Error creating domain 0 > > > > > > > > > > > > (XEN) **************************************** > > > > > > > > > > > > (XEN) > > > > > > > > > > > > (XEN) Reboot in five seconds... > > > > > > > > > > > > > > > > > > > > > > > > I am going to find out how command line arguments passed and parsed. > > > > > > > > > > > > > > > > > > > > > > > > Regards, > > > > > > > > > > > > Oleg > > > > > > > > > > > > > > > > > > > > > > > > ср, 19 апр. 2023 г. в 11:25, Oleg Nikitenko <oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>> > > > > <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>> > > > > > > <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>>> > > > > > > > > > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx> > <mailto:oleshiiwood@xxxxxxxxx <mailto:oleshiiwood@xxxxxxxxx>>> > > > > <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx> <mailto:oleshiiwood@xxxxxxxxx > <mailto:oleshiiwood@xxxxxxxxx>>>>>>>: > > > > > > > > > > > > Hi Michal, > > > > > > > > > > > > > > > > > > > > > > > > You put my nose into the problem. Thank you. > > > > > > > > > > > > I am going to use your point. > > > > > > > > > > > > Let's see what happens. > > > > > > > > > > > > > > > > > > > > > > > > Regards, > > > > > > > > > > > > Oleg > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ср, 19 апр. 2023 г. в 10:37, Michal Orzel <michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>> > > > > <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>>> > > > > > > <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>>>> > > > > > > > > > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>>>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx> > <mailto:michal.orzel@xxxxxxx <mailto:michal.orzel@xxxxxxx>>> > > > > <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx> <mailto:michal.orzel@xxxxxxx > <mailto:michal.orzel@xxxxxxx>>>>>>>: > > > > > > > > > > > > Hi Oleg, > > > > > > > > > > > > > > > > > > > > > > > > On 19/04/2023 09:03, Oleg Nikitenko wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Hello Stefano, > > > > > > > > > > > > > > > > > > > > > > > > > > Thanks for the clarification. > > > > > > > > > > > > > My company uses yocto for image generation. > > > > > > > > > > > > > What kind of information do you need to consult me in this > > > > case ? > > > > > > > > > > > > > > > > > > > > > > > > > > Maybe modules sizes/addresses which were mentioned by @Julien > > > > Grall > > > > > > > <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx>>> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>>>> > > > > > > > > > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx>> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>>> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx>>>>> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>>> > > > > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>>>> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx>> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>>> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx>> > <mailto:julien@xxxxxxx <mailto:julien@xxxxxxx> <mailto:julien@xxxxxxx > <mailto:julien@xxxxxxx>>>>>>> ? > > > > > > > > > > > > > > > > > > > > > > > > Sorry for jumping into discussion, but FWICS the Xen command > > > > line you provided > > > > > > seems to be > > > > > > > not the > > > > > > > > one > > > > > > > > > > > > Xen booted with. The error you are observing most likely is due > > > > to dom0 colors > > > > > > > configuration not > > > > > > > > being > > > > > > > > > > > > specified (i.e. lack of dom0_colors=<> parameter). Although in > > > > the command line you > > > > > > > provided, this > > > > > > > > parameter > > > > > > > > > > > > is set, I strongly doubt that this is the actual command line > > > > in use. > > > > > > > > > > > > > > > > > > > > > > > > You wrote: > > > > > > > > > > > > xen,xen-bootargs = "console=dtuart dtuart=serial0 > > > > dom0_mem=1600M dom0_max_vcpus=2 > > > > > > > dom0_vcpus_pin > > > > > > > > bootscrub=0 > vwfi=native > > > > > > > > > > > > sched=null timer_slop=0 way_szize=65536 xen_colors=0-3 > > > > dom0_colors=4-7"; > > > > > > > > > > > > > > > > > > > > > > > > but: > > > > > > > > > > > > 1) way_szize has a typo > > > > > > > > > > > > 2) you specified 4 colors (0-3) for Xen, but the boot log says > > > > that Xen has only > > > > > > one: > > > > > > > > > > > > (XEN) Xen color(s): [ 0 ] > > > > > > > > > > > > > > > > > > > > > > > > This makes me believe that no colors configuration actually end > > > > up in command line > > > > > > that Xen > > > > > > > booted > > > > > > > > with. > > > > > > > > > > > > Single color for Xen is a "default if not specified" and way > > > > size was probably > > > > > > calculated > > > > > > > by asking > > > > > > > > HW. > > > > > > > > > > > > > > > > > > > > > > > > So I would suggest to first cross-check the command line in > > > > use. > > > > > > > > > > > > > > > > > > > > > > > > ~Michal > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Regards, > > > > > > > > > > > > > Oleg > > > > > > > > > > > > > > > > > > > > > > > > > > вт, 18 апр. 2023 г. в 20:44, Stefano Stabellini > > > > <sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>> > > > > > > > > > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>>> > > > > <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> > > > > <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>>>> > > > > > > > > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>> > > > > > > > > > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx> > <mailto:sstabellini@xxxxxxxxxx <mailto:sstabellini@xxxxxxxxxx>>>>> > > > > <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>> > > > > <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx> <mailto:sstabellini@xxxxxxxxxx > <mailto:sstabellini@xxxxxxxxxx>>>>>>>>: > > > > > > > > > > > > > > > > > > > > > > > > > > On Tue, 18 Apr 2023, Oleg Nikitenko wrote: > > > > > > > > > > > > > > Hi Julien, > > > > > > > > > > > > > > > > > > > > > > > > > > > > >> This feature has not been merged in Xen upstream yet > > > > > > > > > > > > > > > > > > > > > > > > > > > > > would assume that upstream + the series on the ML [1] > > > > work > > > > > > > > > > > > > > > > > > > > > > > > > > > > Please clarify this point. > > > > > > > > > > > > > > Because the two thoughts are controversial. > > > > > > > > > > > > > > > > > > > > > > > > > > Hi Oleg, > > > > > > > > > > > > > > > > > > > > > > > > > > As Julien wrote, there is nothing controversial. As you > > > > are aware, > > > > > > > > > > > > > Xilinx maintains a separate Xen tree specific for Xilinx > > > > here: > > > > > > > > > > > > > https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>> > > > > <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>>>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>>>>> > > > > > > > > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>>> > > > > > > > > > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>>>>> > > > > <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>>>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>>> > > > > <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen> <https://github.com/xilinx/xen > <https://github.com/xilinx/xen>>>>> > > > > > > > > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>>> > > > > > > > > > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen> > <https://github.com/xilinx/xen <https://github.com/xilinx/xen>>>>>>> > > > > > > > > > > > > > > > > > > > > > > > > > > and the branch you are using (xlnx_rebase_4.16) comes > > > > from there. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Instead, the upstream Xen tree lives here: > > > > > > > > > > > > > https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>> > > > > > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>> > > > > > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>>>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>> > > > > > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>> > > > > > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>> > > > > > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary> > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary > <https://xenbits.xen.org/gitweb/?p=xen.git;a=summary>>>>>>> > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > The Cache Coloring feature that you are trying to > > > > configure is present > > > > > > > > > > > > > in xlnx_rebase_4.16, but not yet present upstream (there > > > > is an > > > > > > > > > > > > > outstanding patch series to add cache coloring to Xen > > > > upstream but it > > > > > > > > > > > > > hasn't been merged yet.) > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Anyway, if you are using xlnx_rebase_4.16 it doesn't > > > > matter too much for > > > > > > > > > > > > > you as you already have Cache Coloring as a feature > > > > there. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > I take you are using ImageBuilder to generate the boot > > > > configuration? If > > > > > > > > > > > > > so, please post the ImageBuilder config file that you are > > > > using. > > > > > > > > > > > > > > > > > > > > > > > > > > But from the boot message, it looks like the colors > > > > configuration for > > > > > > > > > > > > > Dom0 is incorrect. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > >
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |