[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] Out sw-iommu space problem
We have dom0 Squeeze with official kernel (2.6.32-35squeeze2) and xen 4.0.2-rc3 from git testing on DELL poweredge T310 with raid controller H200 (all with latest firmware). Some domus linux pv and some windows with gplpv. We randomly running into this error during xendomains stop command (with save of all domus): Sep 14 16:18:40 heliMN02WV kernel: [ 912.336945] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:18:40 heliMN02WV kernel: [ 912.336951] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:18:41 heliMN02WV kernel: [ 912.400331] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:18:41 heliMN02WV kernel: [ 912.400336] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:18:45 heliMN02WV kernel: [ 917.187524] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:18:45 heliMN02WV kernel: [ 917.187533] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:18:51 heliMN02WV kernel: [ 922.454599] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:18:51 heliMN02WV kernel: [ 922.454608] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:18:51 heliMN02WV kernel: [ 922.454694] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:18:51 heliMN02WV kernel: [ 922.454697] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:11 heliMN02WV kernel: [ 942.850048] frontend_changed: backend/vbd/4/768: prepare for reconnect Sep 14 16:19:11 heliMN02WV kernel: [ 942.869244] eth0: port 5(vif4.0) entering disabled state Sep 14 16:19:11 heliMN02WV kernel: [ 942.889192] eth0: port 5(vif4.0) entering disabled state Sep 14 16:19:11 heliMN02WV kernel: [ 943.188048] frontend_changed: backend/vif/4/0: prepare for reconnect Sep 14 16:19:22 heliMN02WV kernel: [ 954.246090] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:19:22 heliMN02WV kernel: [ 954.246095] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:22 heliMN02WV kernel: [ 954.305068] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:19:22 heliMN02WV kernel: [ 954.305074] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:34 heliMN02WV kernel: [ 966.112058] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:19:34 heliMN02WV kernel: [ 966.112064] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:34 heliMN02WV kernel: [ 966.112251] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:19:34 heliMN02WV kernel: [ 966.112255] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:34 heliMN02WV kernel: [ 966.112440] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:19:34 heliMN02WV kernel: [ 966.112443] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:34 heliMN02WV kernel: [ 966.205690] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:19:34 heliMN02WV kernel: [ 966.205693] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:40 heliMN02WV kernel: [ 971.728913] eth0: port 6(vif5.0) entering disabled state Sep 14 16:19:40 heliMN02WV kernel: [ 971.752683] eth0: port 6(vif5.0) entering disabled state Sep 14 16:19:45 heliMN02WV kernel: [ 976.984329] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:19:45 heliMN02WV kernel: [ 976.984333] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:19:49 heliMN02WV kernel: [ 981.288632] eth0: port 7(vif6.0) entering disabled state Sep 14 16:19:49 heliMN02WV kernel: [ 981.304521] eth0: port 7(vif6.0) entering disabled state Sep 14 16:19:50 heliMN02WV kernel: [ 982.329740] frontend_changed: backend/vbd/7/768: prepare for reconnect Sep 14 16:19:50 heliMN02WV kernel: [ 982.372593] eth0: port 8(vif7.0) entering disabled state Sep 14 16:19:51 heliMN02WV kernel: [ 982.416506] eth0: port 8(vif7.0) entering disabled state Sep 14 16:19:51 heliMN02WV kernel: [ 982.744206] frontend_changed: backend/vif/7/0: prepare for reconnect Sep 14 16:20:00 heliMN02WV kernel: [ 991.520780] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:20:00 heliMN02WV kernel: [ 991.520787] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:20:00 heliMN02WV kernel: [ 991.524695] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:20:00 heliMN02WV kernel: [ 991.524698] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:20:00 heliMN02WV kernel: [ 991.525040] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:20:00 heliMN02WV kernel: [ 991.525042] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:20:00 heliMN02WV kernel: [ 991.525371] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:20:00 heliMN02WV kernel: [ 991.525374] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:20:00 heliMN02WV kernel: [ 991.527766] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:20:00 heliMN02WV kernel: [ 991.527769] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:20:01 heliMN02WV kernel: [ 992.493163] mpt2sas 0000:03:00.0: DMA: Out of SW-IOMMU space for 65536 bytes. Sep 14 16:20:01 heliMN02WV kernel: [ 992.493167] sd 0:1:0:0: pci_map_sg failed: request for 524288 bytes! Sep 14 16:20:13 heliMN02WV kernel: [ 1004.938378] tapdisk2[7617]: segfault at 7fff92fb6ff8 ip 0000000000408296 sp 00007fff92fb7000 error 6 in tapdisk2[400000+39000] Sep 14 16:20:13 heliMN02WV kernel: [ 1004.959533] BUG: unable to handle kernel NULL pointer dereference at 0000000000000048 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.959681] IP: [<ffffffff810ce79e>] apply_to_page_range+0x47/0x2f3 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.959773] PGD 3dc5f067 PUD 3db57067 PMD 0 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.959914] Oops: 0000 [#1] SMP Sep 14 16:20:13 heliMN02WV kernel: [ 1004.960026] last sysfs file: /sys/devices/virtual/blktap2/blktap11/remove Sep 14 16:20:13 heliMN02WV kernel: [ 1004.960084] CPU 5 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.960161] Modules linked in: xt_tcpudp tun xt_physdev iptable_filter ip_tables x_tables bridge stp ib_iser rdma_cm ib_cm iw_cm ib_sa ib_mad ib_core ib_addr iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ext2 sha256_generic aes_x86_64 aes_generic cbc blktap xen_evtchn xenfs loop dm_crypt dcdbas snd_pcm snd_timer snd joydev evdev soundcore snd_page_alloc pcspkr power_meter button processor acpi_processor ext4 mbcache jbd2 crc16 dm_mod sd_mod crc_t10dif sg usbhid hid sr_mod cdrom usb_storage ata_generic ehci_hcd ata_piix mpt2sas bnx2 scsi_transport_sas libata usbcore nls_base scsi_mod thermal thermal_sys [last unloaded: scsi_wait_scan] Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962493] Pid: 7617, comm: tapdisk2 Not tainted 2.6.32-5-xen-amd64 #1 PowerEdge T310 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962566] RIP: e030:[<ffffffff810ce79e>] [<ffffffff810ce79e>] apply_to_page_range+0x47/0x2f3 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962671] RSP: e02b:ffff88003dfc9b58 EFLAGS: 00010202 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962726] RAX: 0000000000000880 RBX: ffff88003d8ad000 RCX: ffff88003d8ae000 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962783] RDX: 0000000000000000 RSI: ffff88003d8ad000 RDI: 0000000000000000 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962840] RBP: ffff88003eff2dd0 R08: 0000000000000000 R09: ffff88003f96c180 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962898] R10: 0000000000000002 R11: 0000000000000000 R12: 0000000000000000 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.962955] R13: ffff88003eff2dd0 R14: ffff88003e149000 R15: 0000000000000000 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963014] FS: 00007f6d3c738740(0000) GS:ffff880003782000(0000) knlGS:0000000000000000 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963088] CS: e033 DS: 0000 ES: 0000 CR0: 000000008005003b Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963143] CR2: 0000000000000048 CR3: 000000003f0dc000 CR4: 0000000000002660 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963201] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963258] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963316] Process tapdisk2 (pid: 7617, threadinfo ffff88003dfc8000, task ffff880036a88e20) Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963389] Stack: Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963437] 0000000000000000 ffff88003ea87b40 0000000000000000 0000000000000000 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963583] <0> ffffffffa02f1ee8 0000000000000000 ffffffff8100ece2 ffff880002155480 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.963802] <0> ffff88003d8ae000 0000000000000000 0000000000000000 ffff880002155480 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964066] Call Trace: Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964119] [<ffffffffa02f1ee8>] ? blktap_umap_uaddr_fn+0x0/0x59 [blktap] Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964179] [<ffffffff8100ece2>] ? check_events+0x12/0x20 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964236] [<ffffffffa02f32a5>] ? blktap_device_end_request+0xbd/0x145 [blktap] Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964310] [<ffffffffa02f1743>] ? blktap_ring_vm_close+0x60/0xd1 [blktap] Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964368] [<ffffffff810d13f8>] ? remove_vma+0x2c/0x72 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964423] [<ffffffff810d1567>] ? exit_mmap+0x129/0x148 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964479] [<ffffffff8104cc5d>] ? mmput+0x3c/0xdf Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964534] [<ffffffff81050862>] ? exit_mm+0x102/0x10d Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964592] [<ffffffff8130d0d2>] ? _spin_lock_irq+0x7/0x22 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964648] [<ffffffff81052287>] ? do_exit+0x1f8/0x6c6 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964703] [<ffffffff8105d5a1>] ? __dequeue_signal+0xfb/0x124 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964760] [<ffffffff8100eccf>] ? xen_restore_fl_direct_end+0x0/0x1 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964817] [<ffffffff810e7f35>] ? kmem_cache_free+0x72/0xa3 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964874] [<ffffffff810527cb>] ? do_group_exit+0x76/0x9d Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964930] [<ffffffff8105f0b7>] ? get_signal_to_deliver+0x310/0x339 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.964987] [<ffffffff8101104f>] ? do_notify_resume+0x87/0x73f Sep 14 16:20:13 heliMN02WV kernel: [ 1004.965044] [<ffffffff810d15e1>] ? expand_downwards+0x5b/0x169 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.965101] [<ffffffff8130f589>] ? do_page_fault+0x1f3/0x2f2 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.965157] [<ffffffff810125dc>] ? retint_signal+0x48/0x8c Sep 14 16:20:13 heliMN02WV kernel: [ 1004.965211] Code: 48 89 4c 24 20 4c 89 44 24 18 48 89 54 24 40 72 04 0f 0b eb fe 48 8b 54 24 28 48 89 f0 48 8b 4c 24 40 48 c1 e8 24 25 f8 0f 00 00 <48> 8b 52 48 48 ff c9 48 89 0c 24 48 01 d0 48 89 44 24 30 48 b8 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.967243] RIP [<ffffffff810ce79e>] apply_to_page_range+0x47/0x2f3 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.967336] RSP <ffff88003dfc9b58> Sep 14 16:20:13 heliMN02WV kernel: [ 1004.967388] CR2: 0000000000000048 Sep 14 16:20:13 heliMN02WV kernel: [ 1004.967440] ---[ end trace 78b5f16c10850a91 ]--- Sep 14 16:20:13 heliMN02WV kernel: [ 1004.967495] Fixing recursive fault but reboot is needed! Rebooting the systems doesn't resolve this problem. We have also try to add swiotlb=128 on vmlinuz line but the systems always loops with "out sw-iommu space" message (probably also bug in swiotlb switch on kernel). Can someone help us to solve this problem please? -- View this message in context: http://xen.1045712.n5.nabble.com/Out-sw-iommu-space-problem-tp4803078p4803078.html Sent from the Xen - Dev mailing list archive at Nabble.com. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |