[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-users] Launching a PV Centos6 DomU crashing at "kernel BUG at fs/sysfs/group.c:65!; Kernel panic - not syncing: Fatal exception"



Hi Muriel,

On Tuesday, September 20, 2011 12:09 PM, "Muriel" <mucawhite@xxxxxxxxx>
wrote:
> My installation of SL6 is working with the kernel 32.131.. have you tried 
> this?

I had just download Scientific Linux 6.1 last night thinking it might
make a difference.

Changing my config file to

 name = 'SciLinux61'
 builder = 'linux'
 kernel = '/stor/vmlinuz'
 ramdisk = '/stor/initrd.img'
 disk = [
 'file:/stor/SL-61-x86_64-2011-07-27-Install-DVD.iso,hdc:cdrom,r',
 'phy:/dev/VG0/scilinux,xvda,w',
 ...

Launching the Guest still results in the crash.

Greg


CONSOLE OUTPUT AT CRASH --

xm create -c /stor/centos6_init.cfg

Using config file "/stor/centos6_init.cfg".

Started domain SciLinux61 (id=1)
                                Initializing cgroup subsys cpuset
Initializing cgroup subsys cpu
Linux version 2.6.32-131.0.15.el6.x86_64 (mockbuild@xxxxxxxxxxxx) (gcc
version 4.4.5 20110214 (Red Hat 4.4.5-6) (GCC) ) #1 SMP Sat May 21
10:27:57 CDT 2011
...
Freeing unused kernel memory: 1796k freed
------------[ cut here ]------------
kernel BUG at fs/sysfs/group.c:65!
invalid opcode: 0000 [#1] SMP
last sysfs file: /sys/devices/virtual/block/loop6/removable
CPU 0
Modules linked in: xen_blkfront(+) iscsi_ibft iscsi_boot_sysfs pcspkr
mlx4_ib mlx4_en mlx4_core ib_ipoib ib_cm ib_sa ib_mad ib_core ipv6
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi squashfs cramfs

Modules linked in: xen_blkfront(+) iscsi_ibft iscsi_boot_sysfs pcspkr
mlx4_ib mlx4_en mlx4_core ib_ipoib ib_cm ib_sa ib_mad ib_core ipv6
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi squashfs cramfs
Pid: 18, comm: xenwatch Tainted: G        W  ----------------  
2.6.32-131.0.15.el6.x86_64 #1
RIP: e030:[<ffffffff811e8457>]  [<ffffffff811e8457>]
internal_create_group+0xf7/0x1a0
RSP: e02b:ffff88003ebb1c60  EFLAGS: 00010246
RAX: 00000000ffffffef RBX: ffff88000555c000 RCX: ffff88003eac93c0
RDX: ffffffff81a5b480 RSI: 0000000000000000 RDI: ffff8800056ae870
RBP: ffff88003ebb1cb0 R08: 0000000000000004 R09: 0000000000000000
R10: 000000000000000f R11: 000000000000000f R12: ffff88000555c000
R13: ffff8800056ae870 R14: ffffffff81a5b480 R15: 0000000000000000
FS:  00007f994829b820(0000) GS:ffff880006066000(0000)
knlGS:0000000000000000
CS:  e033 DS: 0000 ES: 0000 CR0: 000000008005003b
CR2: 00007f85753136e8 CR3: 0000000005a5e000 CR4: 0000000000000660
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
Process xenwatch (pid: 18, threadinfo ffff88003ebb0000, task
ffff88003eb7f540)
Stack:
 ffff88003ebb1cc0 0000000081336e91 000032333a323032 ffff88003ebb1cd0
<0> ffff88003ebb1c90 ffff88000555c000 ffff88000555c000 ffff8800056ae860
<0> ffff8800056ae800 0000000000000000 ffff88003ebb1cc0 ffffffff811e8533
Call Trace:
 [<ffffffff811e8533>] sysfs_create_group+0x13/0x20
 [<ffffffff810f7724>] blk_trace_init_sysfs+0x14/0x20
 [<ffffffff8124bb60>] blk_register_queue+0x40/0x100
 [<ffffffff8125116e>] add_disk+0xae/0x160
 [<ffffffffa00c93e4>] backend_changed+0x374/0x700 [xen_blkfront]
 [<ffffffff81007b52>] ? check_events+0x12/0x20
 [<ffffffff812f7a4a>] otherend_changed+0xca/0x180
 [<ffffffff812f613a>] xenwatch_thread+0xaa/0x170
 [<ffffffff8108e100>] ? autoremove_wake_function+0x0/0x40
 [<ffffffff814dcf6c>] ? _spin_unlock_irqrestore+0x1c/0x20
 [<ffffffff812f6090>] ? xenwatch_thread+0x0/0x170
 [<ffffffff8108dd96>] kthread+0x96/0xa0
 [<ffffffff8100c1ca>] child_rip+0xa/0x20
 [<ffffffff8100b393>] ? int_ret_from_sys_call+0x7/0x1b
 [<ffffffff8100bb1d>] ? retint_restore_args+0x5/0x6
 [<ffffffff8100c1c0>] ? child_rip+0x0/0x20
Code: 8b 04 24 48 85 c0 74 27 41 83 c7 01 8b 55 bc 85 d2 74 b1 48 8b 30
48 89 df e8 76 be ff ff eb a4 48 83 7f 30 00 0f 85 49 ff ff ff <0f> 0b
eb fe 48 8b 5d c8 31 d2 48 85 db 74 18 f0 ff 0b 0f 94 c0
RIP  [<ffffffff811e8457>] internal_create_group+0xf7/0x1a0
 RSP <ffff88003ebb1c60>
---[ end trace f46928c89d14ef9b ]---
Kernel panic - not syncing: Fatal exception
Pid: 18, comm: xenwatch Tainted: G      D W  ----------------  
2.6.32-131.0.15.el6.x86_64 #1
Call Trace:
 [<ffffffff814da06e>] ? panic+0x78/0x143
 [<ffffffff814dcf6c>] ? _spin_unlock_irqrestore+0x1c/0x20
 [<ffffffff814de0b4>] ? oops_end+0xe4/0x100
 [<ffffffff8100f2eb>] ? die+0x5b/0x90
 [<ffffffff814dd984>] ? do_trap+0xc4/0x160
 [<ffffffff8100ceb5>] ? do_invalid_op+0x95/0xb0
 [<ffffffff811e8457>] ? internal_create_group+0xf7/0x1a0
 [<ffffffff8118d5aa>] ? ilookup5+0x4a/0x60
 [<ffffffff8100731d>] ? xen_force_evtchn_callback+0xd/0x10
 [<ffffffff81007b52>] ? check_events+0x12/0x20
 [<ffffffff8100bf5b>] ? invalid_op+0x1b/0x20
 [<ffffffff811e8457>] ? internal_create_group+0xf7/0x1a0
 [<ffffffff811e8533>] ? sysfs_create_group+0x13/0x20
 [<ffffffff810f7724>] ? blk_trace_init_sysfs+0x14/0x20
 [<ffffffff8124bb60>] ? blk_register_queue+0x40/0x100
 [<ffffffff8125116e>] ? add_disk+0xae/0x160
 [<ffffffffa00c93e4>] ? backend_changed+0x374/0x700 [xen_blkfront]
 [<ffffffff81007b52>] ? check_events+0x12/0x20
 [<ffffffff812f7a4a>] ? otherend_changed+0xca/0x180
 [<ffffffff812f613a>] ? xenwatch_thread+0xaa/0x170
 [<ffffffff8108e100>] ? autoremove_wake_function+0x0/0x40
 [<ffffffff814dcf6c>] ? _spin_unlock_irqrestore+0x1c/0x20
 [<ffffffff812f6090>] ? xenwatch_thread+0x0/0x170
 [<ffffffff8108dd96>] ? kthread+0x96/0xa0
 [<ffffffff8100c1ca>] ? child_rip+0xa/0x20
 [<ffffffff8100b393>] ? int_ret_from_sys_call+0x7/0x1b
 [<ffffffff8100bb1d>] ? retint_restore_args+0x5/0x6
 [<ffffffff8100c1c0>] ? child_rip+0x0/0x20

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.