|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [BUG] hhvm running on Ubuntu 13.04 with Xen Hypervisor - linux kernel panic
Thanks! But when I run
./configure -DNO_HARDWARE_COUNTERS=1
I get:
Manually-specified variables were not used by the project:
NO_HARDWARE_COUNTERS
On 19/09/2013 10:52, "Wei Liu" <wei.liu2@xxxxxxxxxx> wrote:
>On Wed, Sep 18, 2013 at 11:21:18AM +0000, Craig Carnell wrote:
>> Hi,
>>
>> I am trying out hiphop vm (the php just in time compiler). My setup is
>>a Rackspace Cloud Server running Ubuntu 13.04 with kernel
>>3.8.0-30-generic #44-Ubuntu SMP Thu Aug 22 20:52:24 UTC 2013 x86_64
>>x86_64 x86_64 GNU/Linux
>>
>> The cloud server uses Xen Hypervisor.
>>
>> Hiphopvm is compiled from source using the github repo. When running
>>hhvm from the command line (without any options or php application) the
>>system immediately crashes, throwing linux into a kernel panic and thus
>>death.
>>
>> I have reported this issue on hiphop github issue page:
>>
>> https://github.com/facebook/hiphop-php/issues/1065
>>
>> I am not sure if this is a linux kernel bug or a xen hypervisor bug:
>>
>
>I'm not a expert on VPMU stuffs, but it seems that HHVM makes use of
>(virtual) hardware performance counter which is not well supported at
>the moment, which causes this problem.
>
>Try to compile HHVM without hardware performance counter support might
>solve this problem.
>
> ./configure -DNO_HARDWARE_COUNTERS=1
>
>Wei.
>
>> The output of /var/log/syslog:
>>
>> Sep 18 10:55:58 web kernel: [92118.674736] general protection fault:
>>0000 [#1] SMP
>> Sep 18 10:55:58 web kernel: [92118.674754] Modules linked in: xenfs(F)
>>xen_privcmd(F) xt_tcpudp(F) nf_conntrack_ipv4(F) nf_defrag_ipv4(F)
>>xt_state(F) nf_conntrack(F) xt_comment(F) iptable_filter(F) ip_tables(F)
>>x_tables(F) microcode(F) lp(F) parport(F)
>> Sep 18 10:55:58 web kernel: [92118.674781] CPU 0
>> Sep 18 10:55:58 web kernel: [92118.674787] Pid: 5020, comm: hhvm
>>Tainted: GF 3.8.0-30-generic #44-Ubuntu
>> Sep 18 10:55:58 web kernel: [92118.674795] RIP:
>>e030:[<ffffffff81003046>] [<ffffffff81003046>] native_read_pmc+0x6/0x20
>> Sep 18 10:55:58 web kernel: [92118.674809] RSP: e02b:ffff8800026b9d20
>>EFLAGS: 00010083
>> Sep 18 10:55:58 web kernel: [92118.674814] RAX: ffffffff81c1bd80 RBX:
>>0000000000000000 RCX: 0000000000000000
>> Sep 18 10:55:58 web kernel: [92118.674819] RDX: 0000000000005f6c RSI:
>>ffff8800f7c81900 RDI: 0000000000000000
>> Sep 18 10:55:58 web kernel: [92118.674824] RBP: ffff8800026b9d20 R08:
>>00000000000337d8 R09: ffff8800e933dcc0
>> Sep 18 10:55:58 web kernel: [92118.674830] R10: 00007fff2d3caea0 R11:
>>0000000000000246 R12: ffff8800f87ecc00
>> Sep 18 10:55:58 web kernel: [92118.674835] R13: ffff800000000001 R14:
>>ffff8800f87ecd70 R15: 0000000000000010
>> Sep 18 10:55:58 web kernel: [92118.674844] FS: 00007f43d4c9b180(0000)
>>GS:ffff8800ffc00000(0000) knlGS:0000000000000000
>> Sep 18 10:55:58 web kernel: [92118.674850] CS: e033 DS: 0000 ES: 0000
>>CR0: 000000008005003b
>> Sep 18 10:55:58 web kernel: [92118.674855] CR2: 000000000105ebc0 CR3:
>>00000000025cd000 CR4: 0000000000000660
>> Sep 18 10:55:58 web kernel: [92118.674861] DR0: 0000000000000000 DR1:
>>0000000000000000 DR2: 0000000000000000
>> Sep 18 10:55:58 web kernel: [92118.674867] DR3: 0000000000000000 DR6:
>>00000000ffff0ff0 DR7: 0000000000000400
>> Sep 18 10:55:58 web kernel: [92118.674872] Process hhvm (pid: 5020,
>>threadinfo ffff8800026b8000, task ffff8800f7cfc5c0)
>> Sep 18 10:55:58 web kernel: [92118.674879] Stack:
>> Sep 18 10:55:58 web kernel: [92118.674882] ffff8800026b9d58
>>ffffffff81024625 0000000000000000 ffff8800f87ecc00
>> Sep 18 10:55:58 web kernel: [92118.674893] ffff8800f7c8190c
>>ffffffff811231a0 0000000000000005 ffff8800026b9d68
>> Sep 18 10:55:58 web kernel: [92118.674902] ffffffff81024689
>>ffff8800026b9d90 ffffffff811232a6 00000000ffff02ff
>> Sep 18 10:55:58 web kernel: [92118.674911] Call Trace:
>> Sep 18 10:55:58 web kernel: [92118.674920] [<ffffffff81024625>]
>>x86_perf_event_update+0x55/0xb0
>> Sep 18 10:55:58 web kernel: [92118.674929] [<ffffffff811231a0>] ?
>>perf_read+0x2f0/0x2f0
>> Sep 18 10:55:58 web kernel: [92118.674936] [<ffffffff81024689>]
>>x86_pmu_read+0x9/0x10
>> Sep 18 10:55:58 web kernel: [92118.674942] [<ffffffff811232a6>]
>>__perf_event_read+0x106/0x110
>> Sep 18 10:55:58 web kernel: [92118.674951] [<ffffffff810b9987>]
>>smp_call_function_single+0x147/0x170
>> Sep 18 10:55:58 web kernel: [92118.674959] [<ffffffff811240d0>] ?
>>perf_mmap+0x2f0/0x2f0
>> Sep 18 10:55:58 web kernel: [92118.674966] [<ffffffff81122dda>]
>>perf_event_read+0x10a/0x110
>> Sep 18 10:55:58 web kernel: [92118.674972] [<ffffffff811240d0>] ?
>>perf_mmap+0x2f0/0x2f0
>> Sep 18 10:55:58 web kernel: [92118.674979] [<ffffffff811240dd>]
>>perf_event_reset+0xd/0x20
>> Sep 18 10:55:58 web kernel: [92118.674987] [<ffffffff8111ff08>]
>>perf_event_for_each_child+0x38/0xa0
>> Sep 18 10:55:58 web kernel: [92118.674994] [<ffffffff811240d0>] ?
>>perf_mmap+0x2f0/0x2f0
>> Sep 18 10:55:58 web kernel: [92118.675001] [<ffffffff8112255a>]
>>perf_ioctl+0xba/0x340
>> Sep 18 10:55:58 web kernel: [92118.675009] [<ffffffff811b1885>] ?
>>fd_install+0x25/0x30
>> Sep 18 10:55:58 web kernel: [92118.675016] [<ffffffff811a60e9>]
>>do_vfs_ioctl+0x99/0x570
>> Sep 18 10:55:58 web kernel: [92118.675023] [<ffffffff811a6651>]
>>sys_ioctl+0x91/0xb0
>> Sep 18 10:55:58 web kernel: [92118.675031] [<ffffffff816d575d>]
>>system_call_fastpath+0x1a/0x1f
>> Sep 18 10:55:58 web kernel: [92118.675036] Code: 00 00 00 00 00 55 89
>>f9 48 89 e5 0f 32 31 ff 89 c0 48 c1 e2 20 89 3e 48 09 c2 48 89 d0 5d c3
>>66 0f 1f 44 00 00 55 89 f9 48 89 e5 <0f> 33 89 c0 48 c1 e2 20 48 09 c2
>>48 89 d0 5d c3 66 2e 0f 1f 84
>> Sep 18 10:55:58 web kernel: [92118.675103] RIP [<ffffffff81003046>]
>>native_read_pmc+0x6/0x20
>> Sep 18 10:55:58 web kernel: [92118.675110] RSP <ffff8800026b9d20>
>> Sep 18 10:55:58 web kernel: [92118.675118] ---[ end trace
>>1a73231ba5f74716 ]---
>>
>
>> _______________________________________________
>> Xen-devel mailing list
>> Xen-devel@xxxxxxxxxxxxx
>> http://lists.xen.org/xen-devel
>
>
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |