[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH 12/16] x86/extable: Adjust extable handling to be shadow stack compatible


  • To: Jan Beulich <jbeulich@xxxxxxxx>
  • From: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
  • Date: Tue, 12 May 2020 17:14:30 +0100
  • Authentication-results: esa4.hc3370-68.iphmx.com; dkim=none (message not signed) header.i=none; spf=None smtp.pra=andrew.cooper3@xxxxxxxxxx; spf=Pass smtp.mailfrom=Andrew.Cooper3@xxxxxxxxxx; spf=None smtp.helo=postmaster@xxxxxxxxxxxxxxx; dmarc=pass (p=none dis=none) d=citrix.com
  • Cc: Xen-devel <xen-devel@xxxxxxxxxxxxxxxxxxxx>, Wei Liu <wl@xxxxxxx>, Roger Pau Monné <roger.pau@xxxxxxxxxx>
  • Delivery-date: Tue, 12 May 2020 16:14:38 +0000
  • Ironport-sdr: mgzBGGFjidN/99PQ2OrbS9ToHCO1nogmQvIc/fAxbuJRXysrFvTimlBMYM5sHZUaIFXOuaiSej Po6hdEOPLVnqE0ZhXtUAqJ/SUUhyqyIxWfBJ139LS0BrK8bKE6WTRdyL+/TqrN/XHnk4l8N16r vamD8OQqXl4P3ukWK2vtoFyf9gJufgNXoVWVM8h/RVeHUfYtuvSlbfyf4SRXQk9g72kJMIi2as NmUpksrptk1aEzggKrNVT2EpNe7BtWg02AqGPzLrA+J/sfbfghqNM5NAbN2T3bn2aqR97stUUL HQ4=
  • List-id: Xen developer discussion <xen-devel.lists.xenproject.org>

On 12/05/2020 15:31, Jan Beulich wrote:
> On 11.05.2020 23:09, Andrew Cooper wrote:
>> On 07/05/2020 14:35, Jan Beulich wrote:
>>> On 02.05.2020 00:58, Andrew Cooper wrote:
>>>> --- a/xen/arch/x86/traps.c
>>>> +++ b/xen/arch/x86/traps.c
>>>> @@ -778,6 +778,28 @@ static bool exception_fixup(struct cpu_user_regs 
>>>> *regs, bool print)
>>>>                 vec_name(regs->entry_vector), regs->error_code,
>>>>                 _p(regs->rip), _p(regs->rip), _p(fixup));
>>>>  
>>>> +    if ( IS_ENABLED(CONFIG_XEN_SHSTK) )
>>>> +    {
>>>> +        unsigned long ssp;
>>>> +
>>>> +        asm ("rdsspq %0" : "=r" (ssp) : "0" (1) );
>>>> +        if ( ssp != 1 )
>>>> +        {
>>>> +            unsigned long *ptr = _p(ssp);
>>>> +
>>>> +            /* Search for %rip in the shadow stack, ... */
>>>> +            while ( *ptr != regs->rip )
>>>> +                ptr++;
>>> Wouldn't it be better to bound the loop, as it shouldn't search past
>>> (strictly speaking not even to) the next page boundary? Also you
>>> don't care about the top of the stack (being the to be restored SSP),
>>> do you? I.e. maybe
>>>
>>>             while ( *++ptr != regs->rip )
>>>                 ;
>>>
>>> ?
>>>
>>> And then - isn't searching for a specific RIP value alone prone to
>>> error, in case a it matches an ordinary return address? I.e.
>>> wouldn't you better search for a matching RIP accompanied by a
>>> suitable pointer into the shadow stack and a matching CS value?
>>> Otherwise, ...
>>>
>>>> +            ASSERT(ptr[1] == __HYPERVISOR_CS);
>>> ... also assert that ptr[-1] points into the shadow stack?
>> So this is the problem I was talking about that the previous contexts
>> SSP isn't stored anywhere helpful.
>>
>> What we are in practice doing is looking 2 or 3 words up the shadow
>> stack (depending on exactly how deep our call graph is), to the shadow
>> IRET frame matching the real IRET frame which regs is pointing to.
>>
>> Both IRET frames were pushed in the process of generating the exception,
>> and we've already matched regs->rip to the exception table record.  We
>> need to fix up regs->rip and the shadow lip to the fixup address.
>>
>> As we are always fixing up an exception generated from Xen context, we
>> know that ptr[1] == __HYPERVISOR_CS, and *ptr[-1] = &ptr[2], as we
>> haven't switched shadow stack as part of taking this exception. 
>> However, this second point is fragile to exception handlers moving onto IST.
>>
>> We can't encounter regs->rip in the shadow stack between the current SSP
>> and the IRET frame we're looking to fix up, or we would have taken a
>> recursive fault and not reached exception_fixup() to begin with.
> I'm afraid I don't follow here. Consider a function (also)
> involved in exception handling having this code sequence:
>
>     call    func
>     mov     (%rax), %eax
>
> If the fault we're handling occured on the MOV and
> exception_fixup() is a descendant of func(), then the first
> instance of an address on the shadow stack pointing at this
> MOV is going to be the one which did not fault.

No.  The moment `call func` returns, the address you're looking to match
is rubble no longer on the stack.  Numerically, it will be located at
SSP-8 when the fault for MOV is generated.

In this exact case, it would be clobbered by the shadow IRET frame, but
even if it was deeper in the call tree, we would still never encounter
it from waking up the shadow stack from SSP.

The only things you will find on the shadow stack is the shadow IRET
frame, handle_exception_saved(), do_*(), fixup_exception(), except that
I'd not like to fix the behaviour to require exactly two function calls
of depth for fixup_exception().

>> Therefore, the loop is reasonably bounded in all cases.
>>
>> Sadly, there is no RDSS instruction, so we can't actually use shadow
>> stack reads to spot if we underflowed the shadow stack, and there is no
>> useful alternative to panic() if we fail to find the shadow IRET frame.
> But afaics you don't panic() in this case. Instead you continue
> looping until (presumably) you hit some form of fault.
>
>>>> --- a/xen/arch/x86/x86_64/entry.S
>>>> +++ b/xen/arch/x86/x86_64/entry.S
>>>> @@ -708,7 +708,16 @@ exception_with_ints_disabled:
>>>>          call  search_pre_exception_table
>>>>          testq %rax,%rax                 # no fixup code for faulting EIP?
>>>>          jz    1b
>>>> -        movq  %rax,UREGS_rip(%rsp)
>>>> +        movq  %rax,UREGS_rip(%rsp)      # fixup regular stack
>>>> +
>>>> +#ifdef CONFIG_XEN_SHSTK
>>>> +        mov    $1, %edi
>>>> +        rdsspq %rdi
>>>> +        cmp    $1, %edi
>>>> +        je     .L_exn_shstk_done
>>>> +        wrssq  %rax, (%rdi)             # fixup shadow stack
>>>> +.L_exn_shstk_done:
>>>> +#endif
>>> Again avoid the conditional jump by using alternatives patching?
>> Well - that depends on whether we're likely to gain any new content in
>> the pre exception table.
>>
>> As it stands, it is only the IRET(s) to userspace so would be safe to
>> turn this into an unconditional alternative.  Even in the crash case, we
>> won't be returning to guest context after having started the crash
>> teardown path.
> Ah, right - perhaps indeed better keep it as is then.

That was my reasoning.  It is a path we expect to execute never with
well behaved guests, so I erred on the safe side.

~Andrew



 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.