LKML Archive on lore.kernel.org
help / color / mirror / Atom feed
From: Jan Beulich <jbeulich@suse.com>
To: Miroslav Benes <mbenes@suse.cz>
Cc: boris.ostrovsky@oracle.com, jgross@suse.com,
	sstabellini@kernel.org, tglx@linutronix.de, mingo@redhat.com,
	bp@alien8.de, hpa@zytor.com, jpoimboe@redhat.com,
	andrew.cooper3@citrix.com, x86@kernel.org,
	linux-kernel@vger.kernel.org, live-patching@vger.kernel.org,
	xen-devel@lists.xenproject.org, jslaby@suse.cz
Subject: Re: [PATCH v2 1/2] x86/xen: Make the boot CPU idle task reliable
Date: Thu, 19 Mar 2020 11:01:21 +0100	[thread overview]
Message-ID: <71c4eeaf-958a-b215-3033-c3e0d74a9cfa@suse.com> (raw)
In-Reply-To: <20200319095606.23627-2-mbenes@suse.cz>

On 19.03.2020 10:56, Miroslav Benes wrote:
> The unwinder reports the boot CPU idle task's stack on XEN PV as
> unreliable, which affects at least live patching. There are two reasons
> for this. First, the task does not follow the x86 convention that its
> stack starts at the offset right below saved pt_regs. It allows the
> unwinder to easily detect the end of the stack and verify it. Second,
> startup_xen() function does not store the return address before jumping
> to xen_start_kernel() which confuses the unwinder.
> 
> Amend both issues by moving the starting point of initial stack in
> startup_xen() and storing the return address before the jump, which is
> exactly what call instruction does.
> 
> Signed-off-by: Miroslav Benes <mbenes@suse.cz>
> ---
>  arch/x86/xen/xen-head.S | 8 ++++++--
>  1 file changed, 6 insertions(+), 2 deletions(-)
> 
> diff --git a/arch/x86/xen/xen-head.S b/arch/x86/xen/xen-head.S
> index 1d0cee3163e4..edc776af0e0a 100644
> --- a/arch/x86/xen/xen-head.S
> +++ b/arch/x86/xen/xen-head.S
> @@ -35,7 +35,11 @@ SYM_CODE_START(startup_xen)
>  	rep __ASM_SIZE(stos)
>  
>  	mov %_ASM_SI, xen_start_info
> -	mov $init_thread_union+THREAD_SIZE, %_ASM_SP
> +#ifdef CONFIG_X86_64
> +	mov initial_stack(%rip), %_ASM_SP
> +#else
> +	mov pa(initial_stack), %_ASM_SP
> +#endif

If you need to distinguish the two anyway, why not use %rsp and
%esp respectively?

Jan

  reply	other threads:[~2020-03-19 10:01 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-03-19  9:56 [PATCH v2 0/2] x86/xen: Make idle tasks reliable Miroslav Benes
2020-03-19  9:56 ` [PATCH v2 1/2] x86/xen: Make the boot CPU idle task reliable Miroslav Benes
2020-03-19 10:01   ` Jan Beulich [this message]
2020-03-19 10:31     ` Miroslav Benes
2020-03-19  9:56 ` [PATCH v2 2/2] x86/xen: Make the secondary CPU idle tasks reliable Miroslav Benes
2020-03-19 10:03   ` Jan Beulich
2020-03-19 10:38     ` Miroslav Benes

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=71c4eeaf-958a-b215-3033-c3e0d74a9cfa@suse.com \
    --to=jbeulich@suse.com \
    --cc=andrew.cooper3@citrix.com \
    --cc=boris.ostrovsky@oracle.com \
    --cc=bp@alien8.de \
    --cc=hpa@zytor.com \
    --cc=jgross@suse.com \
    --cc=jpoimboe@redhat.com \
    --cc=jslaby@suse.cz \
    --cc=linux-kernel@vger.kernel.org \
    --cc=live-patching@vger.kernel.org \
    --cc=mbenes@suse.cz \
    --cc=mingo@redhat.com \
    --cc=sstabellini@kernel.org \
    --cc=tglx@linutronix.de \
    --cc=x86@kernel.org \
    --cc=xen-devel@lists.xenproject.org \
    --subject='Re: [PATCH v2 1/2] x86/xen: Make the boot CPU idle task reliable' \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).