LKML Archive on lore.kernel.org
help / color / mirror / Atom feed
From: Suresh Siddha <suresh.b.siddha@intel.com>
To: Ingo Molnar <mingo@elte.hu>
Cc: Suresh Siddha <suresh.b.siddha@intel.com>,
	hpa@zytor.com, tglx@linutronix.de, andi@firstfloor.org,
	hch@infradead.org, linux-kernel@vger.kernel.org,
	Arjan van de Ven <arjan@linux.intel.com>
Subject: Re: [patch 1/2] x86, fpu: split FPU state from task struct - v3
Date: Tue, 4 Mar 2008 09:59:10 -0800	[thread overview]
Message-ID: <20080304175909.GE28006@linux-os.sc.intel.com> (raw)
In-Reply-To: <20080304102804.GA17681@elte.hu>

On Tue, Mar 04, 2008 at 11:28:04AM +0100, Ingo Molnar wrote:
> 
> * Suresh Siddha <suresh.b.siddha@intel.com> wrote:
> 
> > Split the FPU save area from the task struct. This allows easy 
> > migration of FPU context, and it's generally cleaner. It also allows 
> > the following two optimizations:
> 
> hm, i didnt get a reply from you to:
> 
>     http://lkml.org/lkml/2008/3/3/47
> 
> I'm uneasy to apply patches without knowing whether previous problems 
> are fixed. In the changelog there's a mention of:
> 
>   v3: Fixed the non-atomic calling sequence in atomic context.
> 
> is that the same bug?

Yes. Sorry for not pointing out explicitly. Essentially 32bit kernel with
CONFIG_PREEMPT is calling math_state_restore() with interrupts disabled.
Handled this by enabling/disabling the interrupts around the blocking call.
with small code changes ensuring that it is safe to enable/disable interrupts
at this point.

Do you want a v4 or separate patch for handling the kmem_cache_alloc() failure

thanks,
suresh

  reply	other threads:[~2008-03-04 18:02 UTC|newest]

Thread overview: 26+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2008-03-03 23:02 Suresh Siddha
2008-03-03 23:02 ` [patch 2/2] x86, fpu: lazy allocation of FPU area " Suresh Siddha
2008-03-04  1:20   ` Christoph Hellwig
2008-03-04  1:43     ` Suresh Siddha
2008-03-04 10:32       ` Ingo Molnar
2008-03-04 17:55         ` Suresh Siddha
2008-03-05 19:47           ` Pavel Machek
2008-03-06 15:51             ` Ingo Molnar
2008-03-06 19:10               ` Suresh Siddha
2008-03-06 20:24               ` Pavel Machek
2008-03-06 20:52                 ` Andi Kleen
2008-03-07 12:29                 ` H. Peter Anvin
2008-03-07 13:06                   ` Arjan van de Ven
2008-03-07 13:18                     ` Andi Kleen
2008-03-07 13:20                       ` Arjan van de Ven
2008-03-07 13:27                         ` Andi Kleen
2008-03-05 19:48   ` Pavel Machek
2008-03-06 19:26     ` Suresh Siddha
2008-03-06 21:21       ` Pavel Machek
2008-03-04  1:18 ` [patch 1/2] x86, fpu: split FPU state from task struct " Christoph Hellwig
2008-03-04  1:36   ` Suresh Siddha
2008-03-04  8:20     ` Ingo Molnar
2008-03-06 12:39     ` Christoph Hellwig
2008-03-04 10:28 ` Ingo Molnar
2008-03-04 17:59   ` Suresh Siddha [this message]
2008-03-04 20:53     ` Ingo Molnar

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20080304175909.GE28006@linux-os.sc.intel.com \
    --to=suresh.b.siddha@intel.com \
    --cc=andi@firstfloor.org \
    --cc=arjan@linux.intel.com \
    --cc=hch@infradead.org \
    --cc=hpa@zytor.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=mingo@elte.hu \
    --cc=tglx@linutronix.de \
    --subject='Re: [patch 1/2] x86, fpu: split FPU state from task struct - v3' \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).