LKML Archive on lore.kernel.org
help / color / mirror / Atom feed
From: Sean Christopherson <seanjc@google.com>
To: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: kvm@vger.kernel.org, Paolo Bonzini <pbonzini@redhat.com>,
Wanpeng Li <wanpengli@tencent.com>,
Jim Mattson <jmattson@google.com>,
"Dr. David Alan Gilbert" <dgilbert@redhat.com>,
Nitesh Narayan Lal <nitesh@redhat.com>,
Lai Jiangshan <jiangshanlai@gmail.com>,
Maxim Levitsky <mlevitsk@redhat.com>,
Eduardo Habkost <ehabkost@redhat.com>,
linux-kernel@vger.kernel.org
Subject: Re: [PATCH v4 3/8] KVM: x86: hyper-v: Avoid calling kvm_make_vcpus_request_mask() with vcpu_mask==NULL
Date: Thu, 2 Sep 2021 20:57:10 +0000 [thread overview]
Message-ID: <YTE6pk22mslvIDr6@google.com> (raw)
In-Reply-To: <20210827092516.1027264-4-vkuznets@redhat.com>
On Fri, Aug 27, 2021, Vitaly Kuznetsov wrote:
> In preparation to making kvm_make_vcpus_request_mask() use for_each_set_bit()
> switch kvm_hv_flush_tlb() to calling kvm_make_all_cpus_request() for 'all cpus'
> case.
>
> Note: kvm_make_all_cpus_request() (unlike kvm_make_vcpus_request_mask())
> currently dynamically allocates cpumask on each call and this is suboptimal.
> Both kvm_make_all_cpus_request() and kvm_make_vcpus_request_mask() are
> going to be switched to using pre-allocated per-cpu masks.
>
> Signed-off-by: Vitaly Kuznetsov <vkuznets@redhat.com>
> ---
Reviewed-by: Sean Christopherson <seanjc@google.com>
next prev parent reply other threads:[~2021-09-02 20:57 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-08-27 9:25 [PATCH v4 0/8] KVM: Various fixes and improvements around kicking vCPUs Vitaly Kuznetsov
2021-08-27 9:25 ` [PATCH v4 1/8] KVM: Clean up benign vcpu->cpu data races when " Vitaly Kuznetsov
2021-08-27 9:25 ` [PATCH v4 2/8] KVM: KVM: Use cpumask_available() to check for NULL cpumask " Vitaly Kuznetsov
2021-08-27 9:25 ` [PATCH v4 3/8] KVM: x86: hyper-v: Avoid calling kvm_make_vcpus_request_mask() with vcpu_mask==NULL Vitaly Kuznetsov
2021-09-02 20:57 ` Sean Christopherson [this message]
2021-08-27 9:25 ` [PATCH v4 4/8] KVM: Optimize kvm_make_vcpus_request_mask() a bit Vitaly Kuznetsov
2021-09-02 21:00 ` Sean Christopherson
2021-08-27 9:25 ` [PATCH v4 5/8] KVM: Drop 'except' parameter from kvm_make_vcpus_request_mask() Vitaly Kuznetsov
2021-09-02 21:00 ` Sean Christopherson
2021-08-27 9:25 ` [PATCH v4 6/8] KVM: x86: Fix stack-out-of-bounds memory access from ioapic_write_indirect() Vitaly Kuznetsov
2021-08-27 9:25 ` [PATCH v4 7/8] KVM: Pre-allocate cpumasks for kvm_make_all_cpus_request_except() Vitaly Kuznetsov
2021-09-02 21:08 ` Sean Christopherson
2021-09-03 7:20 ` Vitaly Kuznetsov
2021-09-03 14:54 ` Sean Christopherson
2021-08-27 9:25 ` [PATCH v4 8/8] KVM: Make kvm_make_vcpus_request_mask() use pre-allocated cpu_kick_mask Vitaly Kuznetsov
2021-09-02 21:19 ` Sean Christopherson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=YTE6pk22mslvIDr6@google.com \
--to=seanjc@google.com \
--cc=dgilbert@redhat.com \
--cc=ehabkost@redhat.com \
--cc=jiangshanlai@gmail.com \
--cc=jmattson@google.com \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mlevitsk@redhat.com \
--cc=nitesh@redhat.com \
--cc=pbonzini@redhat.com \
--cc=vkuznets@redhat.com \
--cc=wanpengli@tencent.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).