LKML Archive on lore.kernel.org
help / color / mirror / Atom feed
* [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL
@ 2021-08-23 20:42 Ritesh Harjani
  2021-08-23 20:42 ` [PATCHv2 2/2] lib/percpu_test: Add extra tests in percpu_test Ritesh Harjani
  2021-08-24  2:06 ` [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Dennis Zhou
  0 siblings, 2 replies; 7+ messages in thread
From: Ritesh Harjani @ 2021-08-23 20:42 UTC (permalink / raw)
  To: linux-mm
  Cc: linux-kernel, Dennis Zhou, Tejun Heo, Christoph Lameter, Ritesh Harjani

Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL

Signed-off-by: Ritesh Harjani <riteshh@linux.ibm.com>
---
[v1 -> v2]: Use EXPORT_SYMBOL_GPL instead of EXPORT_SYMBOL

 kernel/workqueue.c | 1 +
 1 file changed, 1 insertion(+)

diff --git a/kernel/workqueue.c b/kernel/workqueue.c
index f148eacda55a..993f8983186d 100644
--- a/kernel/workqueue.c
+++ b/kernel/workqueue.c
@@ -3309,6 +3309,7 @@ int schedule_on_each_cpu(work_func_t func)
 	free_percpu(works);
 	return 0;
 }
+EXPORT_SYMBOL_GPL(schedule_on_each_cpu);

 /**
  * execute_in_process_context - reliably execute the routine with user context
--
2.31.1


^ permalink raw reply	[flat|nested] 7+ messages in thread

* [PATCHv2 2/2] lib/percpu_test: Add extra tests in percpu_test
  2021-08-23 20:42 [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Ritesh Harjani
@ 2021-08-23 20:42 ` Ritesh Harjani
  2021-08-24  2:14   ` Dennis Zhou
  2021-08-24  2:06 ` [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Dennis Zhou
  1 sibling, 1 reply; 7+ messages in thread
From: Ritesh Harjani @ 2021-08-23 20:42 UTC (permalink / raw)
  To: linux-mm
  Cc: linux-kernel, Dennis Zhou, Tejun Heo, Christoph Lameter,
	Ritesh Harjani, Aneesh Kumar K . V, Vaibhav Jain,
	kernel test robot

While debugging a issue, we needed to stress test the percpu alloc/free
path. Hence added some tests in lib/percpu_test to stress test
percpu subsystem for allocation with different sizes.

This patch keeps the default behavior of insmod module same for default
test. But when given insmod with different option, it can run a
percpu_stressd daemon (percpu_test_num=2) which does a stress test
evey 10secs unless the module is unloaded.

We found this to be helpful in our testing, since with this we could
easily excercise percpu allo/free path. Hence cleaned this up for
inclusion in percpu_test module.

Logs
======
qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=0
[  334.362973] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_verify
[  334.364946] TEST Starts: percpu_verify
[  334.365601] TEST Completed: percpu_verify
insmod: ERROR: could not insert module /mnt/percpu_test.ko: Resource temporarily unavailable

qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=1
[  336.556464] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_stress
[  336.558388] TEST Starts: percpu_stress
[  336.560611] TEST Completed: percpu_stress
insmod: ERROR: could not insert module /mnt/percpu_test.ko: Resource temporarily unavailable

qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=2
[  339.164406] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_stressd
[  339.165935] TEST Starts: percpu_stressd
[  339.167033] TEST Completed: percpu_stressd
[  339.167082] DAEMON: starts percpu_stressd
[  339.168498] TEST Starts: percpu_stressd: iter (1)
[  339.182530] TEST Completed: percpu_stressd: iter (1)
[  349.341109] TEST Starts: percpu_stressd: iter (2)
[  349.344447] TEST Completed: percpu_stressd: iter (2)
[  359.580829] TEST Starts: percpu_stressd: iter (3)
[  359.584315] TEST Completed: percpu_stressd: iter (3)
[  369.820471] TEST Starts: percpu_stressd: iter (4)
[  369.844402] TEST Completed: percpu_stressd: iter (4)

qemu-> sudo rmmod percpu_test
[  375.001098] percpu_test: EXIT
[qemu][~]

Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Vaibhav Jain <vaibhav@linux.ibm.com>
Reported-by: kernel test robot <lkp@intel.com>
Signed-off-by: Ritesh Harjani <riteshh@linux.ibm.com>
---
[v1 -> v2]: Fix warnings from kernel test robot

 lib/percpu_test.c | 240 ++++++++++++++++++++++++++++++++++++----------
 1 file changed, 191 insertions(+), 49 deletions(-)

diff --git a/lib/percpu_test.c b/lib/percpu_test.c
index 4a3d70bbc1a0..68c57c288dc6 100644
--- a/lib/percpu_test.c
+++ b/lib/percpu_test.c
@@ -1,4 +1,7 @@
 // SPDX-License-Identifier: GPL-2.0-only
+#include <linux/workqueue.h>
+#include <linux/kthread.h>
+#include <linux/cpu.h>
 #include <linux/module.h>

 /* validate @native and @pcp counter values match @expected */
@@ -14,10 +17,25 @@
 		     (long long)(expected), (long long)(expected));	\
 	} while (0)

-static DEFINE_PER_CPU(long, long_counter);
-static DEFINE_PER_CPU(unsigned long, ulong_counter);
+/* upto max alloc size tests for percpu var */
+static char __percpu *counters[1 << PAGE_SHIFT];
+static struct task_struct *percpu_stressd_thread;

-static int __init percpu_test_init(void)
+/* let's not trigger OOM */
+int percpu_alloc_max_size_shift = PAGE_SHIFT - 3;
+module_param(percpu_alloc_max_size_shift, int, 0644);
+MODULE_PARM_DESC(percpu_alloc_max_size_shift, "max size of allocation in stress test will be upto 1 << percpu_alloc_max_size_shift");
+
+static long percpu_stressd_interval = 1 * 10 * HZ;
+module_param(percpu_stressd_interval, long, 0644);
+MODULE_PARM_DESC(percpu_stressd_interval, "percpu_stressd internal");
+
+/* keep the default test same */
+static int percpu_test_num;
+module_param(percpu_test_num, int, 0644);
+MODULE_PARM_DESC(percpu_test_num, "Test number percpu_test_num");
+
+static int percpu_test_verify(void)
 {
 	/*
 	 * volatile prevents compiler from optimizing it uses, otherwise the
@@ -26,109 +44,233 @@ static int __init percpu_test_init(void)
 	volatile unsigned int ui_one = 1;
 	long l = 0;
 	unsigned long ul = 0;
+	long __percpu *long_counter = alloc_percpu(long);
+	unsigned long __percpu *ulong_counter = alloc_percpu(unsigned long);

-	pr_info("percpu test start\n");
+	if (!long_counter || !ulong_counter)
+		goto out;
+
+	pr_debug("percpu_test: %s start cpu: %d\n", __func__, smp_processor_id());

 	preempt_disable();

 	l += -1;
-	__this_cpu_add(long_counter, -1);
-	CHECK(l, long_counter, -1);
+	__this_cpu_add(*long_counter, -1);
+	CHECK(l, *long_counter, -1);

 	l += 1;
-	__this_cpu_add(long_counter, 1);
-	CHECK(l, long_counter, 0);
+	__this_cpu_add(*long_counter, 1);
+	CHECK(l, *long_counter, 0);

 	ul = 0;
-	__this_cpu_write(ulong_counter, 0);
+	__this_cpu_write(*ulong_counter, 0);

 	ul += 1UL;
-	__this_cpu_add(ulong_counter, 1UL);
-	CHECK(ul, ulong_counter, 1);
+	__this_cpu_add(*ulong_counter, 1UL);
+	CHECK(ul, *ulong_counter, 1);

 	ul += -1UL;
-	__this_cpu_add(ulong_counter, -1UL);
-	CHECK(ul, ulong_counter, 0);
+	__this_cpu_add(*ulong_counter, -1UL);
+	CHECK(ul, *ulong_counter, 0);

 	ul += -(unsigned long)1;
-	__this_cpu_add(ulong_counter, -(unsigned long)1);
-	CHECK(ul, ulong_counter, -1);
+	__this_cpu_add(*ulong_counter, -(unsigned long)1);
+	CHECK(ul, *ulong_counter, -1);

 	ul = 0;
-	__this_cpu_write(ulong_counter, 0);
+	__this_cpu_write(*ulong_counter, 0);

 	ul -= 1;
-	__this_cpu_dec(ulong_counter);
-	CHECK(ul, ulong_counter, -1);
-	CHECK(ul, ulong_counter, ULONG_MAX);
+	__this_cpu_dec(*ulong_counter);
+	CHECK(ul, *ulong_counter, -1);
+	CHECK(ul, *ulong_counter, ULONG_MAX);

 	l += -ui_one;
-	__this_cpu_add(long_counter, -ui_one);
-	CHECK(l, long_counter, 0xffffffff);
+	__this_cpu_add(*long_counter, -ui_one);
+	CHECK(l, *long_counter, 0xffffffff);

 	l += ui_one;
-	__this_cpu_add(long_counter, ui_one);
-	CHECK(l, long_counter, (long)0x100000000LL);
+	__this_cpu_add(*long_counter, ui_one);
+	CHECK(l, *long_counter, (long)0x100000000LL);


 	l = 0;
-	__this_cpu_write(long_counter, 0);
+	__this_cpu_write(*long_counter, 0);

 	l -= ui_one;
-	__this_cpu_sub(long_counter, ui_one);
-	CHECK(l, long_counter, -1);
+	__this_cpu_sub(*long_counter, ui_one);
+	CHECK(l, *long_counter, -1);

 	l = 0;
-	__this_cpu_write(long_counter, 0);
+	__this_cpu_write(*long_counter, 0);

 	l += ui_one;
-	__this_cpu_add(long_counter, ui_one);
-	CHECK(l, long_counter, 1);
+	__this_cpu_add(*long_counter, ui_one);
+	CHECK(l, *long_counter, 1);

 	l += -ui_one;
-	__this_cpu_add(long_counter, -ui_one);
-	CHECK(l, long_counter, (long)0x100000000LL);
+	__this_cpu_add(*long_counter, -ui_one);
+	CHECK(l, *long_counter, (long)0x100000000LL);

 	l = 0;
-	__this_cpu_write(long_counter, 0);
+	__this_cpu_write(*long_counter, 0);

 	l -= ui_one;
-	this_cpu_sub(long_counter, ui_one);
-	CHECK(l, long_counter, -1);
-	CHECK(l, long_counter, ULONG_MAX);
+	this_cpu_sub(*long_counter, ui_one);
+	CHECK(l, *long_counter, -1);
+	CHECK(l, *long_counter, ULONG_MAX);

 	ul = 0;
-	__this_cpu_write(ulong_counter, 0);
+	__this_cpu_write(*ulong_counter, 0);

 	ul += ui_one;
-	__this_cpu_add(ulong_counter, ui_one);
-	CHECK(ul, ulong_counter, 1);
+	__this_cpu_add(*ulong_counter, ui_one);
+	CHECK(ul, *ulong_counter, 1);

 	ul = 0;
-	__this_cpu_write(ulong_counter, 0);
+	__this_cpu_write(*ulong_counter, 0);

 	ul -= ui_one;
-	__this_cpu_sub(ulong_counter, ui_one);
-	CHECK(ul, ulong_counter, -1);
-	CHECK(ul, ulong_counter, ULONG_MAX);
+	__this_cpu_sub(*ulong_counter, ui_one);
+	CHECK(ul, *ulong_counter, -1);
+	CHECK(ul, *ulong_counter, ULONG_MAX);

 	ul = 3;
-	__this_cpu_write(ulong_counter, 3);
+	__this_cpu_write(*ulong_counter, 3);

-	ul = this_cpu_sub_return(ulong_counter, ui_one);
-	CHECK(ul, ulong_counter, 2);
+	ul = this_cpu_sub_return(*ulong_counter, ui_one);
+	CHECK(ul, *ulong_counter, 2);

-	ul = __this_cpu_sub_return(ulong_counter, ui_one);
-	CHECK(ul, ulong_counter, 1);
+	ul = __this_cpu_sub_return(*ulong_counter, ui_one);
+	CHECK(ul, *ulong_counter, 1);

 	preempt_enable();

-	pr_info("percpu test done\n");
-	return -EAGAIN;  /* Fail will directly unload the module */
+out:
+	free_percpu(long_counter);
+	free_percpu(ulong_counter);
+	pr_debug("percpu_test: %s done cpu: %d\n", __func__, smp_processor_id());
+
+	/*
+	 * Keep the default functionality same.
+	 * Fail will directly unload this module.
+	 */
+	return -EAGAIN;
+}
+
+static void percpu_test_verify_work(struct work_struct *work)
+{
+	percpu_test_verify();
+}
+
+static int percpu_test_stress(void)
+{
+	int i;
+
+	for (i = 1; i < (1 << percpu_alloc_max_size_shift); i++) {
+		size_t size = i;
+
+		if (size > PCPU_MIN_ALLOC_SIZE)
+			break;
+		counters[i] = (char __percpu *)__alloc_percpu(size, __alignof__(char));
+		if (!counters[i])
+			break;
+		cond_resched();
+	}
+
+	schedule_on_each_cpu(percpu_test_verify_work);
+
+	for (i = 0; i < (1 << percpu_alloc_max_size_shift); i++) {
+		free_percpu(counters[i]);
+		cond_resched();
+	}
+	return -EAGAIN;
+}
+
+static int percpu_stressd(void *v)
+{
+	int iter = 0;
+
+	pr_info("DAEMON: starts %s\n", __func__);
+	do {
+		if (kthread_should_stop())
+			break;
+		iter++;
+		pr_info("TEST Starts: %s: iter (%d)\n", __func__, iter);
+		percpu_test_stress();
+		pr_info("TEST Completed: %s: iter (%d)\n", __func__, iter);
+
+		set_current_state(TASK_INTERRUPTIBLE);
+		schedule_timeout(percpu_stressd_interval);
+	} while (1);
+
+	return 0;
+}
+
+static int percpu_test_stressd(void)
+{
+	percpu_stressd_thread = kthread_run(percpu_stressd, NULL, "percpu_stressd");
+	if (IS_ERR(percpu_stressd_thread))
+		percpu_stressd_thread = NULL;
+	return 0;
+}
+
+enum test_type {
+	PERCPU_VERIFY,
+	PERCPU_STRESS,
+	PERCPU_STRESSD,
+	NR_TESTS,
+};
+
+const char *test_names[NR_TESTS] = {
+	[PERCPU_VERIFY] = "percpu_verify",
+	[PERCPU_STRESS] = "percpu_stress",
+	[PERCPU_STRESSD] = "percpu_stressd",
+};
+
+static int __init percpu_test_init(void)
+{
+	int i, ret = 0;
+	typedef int (*percpu_tests)(void);
+	const percpu_tests test_funcs[NR_TESTS] = {
+		[PERCPU_VERIFY] = percpu_test_verify,
+		[PERCPU_STRESS] = percpu_test_stress,
+		[PERCPU_STRESSD] = percpu_test_stressd,
+	};
+
+	/* sanity checks */
+	if (percpu_alloc_max_size_shift > PAGE_SHIFT)
+		percpu_alloc_max_size_shift = PAGE_SHIFT;
+	if (percpu_test_num > NR_TESTS)
+		percpu_test_num = NR_TESTS;
+
+	pr_info("percpu_test: INIT, interval: %ld, max_shift: %d, run_tests: %s\n",
+			percpu_stressd_interval, percpu_alloc_max_size_shift,
+			percpu_test_num == NR_TESTS ? "run all tests" :
+			test_names[percpu_test_num]);
+
+	/* run a given test */
+	if (percpu_test_num < NR_TESTS) {
+		pr_info("TEST Starts: %s\n", test_names[percpu_test_num]);
+		ret = test_funcs[percpu_test_num]();
+		pr_info("TEST Completed: %s\n", test_names[percpu_test_num]);
+		goto out;
+	}
+
+	for (i = 0; i < NR_TESTS; i++) {
+		pr_info("TEST Starts: %s\n", test_names[i]);
+		test_funcs[i]();
+		pr_info("TEST Completed: %s\n", test_names[i]);
+	}
+out:
+	return ret;
 }

 static void __exit percpu_test_exit(void)
 {
+	if (percpu_stressd_thread)
+		kthread_stop(percpu_stressd_thread);
+	pr_info("percpu_test: EXIT\n");
 }

 module_init(percpu_test_init)
--
2.31.1


^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL
  2021-08-23 20:42 [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Ritesh Harjani
  2021-08-23 20:42 ` [PATCHv2 2/2] lib/percpu_test: Add extra tests in percpu_test Ritesh Harjani
@ 2021-08-24  2:06 ` Dennis Zhou
  2021-08-24  4:24   ` Ritesh Harjani
  2021-08-24 11:12   ` Vlastimil Babka
  1 sibling, 2 replies; 7+ messages in thread
From: Dennis Zhou @ 2021-08-24  2:06 UTC (permalink / raw)
  To: Ritesh Harjani
  Cc: linux-mm, linux-kernel, Dennis Zhou, Tejun Heo, Christoph Lameter

Hello,

On Tue, Aug 24, 2021 at 02:12:29AM +0530, Ritesh Harjani wrote:
> Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL
> 
> Signed-off-by: Ritesh Harjani <riteshh@linux.ibm.com>
> ---
> [v1 -> v2]: Use EXPORT_SYMBOL_GPL instead of EXPORT_SYMBOL
> 
>  kernel/workqueue.c | 1 +
>  1 file changed, 1 insertion(+)
> 
> diff --git a/kernel/workqueue.c b/kernel/workqueue.c
> index f148eacda55a..993f8983186d 100644
> --- a/kernel/workqueue.c
> +++ b/kernel/workqueue.c
> @@ -3309,6 +3309,7 @@ int schedule_on_each_cpu(work_func_t func)
>  	free_percpu(works);
>  	return 0;
>  }
> +EXPORT_SYMBOL_GPL(schedule_on_each_cpu);
> 
>  /**
>   * execute_in_process_context - reliably execute the routine with user context
> --
> 2.31.1
> 

I think you missed the point of Christoph's comment. I agree with him
and don't think a test justifies exporting of this particular function.

Thanks,
Dennis

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCHv2 2/2] lib/percpu_test: Add extra tests in percpu_test
  2021-08-23 20:42 ` [PATCHv2 2/2] lib/percpu_test: Add extra tests in percpu_test Ritesh Harjani
@ 2021-08-24  2:14   ` Dennis Zhou
  2021-08-24  4:23     ` Ritesh Harjani
  0 siblings, 1 reply; 7+ messages in thread
From: Dennis Zhou @ 2021-08-24  2:14 UTC (permalink / raw)
  To: Ritesh Harjani
  Cc: linux-mm, linux-kernel, Tejun Heo, Christoph Lameter,
	Aneesh Kumar K . V, Vaibhav Jain, kernel test robot

Hello,

On Tue, Aug 24, 2021 at 02:12:30AM +0530, Ritesh Harjani wrote:
> While debugging a issue, we needed to stress test the percpu alloc/free
> path. Hence added some tests in lib/percpu_test to stress test
> percpu subsystem for allocation with different sizes.
> 

Can you explain more about the problem you are trying to debug and why
it required stressing the percpu allocator?

> This patch keeps the default behavior of insmod module same for default
> test. But when given insmod with different option, it can run a
> percpu_stressd daemon (percpu_test_num=2) which does a stress test
> evey 10secs unless the module is unloaded.
> 
> We found this to be helpful in our testing, since with this we could
> easily excercise percpu allo/free path. Hence cleaned this up for
> inclusion in percpu_test module.
> 
> Logs
> ======
> qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=0
> [  334.362973] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_verify
> [  334.364946] TEST Starts: percpu_verify
> [  334.365601] TEST Completed: percpu_verify
> insmod: ERROR: could not insert module /mnt/percpu_test.ko: Resource temporarily unavailable

^ this seems wrong? What am I missing.

> 
> qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=1
> [  336.556464] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_stress
> [  336.558388] TEST Starts: percpu_stress
> [  336.560611] TEST Completed: percpu_stress
> insmod: ERROR: could not insert module /mnt/percpu_test.ko: Resource temporarily unavailable
> 
> qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=2
> [  339.164406] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_stressd
> [  339.165935] TEST Starts: percpu_stressd
> [  339.167033] TEST Completed: percpu_stressd
> [  339.167082] DAEMON: starts percpu_stressd
> [  339.168498] TEST Starts: percpu_stressd: iter (1)
> [  339.182530] TEST Completed: percpu_stressd: iter (1)
> [  349.341109] TEST Starts: percpu_stressd: iter (2)
> [  349.344447] TEST Completed: percpu_stressd: iter (2)
> [  359.580829] TEST Starts: percpu_stressd: iter (3)
> [  359.584315] TEST Completed: percpu_stressd: iter (3)
> [  369.820471] TEST Starts: percpu_stressd: iter (4)
> [  369.844402] TEST Completed: percpu_stressd: iter (4)
> 
> qemu-> sudo rmmod percpu_test
> [  375.001098] percpu_test: EXIT
> [qemu][~]
> 
> Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
> Cc: Vaibhav Jain <vaibhav@linux.ibm.com>
> Reported-by: kernel test robot <lkp@intel.com>
> Signed-off-by: Ritesh Harjani <riteshh@linux.ibm.com>
> ---
> [v1 -> v2]: Fix warnings from kernel test robot
> 
>  lib/percpu_test.c | 240 ++++++++++++++++++++++++++++++++++++----------
>  1 file changed, 191 insertions(+), 49 deletions(-)
> 
> diff --git a/lib/percpu_test.c b/lib/percpu_test.c
> index 4a3d70bbc1a0..68c57c288dc6 100644
> --- a/lib/percpu_test.c
> +++ b/lib/percpu_test.c
> @@ -1,4 +1,7 @@
>  // SPDX-License-Identifier: GPL-2.0-only
> +#include <linux/workqueue.h>
> +#include <linux/kthread.h>
> +#include <linux/cpu.h>
>  #include <linux/module.h>
> 
>  /* validate @native and @pcp counter values match @expected */
> @@ -14,10 +17,25 @@
>  		     (long long)(expected), (long long)(expected));	\
>  	} while (0)
> 
> -static DEFINE_PER_CPU(long, long_counter);
> -static DEFINE_PER_CPU(unsigned long, ulong_counter);

^ These are good to keep fwiw as it's one of the few instances of using
the reserved region of the first chunk of percpu memory.

> +/* upto max alloc size tests for percpu var */
> +static char __percpu *counters[1 << PAGE_SHIFT];
> +static struct task_struct *percpu_stressd_thread;
> 
> -static int __init percpu_test_init(void)
> +/* let's not trigger OOM */
> +int percpu_alloc_max_size_shift = PAGE_SHIFT - 3;
> +module_param(percpu_alloc_max_size_shift, int, 0644);
> +MODULE_PARM_DESC(percpu_alloc_max_size_shift, "max size of allocation in stress test will be upto 1 << percpu_alloc_max_size_shift");
> +
> +static long percpu_stressd_interval = 1 * 10 * HZ;
> +module_param(percpu_stressd_interval, long, 0644);
> +MODULE_PARM_DESC(percpu_stressd_interval, "percpu_stressd internal");
> +
> +/* keep the default test same */
> +static int percpu_test_num;
> +module_param(percpu_test_num, int, 0644);
> +MODULE_PARM_DESC(percpu_test_num, "Test number percpu_test_num");
> +
> +static int percpu_test_verify(void)
>  {
>  	/*
>  	 * volatile prevents compiler from optimizing it uses, otherwise the
> @@ -26,109 +44,233 @@ static int __init percpu_test_init(void)
>  	volatile unsigned int ui_one = 1;
>  	long l = 0;
>  	unsigned long ul = 0;
> +	long __percpu *long_counter = alloc_percpu(long);
> +	unsigned long __percpu *ulong_counter = alloc_percpu(unsigned long);
> 
> -	pr_info("percpu test start\n");
> +	if (!long_counter || !ulong_counter)
> +		goto out;
> +
> +	pr_debug("percpu_test: %s start cpu: %d\n", __func__, smp_processor_id());
> 
>  	preempt_disable();
> 
>  	l += -1;
> -	__this_cpu_add(long_counter, -1);
> -	CHECK(l, long_counter, -1);
> +	__this_cpu_add(*long_counter, -1);
> +	CHECK(l, *long_counter, -1);
> 
>  	l += 1;
> -	__this_cpu_add(long_counter, 1);
> -	CHECK(l, long_counter, 0);
> +	__this_cpu_add(*long_counter, 1);
> +	CHECK(l, *long_counter, 0);
> 
>  	ul = 0;
> -	__this_cpu_write(ulong_counter, 0);
> +	__this_cpu_write(*ulong_counter, 0);
> 
>  	ul += 1UL;
> -	__this_cpu_add(ulong_counter, 1UL);
> -	CHECK(ul, ulong_counter, 1);
> +	__this_cpu_add(*ulong_counter, 1UL);
> +	CHECK(ul, *ulong_counter, 1);
> 
>  	ul += -1UL;
> -	__this_cpu_add(ulong_counter, -1UL);
> -	CHECK(ul, ulong_counter, 0);
> +	__this_cpu_add(*ulong_counter, -1UL);
> +	CHECK(ul, *ulong_counter, 0);
> 
>  	ul += -(unsigned long)1;
> -	__this_cpu_add(ulong_counter, -(unsigned long)1);
> -	CHECK(ul, ulong_counter, -1);
> +	__this_cpu_add(*ulong_counter, -(unsigned long)1);
> +	CHECK(ul, *ulong_counter, -1);
> 
>  	ul = 0;
> -	__this_cpu_write(ulong_counter, 0);
> +	__this_cpu_write(*ulong_counter, 0);
> 
>  	ul -= 1;
> -	__this_cpu_dec(ulong_counter);
> -	CHECK(ul, ulong_counter, -1);
> -	CHECK(ul, ulong_counter, ULONG_MAX);
> +	__this_cpu_dec(*ulong_counter);
> +	CHECK(ul, *ulong_counter, -1);
> +	CHECK(ul, *ulong_counter, ULONG_MAX);
> 
>  	l += -ui_one;
> -	__this_cpu_add(long_counter, -ui_one);
> -	CHECK(l, long_counter, 0xffffffff);
> +	__this_cpu_add(*long_counter, -ui_one);
> +	CHECK(l, *long_counter, 0xffffffff);
> 
>  	l += ui_one;
> -	__this_cpu_add(long_counter, ui_one);
> -	CHECK(l, long_counter, (long)0x100000000LL);
> +	__this_cpu_add(*long_counter, ui_one);
> +	CHECK(l, *long_counter, (long)0x100000000LL);
> 
> 
>  	l = 0;
> -	__this_cpu_write(long_counter, 0);
> +	__this_cpu_write(*long_counter, 0);
> 
>  	l -= ui_one;
> -	__this_cpu_sub(long_counter, ui_one);
> -	CHECK(l, long_counter, -1);
> +	__this_cpu_sub(*long_counter, ui_one);
> +	CHECK(l, *long_counter, -1);
> 
>  	l = 0;
> -	__this_cpu_write(long_counter, 0);
> +	__this_cpu_write(*long_counter, 0);
> 
>  	l += ui_one;
> -	__this_cpu_add(long_counter, ui_one);
> -	CHECK(l, long_counter, 1);
> +	__this_cpu_add(*long_counter, ui_one);
> +	CHECK(l, *long_counter, 1);
> 
>  	l += -ui_one;
> -	__this_cpu_add(long_counter, -ui_one);
> -	CHECK(l, long_counter, (long)0x100000000LL);
> +	__this_cpu_add(*long_counter, -ui_one);
> +	CHECK(l, *long_counter, (long)0x100000000LL);
> 
>  	l = 0;
> -	__this_cpu_write(long_counter, 0);
> +	__this_cpu_write(*long_counter, 0);
> 
>  	l -= ui_one;
> -	this_cpu_sub(long_counter, ui_one);
> -	CHECK(l, long_counter, -1);
> -	CHECK(l, long_counter, ULONG_MAX);
> +	this_cpu_sub(*long_counter, ui_one);
> +	CHECK(l, *long_counter, -1);
> +	CHECK(l, *long_counter, ULONG_MAX);
> 
>  	ul = 0;
> -	__this_cpu_write(ulong_counter, 0);
> +	__this_cpu_write(*ulong_counter, 0);
> 
>  	ul += ui_one;
> -	__this_cpu_add(ulong_counter, ui_one);
> -	CHECK(ul, ulong_counter, 1);
> +	__this_cpu_add(*ulong_counter, ui_one);
> +	CHECK(ul, *ulong_counter, 1);
> 
>  	ul = 0;
> -	__this_cpu_write(ulong_counter, 0);
> +	__this_cpu_write(*ulong_counter, 0);
> 
>  	ul -= ui_one;
> -	__this_cpu_sub(ulong_counter, ui_one);
> -	CHECK(ul, ulong_counter, -1);
> -	CHECK(ul, ulong_counter, ULONG_MAX);
> +	__this_cpu_sub(*ulong_counter, ui_one);
> +	CHECK(ul, *ulong_counter, -1);
> +	CHECK(ul, *ulong_counter, ULONG_MAX);
> 
>  	ul = 3;
> -	__this_cpu_write(ulong_counter, 3);
> +	__this_cpu_write(*ulong_counter, 3);
> 
> -	ul = this_cpu_sub_return(ulong_counter, ui_one);
> -	CHECK(ul, ulong_counter, 2);
> +	ul = this_cpu_sub_return(*ulong_counter, ui_one);
> +	CHECK(ul, *ulong_counter, 2);
> 
> -	ul = __this_cpu_sub_return(ulong_counter, ui_one);
> -	CHECK(ul, ulong_counter, 1);
> +	ul = __this_cpu_sub_return(*ulong_counter, ui_one);
> +	CHECK(ul, *ulong_counter, 1);
> 
>  	preempt_enable();
> 
> -	pr_info("percpu test done\n");
> -	return -EAGAIN;  /* Fail will directly unload the module */
> +out:
> +	free_percpu(long_counter);
> +	free_percpu(ulong_counter);
> +	pr_debug("percpu_test: %s done cpu: %d\n", __func__, smp_processor_id());
> +
> +	/*
> +	 * Keep the default functionality same.
> +	 * Fail will directly unload this module.
> +	 */
> +	return -EAGAIN;
> +}
> +
> +static void percpu_test_verify_work(struct work_struct *work)
> +{
> +	percpu_test_verify();
> +}
> +
> +static int percpu_test_stress(void)
> +{
> +	int i;
> +
> +	for (i = 1; i < (1 << percpu_alloc_max_size_shift); i++) {
> +		size_t size = i;
> +
> +		if (size > PCPU_MIN_ALLOC_SIZE)
> +			break;
> +		counters[i] = (char __percpu *)__alloc_percpu(size, __alignof__(char));
> +		if (!counters[i])
> +			break;
> +		cond_resched();
> +	}
> +
> +	schedule_on_each_cpu(percpu_test_verify_work);

I'm not understanding the value of scheduling on each cpu here? Each
CHECK() call is accessing the value via __this_cpu_read() meaning it has
nothing to do with any other cpu's copy of the variable?

> +
> +	for (i = 0; i < (1 << percpu_alloc_max_size_shift); i++) {
> +		free_percpu(counters[i]);
> +		cond_resched();
> +	}
> +	return -EAGAIN;
> +}
> +
> +static int percpu_stressd(void *v)
> +{
> +	int iter = 0;
> +
> +	pr_info("DAEMON: starts %s\n", __func__);
> +	do {
> +		if (kthread_should_stop())
> +			break;
> +		iter++;
> +		pr_info("TEST Starts: %s: iter (%d)\n", __func__, iter);
> +		percpu_test_stress();
> +		pr_info("TEST Completed: %s: iter (%d)\n", __func__, iter);
> +
> +		set_current_state(TASK_INTERRUPTIBLE);
> +		schedule_timeout(percpu_stressd_interval);
> +	} while (1);
> +
> +	return 0;
> +}
> +
> +static int percpu_test_stressd(void)
> +{
> +	percpu_stressd_thread = kthread_run(percpu_stressd, NULL, "percpu_stressd");
> +	if (IS_ERR(percpu_stressd_thread))
> +		percpu_stressd_thread = NULL;
> +	return 0;
> +}
> +
> +enum test_type {
> +	PERCPU_VERIFY,
> +	PERCPU_STRESS,
> +	PERCPU_STRESSD,
> +	NR_TESTS,
> +};
> +
> +const char *test_names[NR_TESTS] = {
> +	[PERCPU_VERIFY] = "percpu_verify",
> +	[PERCPU_STRESS] = "percpu_stress",
> +	[PERCPU_STRESSD] = "percpu_stressd",
> +};
> +
> +static int __init percpu_test_init(void)
> +{
> +	int i, ret = 0;
> +	typedef int (*percpu_tests)(void);
> +	const percpu_tests test_funcs[NR_TESTS] = {
> +		[PERCPU_VERIFY] = percpu_test_verify,
> +		[PERCPU_STRESS] = percpu_test_stress,
> +		[PERCPU_STRESSD] = percpu_test_stressd,
> +	};
> +
> +	/* sanity checks */
> +	if (percpu_alloc_max_size_shift > PAGE_SHIFT)
> +		percpu_alloc_max_size_shift = PAGE_SHIFT;
> +	if (percpu_test_num > NR_TESTS)
> +		percpu_test_num = NR_TESTS;
> +
> +	pr_info("percpu_test: INIT, interval: %ld, max_shift: %d, run_tests: %s\n",
> +			percpu_stressd_interval, percpu_alloc_max_size_shift,
> +			percpu_test_num == NR_TESTS ? "run all tests" :
> +			test_names[percpu_test_num]);
> +
> +	/* run a given test */
> +	if (percpu_test_num < NR_TESTS) {
> +		pr_info("TEST Starts: %s\n", test_names[percpu_test_num]);
> +		ret = test_funcs[percpu_test_num]();
> +		pr_info("TEST Completed: %s\n", test_names[percpu_test_num]);
> +		goto out;
> +	}
> +
> +	for (i = 0; i < NR_TESTS; i++) {
> +		pr_info("TEST Starts: %s\n", test_names[i]);
> +		test_funcs[i]();
> +		pr_info("TEST Completed: %s\n", test_names[i]);
> +	}
> +out:
> +	return ret;
>  }
> 
>  static void __exit percpu_test_exit(void)
>  {
> +	if (percpu_stressd_thread)
> +		kthread_stop(percpu_stressd_thread);
> +	pr_info("percpu_test: EXIT\n");
>  }
> 
>  module_init(percpu_test_init)
> --
> 2.31.1
> 

Thanks,
Dennis

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCHv2 2/2] lib/percpu_test: Add extra tests in percpu_test
  2021-08-24  2:14   ` Dennis Zhou
@ 2021-08-24  4:23     ` Ritesh Harjani
  0 siblings, 0 replies; 7+ messages in thread
From: Ritesh Harjani @ 2021-08-24  4:23 UTC (permalink / raw)
  To: Dennis Zhou
  Cc: linux-mm, linux-kernel, Tejun Heo, Christoph Lameter,
	Aneesh Kumar K . V, Vaibhav Jain, kernel test robot

On 21/08/23 10:14PM, Dennis Zhou wrote:
> Hello,
>
> On Tue, Aug 24, 2021 at 02:12:30AM +0530, Ritesh Harjani wrote:
> > While debugging a issue, we needed to stress test the percpu alloc/free
> > path. Hence added some tests in lib/percpu_test to stress test
> > percpu subsystem for allocation with different sizes.
> >
>
> Can you explain more about the problem you are trying to debug and why
> it required stressing the percpu allocator?

We saw a BUG_ON() in older implementation of percpu code from pcpu_free_area()
function, since it wasn't able to find an entry in the pcpu_chunk->map[].
The problem eventually ended up in a different subsystem which lead to
pcpu_chunk->map[] getting corrupted and hence hit that BUG_ON().
But at the time, we wanted to test percpu_alloc/free with all possible sizes,
before ruling out anything.

Since the problem was also seen in >4+ hrs (sometimes even more), we wrote this
test to stress more on percpu alloc and free path, to see the problem early
along with some other debugging instrumentation.

>
> > This patch keeps the default behavior of insmod module same for default
> > test. But when given insmod with different option, it can run a
> > percpu_stressd daemon (percpu_test_num=2) which does a stress test
> > evey 10secs unless the module is unloaded.
> >
> > We found this to be helpful in our testing, since with this we could
> > easily excercise percpu allo/free path. Hence cleaned this up for
> > inclusion in percpu_test module.
> >
> > Logs
> > ======
> > qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=0
> > [  334.362973] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_verify
> > [  334.364946] TEST Starts: percpu_verify
> > [  334.365601] TEST Completed: percpu_verify
> > insmod: ERROR: could not insert module /mnt/percpu_test.ko: Resource temporarily unavailable
>
> ^ this seems wrong? What am I missing.

Original test itself in this module is written in the way that can unload
itself upon completion. So for percpu_test_num = 0 & 1, it does automatic
unload, whereas for percpu_test_num=2, tester need to unload the module.
Unload will happen if the individual test returns an error value which init
function returns too and the kernel automatically unloads this test module.

>
>
> >
> > qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=1
> > [  336.556464] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_stress
> > [  336.558388] TEST Starts: percpu_stress
> > [  336.560611] TEST Completed: percpu_stress
> > insmod: ERROR: could not insert module /mnt/percpu_test.ko: Resource temporarily unavailable
> >
> > qemu-> sudo insmod /mnt/percpu_test.ko percpu_test_num=2
> > [  339.164406] percpu_test: INIT, interval: 1000, max_shift: 13, run_tests: percpu_stressd
> > [  339.165935] TEST Starts: percpu_stressd
> > [  339.167033] TEST Completed: percpu_stressd
> > [  339.167082] DAEMON: starts percpu_stressd
> > [  339.168498] TEST Starts: percpu_stressd: iter (1)
> > [  339.182530] TEST Completed: percpu_stressd: iter (1)
> > [  349.341109] TEST Starts: percpu_stressd: iter (2)
> > [  349.344447] TEST Completed: percpu_stressd: iter (2)
> > [  359.580829] TEST Starts: percpu_stressd: iter (3)
> > [  359.584315] TEST Completed: percpu_stressd: iter (3)
> > [  369.820471] TEST Starts: percpu_stressd: iter (4)
> > [  369.844402] TEST Completed: percpu_stressd: iter (4)
> >
> > qemu-> sudo rmmod percpu_test
> > [  375.001098] percpu_test: EXIT
> > [qemu][~]
> >
> > Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
> > Cc: Vaibhav Jain <vaibhav@linux.ibm.com>
> > Reported-by: kernel test robot <lkp@intel.com>
> > Signed-off-by: Ritesh Harjani <riteshh@linux.ibm.com>
> > ---
> > [v1 -> v2]: Fix warnings from kernel test robot
> >
> >  lib/percpu_test.c | 240 ++++++++++++++++++++++++++++++++++++----------
> >  1 file changed, 191 insertions(+), 49 deletions(-)
> >
> > diff --git a/lib/percpu_test.c b/lib/percpu_test.c
> > index 4a3d70bbc1a0..68c57c288dc6 100644
> > --- a/lib/percpu_test.c
> > +++ b/lib/percpu_test.c
> > @@ -1,4 +1,7 @@
> >  // SPDX-License-Identifier: GPL-2.0-only
> > +#include <linux/workqueue.h>
> > +#include <linux/kthread.h>
> > +#include <linux/cpu.h>
> >  #include <linux/module.h>
> >
> >  /* validate @native and @pcp counter values match @expected */
> > @@ -14,10 +17,25 @@
> >  		     (long long)(expected), (long long)(expected));	\
> >  	} while (0)
> >
> > -static DEFINE_PER_CPU(long, long_counter);
> > -static DEFINE_PER_CPU(unsigned long, ulong_counter);
>
> ^ These are good to keep fwiw as it's one of the few instances of using
> the reserved region of the first chunk of percpu memory.

Yes, sure. I can make changes to retain this.


>
> > +/* upto max alloc size tests for percpu var */
> > +static char __percpu *counters[1 << PAGE_SHIFT];
> > +static struct task_struct *percpu_stressd_thread;
> >
> > -static int __init percpu_test_init(void)
> > +/* let's not trigger OOM */
> > +int percpu_alloc_max_size_shift = PAGE_SHIFT - 3;
> > +module_param(percpu_alloc_max_size_shift, int, 0644);
> > +MODULE_PARM_DESC(percpu_alloc_max_size_shift, "max size of allocation in stress test will be upto 1 << percpu_alloc_max_size_shift");
> > +
> > +static long percpu_stressd_interval = 1 * 10 * HZ;
> > +module_param(percpu_stressd_interval, long, 0644);
> > +MODULE_PARM_DESC(percpu_stressd_interval, "percpu_stressd internal");
> > +
> > +/* keep the default test same */
> > +static int percpu_test_num;
> > +module_param(percpu_test_num, int, 0644);
> > +MODULE_PARM_DESC(percpu_test_num, "Test number percpu_test_num");
> > +
> > +static int percpu_test_verify(void)
> >  {
> >  	/*
> >  	 * volatile prevents compiler from optimizing it uses, otherwise the
> > @@ -26,109 +44,233 @@ static int __init percpu_test_init(void)
> >  	volatile unsigned int ui_one = 1;
> >  	long l = 0;
> >  	unsigned long ul = 0;
> > +	long __percpu *long_counter = alloc_percpu(long);
> > +	unsigned long __percpu *ulong_counter = alloc_percpu(unsigned long);
> >
> > -	pr_info("percpu test start\n");
> > +	if (!long_counter || !ulong_counter)
> > +		goto out;
> > +
> > +	pr_debug("percpu_test: %s start cpu: %d\n", __func__, smp_processor_id());
> >
> >  	preempt_disable();
> >
> >  	l += -1;
> > -	__this_cpu_add(long_counter, -1);
> > -	CHECK(l, long_counter, -1);
> > +	__this_cpu_add(*long_counter, -1);
> > +	CHECK(l, *long_counter, -1);
> >
> >  	l += 1;
> > -	__this_cpu_add(long_counter, 1);
> > -	CHECK(l, long_counter, 0);
> > +	__this_cpu_add(*long_counter, 1);
> > +	CHECK(l, *long_counter, 0);
> >
> >  	ul = 0;
> > -	__this_cpu_write(ulong_counter, 0);
> > +	__this_cpu_write(*ulong_counter, 0);
> >
> >  	ul += 1UL;
> > -	__this_cpu_add(ulong_counter, 1UL);
> > -	CHECK(ul, ulong_counter, 1);
> > +	__this_cpu_add(*ulong_counter, 1UL);
> > +	CHECK(ul, *ulong_counter, 1);
> >
> >  	ul += -1UL;
> > -	__this_cpu_add(ulong_counter, -1UL);
> > -	CHECK(ul, ulong_counter, 0);
> > +	__this_cpu_add(*ulong_counter, -1UL);
> > +	CHECK(ul, *ulong_counter, 0);
> >
> >  	ul += -(unsigned long)1;
> > -	__this_cpu_add(ulong_counter, -(unsigned long)1);
> > -	CHECK(ul, ulong_counter, -1);
> > +	__this_cpu_add(*ulong_counter, -(unsigned long)1);
> > +	CHECK(ul, *ulong_counter, -1);
> >
> >  	ul = 0;
> > -	__this_cpu_write(ulong_counter, 0);
> > +	__this_cpu_write(*ulong_counter, 0);
> >
> >  	ul -= 1;
> > -	__this_cpu_dec(ulong_counter);
> > -	CHECK(ul, ulong_counter, -1);
> > -	CHECK(ul, ulong_counter, ULONG_MAX);
> > +	__this_cpu_dec(*ulong_counter);
> > +	CHECK(ul, *ulong_counter, -1);
> > +	CHECK(ul, *ulong_counter, ULONG_MAX);
> >
> >  	l += -ui_one;
> > -	__this_cpu_add(long_counter, -ui_one);
> > -	CHECK(l, long_counter, 0xffffffff);
> > +	__this_cpu_add(*long_counter, -ui_one);
> > +	CHECK(l, *long_counter, 0xffffffff);
> >
> >  	l += ui_one;
> > -	__this_cpu_add(long_counter, ui_one);
> > -	CHECK(l, long_counter, (long)0x100000000LL);
> > +	__this_cpu_add(*long_counter, ui_one);
> > +	CHECK(l, *long_counter, (long)0x100000000LL);
> >
> >
> >  	l = 0;
> > -	__this_cpu_write(long_counter, 0);
> > +	__this_cpu_write(*long_counter, 0);
> >
> >  	l -= ui_one;
> > -	__this_cpu_sub(long_counter, ui_one);
> > -	CHECK(l, long_counter, -1);
> > +	__this_cpu_sub(*long_counter, ui_one);
> > +	CHECK(l, *long_counter, -1);
> >
> >  	l = 0;
> > -	__this_cpu_write(long_counter, 0);
> > +	__this_cpu_write(*long_counter, 0);
> >
> >  	l += ui_one;
> > -	__this_cpu_add(long_counter, ui_one);
> > -	CHECK(l, long_counter, 1);
> > +	__this_cpu_add(*long_counter, ui_one);
> > +	CHECK(l, *long_counter, 1);
> >
> >  	l += -ui_one;
> > -	__this_cpu_add(long_counter, -ui_one);
> > -	CHECK(l, long_counter, (long)0x100000000LL);
> > +	__this_cpu_add(*long_counter, -ui_one);
> > +	CHECK(l, *long_counter, (long)0x100000000LL);
> >
> >  	l = 0;
> > -	__this_cpu_write(long_counter, 0);
> > +	__this_cpu_write(*long_counter, 0);
> >
> >  	l -= ui_one;
> > -	this_cpu_sub(long_counter, ui_one);
> > -	CHECK(l, long_counter, -1);
> > -	CHECK(l, long_counter, ULONG_MAX);
> > +	this_cpu_sub(*long_counter, ui_one);
> > +	CHECK(l, *long_counter, -1);
> > +	CHECK(l, *long_counter, ULONG_MAX);
> >
> >  	ul = 0;
> > -	__this_cpu_write(ulong_counter, 0);
> > +	__this_cpu_write(*ulong_counter, 0);
> >
> >  	ul += ui_one;
> > -	__this_cpu_add(ulong_counter, ui_one);
> > -	CHECK(ul, ulong_counter, 1);
> > +	__this_cpu_add(*ulong_counter, ui_one);
> > +	CHECK(ul, *ulong_counter, 1);
> >
> >  	ul = 0;
> > -	__this_cpu_write(ulong_counter, 0);
> > +	__this_cpu_write(*ulong_counter, 0);
> >
> >  	ul -= ui_one;
> > -	__this_cpu_sub(ulong_counter, ui_one);
> > -	CHECK(ul, ulong_counter, -1);
> > -	CHECK(ul, ulong_counter, ULONG_MAX);
> > +	__this_cpu_sub(*ulong_counter, ui_one);
> > +	CHECK(ul, *ulong_counter, -1);
> > +	CHECK(ul, *ulong_counter, ULONG_MAX);
> >
> >  	ul = 3;
> > -	__this_cpu_write(ulong_counter, 3);
> > +	__this_cpu_write(*ulong_counter, 3);
> >
> > -	ul = this_cpu_sub_return(ulong_counter, ui_one);
> > -	CHECK(ul, ulong_counter, 2);
> > +	ul = this_cpu_sub_return(*ulong_counter, ui_one);
> > +	CHECK(ul, *ulong_counter, 2);
> >
> > -	ul = __this_cpu_sub_return(ulong_counter, ui_one);
> > -	CHECK(ul, ulong_counter, 1);
> > +	ul = __this_cpu_sub_return(*ulong_counter, ui_one);
> > +	CHECK(ul, *ulong_counter, 1);
> >
> >  	preempt_enable();
> >
> > -	pr_info("percpu test done\n");
> > -	return -EAGAIN;  /* Fail will directly unload the module */
> > +out:
> > +	free_percpu(long_counter);
> > +	free_percpu(ulong_counter);
> > +	pr_debug("percpu_test: %s done cpu: %d\n", __func__, smp_processor_id());
> > +
> > +	/*
> > +	 * Keep the default functionality same.
> > +	 * Fail will directly unload this module.
> > +	 */
> > +	return -EAGAIN;
> > +}
> > +
> > +static void percpu_test_verify_work(struct work_struct *work)
> > +{
> > +	percpu_test_verify();
> > +}
> > +
> > +static int percpu_test_stress(void)
> > +{
> > +	int i;
> > +
> > +	for (i = 1; i < (1 << percpu_alloc_max_size_shift); i++) {
> > +		size_t size = i;
> > +
> > +		if (size > PCPU_MIN_ALLOC_SIZE)
> > +			break;
> > +		counters[i] = (char __percpu *)__alloc_percpu(size, __alignof__(char));
> > +		if (!counters[i])
> > +			break;
> > +		cond_resched();
> > +	}
> > +
> > +	schedule_on_each_cpu(percpu_test_verify_work);
>
> I'm not understanding the value of scheduling on each cpu here? Each
> CHECK() call is accessing the value via __this_cpu_read() meaning it has
> nothing to do with any other cpu's copy of the variable?

Yes, that is true. This I wrote, since we just wanted some work to happen on
each cpu before freeing all the percpu vars, hence scheduled
percpu_test_verify function to run and verify the original test on each cpu.

Dennis,
Let me know in case if you want me to make any additions to this stress test.
I can make such changes too which we think will be more useful.

This stress test was useful at the time of initial debugging for us.
But if you find something to be added which can add more value to this, I can
certainly do that.

Thanks
-ritesh

>
> > +
> > +	for (i = 0; i < (1 << percpu_alloc_max_size_shift); i++) {
> > +		free_percpu(counters[i]);
> > +		cond_resched();
> > +	}
> > +	return -EAGAIN;
> > +}
> > +
> > +static int percpu_stressd(void *v)
> > +{
> > +	int iter = 0;
> > +
> > +	pr_info("DAEMON: starts %s\n", __func__);
> > +	do {
> > +		if (kthread_should_stop())
> > +			break;
> > +		iter++;
> > +		pr_info("TEST Starts: %s: iter (%d)\n", __func__, iter);
> > +		percpu_test_stress();
> > +		pr_info("TEST Completed: %s: iter (%d)\n", __func__, iter);
> > +
> > +		set_current_state(TASK_INTERRUPTIBLE);
> > +		schedule_timeout(percpu_stressd_interval);
> > +	} while (1);
> > +
> > +	return 0;
> > +}
> > +
> > +static int percpu_test_stressd(void)
> > +{
> > +	percpu_stressd_thread = kthread_run(percpu_stressd, NULL, "percpu_stressd");
> > +	if (IS_ERR(percpu_stressd_thread))
> > +		percpu_stressd_thread = NULL;
> > +	return 0;
> > +}
> > +
> > +enum test_type {
> > +	PERCPU_VERIFY,
> > +	PERCPU_STRESS,
> > +	PERCPU_STRESSD,
> > +	NR_TESTS,
> > +};
> > +
> > +const char *test_names[NR_TESTS] = {
> > +	[PERCPU_VERIFY] = "percpu_verify",
> > +	[PERCPU_STRESS] = "percpu_stress",
> > +	[PERCPU_STRESSD] = "percpu_stressd",
> > +};
> > +
> > +static int __init percpu_test_init(void)
> > +{
> > +	int i, ret = 0;
> > +	typedef int (*percpu_tests)(void);
> > +	const percpu_tests test_funcs[NR_TESTS] = {
> > +		[PERCPU_VERIFY] = percpu_test_verify,
> > +		[PERCPU_STRESS] = percpu_test_stress,
> > +		[PERCPU_STRESSD] = percpu_test_stressd,
> > +	};
> > +
> > +	/* sanity checks */
> > +	if (percpu_alloc_max_size_shift > PAGE_SHIFT)
> > +		percpu_alloc_max_size_shift = PAGE_SHIFT;
> > +	if (percpu_test_num > NR_TESTS)
> > +		percpu_test_num = NR_TESTS;
> > +
> > +	pr_info("percpu_test: INIT, interval: %ld, max_shift: %d, run_tests: %s\n",
> > +			percpu_stressd_interval, percpu_alloc_max_size_shift,
> > +			percpu_test_num == NR_TESTS ? "run all tests" :
> > +			test_names[percpu_test_num]);
> > +
> > +	/* run a given test */
> > +	if (percpu_test_num < NR_TESTS) {
> > +		pr_info("TEST Starts: %s\n", test_names[percpu_test_num]);
> > +		ret = test_funcs[percpu_test_num]();
> > +		pr_info("TEST Completed: %s\n", test_names[percpu_test_num]);
> > +		goto out;
> > +	}
> > +
> > +	for (i = 0; i < NR_TESTS; i++) {
> > +		pr_info("TEST Starts: %s\n", test_names[i]);
> > +		test_funcs[i]();
> > +		pr_info("TEST Completed: %s\n", test_names[i]);
> > +	}
> > +out:
> > +	return ret;
> >  }
> >
> >  static void __exit percpu_test_exit(void)
> >  {
> > +	if (percpu_stressd_thread)
> > +		kthread_stop(percpu_stressd_thread);
> > +	pr_info("percpu_test: EXIT\n");
> >  }
> >
> >  module_init(percpu_test_init)
> > --
> > 2.31.1
> >
>
> Thanks,
> Dennis

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL
  2021-08-24  2:06 ` [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Dennis Zhou
@ 2021-08-24  4:24   ` Ritesh Harjani
  2021-08-24 11:12   ` Vlastimil Babka
  1 sibling, 0 replies; 7+ messages in thread
From: Ritesh Harjani @ 2021-08-24  4:24 UTC (permalink / raw)
  To: Dennis Zhou; +Cc: linux-mm, linux-kernel, Tejun Heo, Christoph Lameter

On 21/08/23 10:06PM, Dennis Zhou wrote:
> Hello,
>
> On Tue, Aug 24, 2021 at 02:12:29AM +0530, Ritesh Harjani wrote:
> > Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL
> >
> > Signed-off-by: Ritesh Harjani <riteshh@linux.ibm.com>
> > ---
> > [v1 -> v2]: Use EXPORT_SYMBOL_GPL instead of EXPORT_SYMBOL
> >
> >  kernel/workqueue.c | 1 +
> >  1 file changed, 1 insertion(+)
> >
> > diff --git a/kernel/workqueue.c b/kernel/workqueue.c
> > index f148eacda55a..993f8983186d 100644
> > --- a/kernel/workqueue.c
> > +++ b/kernel/workqueue.c
> > @@ -3309,6 +3309,7 @@ int schedule_on_each_cpu(work_func_t func)
> >  	free_percpu(works);
> >  	return 0;
> >  }
> > +EXPORT_SYMBOL_GPL(schedule_on_each_cpu);
> >
> >  /**
> >   * execute_in_process_context - reliably execute the routine with user context
> > --
> > 2.31.1
> >
>
> I think you missed the point of Christoph's comment. I agree with him
> and don't think a test justifies exporting of this particular function.

Got it. Based on the comments from 2nd patch, if in case this will be required
I can open code this. (I initially had this open coded, but then I found this).

-ritesh

>
> Thanks,
> Dennis
>

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL
  2021-08-24  2:06 ` [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Dennis Zhou
  2021-08-24  4:24   ` Ritesh Harjani
@ 2021-08-24 11:12   ` Vlastimil Babka
  1 sibling, 0 replies; 7+ messages in thread
From: Vlastimil Babka @ 2021-08-24 11:12 UTC (permalink / raw)
  To: Dennis Zhou, Ritesh Harjani
  Cc: linux-mm, linux-kernel, Tejun Heo, Christoph Lameter, Christoph Hellwig

On 8/24/21 04:06, Dennis Zhou wrote:
> Hello,
> 
> On Tue, Aug 24, 2021 at 02:12:29AM +0530, Ritesh Harjani wrote:
>> Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL
>> 
>> Signed-off-by: Ritesh Harjani <riteshh@linux.ibm.com>
>> ---
>> [v1 -> v2]: Use EXPORT_SYMBOL_GPL instead of EXPORT_SYMBOL
>> 
>>  kernel/workqueue.c | 1 +
>>  1 file changed, 1 insertion(+)
>> 
>> diff --git a/kernel/workqueue.c b/kernel/workqueue.c
>> index f148eacda55a..993f8983186d 100644
>> --- a/kernel/workqueue.c
>> +++ b/kernel/workqueue.c
>> @@ -3309,6 +3309,7 @@ int schedule_on_each_cpu(work_func_t func)
>>  	free_percpu(works);
>>  	return 0;
>>  }
>> +EXPORT_SYMBOL_GPL(schedule_on_each_cpu);
>> 
>>  /**
>>   * execute_in_process_context - reliably execute the routine with user context
>> --
>> 2.31.1
>> 
> 
> I think you missed the point of Christoph's comment. I agree with him
> and don't think a test justifies exporting of this particular function.

Could symbol namespaces help in these cases? It feels wrong to limit in-tree
test modules this way.
https://www.kernel.org/doc/html/latest/core-api/symbol-namespaces.html

> Thanks,
> Dennis
> 


^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2021-08-24 11:13 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-08-23 20:42 [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Ritesh Harjani
2021-08-23 20:42 ` [PATCHv2 2/2] lib/percpu_test: Add extra tests in percpu_test Ritesh Harjani
2021-08-24  2:14   ` Dennis Zhou
2021-08-24  4:23     ` Ritesh Harjani
2021-08-24  2:06 ` [PATCHv2 1/2] kernel/workqueue: Make schedule_on_each_cpu as EXPORT_SYMBOL_GPL Dennis Zhou
2021-08-24  4:24   ` Ritesh Harjani
2021-08-24 11:12   ` Vlastimil Babka

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).