LKML Archive on lore.kernel.org
help / color / mirror / Atom feed
* [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation
@ 2008-02-14 23:02 Martin Devera
  2008-02-18  7:28 ` David Miller
  0 siblings, 1 reply; 6+ messages in thread
From: Martin Devera @ 2008-02-14 23:02 UTC (permalink / raw)
  To: David Miller; +Cc: linux-kernel, kaber, netdev

From: Martin Devera <devik@cdi.cz>

HTB is event driven algorithm and part of its work is to apply
scheduled events at proper times. It tried to defend itself from
livelock by processing only limited number of events per dequeue.
Because of faster computers some users already hit this hardcoded
limit.
This patch uses loops_per_jiffy variable to limit event processing
up to single jiffy interval and then delay remainder to other
jiffy.

Signed-off-by: Martin Devera <devik@cdi.cz>
---
BTW, from my measurement is seems that value 500 was good one
for my first 600MHz machine :-)
Maybe I can make something self-converging (using tasklets probably)
but I'm not sure if it is worth of the complexity.

--- a/net/sched/sch_htb.c	2008-02-14 22:56:48.000000000 +0100
+++ b/net/sched/sch_htb.c	2008-02-14 23:37:02.000000000 +0100
@@ -704,13 +704,17 @@ static void htb_charge_class(struct htb_
  *
  * Scans event queue for pending events and applies them. Returns time of
  * next pending event (0 for no event in pq).
+ * One event costs about 1300 cycles on x86_64, let's be conservative
+ * and round it to 4096. We will allow only loops_per_jiffy/4096 events
+ * in one call to prevent us from livelock.
  * Note: Applied are events whose have cl->pq_key <= q->now.
  */
+#define HTB_EVENT_COST_SHIFTS 12
 static psched_time_t htb_do_events(struct htb_sched *q, int level)
 {
-	int i;
-
-	for (i = 0; i < 500; i++) {
+	int i, max_events = loops_per_jiffy >> HTB_EVENT_COST_SHIFTS;
+	/* <= below is just for case where max_events==0 (unlikely) */
+	for (i = 0; i <= max_events; i++) {
 		struct htb_class *cl;
 		long diff;
 		struct rb_node *p = rb_first(&q->wait_pq[level]);
@@ -728,9 +732,8 @@ static psched_time_t htb_do_events(struc
 		if (cl->cmode != HTB_CAN_SEND)
 			htb_add_to_wait_tree(q, cl, diff);
 	}
-	if (net_ratelimit())
-		printk(KERN_WARNING "htb: too many events !\n");
-	return q->now + PSCHED_TICKS_PER_SEC / 10;
+	/* too much load - let's continue on next tick */
+	return q->now + PSCHED_TICKS_PER_SEC / HZ;
 }
 
 /* Returns class->node+prio from id-tree where classe's id is >= id. NULL

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation
  2008-02-14 23:02 [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation Martin Devera
@ 2008-02-18  7:28 ` David Miller
  2008-02-18  8:03   ` Martin Devera
  0 siblings, 1 reply; 6+ messages in thread
From: David Miller @ 2008-02-18  7:28 UTC (permalink / raw)
  Cc: linux-kernel, kaber, netdev

From: Martin Devera <devik@cdi.cz>
Date: Fri, 15 Feb 2008 00:02:56 +0100

> From: Martin Devera <devik@cdi.cz>
> 
> HTB is event driven algorithm and part of its work is to apply
> scheduled events at proper times. It tried to defend itself from
> livelock by processing only limited number of events per dequeue.
> Because of faster computers some users already hit this hardcoded
> limit.
> This patch uses loops_per_jiffy variable to limit event processing
> up to single jiffy interval and then delay remainder to other
> jiffy.
> 
> Signed-off-by: Martin Devera <devik@cdi.cz>

I think we would be wise to use something other than loops_per_jiffy.

Depending upon the loop calibration method used by a particular
architecture it can me one of many different things.

Some platforms don't even make use of it and thus leave it at it's
default value of "1<<12", so using it as a heuristic here is arbitrary
at best.

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation
  2008-02-18  7:28 ` David Miller
@ 2008-02-18  8:03   ` Martin Devera
  2008-02-18  8:17     ` David Miller
  0 siblings, 1 reply; 6+ messages in thread
From: Martin Devera @ 2008-02-18  8:03 UTC (permalink / raw)
  To: David Miller; +Cc: linux-kernel, kaber, netdev

>> up to single jiffy interval and then delay remainder to other
>> jiffy.
>>
>> Signed-off-by: Martin Devera <devik@cdi.cz>
> 
> I think we would be wise to use something other than loops_per_jiffy.
> 
> Depending upon the loop calibration method used by a particular
> architecture it can me one of many different things.
> 
> Some platforms don't even make use of it and thus leave it at it's

aha, ok, I'm not so informed about crossplatform issues.
I was also thining about looking at jiffies value and stop once
it is startjiffy+2, but with NO_HZ introduction, are jiffies
still incremented ?

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation
  2008-02-18  8:03   ` Martin Devera
@ 2008-02-18  8:17     ` David Miller
  2008-02-18 10:08       ` Martin Devera
  0 siblings, 1 reply; 6+ messages in thread
From: David Miller @ 2008-02-18  8:17 UTC (permalink / raw)
  To: devik; +Cc: linux-kernel, kaber, netdev

From: Martin Devera <devik@cdi.cz>
Date: Mon, 18 Feb 2008 09:03:52 +0100

> aha, ok, I'm not so informed about crossplatform issues.
> I was also thining about looking at jiffies value and stop once
> it is startjiffy+2, but with NO_HZ introduction, are jiffies
> still incremented ?

There should always be at least once cpu tasked with incrementing
jiffies.  Lots of stuff would break if not :-)


^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation
  2008-02-18  8:17     ` David Miller
@ 2008-02-18 10:08       ` Martin Devera
  2008-02-18 10:43         ` David Miller
  0 siblings, 1 reply; 6+ messages in thread
From: Martin Devera @ 2008-02-18 10:08 UTC (permalink / raw)
  To: David Miller; +Cc: linux-kernel, kaber, netdev

David Miller wrote:
> From: Martin Devera <devik@cdi.cz>
> Date: Mon, 18 Feb 2008 09:03:52 +0100
> 
>> aha, ok, I'm not so informed about crossplatform issues.
>> I was also thining about looking at jiffies value and stop once
>> it is startjiffy+2, but with NO_HZ introduction, are jiffies
>> still incremented ?
> 
> There should always be at least once cpu tasked with incrementing
> jiffies.  Lots of stuff would break if not :-)
> 

Aha ok, so that when (at least one) cpu is busy then I can count on
jiffies incrementing via do_timer, can't I ?
So that I'd remove the loop limit altogether but limiting it to
1 or 2 jiffies to prevent livelock.
Like
max_jiff = jiffies+2; /* not +1 at we could be at +0.9999 now */
while (jiffies<max_jiff) do_hard_potentionaly_long_work();
if (more_work) schedule_to_next_jiffie();

This will keep event queue work load under 66% of system load which
seems reasonable to me.

Would you accept such solution ?

^ permalink raw reply	[flat|nested] 6+ messages in thread

* Re: [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation
  2008-02-18 10:08       ` Martin Devera
@ 2008-02-18 10:43         ` David Miller
  0 siblings, 0 replies; 6+ messages in thread
From: David Miller @ 2008-02-18 10:43 UTC (permalink / raw)
  To: devik; +Cc: linux-kernel, kaber, netdev

From: Martin Devera <devik@cdi.cz>
Date: Mon, 18 Feb 2008 11:08:09 +0100

> Like
> max_jiff = jiffies+2; /* not +1 at we could be at +0.9999 now */
> while (jiffies<max_jiff) do_hard_potentionaly_long_work();
> if (more_work) schedule_to_next_jiffie();
> 
> This will keep event queue work load under 66% of system load which
> seems reasonable to me.
> 
> Would you accept such solution ?

Sure.

^ permalink raw reply	[flat|nested] 6+ messages in thread

end of thread, other threads:[~2008-02-18 10:43 UTC | newest]

Thread overview: 6+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2008-02-14 23:02 [PATCH 2.6.24 1/1] sch_htb: fix "too many events" situation Martin Devera
2008-02-18  7:28 ` David Miller
2008-02-18  8:03   ` Martin Devera
2008-02-18  8:17     ` David Miller
2008-02-18 10:08       ` Martin Devera
2008-02-18 10:43         ` David Miller

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).