From: "Ahmed S. Darwish" <a.darwish@linutronix.de> To: Peter Zijlstra <peterz@infradead.org>, Ingo Molnar <mingo@redhat.com>, Will Deacon <will@kernel.org> Cc: Thomas Gleixner <tglx@linutronix.de>, "Paul E. McKenney" <paulmck@kernel.org>, "Sebastian A. Siewior" <bigeasy@linutronix.de>, Steven Rostedt <rostedt@goodmis.org>, LKML <linux-kernel@vger.kernel.org>, "Ahmed S. Darwish" <a.darwish@linutronix.de>, Jonathan Corbet <corbet@lwn.net>, linux-doc@vger.kernel.org Subject: [PATCH v1 11/25] seqlock: Add missing kernel-doc annotations Date: Tue, 19 May 2020 23:45:33 +0200 Message-ID: <20200519214547.352050-12-a.darwish@linutronix.de> (raw) In-Reply-To: <20200519214547.352050-1-a.darwish@linutronix.de> A small number of the the exported seqlock.h functions are kernel-doc annotated. Since seqlock.h is now included by the kernel's RST documentation, add kernel-doc annotations for all of the remaining functions. Signed-off-by: Ahmed S. Darwish <a.darwish@linutronix.de> --- include/linux/seqlock.h | 414 +++++++++++++++++++++++++++++++++++----- 1 file changed, 361 insertions(+), 53 deletions(-) diff --git a/include/linux/seqlock.h b/include/linux/seqlock.h index dfec0c9c19c4..dd55555ff607 100644 --- a/include/linux/seqlock.h +++ b/include/linux/seqlock.h @@ -57,6 +57,10 @@ static inline void __seqcount_init(seqcount_t *s, const char *name, # define SEQCOUNT_DEP_MAP_INIT(lockname) \ .dep_map = { .name = #lockname } \ +/** + * seqcount_init() - runtime initializer for seqcount_t + * @s: Pointer to the &typedef seqcount_t instance + */ # define seqcount_init(s) \ do { \ static struct lock_class_key __key; \ @@ -80,13 +84,17 @@ static inline void seqcount_lockdep_reader_access(const seqcount_t *s) # define seqcount_lockdep_reader_access(x) #endif -#define SEQCNT_ZERO(lockname) { .sequence = 0, SEQCOUNT_DEP_MAP_INIT(lockname)} +/** + * SEQCNT_ZERO() - static initializer for seqcount_t + * @name: Name of the &typedef seqcount_t instance + */ +#define SEQCNT_ZERO(name) { .sequence = 0, SEQCOUNT_DEP_MAP_INIT(name) } /** - * __read_seqcount_begin - begin a seq-read critical section (without barrier) - * @s: pointer to seqcount_t - * Returns: count to be passed to read_seqcount_retry + * __read_seqcount_begin() - begin a seq-read critical section (without barrier) + * @s: Pointer to &typedef seqcount_t + * Returns: count to be passed to read_seqcount_retry() * * __read_seqcount_begin is like read_seqcount_begin, but has no smp_rmb() * barrier. Callers should ensure that smp_rmb() or equivalent ordering is @@ -110,9 +118,9 @@ static inline unsigned __read_seqcount_begin(const seqcount_t *s) } /** - * raw_read_seqcount - Read the raw seqcount - * @s: pointer to seqcount_t - * Returns: count to be passed to read_seqcount_retry + * raw_read_seqcount() - Read the raw seqcount + * @s: Pointer to &typedef seqcount_t + * Returns: count to be passed to read_seqcount_retry() * * raw_read_seqcount opens a read critical section of the given * seqcount without any lockdep checking and without checking or @@ -126,13 +134,13 @@ static inline unsigned raw_read_seqcount(const seqcount_t *s) } /** - * raw_read_seqcount_begin - start seq-read critical section w/o lockdep - * @s: pointer to seqcount_t - * Returns: count to be passed to read_seqcount_retry + * raw_read_seqcount_begin() - start seq-read critical section w/o lockdep + * @s: Pointer to &typedef seqcount_t + * Returns: count to be passed to read_seqcount_retry() * * raw_read_seqcount_begin opens a read critical section of the given * seqcount, but without any lockdep checking. Validity of the critical - * section is tested by checking read_seqcount_retry function. + * section is tested by calling read_seqcount_retry(). */ static inline unsigned raw_read_seqcount_begin(const seqcount_t *s) { @@ -142,13 +150,13 @@ static inline unsigned raw_read_seqcount_begin(const seqcount_t *s) } /** - * read_seqcount_begin - begin a seq-read critical section - * @s: pointer to seqcount_t - * Returns: count to be passed to read_seqcount_retry + * read_seqcount_begin() - begin a seq-read critical section + * @s: Pointer to &typedef seqcount_t + * Returns: count to be passed to read_seqcount_retry() * - * read_seqcount_begin opens a read critical section of the given seqcount. - * Validity of the critical section is tested by checking read_seqcount_retry - * function. + * read_seqcount_begin opens a read critical section of the given + * seqcount_t. Validity of the critical section is tested by calling + * read_seqcount_retry(). */ static inline unsigned read_seqcount_begin(const seqcount_t *s) { @@ -157,8 +165,8 @@ static inline unsigned read_seqcount_begin(const seqcount_t *s) } /** - * raw_seqcount_begin - begin a seq-read critical section - * @s: pointer to seqcount_t + * raw_seqcount_begin() - begin a seq-read critical section + * @s: Pointer to &typedef seqcount_t * Returns: count to be passed to read_seqcount_retry * * raw_seqcount_begin opens a read critical section of the given seqcount. @@ -178,8 +186,8 @@ static inline unsigned raw_seqcount_begin(const seqcount_t *s) } /** - * __read_seqcount_retry - end a seq-read critical section (without barrier) - * @s: pointer to seqcount_t + * __read_seqcount_retry() - end a seq-read critical section (without barrier) + * @s: Pointer to &typedef seqcount_t * @start: count, from read_seqcount_begin * Returns: 1 if retry is required, else 0 * @@ -197,8 +205,8 @@ static inline int __read_seqcount_retry(const seqcount_t *s, unsigned start) } /** - * read_seqcount_retry - end a seq-read critical section - * @s: pointer to seqcount_t + * read_seqcount_retry() - end a seq-read critical section + * @s: Pointer to &typedef seqcount_t * @start: count, from read_seqcount_begin * Returns: 1 if retry is required, else 0 * @@ -225,8 +233,8 @@ static inline void raw_write_seqcount_end(seqcount_t *s) } /** - * raw_write_seqcount_barrier - do a seq write barrier - * @s: pointer to seqcount_t + * raw_write_seqcount_barrier() - do a seq write barrier + * @s: Pointer to &typedef seqcount_t * * This can be used to provide an ordering guarantee instead of the * usual consistency guarantee. It is one wmb cheaper, because we can @@ -267,6 +275,21 @@ static inline void raw_write_seqcount_barrier(seqcount_t *s) s->sequence++; } +/** + * raw_read_seqcount_latch() - pick even or odd seqcount latch data copy + * @s: Pointer to &typedef seqcount_t + * + * Use seqcount latching to switch between two storage places with + * sequence protection to allow interruptible, preemptible, writer + * sections. + * + * Check raw_write_seqcount_latch() for more details and a full reader + * and writer usage example. + * + * Return: sequence counter. Use the lowest bit as index for picking + * which data copy to read. Full counter must then be passed to + * read_seqcount_retry(). + */ static inline int raw_read_seqcount_latch(seqcount_t *s) { /* Pairs with the first smp_wmb() in raw_write_seqcount_latch() */ @@ -275,8 +298,8 @@ static inline int raw_read_seqcount_latch(seqcount_t *s) } /** - * raw_write_seqcount_latch - redirect readers to even/odd copy - * @s: pointer to seqcount_t + * raw_write_seqcount_latch() - redirect readers to even/odd copy + * @s: Pointer to &typedef seqcount_t * * The latch technique is a multiversion concurrency control method that allows * queries during non-atomic modifications. If you can guarantee queries never @@ -336,8 +359,8 @@ static inline int raw_read_seqcount_latch(seqcount_t *s) * idx = seq & 0x01; * entry = data_query(latch->data[idx], ...); * - * smp_rmb(); - * } while (seq != latch->seq); + * // read_seqcount_retry() includes necessary smp_rmb() + * } while (read_seqcount_retry(&latch->seq, seq); * * return entry; * } @@ -391,11 +414,26 @@ static inline void __write_seqcount_begin(seqcount_t *s) __write_seqcount_begin_nested(s, 0); } +/** + * write_seqcount_begin() - start a seqcount write-side critical section + * @s: Pointer to &typedef seqcount_t + * + * write_seqcount_begin opens a write-side critical section of the given + * seqcount. Seqcount write-side critical sections must be externally + * serialized and non-preemptible. + */ static inline void write_seqcount_begin(seqcount_t *s) { write_seqcount_begin_nested(s, 0); } +/** + * write_seqcount_end() - end a seqcount write-side critical section + * @s: Pointer to &typedef seqcount_t + * + * write_seqcount_end closes a write-side critical section of the given + * seqcount. + */ static inline void write_seqcount_end(seqcount_t *s) { seqcount_release(&s->dep_map, _RET_IP_); @@ -403,8 +441,8 @@ static inline void write_seqcount_end(seqcount_t *s) } /** - * write_seqcount_invalidate - invalidate in-progress read-side seq operations - * @s: pointer to seqcount_t + * write_seqcount_invalidate() - invalidate in-progress read-side seq operations + * @s: Pointer to &typedef seqcount_t * * After write_seqcount_invalidate, no read-side seq operations will complete * successfully and see data older than this. @@ -435,32 +473,68 @@ typedef struct { .lock = __SPIN_LOCK_UNLOCKED(lockname) \ } -#define seqlock_init(x) \ +/** + * seqlock_init() - dynamic initializer for seqlock_t + * @sl: Pointer to the seqlock_t instance + */ +#define seqlock_init(sl) \ do { \ - seqcount_init(&(x)->seqcount); \ - spin_lock_init(&(x)->lock); \ + seqcount_init(&(sl)->seqcount); \ + spin_lock_init(&(sl)->lock); \ } while (0) -#define DEFINE_SEQLOCK(x) \ - seqlock_t x = __SEQLOCK_UNLOCKED(x) +/** + * DEFINE_SEQLOCK() - Define a statically-allocated seqlock_t + * @sl: Name of the &typedef seqlock_t instance + */ +#define DEFINE_SEQLOCK(sl) \ + seqlock_t sl = __SEQLOCK_UNLOCKED(sl) -/* - * Read side functions for starting and finalizing a read side section. +/** + * read_seqbegin() - start a seqlock_t read-side critical section + * @sl: Pointer to &typedef seqlock_t + * + * read_seqbegin opens a read side critical section of the given + * seqlock_t. Validity of the critical section is tested by checking + * read_seqretry(). + * + * Return: count to be passed to read_seqretry() */ static inline unsigned read_seqbegin(const seqlock_t *sl) { return read_seqcount_begin(&sl->seqcount); } +/** + * read_seqretry() - end a seqlock_t read side critical section + * @sl: Pointer to &typedef seqlock_t + * @start: count, from read_seqbegin() + * + * read_seqretry closes a read side critical section of the given + * seqlock_t. If the read side critical section was invalid, it must be + * ignored and retried. + * + * Return: 1 if a retry is required, 0 otherwise + */ static inline unsigned read_seqretry(const seqlock_t *sl, unsigned start) { return read_seqcount_retry(&sl->seqcount, start); } -/* - * Lock out other writers and update the count. - * Acts like a normal spin_lock/unlock. - * Don't need preempt_disable() because that is in the spin_lock already. +/** + * write_seqlock() - start a seqlock_t write side critical section + * @sl: Pointer to &typedef seqlock_t + * + * write_seqlock opens a write side critical section of the given + * seqlock_t. It also acquires the spinlock embedded inside the + * sequential lock. All seqlock_t write side critical sections are thus + * automatically serialized and non-preemptible. + * + * If the seqlock_t read side section can be invoked from a hardirq or + * softirq context, the ``_irqsave`` and ``_bh`` variants of this + * function must be respectively used instead. + * + * The opened write side section must be closed with write_sequnlock(). */ static inline void write_seqlock(seqlock_t *sl) { @@ -468,30 +542,74 @@ static inline void write_seqlock(seqlock_t *sl) __write_seqcount_begin(&sl->seqcount); } +/** + * write_sequnlock() - end a seqlock_t write side critical section + * @sl: Pointer to &typedef seqlock_t + * + * write_sequnlock closes the (serialized and non-preemptible) write + * side critical section of the given seqlock_t. + */ static inline void write_sequnlock(seqlock_t *sl) { write_seqcount_end(&sl->seqcount); spin_unlock(&sl->lock); } +/** + * write_seqlock_bh() - start a softirqs-disabled seqlock_t write section + * @sl: Pointer to &typedef seqlock_t + * + * write_seqlock_bh is a write_seqlock() variant that disables softirqs + * before opening the serialized seqlock_t write side critical section. + * Use it only if the read side section, or other writers, can be + * invoked from a softirq context. + * + * The opened write section must be closed with write_sequnlock_bh(). + */ static inline void write_seqlock_bh(seqlock_t *sl) { spin_lock_bh(&sl->lock); __write_seqcount_begin(&sl->seqcount); } +/** + * write_sequnlock_bh() - end a softirqs-disabled seqlock_t write section + * @sl: Pointer to &typedef seqlock_t + * + * write_sequnlock_bh closes the serialized, softirqs-disabled, + * seqlock_t write side critical section. It enables softirqs if they + * were already enabled before calling the paired write_seqlock_bh(). + */ static inline void write_sequnlock_bh(seqlock_t *sl) { write_seqcount_end(&sl->seqcount); spin_unlock_bh(&sl->lock); } +/** + * write_seqlock_irq() - start a non-interruptible seqlock_t write side section + * @sl: Pointer to &typedef seqlock_t + * + * write_seqlock_irq is a write_seqlock() variant where hardirqs are + * disabled before opening the serialized and non-preemptible seqlock_t + * write side critical section. + */ static inline void write_seqlock_irq(seqlock_t *sl) { spin_lock_irq(&sl->lock); __write_seqcount_begin(&sl->seqcount); } +/** + * write_sequnlock_irq() - end a non-interruptible seqlock_t write side section + * @sl: Pointer to &typedef seqlock_t + * + * write_sequnlock_irq closes the serialized and non-interruptible write + * side critical section of the given seqlock_t. It enables local + * interrupts afterwards. + * + * The write critical section must've been opened with write_seqlock_irq(). + */ static inline void write_sequnlock_irq(seqlock_t *sl) { write_seqcount_end(&sl->seqcount); @@ -507,9 +625,36 @@ static inline unsigned long __write_seqlock_irqsave(seqlock_t *sl) return flags; } +/** + * write_seqlock_irqsave() - start a non-interruptible seqlock_t write section + * @lock: Pointer to &typedef seqlock_t + * @flags: Stack-allocated storage for saving caller's local interrupt + * state, to be passed to write_sequnlock_irqrestore(). + * + * write_seqlock_irqsave is a write_seqlock() variant where the caller's + * local interrupts state is saved, then local interrupts are disabled, + * before opening the serialized and non-preemptible seqlock_t write + * side critical section. + * + * Use this only if the read side section can be invoked from a hardirq + * context. + * + * The opened write section must be closed with write_sequnlock_irqrestore(). + */ #define write_seqlock_irqsave(lock, flags) \ do { flags = __write_seqlock_irqsave(lock); } while (0) +/** + * write_sequnlock_irqrestore() - end non-interruptible seqlock_t write section + * @sl: Pointer to &typedef seqlock_t + * @flags: Caller's saved interrupt state, from write_seqlock_irqsave() + * + * write_sequnlock_irq closes the serialized and non-interruptible write + * side critical section of the given seqlock_t. It then restores the + * caller's local interrupts saved state. + * + * The write section must've been opened with write_seqlock_irqsave(). + */ static inline void write_sequnlock_irqrestore(seqlock_t *sl, unsigned long flags) { @@ -517,30 +662,61 @@ write_sequnlock_irqrestore(seqlock_t *sl, unsigned long flags) spin_unlock_irqrestore(&sl->lock, flags); } -/* - * A locking reader exclusively locks out other writers and locking readers, - * but doesn't update the sequence number. Acts like a normal spin_lock/unlock. - * Don't need preempt_disable() because that is in the spin_lock already. +/** + * read_seqlock_excl() - begin a seqlock_t locking reader critical section + * @sl: Pointer to &typedef seqlock_t + * + * read_seqlock_excl opens a locking reader critical section for the + * given seqlock_t. A locking reader exclusively locks out other writers + * and other locking readers, but doesn't update the sequence number. + * + * Locking readers act like a normal spin_lock()/spin_unlock(). + * + * The opened read side section must be closed with read_sequnlock_excl(). */ static inline void read_seqlock_excl(seqlock_t *sl) { spin_lock(&sl->lock); } +/** + * read_sequnlock_excl() - end a seqlock_t locking reader critical section + * @sl: Pointer to &typedef seqlock_t + * + * read_sequnlock_excl closes a locking reader critical section. The + * read section must've been opened with read_seqlock_excl(). + */ static inline void read_sequnlock_excl(seqlock_t *sl) { spin_unlock(&sl->lock); } /** - * read_seqbegin_or_lock - begin a sequence number check or locking block - * @lock: sequence lock - * @seq : sequence number to be checked + * read_seqbegin_or_lock() - begin a seqlock_t lockless or locking reader + * @lock: Pointer to &typedef seqlock_t + * @seq : Marker and return parameter. If the passed value is even, the + * reader will become a *lockless* seqlock_t sequence counter reader as + * in read_seqbegin(). If the passed value is odd, the reader will + * become a fully locking reader, as in read_seqlock_excl(). In the + * first call to read_seqbegin_or_lock(), the caller **must** initialize + * and pass an even value in @seq so a lockless read is optimistically + * tried first. * - * First try it once optimistically without taking the lock. If that fails, - * take the lock. The sequence number is also used as a marker for deciding - * whether to be a reader (even) or writer (odd). - * N.B. seq must be initialized to an even number to begin with. + * read_seqbegin_or_lock optimistically tries a lockless seqlock_t + * sequence counter read first. If an odd counter is found, the lockless + * read trial has failed, and the reader transforms to a full seqlock_t + * locking reader as in read_seqlock_excl(). This is typically used to + * avoid lockless seqlock_t readers starvation (too much retry loops) in + * the case of a sharp spike in write activity. + * + * The opened read section must be closed with done_seqretry(). Check + * Documentation/locking/seqlock.rst for template example code. + * + * Return: The read critical section status is returned through @seq, + * which is overloaded as a return parameter. This value must be passed + * to need_seqretry() to check the validity of the tried seqlock_t read + * section. If the read section must be retried, the returned value must + * also be passed to the next iteration of read_seqbegin_or_lock(). */ static inline void read_seqbegin_or_lock(seqlock_t *lock, int *seq) { @@ -550,32 +726,98 @@ static inline void read_seqbegin_or_lock(seqlock_t *lock, int *seq) read_seqlock_excl(lock); } +/** + * need_seqretry() - validate seqlock_t "locking or lockless" reader section + * @lock: Pointer to &typedef seqlock_t + * @seq: count, from read_seqbegin_or_lock() + * + * need_seqretry checks if the seqlock_t read-side critical section + * started with read_seqbegin_or_lock() is valid. If it was not, the + * caller must retry the read-side section. + * + * Return: 1 if a retry is required, 0 otherwise + */ static inline int need_seqretry(seqlock_t *lock, int seq) { return !(seq & 1) && read_seqretry(lock, seq); } +/** + * done_seqretry() - end seqlock_t "locking or lockless" reader section + * @lock: Pointer to &typedef seqlock_t + * @seq: count, from read_seqbegin_or_lock() + * + * done_seqretry finishes the seqlock_t read side critical section + * started by read_seqbegin_or_lock(). Before finishing the critical + * section, the validity of the read side section must've been already + * verified with need_seqretry(). + */ static inline void done_seqretry(seqlock_t *lock, int seq) { if (seq & 1) read_sequnlock_excl(lock); } +/** + * read_seqlock_excl_bh() - start a locking reader seqlock_t section + * with softirqs disabled + * @sl: Pointer to &typedef seqlock_t + * + * read_seqlock_excl_bh is a variant of read_seqlock_excl() that saves + * softirqs state, then disables softirqs, before starting the locking + * reader read side section. Only use this variant if the seqlock_t + * write side section, *or other read sections*, can be invoked from a + * softirq context + * + * The opened section must be closed with read_sequnlock_excl_bh(). + */ static inline void read_seqlock_excl_bh(seqlock_t *sl) { spin_lock_bh(&sl->lock); } +/** + * read_sequnlock_excl_bh() - stop a seqlock_t softirq-disabled locking + * reader section + * @sl: Pointer to &typedef seqlock_t + * + * read_sequnlock_excl_bh ends the softirq-disabled seqlock_t locking + * reader read side section. It restores the softirqs state saved by + * read_seqlock_excl_bh() afterwards. + */ static inline void read_sequnlock_excl_bh(seqlock_t *sl) { spin_unlock_bh(&sl->lock); } +/** + * read_seqlock_excl_irq() - start a non-interruptible seqlock_t locking + * reader section + * @sl: Pointer to &typedef seqlock_t + * + * read_seqlock_excl_irq is a variant of read_seqlock_excl() that + * disables interrupts before starting the locking reader read side + * section. Only use this variant if the seqlock_t write side section, + * *or other read sections*, can be invoked from a hardirq context + * + * The opened read section must be closed with read_sequnlock_excl_irq(). + */ static inline void read_seqlock_excl_irq(seqlock_t *sl) { spin_lock_irq(&sl->lock); } +/** + * read_sequnlock_excl_irq() - end an interrupts-disabled seqlock_t + * locking reader section + * @sl: Pointer to &typedef seqlock_t + * + * read_sequnlock_excl_irq ends the interrupts-disabled seqlock_t + * locking reader read side critical section. It enables local + * interrupts afterwards. + * + * The read section must've been started with read_seqlock_excl_irq(). + */ static inline void read_sequnlock_excl_irq(seqlock_t *sl) { spin_unlock_irq(&sl->lock); @@ -589,15 +831,68 @@ static inline unsigned long __read_seqlock_excl_irqsave(seqlock_t *sl) return flags; } +/** + * read_seqlock_excl_irqsave() - start a non-interruptible seqlock_t + * locking reader section + * @lock: Pointer to &typedef seqlock_t + * @flags: Stack-allocated storage for saving caller's local interrupt + * state, to be passed to read_sequnlock_excl_irqrestore(). + * + * read_seqlock_excl_irqsave is a read_seqlock_excl() variant which + * saves the caller's local interrupts state, then disables local + * interrupts, before opening the seqlock_t locking reader critical + * section. + * + * Use this only if the seqlock_t write side critical section, or other + * read side sections, can be invoked from a hardirq context. + * + * The opened locking reader critical section must be closed with + * read_sequnlock_excl_irqrestore(). + */ #define read_seqlock_excl_irqsave(lock, flags) \ do { flags = __read_seqlock_excl_irqsave(lock); } while (0) +/** + * read_sequnlock_excl_irqrestore() - end non-interruptible seqlock_t + * locking reader section + * @sl: Pointer to &typedef seqlock_t + * @flags: Caller's saved interrupt state, from + * read_seqlock_excl_irqsave() + * + * read_sequnlock_excl_irqrestore closes the non-interruptible seqlock_t + * locking reader section. It then restores the caller's local + * interrupts saved state. + * + * The read section must've been opened with read_seqlock_excl_irqsave(). + */ static inline void read_sequnlock_excl_irqrestore(seqlock_t *sl, unsigned long flags) { spin_unlock_irqrestore(&sl->lock, flags); } +/** + * read_seqbegin_or_lock_irqsave() - begin a seqlock_t lockless reader, or + * a non-interruptible locking reader + * @lock: Pointer to &typedef seqlock_t + * @seq: Marker and return parameter. Check read_seqbegin_or_lock(). + * + * read_seqbegin_or_lock_irqsave is a variant of read_seqbegin_or_lock() + * which saves the local interrupts state, then disables local + * interrupts, before opening a seqlock_t *locking reader* critical + * section. + * + * The opened section must be closed with done_seqretry_irqrestore(). + * + * Return: + * + * 1. The saved local interrupts state in case of a locking reader, to + * be passed to done_seqretry_irqrestore(). + * + * 2. The read critical section status, returned through @seq which is + * overloaded as a return parameter. Check read_seqbegin_or_lock() + * for more info. + */ static inline unsigned long read_seqbegin_or_lock_irqsave(seqlock_t *lock, int *seq) { @@ -611,6 +906,19 @@ read_seqbegin_or_lock_irqsave(seqlock_t *lock, int *seq) return flags; } +/** + * done_seqretry_irqrestore() - end a seqlock_t lockless reader, or a + * non-interruptible locking reader section + * @lock: Pointer to &typedef seqlock_t + * @seq: Count, from read_seqbegin_or_lock_irqsave() + * @flags: Caller's saved local interrupt state in case of a locking + * reader, also from read_seqbegin_or_lock_irqsave() + * + * done_seqretry_irqrestore is a variant of done_seqretry() which + * restores the callers saved local interrupts state in case of a + * locking reader. Check done_seqretry() for more information. The read + * section must've been opened with read_seqbegin_or_lock_irqsave(). + */ static inline void done_seqretry_irqrestore(seqlock_t *lock, int seq, unsigned long flags) { -- 2.20.1
next prev parent reply index Thread overview: 258+ messages / expand[flat|nested] mbox.gz Atom feed top 2020-05-19 21:45 [PATCH v1 00/25] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 01/25] net: core: device_rename: Use rwsem instead of a seqcount Ahmed S. Darwish 2020-05-19 22:01 ` Stephen Hemminger 2020-05-19 22:23 ` Thomas Gleixner 2020-05-19 23:11 ` Stephen Hemminger 2020-05-19 23:42 ` Thomas Gleixner 2020-05-20 0:06 ` Stephen Hemminger 2020-05-20 1:55 ` Thomas Gleixner 2020-05-20 2:57 ` David Miller 2020-05-20 3:18 ` Eric Dumazet 2020-05-20 4:36 ` Stephen Hemminger 2020-05-20 19:37 ` Thomas Gleixner 2020-05-20 21:36 ` Stephen Hemminger 2020-05-20 2:01 ` Eric Dumazet 2020-05-20 6:42 ` Ahmed S. Darwish 2020-05-20 12:51 ` Eric Dumazet 2020-06-03 14:33 ` Ahmed S. Darwish 2020-05-20 14:37 ` Dan Carpenter 2020-05-25 16:22 ` Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 02/25] mm/swap: Don't abuse the seqcount latching API Ahmed S. Darwish 2020-05-20 12:22 ` Konstantin Khlebnikov 2020-05-20 13:05 ` Peter Zijlstra 2020-05-22 14:57 ` Peter Zijlstra 2020-05-22 15:17 ` Sebastian A. Siewior 2020-05-22 16:23 ` Peter Zijlstra 2020-05-25 15:24 ` Ahmed S. Darwish 2020-05-25 15:45 ` Peter Zijlstra 2020-05-25 16:10 ` John Ogness 2020-09-10 15:08 ` [tip: locking/core] mm/swap: Do not abuse the seqcount_t " tip-bot2 for Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 03/25] net: phy: fixed_phy: Remove unused seqcount Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 04/25] block: nr_sects_write(): Disable preemption on seqcount write Ahmed S. Darwish 2020-05-22 16:39 ` Peter Zijlstra 2020-05-25 9:56 ` Ahmed S. Darwish [not found] ` <20200522001237.A00E8206BE@mail.kernel.org> 2020-05-25 10:12 ` Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 05/25] u64_stats: Document writer non-preemptibility requirement Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 06/25] dma-buf: Remove custom seqcount lockdep class key Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 07/25] lockdep: Add preemption disabled assertion API Ahmed S. Darwish 2020-05-22 17:55 ` Peter Zijlstra 2020-05-23 14:59 ` Sebastian A. Siewior 2020-05-23 22:41 ` Peter Zijlstra 2020-05-24 10:50 ` Sebastian A. Siewior 2020-05-25 10:22 ` Peter Zijlstra 2020-05-26 0:52 ` Ahmed S. Darwish 2020-05-26 8:13 ` Peter Zijlstra 2020-05-26 9:45 ` Ahmed S. Darwish 2020-06-03 15:30 ` Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 08/25] seqlock: lockdep assert non-preemptibility on seqcount_t write Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 09/25] Documentation: locking: Describe seqlock design and usage Ahmed S. Darwish 2020-05-22 18:01 ` Peter Zijlstra 2020-05-22 22:24 ` Steven Rostedt 2020-05-25 10:50 ` Ahmed S. Darwish 2020-05-25 11:02 ` Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 10/25] seqlock: Add RST directives to kernel-doc code samples and notes Ahmed S. Darwish 2020-05-22 18:02 ` Peter Zijlstra 2020-05-22 18:03 ` Peter Zijlstra 2020-05-22 18:26 ` Thomas Gleixner 2020-05-22 18:32 ` Peter Zijlstra 2020-05-25 9:36 ` Ahmed S. Darwish 2020-05-25 13:44 ` Peter Zijlstra 2020-05-25 14:07 ` Peter Zijlstra 2020-05-19 21:45 ` Ahmed S. Darwish [this message] 2020-05-19 21:45 ` [PATCH v1 12/25] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 13/25] dma-buf: Use sequence counter with associated wound/wait mutex Ahmed S. Darwish 2020-05-20 10:48 ` Christian König 2020-05-21 0:09 ` Ahmed S. Darwish 2020-05-21 13:20 ` Christian König 2020-05-19 21:45 ` [PATCH v1 14/25] sched: tasks: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 15/25] netfilter: conntrack: " Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 16/25] netfilter: nft_set_rbtree: Use sequence counter with associated rwlock Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 17/25] xfrm: policy: Use sequence counters with associated lock Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 18/25] timekeeping: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 19/25] vfs: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 20/25] raid5: " Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 21/25] iocost: " Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 22/25] NFSv4: " Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 23/25] userfaultfd: " Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 24/25] kvm/eventfd: " Ahmed S. Darwish 2020-05-19 21:45 ` [PATCH v1 25/25] hrtimer: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 00/18] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 01/18] Documentation: locking: Describe seqlock design and usage Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 02/18] seqlock: Properly format kernel-doc code samples Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 03/18] seqlock: Add missing kernel-doc annotations Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 04/18] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 05/18] dma-buf: Remove custom seqcount lockdep class key Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 06/18] dma-buf: Use sequence counter with associated wound/wait mutex Ahmed S. Darwish 2020-06-08 14:32 ` Daniel Vetter 2020-06-08 0:57 ` [PATCH v2 07/18] sched: tasks: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 08/18] netfilter: conntrack: " Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 09/18] netfilter: nft_set_rbtree: Use sequence counter with associated rwlock Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 10/18] xfrm: policy: Use sequence counters with associated lock Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 11/18] timekeeping: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 12/18] vfs: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 13/18] raid5: " Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 14/18] iocost: " Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 15/18] NFSv4: " Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 16/18] userfaultfd: " Ahmed S. Darwish 2020-06-08 0:57 ` [PATCH v2 17/18] kvm/eventfd: " Ahmed S. Darwish 2020-06-08 12:57 ` Paolo Bonzini 2020-06-08 0:57 ` [PATCH v2 18/18] hrtimer: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 00/20] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 01/20] Documentation: locking: Describe seqlock design and usage Ahmed S. Darwish 2020-07-06 21:04 ` Peter Zijlstra 2020-07-06 21:12 ` Jonathan Corbet 2020-07-06 21:16 ` Peter Zijlstra 2020-07-07 10:12 ` Ahmed S. Darwish 2020-07-07 12:47 ` Peter Zijlstra 2020-06-30 5:44 ` [PATCH v3 02/20] seqlock: Properly format kernel-doc code samples Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 03/20] seqlock: Add missing kernel-doc annotations Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 04/20] lockdep: Add preemption enabled/disabled assertion APIs Ahmed S. Darwish 2020-07-06 20:50 ` Peter Zijlstra 2020-07-07 7:34 ` Sebastian A. Siewior 2020-06-30 5:44 ` [PATCH v3 05/20] seqlock: lockdep assert non-preemptibility on seqcount_t write Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 06/20] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-07-06 21:21 ` Peter Zijlstra 2020-07-07 8:40 ` Ahmed S. Darwish 2020-07-07 13:04 ` Peter Zijlstra 2020-07-07 14:37 ` Peter Zijlstra 2020-07-08 9:12 ` Peter Zijlstra 2020-07-08 10:43 ` Ahmed S. Darwish 2020-07-08 10:33 ` Ahmed S. Darwish 2020-07-08 12:29 ` Peter Zijlstra 2020-07-08 14:13 ` Peter Zijlstra 2020-07-08 14:25 ` Peter Zijlstra 2020-07-08 15:09 ` Ahmed S. Darwish 2020-07-08 15:35 ` Peter Zijlstra 2020-07-08 15:58 ` Ahmed S. Darwish 2020-07-08 16:16 ` Peter Zijlstra 2020-07-08 16:18 ` Peter Zijlstra 2020-07-08 16:01 ` Peter Zijlstra 2020-06-30 5:44 ` [PATCH v3 07/20] dma-buf: Remove custom seqcount lockdep class key Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 08/20] dma-buf: Use sequence counter with associated wound/wait mutex Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 09/20] sched: tasks: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 10/20] netfilter: conntrack: " Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 11/20] netfilter: nft_set_rbtree: Use sequence counter with associated rwlock Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 12/20] xfrm: policy: Use sequence counters with associated lock Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 13/20] timekeeping: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 14/20] vfs: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 15/20] raid5: " Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 16/20] iocost: " Ahmed S. Darwish 2020-06-30 7:11 ` Daniel Wagner 2020-06-30 5:44 ` [PATCH v3 17/20] NFSv4: " Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 18/20] userfaultfd: " Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 19/20] kvm/eventfd: " Ahmed S. Darwish 2020-06-30 5:44 ` [PATCH v3 20/20] hrtimer: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 00/24] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 01/24] Documentation: locking: Describe seqlock design and usage Ahmed S. Darwish 2020-07-21 1:35 ` Steven Rostedt 2020-07-21 1:37 ` Steven Rostedt 2020-07-21 5:34 ` Ahmed S. Darwish 2020-07-21 1:44 ` Steven Rostedt 2020-07-21 1:51 ` Steven Rostedt 2020-07-21 7:15 ` Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 02/24] seqlock: Properly format kernel-doc code samples Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 03/24] seqlock: seqcount_t latch: End read sections with read_seqcount_retry() Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 04/24] seqlock: Reorder seqcount_t and seqlock_t API definitions Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 05/24] seqlock: Add kernel-doc for seqcount_t and seqlock_t APIs Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 06/24] seqlock: Implement raw_seqcount_begin() in terms of raw_read_seqcount() Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 07/24] lockdep: Add preemption enabled/disabled assertion APIs Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 08/24] seqlock: lockdep assert non-preemptibility on seqcount_t write Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-08-08 23:21 ` [PATCH v4 08/24] " Guenter Roeck 2020-08-08 23:23 ` Guenter Roeck 2020-08-09 18:42 ` Ahmed S. Darwish 2020-08-10 8:59 ` Greg KH 2020-08-10 9:48 ` peterz 2020-08-10 10:03 ` Greg KH 2020-08-10 9:54 ` [PATCH] Revert "seqlock: lockdep assert non-preemptibility on seqcount_t write" Ahmed S. Darwish 2020-08-10 10:05 ` Greg KH 2020-08-10 10:35 ` Ahmed S. Darwish 2020-08-10 14:10 ` Guenter Roeck 2020-08-18 22:51 ` Valdis Klētnieks 2020-08-19 0:56 ` Guenter Roeck 2020-08-19 7:00 ` Sebastian Andrzej Siewior 2020-08-19 7:34 ` Valdis Klētnieks 2020-08-19 16:15 ` Guenter Roeck 2020-08-10 19:55 ` [PATCH v4 08/24] seqlock: lockdep assert non-preemptibility on seqcount_t write Thomas Gleixner 2020-08-11 10:06 ` Greg KH 2020-07-20 15:55 ` [PATCH v4 09/24] seqlock: Extend seqcount API with associated locks Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 10/24] seqlock: Align multi-line macros newline escapes at 72 columns Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 11/24] dma-buf: Remove custom seqcount lockdep class key Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 12/24] dma-buf: Use sequence counter with associated wound/wait mutex Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 13/24] sched: tasks: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 14/24] netfilter: conntrack: " Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 15/24] netfilter: nft_set_rbtree: Use sequence counter with associated rwlock Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 16/24] xfrm: policy: Use sequence counters with associated lock Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 17/24] timekeeping: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 18/24] vfs: Use sequence counter with associated spinlock Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 19/24] raid5: " Ahmed S. Darwish 2020-07-22 6:40 ` Song Liu 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 20/24] iocost: " Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 21/24] NFSv4: " Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 22/24] userfaultfd: " Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 23/24] kvm/eventfd: " Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 15:55 ` [PATCH v4 24/24] hrtimer: Use sequence counter with associated raw spinlock Ahmed S. Darwish 2020-07-29 14:33 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-07-20 16:49 ` [PATCH v4 00/24] seqlock: Extend seqcount API with associated locks Eric Biggers 2020-07-20 17:33 ` Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 0/8] seqlock: Introduce seqcount_latch_t Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 1/8] time/sched_clock: Use raw_read_seqcount_latch() during suspend Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 2/8] mm/swap: Do not abuse the seqcount_t latching API Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 3/8] seqlock: Introduce seqcount_latch_t Ahmed S. Darwish 2020-09-10 15:08 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 4/8] time/sched_clock: Use seqcount_latch_t Ahmed S. Darwish 2020-09-10 15:08 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 5/8] timekeeping: " Ahmed S. Darwish 2020-09-10 15:08 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 6/8] x86/tsc: " Ahmed S. Darwish 2020-09-04 7:41 ` peterz 2020-09-04 8:03 ` peterz 2020-09-07 16:29 ` Ahmed S. Darwish 2020-09-07 17:30 ` peterz 2020-09-08 6:23 ` Ahmed S. Darwish 2020-09-10 15:08 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 7/8] rbtree_latch: " Ahmed S. Darwish 2020-09-10 15:08 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-08-27 11:40 ` [PATCH v1 8/8] seqlock: seqcount latch APIs: Only allow seqcount_latch_t Ahmed S. Darwish 2020-09-10 15:08 ` [tip: locking/core] " tip-bot2 for Ahmed S. Darwish 2020-08-28 1:07 ` [PATCH v1 0/5] seqlock: Introduce PREEMPT_RT support Ahmed S. Darwish 2020-08-28 1:07 ` [PATCH v1 1/5] seqlock: seqcount_LOCKTYPE_t: Standardize naming convention Ahmed S. Darwish 2020-08-28 8:18 ` peterz 2020-08-28 8:24 ` Ahmed S. Darwish 2020-08-28 1:07 ` [PATCH v1 2/5] seqlock: Use unique prefix for seqcount_t property accessors Ahmed S. Darwish 2020-08-28 8:27 ` peterz 2020-08-28 8:59 ` Ahmed S. Darwish 2020-08-28 1:07 ` [PATCH v1 3/5] seqlock: seqcount_t: Implement all read APIs as statement expressions Ahmed S. Darwish 2020-08-28 8:30 ` peterz 2020-08-28 8:37 ` Ahmed S. Darwish 2020-08-28 1:07 ` [PATCH v1 4/5] seqlock: seqcount_LOCKTYPE_t: Introduce PREEMPT_RT support Ahmed S. Darwish 2020-08-28 8:57 ` peterz 2020-08-28 8:59 ` peterz 2020-08-28 9:31 ` Ahmed S. Darwish 2020-08-28 14:36 ` Ahmed S. Darwish 2020-08-28 1:07 ` [PATCH v1 5/5] seqlock: PREEMPT_RT: Do not starve seqlock_t writers Ahmed S. Darwish 2020-09-04 6:52 ` [PATCH v1 0/5] seqlock: Introduce PREEMPT_RT support peterz 2020-09-04 7:30 ` Ahmed S. Darwish 2020-09-10 15:08 ` [tip: locking/core] seqlock: seqcount_LOCKNAME_t: " tip-bot2 for Ahmed S. Darwish
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20200519214547.352050-12-a.darwish@linutronix.de \ --to=a.darwish@linutronix.de \ --cc=bigeasy@linutronix.de \ --cc=corbet@lwn.net \ --cc=linux-doc@vger.kernel.org \ --cc=linux-kernel@vger.kernel.org \ --cc=mingo@redhat.com \ --cc=paulmck@kernel.org \ --cc=peterz@infradead.org \ --cc=rostedt@goodmis.org \ --cc=tglx@linutronix.de \ --cc=will@kernel.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: link
LKML Archive on lore.kernel.org Archives are clonable: git clone --mirror https://lkml.kernel.org/lkml/0 lkml/git/0.git git clone --mirror https://lkml.kernel.org/lkml/1 lkml/git/1.git git clone --mirror https://lkml.kernel.org/lkml/2 lkml/git/2.git git clone --mirror https://lkml.kernel.org/lkml/3 lkml/git/3.git git clone --mirror https://lkml.kernel.org/lkml/4 lkml/git/4.git git clone --mirror https://lkml.kernel.org/lkml/5 lkml/git/5.git git clone --mirror https://lkml.kernel.org/lkml/6 lkml/git/6.git git clone --mirror https://lkml.kernel.org/lkml/7 lkml/git/7.git git clone --mirror https://lkml.kernel.org/lkml/8 lkml/git/8.git git clone --mirror https://lkml.kernel.org/lkml/9 lkml/git/9.git # If you have public-inbox 1.1+ installed, you may # initialize and index your mirror using the following commands: public-inbox-init -V2 lkml lkml/ https://lkml.kernel.org/lkml \ linux-kernel@vger.kernel.org public-inbox-index lkml Example config snippet for mirrors Newsgroup available over NNTP: nntp://nntp.lore.kernel.org/org.kernel.vger.linux-kernel AGPL code for this site: git clone https://public-inbox.org/public-inbox.git