[lttng-dev] [RFC PATCH urcu] Document uatomic operations
Mathieu Desnoyers
mathieu.desnoyers at efficios.com
Tue May 15 08:10:03 EDT 2012
* Paul E. McKenney (paulmck at linux.vnet.ibm.com) wrote:
> On Mon, May 14, 2012 at 10:39:01PM -0400, Mathieu Desnoyers wrote:
> > Document each atomic operation provided by urcu/uatomic.h, along with
> > their memory barrier guarantees.
>
> Great to see the documentation!!! Some comments below.
>
> Thanx, Paul
>
> > Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers at efficios.com>
> > ---
> > diff --git a/doc/Makefile.am b/doc/Makefile.am
> > index bec1d7c..db9811c 100644
> > --- a/doc/Makefile.am
> > +++ b/doc/Makefile.am
> > @@ -1 +1 @@
> > -dist_doc_DATA = rcu-api.txt
> > +dist_doc_DATA = rcu-api.txt uatomic-api.txt
> > diff --git a/doc/uatomic-api.txt b/doc/uatomic-api.txt
> > new file mode 100644
> > index 0000000..3605acf
> > --- /dev/null
> > +++ b/doc/uatomic-api.txt
> > @@ -0,0 +1,80 @@
> > +Userspace RCU Atomic Operations API
> > +by Mathieu Desnoyers and Paul E. McKenney
> > +
> > +
> > +This document describes the <urcu/uatomic.h> API. Those are the atomic
> > +operations provided by the Userspace RCU library. The general rule
> > +regarding memory barriers is that only uatomic_xchg(),
> > +uatomic_cmpxchg(), uatomic_add_return(), and uatomic_sub_return() imply
> > +full memory barriers before and after the atomic operation. Other
> > +primitives don't guarantee any memory barrier.
> > +
> > +Only atomic operations performed on integers ("int" and "long", signed
> > +and unsigned) are supported on all architectures. Some architectures
> > +also support 1-byte and 2-byte atomic operations. Those respectively
> > +have UATOMIC_HAS_ATOMIC_BYTE and UATOMIC_HAS_ATOMIC_SHORT defined when
> > +uatomic.h is included. An architecture trying to perform an atomic write
> > +to a type size not supported by the architecture will trigger an illegal
> > +instruction.
> > +
> > +In the description below, "type" is a type that can be atomically
> > +written to by the architecture. It needs to be at most word-sized, and
> > +its alignment needs to greater or equal to its size.
> > +
> > +type uatomic_set(type *addr, type v)
> > +
> > + Atomically write @v into @addr.
>
> Wouldn't this instead be "void uatomic_set(type *addr, type v)"?
Well, in that case, we'd need to change the macro. Currently,
_uatomic_set maps directly to:
#define _uatomic_set(addr, v) CMM_STORE_SHARED(*(addr), (v))
and CMM_STORE_SHARED returns v. The question becomes: should we change
_uatomic_set or CMM_STORE_SHARED so they don't return anything, or
document that they return something ?
One thing I noticed is that linters often complain that the return value
of CMM_STORE_SHARED is never used. One thing we could look into is to
try some gcc attributes and/or linter annotations to flag this return
value as possibly unused. Thoughts ?
>
> By "Atomically write @v into @addr", what is meant is that no concurrent
> operation that reads from addr will see partial effects of uatomic_set(),
> correct? In other words, the concurrent read will either see v or
> the old value, not a mush of the two.
yep. I added that clarification.
>
> > +
> > +type uatomic_read(type *addr)
> > +
> > + Atomically read @v from @addr.
>
> Similar comments on the meaning of "atomically". This may sound picky,
> but people coming from an x86 environment might otherwise assume that
> there is lock prefix involved...
same.
>
> > +
> > +type uatomic_cmpxchg(type *addr, type old, type new)
> > +
> > + Atomically check if @addr contains @old. If true, then replace
> > + the content of @addr by @new. Return the value previously
> > + contained by @addr. This function imply a full memory barrier
> > + before and after the atomic operation.
>
> Suggest "then atomically replace" or some such. It might not hurt
> to add that this is an atomic read-modify-write operation.
Updated to:
type uatomic_cmpxchg(type *addr, type old, type new)
An atomic read-modify-write operation that performs this
sequence of operations atomically: check if @addr contains @old.
If true, then replace the content of @addr by @new. Return the
value previously contained by @addr. This function imply a full
memory barrier before and after the atomic operation.
>
> Similar comments on the other value-returning atomics.
Will do something similar.
>
> > +
> > +type uatomic_xchg(type *addr, type new)
> > +
> > + Atomically replace the content of @addr by @new, and return the
> > + value previously contained by @addr. This function imply a full
> > + memory barrier before and after the atomic operation.
> > +
> > +type uatomic_add_return(type *addr, type v)
> > +type uatomic_sub_return(type *addr, type v)
> > +
> > + Atomically increment/decrement the content of @addr by @v, and
> > + return the resulting value. This function imply a full memory
> > + barrier before and after the atomic operation.
> > +
> > +void uatomic_and(type *addr, type mask)
> > +void uatomic_or(type *addr, type mask)
> > +
> > + Atomically write the result of bitwise "and"/"or" between the
> > + content of @addr and @mask into @addr. Memory barriers are
> > + provided by explicitly using cmm_smp_mb__before_uatomic_and(),
> > + cmm_smp_mb__after_uatomic_and(),
> > + cmm_smp_mb__before_uatomic_or(), and
> > + cmm_smp_mb__after_uatomic_or().
>
> I suggest replacing "Memory barriers are provided ..." with something like
> "These operations do not necessarily imply memory barriers. If memory
> barriers are needed, they may be provided ...". Then perhaps add a
> sentence stating that the advantage of using the four __before_/__after_
> primitives is that they are no-ops on architectures in which the underlying
> atomic instructions implicitly supply the needed memory barriers.
>
> Simlar comments on the other non-value-returning atomic operations below.
OK, done.
Here is the update:
diff --git a/doc/Makefile.am b/doc/Makefile.am
index 27d3793..3422653 100644
--- a/doc/Makefile.am
+++ b/doc/Makefile.am
@@ -1 +1 @@
-dist_doc_DATA = rcu-api.txt cds-api.txt
+dist_doc_DATA = rcu-api.txt cds-api.txt uatomic-api.txt
diff --git a/doc/uatomic-api.txt b/doc/uatomic-api.txt
new file mode 100644
index 0000000..3ad8fbb
--- /dev/null
+++ b/doc/uatomic-api.txt
@@ -0,0 +1,102 @@
+Userspace RCU Atomic Operations API
+by Mathieu Desnoyers and Paul E. McKenney
+
+
+This document describes the <urcu/uatomic.h> API. Those are the atomic
+operations provided by the Userspace RCU library. The general rule
+regarding memory barriers is that only uatomic_xchg(),
+uatomic_cmpxchg(), uatomic_add_return(), and uatomic_sub_return() imply
+full memory barriers before and after the atomic operation. Other
+primitives don't guarantee any memory barrier.
+
+Only atomic operations performed on integers ("int" and "long", signed
+and unsigned) are supported on all architectures. Some architectures
+also support 1-byte and 2-byte atomic operations. Those respectively
+have UATOMIC_HAS_ATOMIC_BYTE and UATOMIC_HAS_ATOMIC_SHORT defined when
+uatomic.h is included. An architecture trying to perform an atomic write
+to a type size not supported by the architecture will trigger an illegal
+instruction.
+
+In the description below, "type" is a type that can be atomically
+written to by the architecture. It needs to be at most word-sized, and
+its alignment needs to greater or equal to its size.
+
+type uatomic_set(type *addr, type v)
+
+ Atomically write @v into @addr. By "atomically", we mean that no
+ concurrent operation that reads from addr will see partial
+ effects of uatomic_set().
+
+type uatomic_read(type *addr)
+
+ Atomically read @v from @addr. By "atomically", we mean that
+ uatomic_read() cannot see a partial effect of any concurrent
+ uatomic update.
+
+type uatomic_cmpxchg(type *addr, type old, type new)
+
+ An atomic read-modify-write operation that performs this
+ sequence of operations atomically: check if @addr contains @old.
+ If true, then replace the content of @addr by @new. Return the
+ value previously contained by @addr. This function imply a full
+ memory barrier before and after the atomic operation.
+
+type uatomic_xchg(type *addr, type new)
+
+ An atomic read-modify-write operation that performs this sequence
+ of operations atomically: replace the content of @addr by @new,
+ and return the value previously contained by @addr. This
+ function imply a full memory barrier before and after the atomic
+ operation.
+
+type uatomic_add_return(type *addr, type v)
+type uatomic_sub_return(type *addr, type v)
+
+ An atomic read-modify-write operation that performs this
+ sequence of operations atomically: increment/decrement the
+ content of @addr by @v, and return the resulting value. This
+ function imply a full memory barrier before and after the atomic
+ operation.
+
+void uatomic_and(type *addr, type mask)
+void uatomic_or(type *addr, type mask)
+
+ Atomically write the result of bitwise "and"/"or" between the
+ content of @addr and @mask into @addr.
+ These operations do not necessarily imply memory barriers.
+ If memory barriers are needed, they may be provided by
+ explicitly using
+ cmm_smp_mb__before_uatomic_and(),
+ cmm_smp_mb__after_uatomic_and(),
+ cmm_smp_mb__before_uatomic_or(), and
+ cmm_smp_mb__after_uatomic_or(). These explicit barriers are
+ np-ops on architectures in which the underlying atomic
+ instructions implicitly supply the needed memory barriers.
+
+void uatomic_add(type *addr, type v)
+void uatomic_sub(type *addr, type v)
+
+ Atomically increment/decrement the content of @addr by @v.
+ These operations do not necessarily imply memory barriers.
+ If memory barriers are needed, they may be provided by
+ explicitly using
+ cmm_smp_mb__before_uatomic_add(),
+ cmm_smp_mb__after_uatomic_add(),
+ cmm_smp_mb__before_uatomic_sub(), and
+ cmm_smp_mb__after_uatomic_sub(). These explicit barriers are
+ np-ops on architectures in which the underlying atomic
+ instructions implicitly supply the needed memory barriers.
+
+void uatomic_inc(type *addr)
+void uatomic_dec(type *addr)
+
+ Atomically increment/decrement the content of @addr by 1.
+ These operations do not necessarily imply memory barriers.
+ If memory barriers are needed, they may be provided by
+ explicitly using
+ cmm_smp_mb__before_uatomic_inc(),
+ cmm_smp_mb__after_uatomic_inc(),
+ cmm_smp_mb__before_uatomic_dec(), and
+ cmm_smp_mb__after_uatomic_dec(). These explicit barriers are
+ np-ops on architectures in which the underlying atomic
+ instructions implicitly supply the needed memory barriers.
--
Mathieu Desnoyers
Operating System Efficiency R&D Consultant
EfficiOS Inc.
http://www.efficios.com
More information about the lttng-dev
mailing list