[ltt-dev] [PATCH v2] cmm: provide lightweight smp_rmb/smp_wmb on PPC
Paul E. McKenney
paulmck at linux.vnet.ibm.com
Thu Sep 22 10:49:11 EDT 2011
On Thu, Sep 22, 2011 at 10:42:52AM +0200, Paolo Bonzini wrote:
> lwsync orders loads in cacheable memory with respect to other loads,
> and stores in cacheable memory with respect to other stores. Use it
> to implement smp_rmb/smp_wmb.
>
> The heavy-weight sync is still used for the "full" rmb/wmb operations,
> as well as for smp_mb.
>
> Signed-off-by: Paolo Bonzini <pbonzini at redhat.com>
> ---
> urcu/arch/ppc.h | 10 +++++++++-
> 1 files changed, 9 insertions(+), 1 deletions(-)
>
> diff --git a/urcu/arch/ppc.h b/urcu/arch/ppc.h
> index a03d688..05f7db6 100644
> --- a/urcu/arch/ppc.h
> +++ b/urcu/arch/ppc.h
> @@ -32,7 +32,15 @@ extern "C" {
> /* Include size of POWER5+ L3 cache lines: 256 bytes */
> #define CAA_CACHE_LINE_SIZE 256
>
> -#define cmm_mb() asm volatile("sync":::"memory")
> +#define cmm_mb() asm volatile("sync":::"memory")
> +
> +/* lwsync does not preserve ordering of cacheable vs. non-cacheable
> + * accesses, but it is good when MMIO is not in use. An eieio+lwsync
> + * pair is also not enough for rmb, because it will order cacheable
> + * and non-cacheable memory operations separately---i.e. not the latter
> + * against the former. */
> +#define cmm_smp_rmb() asm volatile("lwsync":::"memory")
> +#define cmm_smp_wmb() asm volatile("lwsync":::"memory")
This works for recent Power hardware, and I see no reason to care about
stuff old enough to lack lwsync. I must defer to others on embedded
PowerPC.
Thanx, Paul
More information about the lttng-dev
mailing list