[ltt-dev] [PATCH 08/10] call_rcu: redo futex implementation

Mathieu Desnoyers compudj at krystal.dyndns.org
Thu Jun 9 09:24:26 EDT 2011


* Paolo Bonzini (pbonzini at redhat.com) wrote:
> On 06/09/2011 12:42 AM, Mathieu Desnoyers wrote:
> > I'm tempted to just leave the mutex in place for now until we see a
> > clearly measurable performance gain by going with the "or/and" flag
> > approach compared to the simpler (in terms of memory ordering) mutex
> > implementation.
> 
> I don't think the memory ordering guarantees are much different:
> URCU_CALL_RCU_STOP is always written before a futex system call, and
> URCU_CALL_RCU_STOPPED is always written as the last thing before the
> thread exits. In both cases there is no need for a memory barrier.
> So, the version with atomic accesses seems simpler to me, and I
> generally like patches that remove more lines than they add.

Good point. Also, I notice that your patch fix the flag accesses that
are performed outside of mutex protection (e.g. busy loop waiting for
"stopped"), which busy loop on a non-volatile variable. So yes, your
patch is a clear improvement.

> Also, the mutex
> in code that strives to be lockless (in the RT case) seemed out of
> place...

The "RT" guarantee only apply to the thread calling call_rcu() AFAIK.
The call_rcu worker thread is waiting on a poll() for a certain delay in
RT mode, which makes it entirely non-RT. So the mutex there does not
change the RT/non-RT guarantees provided here.

> 
> One thing is that the RT/non-RT flag is not modifiable. So perhaps it is
> better to move it to a separate variable and limit the uatomic_read noise.
> See the attached patch.

Good point. Will merge.

> 
> > As far as the or/and implementations you provide in separate patches are
> > concerned, I'm tempted to pull them for completeness sake (they might
> > end up being useful). Before I do that though, I'm just curious about
> > the memory ordering guarantees they should provide: no-ordering seems
> > like a good thing for or/and: it would be similar to add/sub.
> 
> Yes, I agree.

Thanks!

Mathieu

> 
> Paolo
> 
> ------------------ 8< --------------------------
> 
> From 9eb4f2821f212705a72ea38568b2a6157665b203 Mon Sep 17 00:00:00 2001
> From: Paolo Bonzini <pbonzini at redhat.com>
> Date: Wed, 8 Jun 2011 09:33:51 +0200
> Subject: [PATCH] call_rcu: drop mutex
> 
> The mutex is being used only to protect OR accesses to the flags.
> Just use atomic operations for that.
> 
> Signed-off-by: Paolo Bonzini <pbonzini at redhat.com>
> ---
>  urcu-call-rcu-impl.h |   28 ++++++++++------------------
>  1 files changed, 10 insertions(+), 18 deletions(-)
> 
> diff --git a/urcu-call-rcu-impl.h b/urcu-call-rcu-impl.h
> index 9beb58c..82fec80 100644
> --- a/urcu-call-rcu-impl.h
> +++ b/urcu-call-rcu-impl.h
> @@ -46,7 +46,6 @@
>  struct call_rcu_data {
>  	struct cds_wfq_queue cbs;
>  	unsigned long flags;
> -	pthread_mutex_t mtx;
>  	int futex;
>  	unsigned long qlen; /* maintained for debugging. */
>  	pthread_t tid;
> @@ -204,6 +203,7 @@ static void *call_rcu_thread(void *arg)
>  	struct cds_wfq_node **cbs_tail;
>  	struct call_rcu_data *crdp = (struct call_rcu_data *)arg;
>  	struct rcu_head *rhp;
> +	int rt = !!(uatomic_read(&crdp->flags) & URCU_CALL_RCU_RT);
>  
>  	if (set_thread_cpu_affinity(crdp) != 0) {
>  		perror("pthread_setaffinity_np");
> @@ -212,7 +212,7 @@ static void *call_rcu_thread(void *arg)
>  
>  	thread_call_rcu_data = crdp;
>  	for (;;) {
> -		if (!(crdp->flags & URCU_CALL_RCU_RT)) {
> +		if (!rt) {
>  			uatomic_dec(&crdp->futex);
>  			/* Decrement futex before reading call_rcu list */
>  			cmm_smp_mb();
> @@ -240,8 +240,8 @@ static void *call_rcu_thread(void *arg)
>  			} while (cbs != NULL);
>  			uatomic_sub(&crdp->qlen, cbcount);
>  		}
> -		if (crdp->flags & URCU_CALL_RCU_STOP) {
> -			if (!(crdp->flags & URCU_CALL_RCU_RT)) {
> +		if (uatomic_read(&crdp->flags) & URCU_CALL_RCU_STOP) {
> +			if (!rt) {
>  				/*
>  				 * Read call_rcu list before write futex.
>  				 */
> @@ -250,15 +250,13 @@ static void *call_rcu_thread(void *arg)
>  			}
>  			break;
>  		}
> -		if (!(crdp->flags & URCU_CALL_RCU_RT)) {
> +		if (!rt) {
>  			if (&crdp->cbs.head == _CMM_LOAD_SHARED(crdp->cbs.tail))
>  				call_rcu_wait(crdp);
>  		}
>  		poll(NULL, 0, 10);
>  	}
> -	call_rcu_lock(&crdp->mtx);
> -	crdp->flags |= URCU_CALL_RCU_STOPPED;
> -	call_rcu_unlock(&crdp->mtx);
> +	uatomic_or(&crdp->flags, URCU_CALL_RCU_STOPPED);
>  	return NULL;
>  }
>  
> @@ -282,10 +280,6 @@ static void call_rcu_data_init(struct call_rcu_data **crdpp,
>  	memset(crdp, '\0', sizeof(*crdp));
>  	cds_wfq_init(&crdp->cbs);
>  	crdp->qlen = 0;
> -	if (pthread_mutex_init(&crdp->mtx, NULL) != 0) {
> -		perror("pthread_mutex_init");
> -		exit(-1);
> -	}
>  	crdp->futex = 0;
>  	crdp->flags = flags;
>  	cds_list_add(&crdp->list, &call_rcu_data_list);
> @@ -568,12 +562,10 @@ void call_rcu_data_free(struct call_rcu_data *crdp)
>  	if (crdp == NULL || crdp == default_call_rcu_data) {
>  		return;
>  	}
> -	if ((crdp->flags & URCU_CALL_RCU_STOPPED) == 0) {
> -		call_rcu_lock(&crdp->mtx);
> -		crdp->flags |= URCU_CALL_RCU_STOP;
> -		call_rcu_unlock(&crdp->mtx);
> +	if ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_STOPPED) == 0) {
> +		uatomic_or(&crdp->flags, URCU_CALL_RCU_STOP);
>  		wake_call_rcu_thread(crdp);
> -		while ((crdp->flags & URCU_CALL_RCU_STOPPED) == 0)
> +		while ((uatomic_read(&crdp->flags) & URCU_CALL_RCU_STOPPED) == 0)
>  			poll(NULL, 0, 1);
>  	}
>  	if (&crdp->cbs.head != _CMM_LOAD_SHARED(crdp->cbs.tail)) {
> @@ -657,7 +649,7 @@ void call_rcu_after_fork_child(void)
>  		if (crdp == default_call_rcu_data)
>  			crdp = cds_list_entry(crdp->list.prev,
>  					      struct call_rcu_data, list);
> -		crdp->flags = URCU_CALL_RCU_STOPPED;
> +		uatomic_set(&crdp->flags, URCU_CALL_RCU_STOPPED);
>  		call_rcu_data_free(crdp);
>  	}
>  }
> -- 
> 1.7.4.4
> 

-- 
Mathieu Desnoyers
Operating System Efficiency R&D Consultant
EfficiOS Inc.
http://www.efficios.com




More information about the lttng-dev mailing list