diff options
| author | Thomas Gleixner <tglx@linutronix.de> | 2009-11-17 18:22:11 +0100 | 
|---|---|---|
| committer | Thomas Gleixner <tglx@linutronix.de> | 2009-12-14 23:55:33 +0100 | 
| commit | d209d74d52ab39dc071656533cac095294f70de7 (patch) | |
| tree | f48ae39908db89df7d93df79ac713318e026f0f7 /kernel/rtmutex.c | |
| parent | 1d615482547584b9a8bb6316a58fed6ce90dd9ff (diff) | |
rtmutes: Convert rtmutex.lock to raw_spinlock
Convert locks which cannot be sleeping locks in preempt-rt to
raw_spinlocks.
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/rtmutex.c')
| -rw-r--r-- | kernel/rtmutex.c | 48 | 
1 files changed, 24 insertions, 24 deletions
| diff --git a/kernel/rtmutex.c b/kernel/rtmutex.c index d33da470f9da..a9604815786a 100644 --- a/kernel/rtmutex.c +++ b/kernel/rtmutex.c @@ -231,7 +231,7 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  		goto out_unlock_pi;  	lock = waiter->lock; -	if (!spin_trylock(&lock->wait_lock)) { +	if (!raw_spin_trylock(&lock->wait_lock)) {  		raw_spin_unlock_irqrestore(&task->pi_lock, flags);  		cpu_relax();  		goto retry; @@ -240,7 +240,7 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  	/* Deadlock detection */  	if (lock == orig_lock || rt_mutex_owner(lock) == top_task) {  		debug_rt_mutex_deadlock(deadlock_detect, orig_waiter, lock); -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		ret = deadlock_detect ? -EDEADLK : 0;  		goto out_unlock_pi;  	} @@ -280,7 +280,7 @@ static int rt_mutex_adjust_prio_chain(struct task_struct *task,  	raw_spin_unlock_irqrestore(&task->pi_lock, flags);  	top_waiter = rt_mutex_top_waiter(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	if (!detect_deadlock && waiter != top_waiter)  		goto out_put_task; @@ -459,12 +459,12 @@ static int task_blocks_on_rt_mutex(struct rt_mutex *lock,  	 */  	get_task_struct(owner); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	res = rt_mutex_adjust_prio_chain(owner, detect_deadlock, lock, waiter,  					 task); -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	return res;  } @@ -575,11 +575,11 @@ static void remove_waiter(struct rt_mutex *lock,  	/* gets dropped in rt_mutex_adjust_prio_chain()! */  	get_task_struct(owner); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	rt_mutex_adjust_prio_chain(owner, 0, lock, NULL, current); -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  }  /* @@ -672,14 +672,14 @@ __rt_mutex_slowlock(struct rt_mutex *lock, int state,  				break;  		} -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		debug_rt_mutex_print_deadlock(waiter);  		if (waiter->task)  			schedule_rt_mutex(lock); -		spin_lock(&lock->wait_lock); +		raw_spin_lock(&lock->wait_lock);  		set_current_state(state);  	} @@ -700,11 +700,11 @@ rt_mutex_slowlock(struct rt_mutex *lock, int state,  	debug_rt_mutex_init_waiter(&waiter);  	waiter.task = NULL; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	/* Try to acquire the lock again: */  	if (try_to_take_rt_mutex(lock)) { -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		return 0;  	} @@ -731,7 +731,7 @@ rt_mutex_slowlock(struct rt_mutex *lock, int state,  	 */  	fixup_rt_mutex_waiters(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	/* Remove pending timer: */  	if (unlikely(timeout)) @@ -758,7 +758,7 @@ rt_mutex_slowtrylock(struct rt_mutex *lock)  {  	int ret = 0; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	if (likely(rt_mutex_owner(lock) != current)) { @@ -770,7 +770,7 @@ rt_mutex_slowtrylock(struct rt_mutex *lock)  		fixup_rt_mutex_waiters(lock);  	} -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	return ret;  } @@ -781,7 +781,7 @@ rt_mutex_slowtrylock(struct rt_mutex *lock)  static void __sched  rt_mutex_slowunlock(struct rt_mutex *lock)  { -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	debug_rt_mutex_unlock(lock); @@ -789,13 +789,13 @@ rt_mutex_slowunlock(struct rt_mutex *lock)  	if (!rt_mutex_has_waiters(lock)) {  		lock->owner = NULL; -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		return;  	}  	wakeup_next_waiter(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	/* Undo pi boosting if necessary: */  	rt_mutex_adjust_prio(current); @@ -970,8 +970,8 @@ EXPORT_SYMBOL_GPL(rt_mutex_destroy);  void __rt_mutex_init(struct rt_mutex *lock, const char *name)  {  	lock->owner = NULL; -	spin_lock_init(&lock->wait_lock); -	plist_head_init(&lock->wait_list, &lock->wait_lock); +	raw_spin_lock_init(&lock->wait_lock); +	plist_head_init_raw(&lock->wait_list, &lock->wait_lock);  	debug_rt_mutex_init(lock, name);  } @@ -1032,7 +1032,7 @@ int rt_mutex_start_proxy_lock(struct rt_mutex *lock,  {  	int ret; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	mark_rt_mutex_waiters(lock); @@ -1040,7 +1040,7 @@ int rt_mutex_start_proxy_lock(struct rt_mutex *lock,  		/* We got the lock for task. */  		debug_rt_mutex_lock(lock);  		rt_mutex_set_owner(lock, task, 0); -		spin_unlock(&lock->wait_lock); +		raw_spin_unlock(&lock->wait_lock);  		rt_mutex_deadlock_account_lock(lock, task);  		return 1;  	} @@ -1056,7 +1056,7 @@ int rt_mutex_start_proxy_lock(struct rt_mutex *lock,  		 */  		ret = 0;  	} -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	debug_rt_mutex_print_deadlock(waiter); @@ -1106,7 +1106,7 @@ int rt_mutex_finish_proxy_lock(struct rt_mutex *lock,  {  	int ret; -	spin_lock(&lock->wait_lock); +	raw_spin_lock(&lock->wait_lock);  	set_current_state(TASK_INTERRUPTIBLE); @@ -1124,7 +1124,7 @@ int rt_mutex_finish_proxy_lock(struct rt_mutex *lock,  	 */  	fixup_rt_mutex_waiters(lock); -	spin_unlock(&lock->wait_lock); +	raw_spin_unlock(&lock->wait_lock);  	/*  	 * Readjust priority, when we did not get the lock. We might have been | 
