mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-23 20:53:53 +08:00
102 lines
2.4 KiB
C
102 lines
2.4 KiB
C
|
#ifndef __ASM_SH64_SEMAPHORE_HELPER_H
|
||
|
#define __ASM_SH64_SEMAPHORE_HELPER_H
|
||
|
|
||
|
/*
|
||
|
* This file is subject to the terms and conditions of the GNU General Public
|
||
|
* License. See the file "COPYING" in the main directory of this archive
|
||
|
* for more details.
|
||
|
*
|
||
|
* include/asm-sh64/semaphore-helper.h
|
||
|
*
|
||
|
* Copyright (C) 2000, 2001 Paolo Alberelli
|
||
|
*
|
||
|
*/
|
||
|
#include <asm/errno.h>
|
||
|
|
||
|
/*
|
||
|
* SMP- and interrupt-safe semaphores helper functions.
|
||
|
*
|
||
|
* (C) Copyright 1996 Linus Torvalds
|
||
|
* (C) Copyright 1999 Andrea Arcangeli
|
||
|
*/
|
||
|
|
||
|
/*
|
||
|
* These two _must_ execute atomically wrt each other.
|
||
|
*
|
||
|
* This is trivially done with load_locked/store_cond,
|
||
|
* which we have. Let the rest of the losers suck eggs.
|
||
|
*/
|
||
|
static __inline__ void wake_one_more(struct semaphore * sem)
|
||
|
{
|
||
|
atomic_inc((atomic_t *)&sem->sleepers);
|
||
|
}
|
||
|
|
||
|
static __inline__ int waking_non_zero(struct semaphore *sem)
|
||
|
{
|
||
|
unsigned long flags;
|
||
|
int ret = 0;
|
||
|
|
||
|
spin_lock_irqsave(&semaphore_wake_lock, flags);
|
||
|
if (sem->sleepers > 0) {
|
||
|
sem->sleepers--;
|
||
|
ret = 1;
|
||
|
}
|
||
|
spin_unlock_irqrestore(&semaphore_wake_lock, flags);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* waking_non_zero_interruptible:
|
||
|
* 1 got the lock
|
||
|
* 0 go to sleep
|
||
|
* -EINTR interrupted
|
||
|
*
|
||
|
* We must undo the sem->count down_interruptible() increment while we are
|
||
|
* protected by the spinlock in order to make atomic this atomic_inc() with the
|
||
|
* atomic_read() in wake_one_more(), otherwise we can race. -arca
|
||
|
*/
|
||
|
static __inline__ int waking_non_zero_interruptible(struct semaphore *sem,
|
||
|
struct task_struct *tsk)
|
||
|
{
|
||
|
unsigned long flags;
|
||
|
int ret = 0;
|
||
|
|
||
|
spin_lock_irqsave(&semaphore_wake_lock, flags);
|
||
|
if (sem->sleepers > 0) {
|
||
|
sem->sleepers--;
|
||
|
ret = 1;
|
||
|
} else if (signal_pending(tsk)) {
|
||
|
atomic_inc(&sem->count);
|
||
|
ret = -EINTR;
|
||
|
}
|
||
|
spin_unlock_irqrestore(&semaphore_wake_lock, flags);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* waking_non_zero_trylock:
|
||
|
* 1 failed to lock
|
||
|
* 0 got the lock
|
||
|
*
|
||
|
* We must undo the sem->count down_trylock() increment while we are
|
||
|
* protected by the spinlock in order to make atomic this atomic_inc() with the
|
||
|
* atomic_read() in wake_one_more(), otherwise we can race. -arca
|
||
|
*/
|
||
|
static __inline__ int waking_non_zero_trylock(struct semaphore *sem)
|
||
|
{
|
||
|
unsigned long flags;
|
||
|
int ret = 1;
|
||
|
|
||
|
spin_lock_irqsave(&semaphore_wake_lock, flags);
|
||
|
if (sem->sleepers <= 0)
|
||
|
atomic_inc(&sem->count);
|
||
|
else {
|
||
|
sem->sleepers--;
|
||
|
ret = 0;
|
||
|
}
|
||
|
spin_unlock_irqrestore(&semaphore_wake_lock, flags);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
#endif /* __ASM_SH64_SEMAPHORE_HELPER_H */
|