mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-05 01:54:09 +08:00
294 lines
12 KiB
Plaintext
294 lines
12 KiB
Plaintext
|
Hardware Spinlock Framework
|
||
|
|
||
|
1. Introduction
|
||
|
|
||
|
Hardware spinlock modules provide hardware assistance for synchronization
|
||
|
and mutual exclusion between heterogeneous processors and those not operating
|
||
|
under a single, shared operating system.
|
||
|
|
||
|
For example, OMAP4 has dual Cortex-A9, dual Cortex-M3 and a C64x+ DSP,
|
||
|
each of which is running a different Operating System (the master, A9,
|
||
|
is usually running Linux and the slave processors, the M3 and the DSP,
|
||
|
are running some flavor of RTOS).
|
||
|
|
||
|
A generic hwspinlock framework allows platform-independent drivers to use
|
||
|
the hwspinlock device in order to access data structures that are shared
|
||
|
between remote processors, that otherwise have no alternative mechanism
|
||
|
to accomplish synchronization and mutual exclusion operations.
|
||
|
|
||
|
This is necessary, for example, for Inter-processor communications:
|
||
|
on OMAP4, cpu-intensive multimedia tasks are offloaded by the host to the
|
||
|
remote M3 and/or C64x+ slave processors (by an IPC subsystem called Syslink).
|
||
|
|
||
|
To achieve fast message-based communications, a minimal kernel support
|
||
|
is needed to deliver messages arriving from a remote processor to the
|
||
|
appropriate user process.
|
||
|
|
||
|
This communication is based on simple data structures that is shared between
|
||
|
the remote processors, and access to it is synchronized using the hwspinlock
|
||
|
module (remote processor directly places new messages in this shared data
|
||
|
structure).
|
||
|
|
||
|
A common hwspinlock interface makes it possible to have generic, platform-
|
||
|
independent, drivers.
|
||
|
|
||
|
2. User API
|
||
|
|
||
|
struct hwspinlock *hwspin_lock_request(void);
|
||
|
- dynamically assign an hwspinlock and return its address, or NULL
|
||
|
in case an unused hwspinlock isn't available. Users of this
|
||
|
API will usually want to communicate the lock's id to the remote core
|
||
|
before it can be used to achieve synchronization.
|
||
|
Can be called from an atomic context (this function will not sleep) but
|
||
|
not from within interrupt context.
|
||
|
|
||
|
struct hwspinlock *hwspin_lock_request_specific(unsigned int id);
|
||
|
- assign a specific hwspinlock id and return its address, or NULL
|
||
|
if that hwspinlock is already in use. Usually board code will
|
||
|
be calling this function in order to reserve specific hwspinlock
|
||
|
ids for predefined purposes.
|
||
|
Can be called from an atomic context (this function will not sleep) but
|
||
|
not from within interrupt context.
|
||
|
|
||
|
int hwspin_lock_free(struct hwspinlock *hwlock);
|
||
|
- free a previously-assigned hwspinlock; returns 0 on success, or an
|
||
|
appropriate error code on failure (e.g. -EINVAL if the hwspinlock
|
||
|
is already free).
|
||
|
Can be called from an atomic context (this function will not sleep) but
|
||
|
not from within interrupt context.
|
||
|
|
||
|
int hwspin_lock_timeout(struct hwspinlock *hwlock, unsigned int timeout);
|
||
|
- lock a previously-assigned hwspinlock with a timeout limit (specified in
|
||
|
msecs). If the hwspinlock is already taken, the function will busy loop
|
||
|
waiting for it to be released, but give up when the timeout elapses.
|
||
|
Upon a successful return from this function, preemption is disabled so
|
||
|
the caller must not sleep, and is advised to release the hwspinlock as
|
||
|
soon as possible, in order to minimize remote cores polling on the
|
||
|
hardware interconnect.
|
||
|
Returns 0 when successful and an appropriate error code otherwise (most
|
||
|
notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs).
|
||
|
The function will never sleep.
|
||
|
|
||
|
int hwspin_lock_timeout_irq(struct hwspinlock *hwlock, unsigned int timeout);
|
||
|
- lock a previously-assigned hwspinlock with a timeout limit (specified in
|
||
|
msecs). If the hwspinlock is already taken, the function will busy loop
|
||
|
waiting for it to be released, but give up when the timeout elapses.
|
||
|
Upon a successful return from this function, preemption and the local
|
||
|
interrupts are disabled, so the caller must not sleep, and is advised to
|
||
|
release the hwspinlock as soon as possible.
|
||
|
Returns 0 when successful and an appropriate error code otherwise (most
|
||
|
notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs).
|
||
|
The function will never sleep.
|
||
|
|
||
|
int hwspin_lock_timeout_irqsave(struct hwspinlock *hwlock, unsigned int to,
|
||
|
unsigned long *flags);
|
||
|
- lock a previously-assigned hwspinlock with a timeout limit (specified in
|
||
|
msecs). If the hwspinlock is already taken, the function will busy loop
|
||
|
waiting for it to be released, but give up when the timeout elapses.
|
||
|
Upon a successful return from this function, preemption is disabled,
|
||
|
local interrupts are disabled and their previous state is saved at the
|
||
|
given flags placeholder. The caller must not sleep, and is advised to
|
||
|
release the hwspinlock as soon as possible.
|
||
|
Returns 0 when successful and an appropriate error code otherwise (most
|
||
|
notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs).
|
||
|
The function will never sleep.
|
||
|
|
||
|
int hwspin_trylock(struct hwspinlock *hwlock);
|
||
|
- attempt to lock a previously-assigned hwspinlock, but immediately fail if
|
||
|
it is already taken.
|
||
|
Upon a successful return from this function, preemption is disabled so
|
||
|
caller must not sleep, and is advised to release the hwspinlock as soon as
|
||
|
possible, in order to minimize remote cores polling on the hardware
|
||
|
interconnect.
|
||
|
Returns 0 on success and an appropriate error code otherwise (most
|
||
|
notably -EBUSY if the hwspinlock was already taken).
|
||
|
The function will never sleep.
|
||
|
|
||
|
int hwspin_trylock_irq(struct hwspinlock *hwlock);
|
||
|
- attempt to lock a previously-assigned hwspinlock, but immediately fail if
|
||
|
it is already taken.
|
||
|
Upon a successful return from this function, preemption and the local
|
||
|
interrupts are disabled so caller must not sleep, and is advised to
|
||
|
release the hwspinlock as soon as possible.
|
||
|
Returns 0 on success and an appropriate error code otherwise (most
|
||
|
notably -EBUSY if the hwspinlock was already taken).
|
||
|
The function will never sleep.
|
||
|
|
||
|
int hwspin_trylock_irqsave(struct hwspinlock *hwlock, unsigned long *flags);
|
||
|
- attempt to lock a previously-assigned hwspinlock, but immediately fail if
|
||
|
it is already taken.
|
||
|
Upon a successful return from this function, preemption is disabled,
|
||
|
the local interrupts are disabled and their previous state is saved
|
||
|
at the given flags placeholder. The caller must not sleep, and is advised
|
||
|
to release the hwspinlock as soon as possible.
|
||
|
Returns 0 on success and an appropriate error code otherwise (most
|
||
|
notably -EBUSY if the hwspinlock was already taken).
|
||
|
The function will never sleep.
|
||
|
|
||
|
void hwspin_unlock(struct hwspinlock *hwlock);
|
||
|
- unlock a previously-locked hwspinlock. Always succeed, and can be called
|
||
|
from any context (the function never sleeps). Note: code should _never_
|
||
|
unlock an hwspinlock which is already unlocked (there is no protection
|
||
|
against this).
|
||
|
|
||
|
void hwspin_unlock_irq(struct hwspinlock *hwlock);
|
||
|
- unlock a previously-locked hwspinlock and enable local interrupts.
|
||
|
The caller should _never_ unlock an hwspinlock which is already unlocked.
|
||
|
Doing so is considered a bug (there is no protection against this).
|
||
|
Upon a successful return from this function, preemption and local
|
||
|
interrupts are enabled. This function will never sleep.
|
||
|
|
||
|
void
|
||
|
hwspin_unlock_irqrestore(struct hwspinlock *hwlock, unsigned long *flags);
|
||
|
- unlock a previously-locked hwspinlock.
|
||
|
The caller should _never_ unlock an hwspinlock which is already unlocked.
|
||
|
Doing so is considered a bug (there is no protection against this).
|
||
|
Upon a successful return from this function, preemption is reenabled,
|
||
|
and the state of the local interrupts is restored to the state saved at
|
||
|
the given flags. This function will never sleep.
|
||
|
|
||
|
int hwspin_lock_get_id(struct hwspinlock *hwlock);
|
||
|
- retrieve id number of a given hwspinlock. This is needed when an
|
||
|
hwspinlock is dynamically assigned: before it can be used to achieve
|
||
|
mutual exclusion with a remote cpu, the id number should be communicated
|
||
|
to the remote task with which we want to synchronize.
|
||
|
Returns the hwspinlock id number, or -EINVAL if hwlock is null.
|
||
|
|
||
|
3. Typical usage
|
||
|
|
||
|
#include <linux/hwspinlock.h>
|
||
|
#include <linux/err.h>
|
||
|
|
||
|
int hwspinlock_example1(void)
|
||
|
{
|
||
|
struct hwspinlock *hwlock;
|
||
|
int ret;
|
||
|
|
||
|
/* dynamically assign a hwspinlock */
|
||
|
hwlock = hwspin_lock_request();
|
||
|
if (!hwlock)
|
||
|
...
|
||
|
|
||
|
id = hwspin_lock_get_id(hwlock);
|
||
|
/* probably need to communicate id to a remote processor now */
|
||
|
|
||
|
/* take the lock, spin for 1 sec if it's already taken */
|
||
|
ret = hwspin_lock_timeout(hwlock, 1000);
|
||
|
if (ret)
|
||
|
...
|
||
|
|
||
|
/*
|
||
|
* we took the lock, do our thing now, but do NOT sleep
|
||
|
*/
|
||
|
|
||
|
/* release the lock */
|
||
|
hwspin_unlock(hwlock);
|
||
|
|
||
|
/* free the lock */
|
||
|
ret = hwspin_lock_free(hwlock);
|
||
|
if (ret)
|
||
|
...
|
||
|
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
int hwspinlock_example2(void)
|
||
|
{
|
||
|
struct hwspinlock *hwlock;
|
||
|
int ret;
|
||
|
|
||
|
/*
|
||
|
* assign a specific hwspinlock id - this should be called early
|
||
|
* by board init code.
|
||
|
*/
|
||
|
hwlock = hwspin_lock_request_specific(PREDEFINED_LOCK_ID);
|
||
|
if (!hwlock)
|
||
|
...
|
||
|
|
||
|
/* try to take it, but don't spin on it */
|
||
|
ret = hwspin_trylock(hwlock);
|
||
|
if (!ret) {
|
||
|
pr_info("lock is already taken\n");
|
||
|
return -EBUSY;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* we took the lock, do our thing now, but do NOT sleep
|
||
|
*/
|
||
|
|
||
|
/* release the lock */
|
||
|
hwspin_unlock(hwlock);
|
||
|
|
||
|
/* free the lock */
|
||
|
ret = hwspin_lock_free(hwlock);
|
||
|
if (ret)
|
||
|
...
|
||
|
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
|
||
|
4. API for implementors
|
||
|
|
||
|
int hwspin_lock_register(struct hwspinlock *hwlock);
|
||
|
- to be called from the underlying platform-specific implementation, in
|
||
|
order to register a new hwspinlock instance. Can be called from an atomic
|
||
|
context (this function will not sleep) but not from within interrupt
|
||
|
context. Returns 0 on success, or appropriate error code on failure.
|
||
|
|
||
|
struct hwspinlock *hwspin_lock_unregister(unsigned int id);
|
||
|
- to be called from the underlying vendor-specific implementation, in order
|
||
|
to unregister an existing (and unused) hwspinlock instance.
|
||
|
Can be called from an atomic context (will not sleep) but not from
|
||
|
within interrupt context.
|
||
|
Returns the address of hwspinlock on success, or NULL on error (e.g.
|
||
|
if the hwspinlock is sill in use).
|
||
|
|
||
|
5. struct hwspinlock
|
||
|
|
||
|
This struct represents an hwspinlock instance. It is registered by the
|
||
|
underlying hwspinlock implementation using the hwspin_lock_register() API.
|
||
|
|
||
|
/**
|
||
|
* struct hwspinlock - vendor-specific hwspinlock implementation
|
||
|
*
|
||
|
* @dev: underlying device, will be used with runtime PM api
|
||
|
* @ops: vendor-specific hwspinlock handlers
|
||
|
* @id: a global, unique, system-wide, index of the lock.
|
||
|
* @lock: initialized and used by hwspinlock core
|
||
|
* @owner: underlying implementation module, used to maintain module ref count
|
||
|
*/
|
||
|
struct hwspinlock {
|
||
|
struct device *dev;
|
||
|
const struct hwspinlock_ops *ops;
|
||
|
int id;
|
||
|
spinlock_t lock;
|
||
|
struct module *owner;
|
||
|
};
|
||
|
|
||
|
The underlying implementation is responsible to assign the dev, ops, id and
|
||
|
owner members. The lock member, OTOH, is initialized and used by the hwspinlock
|
||
|
core.
|
||
|
|
||
|
6. Implementation callbacks
|
||
|
|
||
|
There are three possible callbacks defined in 'struct hwspinlock_ops':
|
||
|
|
||
|
struct hwspinlock_ops {
|
||
|
int (*trylock)(struct hwspinlock *lock);
|
||
|
void (*unlock)(struct hwspinlock *lock);
|
||
|
void (*relax)(struct hwspinlock *lock);
|
||
|
};
|
||
|
|
||
|
The first two callbacks are mandatory:
|
||
|
|
||
|
The ->trylock() callback should make a single attempt to take the lock, and
|
||
|
return 0 on failure and 1 on success. This callback may _not_ sleep.
|
||
|
|
||
|
The ->unlock() callback releases the lock. It always succeed, and it, too,
|
||
|
may _not_ sleep.
|
||
|
|
||
|
The ->relax() callback is optional. It is called by hwspinlock core while
|
||
|
spinning on a lock, and can be used by the underlying implementation to force
|
||
|
a delay between two successive invocations of ->trylock(). It may _not_ sleep.
|