| =========================== |
| Hardware Spinlock Framework |
| =========================== |
| |
| Introduction |
| ============ |
| |
| Hardware spinlock modules provide hardware assistance for synchronization |
| and mutual exclusion between heterogeneous processors and those not operating |
| under a single, shared operating system. |
| |
| For example, OMAP4 has dual Cortex-A9, dual Cortex-M3 and a C64x+ DSP, |
| each of which is running a different Operating System (the master, A9, |
| is usually running Linux and the slave processors, the M3 and the DSP, |
| are running some flavor of RTOS). |
| |
| A generic hwspinlock framework allows platform-independent drivers to use |
| the hwspinlock device in order to access data structures that are shared |
| between remote processors, that otherwise have no alternative mechanism |
| to accomplish synchronization and mutual exclusion operations. |
| |
| This is necessary, for example, for Inter-processor communications: |
| on OMAP4, cpu-intensive multimedia tasks are offloaded by the host to the |
| remote M3 and/or C64x+ slave processors (by an IPC subsystem called Syslink). |
| |
| To achieve fast message-based communications, a minimal kernel support |
| is needed to deliver messages arriving from a remote processor to the |
| appropriate user process. |
| |
| This communication is based on simple data structures that is shared between |
| the remote processors, and access to it is synchronized using the hwspinlock |
| module (remote processor directly places new messages in this shared data |
| structure). |
| |
| A common hwspinlock interface makes it possible to have generic, platform- |
| independent, drivers. |
| |
| User API |
| ======== |
| |
| :: |
| |
| struct hwspinlock *hwspin_lock_request(void); |
| |
| Dynamically assign an hwspinlock and return its address, or NULL |
| in case an unused hwspinlock isn't available. Users of this |
| API will usually want to communicate the lock's id to the remote core |
| before it can be used to achieve synchronization. |
| |
| Should be called from a process context (might sleep). |
| |
| :: |
| |
| struct hwspinlock *hwspin_lock_request_specific(unsigned int id); |
| |
| Assign a specific hwspinlock id and return its address, or NULL |
| if that hwspinlock is already in use. Usually board code will |
| be calling this function in order to reserve specific hwspinlock |
| ids for predefined purposes. |
| |
| Should be called from a process context (might sleep). |
| |
| :: |
| |
| int of_hwspin_lock_get_id(struct device_node *np, int index); |
| |
| Retrieve the global lock id for an OF phandle-based specific lock. |
| This function provides a means for DT users of a hwspinlock module |
| to get the global lock id of a specific hwspinlock, so that it can |
| be requested using the normal hwspin_lock_request_specific() API. |
| |
| The function returns a lock id number on success, -EPROBE_DEFER if |
| the hwspinlock device is not yet registered with the core, or other |
| error values. |
| |
| Should be called from a process context (might sleep). |
| |
| :: |
| |
| int hwspin_lock_free(struct hwspinlock *hwlock); |
| |
| Free a previously-assigned hwspinlock; returns 0 on success, or an |
| appropriate error code on failure (e.g. -EINVAL if the hwspinlock |
| is already free). |
| |
| Should be called from a process context (might sleep). |
| |
| :: |
| |
| int hwspin_lock_bust(struct hwspinlock *hwlock, unsigned int id); |
| |
| After verifying the owner of the hwspinlock, release a previously acquired |
| hwspinlock; returns 0 on success, or an appropriate error code on failure |
| (e.g. -EOPNOTSUPP if the bust operation is not defined for the specific |
| hwspinlock). |
| |
| Should be called from a process context (might sleep). |
| |
| :: |
| |
| int hwspin_lock_timeout(struct hwspinlock *hwlock, unsigned int timeout); |
| |
| Lock a previously-assigned hwspinlock with a timeout limit (specified in |
| msecs). If the hwspinlock is already taken, the function will busy loop |
| waiting for it to be released, but give up when the timeout elapses. |
| Upon a successful return from this function, preemption is disabled so |
| the caller must not sleep, and is advised to release the hwspinlock as |
| soon as possible, in order to minimize remote cores polling on the |
| hardware interconnect. |
| |
| Returns 0 when successful and an appropriate error code otherwise (most |
| notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs). |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_lock_timeout_irq(struct hwspinlock *hwlock, unsigned int timeout); |
| |
| Lock a previously-assigned hwspinlock with a timeout limit (specified in |
| msecs). If the hwspinlock is already taken, the function will busy loop |
| waiting for it to be released, but give up when the timeout elapses. |
| Upon a successful return from this function, preemption and the local |
| interrupts are disabled, so the caller must not sleep, and is advised to |
| release the hwspinlock as soon as possible. |
| |
| Returns 0 when successful and an appropriate error code otherwise (most |
| notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs). |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_lock_timeout_irqsave(struct hwspinlock *hwlock, unsigned int to, |
| unsigned long *flags); |
| |
| Lock a previously-assigned hwspinlock with a timeout limit (specified in |
| msecs). If the hwspinlock is already taken, the function will busy loop |
| waiting for it to be released, but give up when the timeout elapses. |
| Upon a successful return from this function, preemption is disabled, |
| local interrupts are disabled and their previous state is saved at the |
| given flags placeholder. The caller must not sleep, and is advised to |
| release the hwspinlock as soon as possible. |
| |
| Returns 0 when successful and an appropriate error code otherwise (most |
| notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs). |
| |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_lock_timeout_raw(struct hwspinlock *hwlock, unsigned int timeout); |
| |
| Lock a previously-assigned hwspinlock with a timeout limit (specified in |
| msecs). If the hwspinlock is already taken, the function will busy loop |
| waiting for it to be released, but give up when the timeout elapses. |
| |
| Caution: User must protect the routine of getting hardware lock with mutex |
| or spinlock to avoid dead-lock, that will let user can do some time-consuming |
| or sleepable operations under the hardware lock. |
| |
| Returns 0 when successful and an appropriate error code otherwise (most |
| notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs). |
| |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_lock_timeout_in_atomic(struct hwspinlock *hwlock, unsigned int to); |
| |
| Lock a previously-assigned hwspinlock with a timeout limit (specified in |
| msecs). If the hwspinlock is already taken, the function will busy loop |
| waiting for it to be released, but give up when the timeout elapses. |
| |
| This function shall be called only from an atomic context and the timeout |
| value shall not exceed a few msecs. |
| |
| Returns 0 when successful and an appropriate error code otherwise (most |
| notably -ETIMEDOUT if the hwspinlock is still busy after timeout msecs). |
| |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_trylock(struct hwspinlock *hwlock); |
| |
| |
| Attempt to lock a previously-assigned hwspinlock, but immediately fail if |
| it is already taken. |
| |
| Upon a successful return from this function, preemption is disabled so |
| caller must not sleep, and is advised to release the hwspinlock as soon as |
| possible, in order to minimize remote cores polling on the hardware |
| interconnect. |
| |
| Returns 0 on success and an appropriate error code otherwise (most |
| notably -EBUSY if the hwspinlock was already taken). |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_trylock_irq(struct hwspinlock *hwlock); |
| |
| |
| Attempt to lock a previously-assigned hwspinlock, but immediately fail if |
| it is already taken. |
| |
| Upon a successful return from this function, preemption and the local |
| interrupts are disabled so caller must not sleep, and is advised to |
| release the hwspinlock as soon as possible. |
| |
| Returns 0 on success and an appropriate error code otherwise (most |
| notably -EBUSY if the hwspinlock was already taken). |
| |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_trylock_irqsave(struct hwspinlock *hwlock, unsigned long *flags); |
| |
| Attempt to lock a previously-assigned hwspinlock, but immediately fail if |
| it is already taken. |
| |
| Upon a successful return from this function, preemption is disabled, |
| the local interrupts are disabled and their previous state is saved |
| at the given flags placeholder. The caller must not sleep, and is advised |
| to release the hwspinlock as soon as possible. |
| |
| Returns 0 on success and an appropriate error code otherwise (most |
| notably -EBUSY if the hwspinlock was already taken). |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_trylock_raw(struct hwspinlock *hwlock); |
| |
| Attempt to lock a previously-assigned hwspinlock, but immediately fail if |
| it is already taken. |
| |
| Caution: User must protect the routine of getting hardware lock with mutex |
| or spinlock to avoid dead-lock, that will let user can do some time-consuming |
| or sleepable operations under the hardware lock. |
| |
| Returns 0 on success and an appropriate error code otherwise (most |
| notably -EBUSY if the hwspinlock was already taken). |
| The function will never sleep. |
| |
| :: |
| |
| int hwspin_trylock_in_atomic(struct hwspinlock *hwlock); |
| |
| Attempt to lock a previously-assigned hwspinlock, but immediately fail if |
| it is already taken. |
| |
| This function shall be called only from an atomic context. |
| |
| Returns 0 on success and an appropriate error code otherwise (most |
| notably -EBUSY if the hwspinlock was already taken). |
| The function will never sleep. |
| |
| :: |
| |
| void hwspin_unlock(struct hwspinlock *hwlock); |
| |
| Unlock a previously-locked hwspinlock. Always succeed, and can be called |
| from any context (the function never sleeps). |
| |
| .. note:: |
| |
| code should **never** unlock an hwspinlock which is already unlocked |
| (there is no protection against this). |
| |
| :: |
| |
| void hwspin_unlock_irq(struct hwspinlock *hwlock); |
| |
| Unlock a previously-locked hwspinlock and enable local interrupts. |
| The caller should **never** unlock an hwspinlock which is already unlocked. |
| |
| Doing so is considered a bug (there is no protection against this). |
| Upon a successful return from this function, preemption and local |
| interrupts are enabled. This function will never sleep. |
| |
| :: |
| |
| void |
| hwspin_unlock_irqrestore(struct hwspinlock *hwlock, unsigned long *flags); |
| |
| Unlock a previously-locked hwspinlock. |
| |
| The caller should **never** unlock an hwspinlock which is already unlocked. |
| Doing so is considered a bug (there is no protection against this). |
| Upon a successful return from this function, preemption is reenabled, |
| and the state of the local interrupts is restored to the state saved at |
| the given flags. This function will never sleep. |
| |
| :: |
| |
| void hwspin_unlock_raw(struct hwspinlock *hwlock); |
| |
| Unlock a previously-locked hwspinlock. |
| |
| The caller should **never** unlock an hwspinlock which is already unlocked. |
| Doing so is considered a bug (there is no protection against this). |
| This function will never sleep. |
| |
| :: |
| |
| void hwspin_unlock_in_atomic(struct hwspinlock *hwlock); |
| |
| Unlock a previously-locked hwspinlock. |
| |
| The caller should **never** unlock an hwspinlock which is already unlocked. |
| Doing so is considered a bug (there is no protection against this). |
| This function will never sleep. |
| |
| :: |
| |
| int hwspin_lock_get_id(struct hwspinlock *hwlock); |
| |
| Retrieve id number of a given hwspinlock. This is needed when an |
| hwspinlock is dynamically assigned: before it can be used to achieve |
| mutual exclusion with a remote cpu, the id number should be communicated |
| to the remote task with which we want to synchronize. |
| |
| Returns the hwspinlock id number, or -EINVAL if hwlock is null. |
| |
| Typical usage |
| ============= |
| |
| :: |
| |
| #include <linux/hwspinlock.h> |
| #include <linux/err.h> |
| |
| int hwspinlock_example1(void) |
| { |
| struct hwspinlock *hwlock; |
| int ret; |
| |
| /* dynamically assign a hwspinlock */ |
| hwlock = hwspin_lock_request(); |
| if (!hwlock) |
| ... |
| |
| id = hwspin_lock_get_id(hwlock); |
| /* probably need to communicate id to a remote processor now */ |
| |
| /* take the lock, spin for 1 sec if it's already taken */ |
| ret = hwspin_lock_timeout(hwlock, 1000); |
| if (ret) |
| ... |
| |
| /* |
| * we took the lock, do our thing now, but do NOT sleep |
| */ |
| |
| /* release the lock */ |
| hwspin_unlock(hwlock); |
| |
| /* free the lock */ |
| ret = hwspin_lock_free(hwlock); |
| if (ret) |
| ... |
| |
| return ret; |
| } |
| |
| int hwspinlock_example2(void) |
| { |
| struct hwspinlock *hwlock; |
| int ret; |
| |
| /* |
| * assign a specific hwspinlock id - this should be called early |
| * by board init code. |
| */ |
| hwlock = hwspin_lock_request_specific(PREDEFINED_LOCK_ID); |
| if (!hwlock) |
| ... |
| |
| /* try to take it, but don't spin on it */ |
| ret = hwspin_trylock(hwlock); |
| if (!ret) { |
| pr_info("lock is already taken\n"); |
| return -EBUSY; |
| } |
| |
| /* |
| * we took the lock, do our thing now, but do NOT sleep |
| */ |
| |
| /* release the lock */ |
| hwspin_unlock(hwlock); |
| |
| /* free the lock */ |
| ret = hwspin_lock_free(hwlock); |
| if (ret) |
| ... |
| |
| return ret; |
| } |
| |
| |
| API for implementors |
| ==================== |
| |
| :: |
| |
| int hwspin_lock_register(struct hwspinlock_device *bank, struct device *dev, |
| const struct hwspinlock_ops *ops, int base_id, int num_locks); |
| |
| To be called from the underlying platform-specific implementation, in |
| order to register a new hwspinlock device (which is usually a bank of |
| numerous locks). Should be called from a process context (this function |
| might sleep). |
| |
| Returns 0 on success, or appropriate error code on failure. |
| |
| :: |
| |
| int hwspin_lock_unregister(struct hwspinlock_device *bank); |
| |
| To be called from the underlying vendor-specific implementation, in order |
| to unregister an hwspinlock device (which is usually a bank of numerous |
| locks). |
| |
| Should be called from a process context (this function might sleep). |
| |
| Returns the address of hwspinlock on success, or NULL on error (e.g. |
| if the hwspinlock is still in use). |
| |
| Important structs |
| ================= |
| |
| struct hwspinlock_device is a device which usually contains a bank |
| of hardware locks. It is registered by the underlying hwspinlock |
| implementation using the hwspin_lock_register() API. |
| |
| :: |
| |
| /** |
| * struct hwspinlock_device - a device which usually spans numerous hwspinlocks |
| * @dev: underlying device, will be used to invoke runtime PM api |
| * @ops: platform-specific hwspinlock handlers |
| * @base_id: id index of the first lock in this device |
| * @num_locks: number of locks in this device |
| * @lock: dynamically allocated array of 'struct hwspinlock' |
| */ |
| struct hwspinlock_device { |
| struct device *dev; |
| const struct hwspinlock_ops *ops; |
| int base_id; |
| int num_locks; |
| struct hwspinlock lock[0]; |
| }; |
| |
| struct hwspinlock_device contains an array of hwspinlock structs, each |
| of which represents a single hardware lock:: |
| |
| /** |
| * struct hwspinlock - this struct represents a single hwspinlock instance |
| * @bank: the hwspinlock_device structure which owns this lock |
| * @lock: initialized and used by hwspinlock core |
| * @priv: private data, owned by the underlying platform-specific hwspinlock drv |
| */ |
| struct hwspinlock { |
| struct hwspinlock_device *bank; |
| spinlock_t lock; |
| void *priv; |
| }; |
| |
| When registering a bank of locks, the hwspinlock driver only needs to |
| set the priv members of the locks. The rest of the members are set and |
| initialized by the hwspinlock core itself. |
| |
| Implementation callbacks |
| ======================== |
| |
| There are three possible callbacks defined in 'struct hwspinlock_ops':: |
| |
| struct hwspinlock_ops { |
| int (*trylock)(struct hwspinlock *lock); |
| void (*unlock)(struct hwspinlock *lock); |
| void (*relax)(struct hwspinlock *lock); |
| }; |
| |
| The first two callbacks are mandatory: |
| |
| The ->trylock() callback should make a single attempt to take the lock, and |
| return 0 on failure and 1 on success. This callback may **not** sleep. |
| |
| The ->unlock() callback releases the lock. It always succeed, and it, too, |
| may **not** sleep. |
| |
| The ->relax() callback is optional. It is called by hwspinlock core while |
| spinning on a lock, and can be used by the underlying implementation to force |
| a delay between two successive invocations of ->trylock(). It may **not** sleep. |