mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-16 15:34:48 +08:00
5e66a0cb5f
Replace the existing /* fall through */ comments and its variants with the new pseudo-keyword macro fallthrough[1]. Also, remove unnecessary fall-through markings when it is the case. [1] https://www.kernel.org/doc/html/v5.7/process/deprecated.html?highlight=fallthrough#implicit-switch-case-fall-through Signed-off-by: Gustavo A. R. Silva <gustavoars@kernel.org> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://lore.kernel.org/r/20200727224201.GA10133@embeddedor
291 lines
7.2 KiB
C
291 lines
7.2 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* PowerNV OPAL asynchronous completion interfaces
|
|
*
|
|
* Copyright 2013-2017 IBM Corp.
|
|
*/
|
|
|
|
#undef DEBUG
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/init.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/semaphore.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/gfp.h>
|
|
#include <linux/of.h>
|
|
#include <asm/machdep.h>
|
|
#include <asm/opal.h>
|
|
|
|
enum opal_async_token_state {
|
|
ASYNC_TOKEN_UNALLOCATED = 0,
|
|
ASYNC_TOKEN_ALLOCATED,
|
|
ASYNC_TOKEN_DISPATCHED,
|
|
ASYNC_TOKEN_ABANDONED,
|
|
ASYNC_TOKEN_COMPLETED
|
|
};
|
|
|
|
struct opal_async_token {
|
|
enum opal_async_token_state state;
|
|
struct opal_msg response;
|
|
};
|
|
|
|
static DECLARE_WAIT_QUEUE_HEAD(opal_async_wait);
|
|
static DEFINE_SPINLOCK(opal_async_comp_lock);
|
|
static struct semaphore opal_async_sem;
|
|
static unsigned int opal_max_async_tokens;
|
|
static struct opal_async_token *opal_async_tokens;
|
|
|
|
static int __opal_async_get_token(void)
|
|
{
|
|
unsigned long flags;
|
|
int i, token = -EBUSY;
|
|
|
|
spin_lock_irqsave(&opal_async_comp_lock, flags);
|
|
|
|
for (i = 0; i < opal_max_async_tokens; i++) {
|
|
if (opal_async_tokens[i].state == ASYNC_TOKEN_UNALLOCATED) {
|
|
opal_async_tokens[i].state = ASYNC_TOKEN_ALLOCATED;
|
|
token = i;
|
|
break;
|
|
}
|
|
}
|
|
|
|
spin_unlock_irqrestore(&opal_async_comp_lock, flags);
|
|
return token;
|
|
}
|
|
|
|
/*
|
|
* Note: If the returned token is used in an opal call and opal returns
|
|
* OPAL_ASYNC_COMPLETION you MUST call one of opal_async_wait_response() or
|
|
* opal_async_wait_response_interruptible() at least once before calling another
|
|
* opal_async_* function
|
|
*/
|
|
int opal_async_get_token_interruptible(void)
|
|
{
|
|
int token;
|
|
|
|
/* Wait until a token is available */
|
|
if (down_interruptible(&opal_async_sem))
|
|
return -ERESTARTSYS;
|
|
|
|
token = __opal_async_get_token();
|
|
if (token < 0)
|
|
up(&opal_async_sem);
|
|
|
|
return token;
|
|
}
|
|
EXPORT_SYMBOL_GPL(opal_async_get_token_interruptible);
|
|
|
|
static int __opal_async_release_token(int token)
|
|
{
|
|
unsigned long flags;
|
|
int rc;
|
|
|
|
if (token < 0 || token >= opal_max_async_tokens) {
|
|
pr_err("%s: Passed token is out of range, token %d\n",
|
|
__func__, token);
|
|
return -EINVAL;
|
|
}
|
|
|
|
spin_lock_irqsave(&opal_async_comp_lock, flags);
|
|
switch (opal_async_tokens[token].state) {
|
|
case ASYNC_TOKEN_COMPLETED:
|
|
case ASYNC_TOKEN_ALLOCATED:
|
|
opal_async_tokens[token].state = ASYNC_TOKEN_UNALLOCATED;
|
|
rc = 0;
|
|
break;
|
|
/*
|
|
* DISPATCHED and ABANDONED tokens must wait for OPAL to respond.
|
|
* Mark a DISPATCHED token as ABANDONED so that the response handling
|
|
* code knows no one cares and that it can free it then.
|
|
*/
|
|
case ASYNC_TOKEN_DISPATCHED:
|
|
opal_async_tokens[token].state = ASYNC_TOKEN_ABANDONED;
|
|
fallthrough;
|
|
default:
|
|
rc = 1;
|
|
}
|
|
spin_unlock_irqrestore(&opal_async_comp_lock, flags);
|
|
|
|
return rc;
|
|
}
|
|
|
|
int opal_async_release_token(int token)
|
|
{
|
|
int ret;
|
|
|
|
ret = __opal_async_release_token(token);
|
|
if (!ret)
|
|
up(&opal_async_sem);
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_GPL(opal_async_release_token);
|
|
|
|
int opal_async_wait_response(uint64_t token, struct opal_msg *msg)
|
|
{
|
|
if (token >= opal_max_async_tokens) {
|
|
pr_err("%s: Invalid token passed\n", __func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!msg) {
|
|
pr_err("%s: Invalid message pointer passed\n", __func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/*
|
|
* There is no need to mark the token as dispatched, wait_event()
|
|
* will block until the token completes.
|
|
*
|
|
* Wakeup the poller before we wait for events to speed things
|
|
* up on platforms or simulators where the interrupts aren't
|
|
* functional.
|
|
*/
|
|
opal_wake_poller();
|
|
wait_event(opal_async_wait, opal_async_tokens[token].state
|
|
== ASYNC_TOKEN_COMPLETED);
|
|
memcpy(msg, &opal_async_tokens[token].response, sizeof(*msg));
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(opal_async_wait_response);
|
|
|
|
int opal_async_wait_response_interruptible(uint64_t token, struct opal_msg *msg)
|
|
{
|
|
unsigned long flags;
|
|
int ret;
|
|
|
|
if (token >= opal_max_async_tokens) {
|
|
pr_err("%s: Invalid token passed\n", __func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!msg) {
|
|
pr_err("%s: Invalid message pointer passed\n", __func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/*
|
|
* The first time this gets called we mark the token as DISPATCHED
|
|
* so that if wait_event_interruptible() returns not zero and the
|
|
* caller frees the token, we know not to actually free the token
|
|
* until the response comes.
|
|
*
|
|
* Only change if the token is ALLOCATED - it may have been
|
|
* completed even before the caller gets around to calling this
|
|
* the first time.
|
|
*
|
|
* There is also a dirty great comment at the token allocation
|
|
* function that if the opal call returns OPAL_ASYNC_COMPLETION to
|
|
* the caller then the caller *must* call this or the not
|
|
* interruptible version before doing anything else with the
|
|
* token.
|
|
*/
|
|
if (opal_async_tokens[token].state == ASYNC_TOKEN_ALLOCATED) {
|
|
spin_lock_irqsave(&opal_async_comp_lock, flags);
|
|
if (opal_async_tokens[token].state == ASYNC_TOKEN_ALLOCATED)
|
|
opal_async_tokens[token].state = ASYNC_TOKEN_DISPATCHED;
|
|
spin_unlock_irqrestore(&opal_async_comp_lock, flags);
|
|
}
|
|
|
|
/*
|
|
* Wakeup the poller before we wait for events to speed things
|
|
* up on platforms or simulators where the interrupts aren't
|
|
* functional.
|
|
*/
|
|
opal_wake_poller();
|
|
ret = wait_event_interruptible(opal_async_wait,
|
|
opal_async_tokens[token].state ==
|
|
ASYNC_TOKEN_COMPLETED);
|
|
if (!ret)
|
|
memcpy(msg, &opal_async_tokens[token].response, sizeof(*msg));
|
|
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_GPL(opal_async_wait_response_interruptible);
|
|
|
|
/* Called from interrupt context */
|
|
static int opal_async_comp_event(struct notifier_block *nb,
|
|
unsigned long msg_type, void *msg)
|
|
{
|
|
struct opal_msg *comp_msg = msg;
|
|
enum opal_async_token_state state;
|
|
unsigned long flags;
|
|
uint64_t token;
|
|
|
|
if (msg_type != OPAL_MSG_ASYNC_COMP)
|
|
return 0;
|
|
|
|
token = be64_to_cpu(comp_msg->params[0]);
|
|
spin_lock_irqsave(&opal_async_comp_lock, flags);
|
|
state = opal_async_tokens[token].state;
|
|
opal_async_tokens[token].state = ASYNC_TOKEN_COMPLETED;
|
|
spin_unlock_irqrestore(&opal_async_comp_lock, flags);
|
|
|
|
if (state == ASYNC_TOKEN_ABANDONED) {
|
|
/* Free the token, no one else will */
|
|
opal_async_release_token(token);
|
|
return 0;
|
|
}
|
|
memcpy(&opal_async_tokens[token].response, comp_msg, sizeof(*comp_msg));
|
|
wake_up(&opal_async_wait);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct notifier_block opal_async_comp_nb = {
|
|
.notifier_call = opal_async_comp_event,
|
|
.next = NULL,
|
|
.priority = 0,
|
|
};
|
|
|
|
int __init opal_async_comp_init(void)
|
|
{
|
|
struct device_node *opal_node;
|
|
const __be32 *async;
|
|
int err;
|
|
|
|
opal_node = of_find_node_by_path("/ibm,opal");
|
|
if (!opal_node) {
|
|
pr_err("%s: Opal node not found\n", __func__);
|
|
err = -ENOENT;
|
|
goto out;
|
|
}
|
|
|
|
async = of_get_property(opal_node, "opal-msg-async-num", NULL);
|
|
if (!async) {
|
|
pr_err("%s: %pOF has no opal-msg-async-num\n",
|
|
__func__, opal_node);
|
|
err = -ENOENT;
|
|
goto out_opal_node;
|
|
}
|
|
|
|
opal_max_async_tokens = be32_to_cpup(async);
|
|
opal_async_tokens = kcalloc(opal_max_async_tokens,
|
|
sizeof(*opal_async_tokens), GFP_KERNEL);
|
|
if (!opal_async_tokens) {
|
|
err = -ENOMEM;
|
|
goto out_opal_node;
|
|
}
|
|
|
|
err = opal_message_notifier_register(OPAL_MSG_ASYNC_COMP,
|
|
&opal_async_comp_nb);
|
|
if (err) {
|
|
pr_err("%s: Can't register OPAL event notifier (%d)\n",
|
|
__func__, err);
|
|
kfree(opal_async_tokens);
|
|
goto out_opal_node;
|
|
}
|
|
|
|
sema_init(&opal_async_sem, opal_max_async_tokens);
|
|
|
|
out_opal_node:
|
|
of_node_put(opal_node);
|
|
out:
|
|
return err;
|
|
}
|