mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-02 16:44:10 +08:00
b5a9d61eeb
When the computer is turned off, all the processes are killed and then all the filesystems are umounted. OrangeFS should not wait for the userspace daemon to come back in that case. This only works for plain umount(2). To actually take advantage of this interactively, `umount -f' is needed; otherwise umount will issue a statfs first, which will wait for the userspace daemon to come back. Signed-off-by: Martin Brandenburg <martin@omnibond.com> Signed-off-by: Mike Marshall <hubcap@omnibond.com>
365 lines
10 KiB
C
365 lines
10 KiB
C
/*
|
|
* (C) 2001 Clemson University and The University of Chicago
|
|
* (C) 2011 Omnibond Systems
|
|
*
|
|
* Changes by Acxiom Corporation to implement generic service_operation()
|
|
* function, Copyright Acxiom Corporation, 2005.
|
|
*
|
|
* See COPYING in top-level directory.
|
|
*/
|
|
|
|
/*
|
|
* In-kernel waitqueue operations.
|
|
*/
|
|
|
|
#include "protocol.h"
|
|
#include "orangefs-kernel.h"
|
|
#include "orangefs-bufmap.h"
|
|
|
|
static int wait_for_matching_downcall(struct orangefs_kernel_op_s *, long, bool);
|
|
static void orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s *);
|
|
|
|
/*
|
|
* What we do in this function is to walk the list of operations that are
|
|
* present in the request queue and mark them as purged.
|
|
* NOTE: This is called from the device close after client-core has
|
|
* guaranteed that no new operations could appear on the list since the
|
|
* client-core is anyway going to exit.
|
|
*/
|
|
void purge_waiting_ops(void)
|
|
{
|
|
struct orangefs_kernel_op_s *op;
|
|
|
|
spin_lock(&orangefs_request_list_lock);
|
|
list_for_each_entry(op, &orangefs_request_list, list) {
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"pvfs2-client-core: purging op tag %llu %s\n",
|
|
llu(op->tag),
|
|
get_opname_string(op));
|
|
set_op_state_purged(op);
|
|
gossip_debug(GOSSIP_DEV_DEBUG,
|
|
"%s: op:%s: op_state:%d: process:%s:\n",
|
|
__func__,
|
|
get_opname_string(op),
|
|
op->op_state,
|
|
current->comm);
|
|
}
|
|
spin_unlock(&orangefs_request_list_lock);
|
|
}
|
|
|
|
/*
|
|
* submits a ORANGEFS operation and waits for it to complete
|
|
*
|
|
* Note op->downcall.status will contain the status of the operation (in
|
|
* errno format), whether provided by pvfs2-client or a result of failure to
|
|
* service the operation. If the caller wishes to distinguish, then
|
|
* op->state can be checked to see if it was serviced or not.
|
|
*
|
|
* Returns contents of op->downcall.status for convenience
|
|
*/
|
|
int service_operation(struct orangefs_kernel_op_s *op,
|
|
const char *op_name,
|
|
int flags)
|
|
{
|
|
long timeout = MAX_SCHEDULE_TIMEOUT;
|
|
int ret = 0;
|
|
|
|
DEFINE_WAIT(wait_entry);
|
|
|
|
op->upcall.tgid = current->tgid;
|
|
op->upcall.pid = current->pid;
|
|
|
|
retry_servicing:
|
|
op->downcall.status = 0;
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s: %s op:%p: process:%s: pid:%d:\n",
|
|
__func__,
|
|
op_name,
|
|
op,
|
|
current->comm,
|
|
current->pid);
|
|
|
|
/*
|
|
* If ORANGEFS_OP_NO_MUTEX was set in flags, we need to avoid
|
|
* acquiring the request_mutex because we're servicing a
|
|
* high priority remount operation and the request_mutex is
|
|
* already taken.
|
|
*/
|
|
if (!(flags & ORANGEFS_OP_NO_MUTEX)) {
|
|
if (flags & ORANGEFS_OP_INTERRUPTIBLE)
|
|
ret = mutex_lock_interruptible(&orangefs_request_mutex);
|
|
else
|
|
ret = mutex_lock_killable(&orangefs_request_mutex);
|
|
/*
|
|
* check to see if we were interrupted while waiting for
|
|
* mutex
|
|
*/
|
|
if (ret < 0) {
|
|
op->downcall.status = ret;
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s: service_operation interrupted.\n",
|
|
__func__);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
/* queue up the operation */
|
|
spin_lock(&orangefs_request_list_lock);
|
|
spin_lock(&op->lock);
|
|
set_op_state_waiting(op);
|
|
gossip_debug(GOSSIP_DEV_DEBUG,
|
|
"%s: op:%s: op_state:%d: process:%s:\n",
|
|
__func__,
|
|
get_opname_string(op),
|
|
op->op_state,
|
|
current->comm);
|
|
/* add high priority remount op to the front of the line. */
|
|
if (flags & ORANGEFS_OP_PRIORITY)
|
|
list_add(&op->list, &orangefs_request_list);
|
|
else
|
|
list_add_tail(&op->list, &orangefs_request_list);
|
|
spin_unlock(&op->lock);
|
|
wake_up_interruptible(&orangefs_request_list_waitq);
|
|
if (!__is_daemon_in_service()) {
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s:client core is NOT in service.\n",
|
|
__func__);
|
|
/*
|
|
* Don't wait for the userspace component to return if
|
|
* the filesystem is being umounted anyway.
|
|
*/
|
|
if (op->upcall.type == ORANGEFS_VFS_OP_FS_UMOUNT)
|
|
timeout = 0;
|
|
else
|
|
timeout = op_timeout_secs * HZ;
|
|
}
|
|
spin_unlock(&orangefs_request_list_lock);
|
|
|
|
if (!(flags & ORANGEFS_OP_NO_MUTEX))
|
|
mutex_unlock(&orangefs_request_mutex);
|
|
|
|
ret = wait_for_matching_downcall(op, timeout,
|
|
flags & ORANGEFS_OP_INTERRUPTIBLE);
|
|
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s: wait_for_matching_downcall returned %d for %p\n",
|
|
__func__,
|
|
ret,
|
|
op);
|
|
|
|
/* got matching downcall; make sure status is in errno format */
|
|
if (!ret) {
|
|
spin_unlock(&op->lock);
|
|
op->downcall.status =
|
|
orangefs_normalize_to_errno(op->downcall.status);
|
|
ret = op->downcall.status;
|
|
goto out;
|
|
}
|
|
|
|
/* failed to get matching downcall */
|
|
if (ret == -ETIMEDOUT) {
|
|
gossip_err("%s: %s -- wait timed out; aborting attempt.\n",
|
|
__func__,
|
|
op_name);
|
|
}
|
|
|
|
/*
|
|
* remove a waiting op from the request list or
|
|
* remove an in-progress op from the in-progress list.
|
|
*/
|
|
orangefs_clean_up_interrupted_operation(op);
|
|
|
|
op->downcall.status = ret;
|
|
/* retry if operation has not been serviced and if requested */
|
|
if (ret == -EAGAIN) {
|
|
op->attempts++;
|
|
timeout = op_timeout_secs * HZ;
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"orangefs: tag %llu (%s)"
|
|
" -- operation to be retried (%d attempt)\n",
|
|
llu(op->tag),
|
|
op_name,
|
|
op->attempts);
|
|
|
|
/*
|
|
* io ops (ops that use the shared memory buffer) have
|
|
* to be returned to their caller for a retry. Other ops
|
|
* can just be recycled here.
|
|
*/
|
|
if (!op->uses_shared_memory)
|
|
goto retry_servicing;
|
|
}
|
|
|
|
out:
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s: %s returning: %d for %p.\n",
|
|
__func__,
|
|
op_name,
|
|
ret,
|
|
op);
|
|
return ret;
|
|
}
|
|
|
|
/* This can get called on an I/O op if it had a bad service_operation. */
|
|
bool orangefs_cancel_op_in_progress(struct orangefs_kernel_op_s *op)
|
|
{
|
|
u64 tag = op->tag;
|
|
if (!op_state_in_progress(op))
|
|
return false;
|
|
|
|
op->slot_to_free = op->upcall.req.io.buf_index;
|
|
memset(&op->upcall, 0, sizeof(op->upcall));
|
|
memset(&op->downcall, 0, sizeof(op->downcall));
|
|
op->upcall.type = ORANGEFS_VFS_OP_CANCEL;
|
|
op->upcall.req.cancel.op_tag = tag;
|
|
op->downcall.type = ORANGEFS_VFS_OP_INVALID;
|
|
op->downcall.status = -1;
|
|
orangefs_new_tag(op);
|
|
|
|
spin_lock(&orangefs_request_list_lock);
|
|
/* orangefs_request_list_lock is enough of a barrier here */
|
|
if (!__is_daemon_in_service()) {
|
|
spin_unlock(&orangefs_request_list_lock);
|
|
return false;
|
|
}
|
|
spin_lock(&op->lock);
|
|
set_op_state_waiting(op);
|
|
gossip_debug(GOSSIP_DEV_DEBUG,
|
|
"%s: op:%s: op_state:%d: process:%s:\n",
|
|
__func__,
|
|
get_opname_string(op),
|
|
op->op_state,
|
|
current->comm);
|
|
list_add(&op->list, &orangefs_request_list);
|
|
spin_unlock(&op->lock);
|
|
spin_unlock(&orangefs_request_list_lock);
|
|
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"Attempting ORANGEFS operation cancellation of tag %llu\n",
|
|
llu(tag));
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* Change an op to the "given up" state and remove it from its list.
|
|
*/
|
|
static void
|
|
orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s *op)
|
|
{
|
|
/*
|
|
* handle interrupted cases depending on what state we were in when
|
|
* the interruption is detected.
|
|
*
|
|
* Called with op->lock held.
|
|
*/
|
|
|
|
/*
|
|
* List manipulation code elsewhere will ignore ops that
|
|
* have been given up upon.
|
|
*/
|
|
op->op_state |= OP_VFS_STATE_GIVEN_UP;
|
|
|
|
if (list_empty(&op->list)) {
|
|
/* caught copying to/from daemon */
|
|
BUG_ON(op_state_serviced(op));
|
|
spin_unlock(&op->lock);
|
|
wait_for_completion(&op->waitq);
|
|
} else if (op_state_waiting(op)) {
|
|
/*
|
|
* upcall hasn't been read; remove op from upcall request
|
|
* list.
|
|
*/
|
|
spin_unlock(&op->lock);
|
|
spin_lock(&orangefs_request_list_lock);
|
|
list_del_init(&op->list);
|
|
spin_unlock(&orangefs_request_list_lock);
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"Interrupted: Removed op %p from request_list\n",
|
|
op);
|
|
} else if (op_state_in_progress(op)) {
|
|
/* op must be removed from the in progress htable */
|
|
spin_unlock(&op->lock);
|
|
spin_lock(&orangefs_htable_ops_in_progress_lock);
|
|
list_del_init(&op->list);
|
|
spin_unlock(&orangefs_htable_ops_in_progress_lock);
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"Interrupted: Removed op %p"
|
|
" from htable_ops_in_progress\n",
|
|
op);
|
|
} else {
|
|
spin_unlock(&op->lock);
|
|
gossip_err("interrupted operation is in a weird state 0x%x\n",
|
|
op->op_state);
|
|
}
|
|
reinit_completion(&op->waitq);
|
|
}
|
|
|
|
/*
|
|
* Sleeps on waitqueue waiting for matching downcall.
|
|
* If client-core finishes servicing, then we are good to go.
|
|
* else if client-core exits, we get woken up here, and retry with a timeout
|
|
*
|
|
* When this call returns to the caller, the specified op will no
|
|
* longer be in either the in_progress hash table or on the request list.
|
|
*
|
|
* Returns 0 on success and -errno on failure
|
|
* Errors are:
|
|
* EAGAIN in case we want the caller to requeue and try again..
|
|
* EINTR/EIO/ETIMEDOUT indicating we are done trying to service this
|
|
* operation since client-core seems to be exiting too often
|
|
* or if we were interrupted.
|
|
*
|
|
* Returns with op->lock taken.
|
|
*/
|
|
static int wait_for_matching_downcall(struct orangefs_kernel_op_s *op,
|
|
long timeout,
|
|
bool interruptible)
|
|
{
|
|
long n;
|
|
|
|
/*
|
|
* There's a "schedule_timeout" inside of these wait
|
|
* primitives, during which the op is out of the hands of the
|
|
* user process that needs something done and is being
|
|
* manipulated by the client-core process.
|
|
*/
|
|
if (interruptible)
|
|
n = wait_for_completion_interruptible_timeout(&op->waitq,
|
|
timeout);
|
|
else
|
|
n = wait_for_completion_killable_timeout(&op->waitq, timeout);
|
|
|
|
spin_lock(&op->lock);
|
|
|
|
if (op_state_serviced(op))
|
|
return 0;
|
|
|
|
if (unlikely(n < 0)) {
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s: operation interrupted, tag %llu, %p\n",
|
|
__func__,
|
|
llu(op->tag),
|
|
op);
|
|
return -EINTR;
|
|
}
|
|
if (op_state_purged(op)) {
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s: operation purged, tag %llu, %p, %d\n",
|
|
__func__,
|
|
llu(op->tag),
|
|
op,
|
|
op->attempts);
|
|
return (op->attempts < ORANGEFS_PURGE_RETRY_COUNT) ?
|
|
-EAGAIN :
|
|
-EIO;
|
|
}
|
|
/* must have timed out, then... */
|
|
gossip_debug(GOSSIP_WAIT_DEBUG,
|
|
"%s: operation timed out, tag %llu, %p, %d)\n",
|
|
__func__,
|
|
llu(op->tag),
|
|
op,
|
|
op->attempts);
|
|
return -ETIMEDOUT;
|
|
}
|