summaryrefslogtreecommitdiff
path: root/fs/orangefs/waitqueue.c
diff options
context:
space:
mode:
authorAl Viro <viro@zeniv.linux.org.uk>2016-02-11 23:07:19 -0500
committerMike Marshall <hubcap@omnibond.com>2016-02-19 13:45:53 -0500
commit78699e29fd784a4613d254a22627f336c55c4a76 (patch)
treec3a8d93a89dd7abed775cdd27a9349cbbd0e539c /fs/orangefs/waitqueue.c
parent1357d06d49d1f87af48ab768d34af55bff18b0c3 (diff)
orangefs: delay freeing slot until cancel completes
Make cancels reuse the aborted read/write op, to make sure they do not fail on lack of memory. Don't issue a cancel unless the daemon has seen our read/write, has not replied and isn't being shut down. If cancel *is* issued, don't wait for it to complete; stash the slot in there and just have it freed when cancel is finally replied to or purged (and delay dropping the reference until then, obviously). Signed-off-by: Al Viro <viro@zeniv.linux.org.uk> Signed-off-by: Mike Marshall <hubcap@omnibond.com>
Diffstat (limited to 'fs/orangefs/waitqueue.c')
-rw-r--r--fs/orangefs/waitqueue.c133
1 files changed, 40 insertions, 93 deletions
diff --git a/fs/orangefs/waitqueue.c b/fs/orangefs/waitqueue.c
index 191d886ccc57..3ea1665efdf0 100644
--- a/fs/orangefs/waitqueue.c
+++ b/fs/orangefs/waitqueue.c
@@ -16,7 +16,6 @@
#include "orangefs-kernel.h"
#include "orangefs-bufmap.h"
-static int wait_for_cancellation_downcall(struct orangefs_kernel_op_s *);
static int wait_for_matching_downcall(struct orangefs_kernel_op_s *);
/*
@@ -36,25 +35,29 @@ void purge_waiting_ops(void)
"pvfs2-client-core: purging op tag %llu %s\n",
llu(op->tag),
get_opname_string(op));
- spin_lock(&op->lock);
set_op_state_purged(op);
- spin_unlock(&op->lock);
}
spin_unlock(&orangefs_request_list_lock);
}
static inline void
-add_op_to_request_list(struct orangefs_kernel_op_s *op)
+__add_op_to_request_list(struct orangefs_kernel_op_s *op)
{
- spin_lock(&orangefs_request_list_lock);
spin_lock(&op->lock);
set_op_state_waiting(op);
list_add_tail(&op->list, &orangefs_request_list);
- spin_unlock(&orangefs_request_list_lock);
spin_unlock(&op->lock);
wake_up_interruptible(&orangefs_request_list_waitq);
}
+static inline void
+add_op_to_request_list(struct orangefs_kernel_op_s *op)
+{
+ spin_lock(&orangefs_request_list_lock);
+ __add_op_to_request_list(op);
+ spin_unlock(&orangefs_request_list_lock);
+}
+
static inline
void add_priority_op_to_request_list(struct orangefs_kernel_op_s *op)
{
@@ -159,15 +162,7 @@ retry_servicing:
if (flags & ORANGEFS_OP_ASYNC)
return 0;
- if (flags & ORANGEFS_OP_CANCELLATION) {
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:"
- "About to call wait_for_cancellation_downcall.\n",
- __func__);
- ret = wait_for_cancellation_downcall(op);
- } else {
- ret = wait_for_matching_downcall(op);
- }
+ ret = wait_for_matching_downcall(op);
if (ret < 0) {
/* failed to get matching downcall */
@@ -273,6 +268,36 @@ retry_servicing:
return ret;
}
+bool orangefs_cancel_op_in_progress(struct orangefs_kernel_op_s *op)
+{
+ u64 tag = op->tag;
+ if (!op_state_in_progress(op))
+ return false;
+
+ op->slot_to_free = op->upcall.req.io.buf_index;
+ memset(&op->upcall, 0, sizeof(op->upcall));
+ memset(&op->downcall, 0, sizeof(op->downcall));
+ op->upcall.type = ORANGEFS_VFS_OP_CANCEL;
+ op->upcall.req.cancel.op_tag = tag;
+ op->downcall.type = ORANGEFS_VFS_OP_INVALID;
+ op->downcall.status = -1;
+ orangefs_new_tag(op);
+
+ spin_lock(&orangefs_request_list_lock);
+ /* orangefs_request_list_lock is enough of a barrier here */
+ if (!__is_daemon_in_service()) {
+ spin_unlock(&orangefs_request_list_lock);
+ return false;
+ }
+ __add_op_to_request_list(op);
+ spin_unlock(&orangefs_request_list_lock);
+
+ gossip_debug(GOSSIP_UTILS_DEBUG,
+ "Attempting ORANGEFS operation cancellation of tag %llu\n",
+ llu(tag));
+ return true;
+}
+
static void orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s *op)
{
/*
@@ -426,81 +451,3 @@ static int wait_for_matching_downcall(struct orangefs_kernel_op_s *op)
return ret;
}
-
-/*
- * similar to wait_for_matching_downcall(), but used in the special case
- * of I/O cancellations.
- *
- * Note we need a special wait function because if this is called we already
- * know that a signal is pending in current and need to service the
- * cancellation upcall anyway. the only way to exit this is to either
- * timeout or have the cancellation be serviced properly.
- */
-static int wait_for_cancellation_downcall(struct orangefs_kernel_op_s *op)
-{
- int ret = -EINVAL;
- DEFINE_WAIT(wait_entry);
-
- while (1) {
- spin_lock(&op->lock);
- prepare_to_wait(&op->waitq, &wait_entry, TASK_INTERRUPTIBLE);
- if (op_state_serviced(op)) {
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:op-state is SERVICED.\n",
- __func__);
- spin_unlock(&op->lock);
- ret = 0;
- break;
- }
-
- if (signal_pending(current)) {
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:operation interrupted by a signal (tag"
- " %llu, op %p)\n",
- __func__,
- llu(op->tag),
- op);
- orangefs_clean_up_interrupted_operation(op);
- ret = -EINTR;
- break;
- }
-
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:About to call schedule_timeout.\n",
- __func__);
- spin_unlock(&op->lock);
- ret = schedule_timeout(op_timeout_secs * HZ);
-
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:Value returned from schedule_timeout(%d).\n",
- __func__,
- ret);
- if (!ret) {
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:*** operation timed out: %p\n",
- __func__,
- op);
- spin_lock(&op->lock);
- orangefs_clean_up_interrupted_operation(op);
- ret = -ETIMEDOUT;
- break;
- }
-
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:Breaking out of loop, regardless of value returned by schedule_timeout.\n",
- __func__);
- ret = -ETIMEDOUT;
- break;
- }
-
- spin_lock(&op->lock);
- finish_wait(&op->waitq, &wait_entry);
- spin_unlock(&op->lock);
-
- gossip_debug(GOSSIP_WAIT_DEBUG,
- "%s:returning ret(%d)\n",
- __func__,
- ret);
-
- return ret;
-}