]> git.proxmox.com Git - mirror_ubuntu-hirsute-kernel.git/commitdiff
staging: lustre: lnet: remove cfs_block_allsigs calls.
authorNeilBrown <neilb@suse.com>
Tue, 20 Feb 2018 02:23:37 +0000 (13:23 +1100)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Thu, 22 Feb 2018 14:03:23 +0000 (15:03 +0100)
Both places that cfs_block_allsigs() is used here,
the goal is to turn an interruptible wait into an
uninterruptible way.
So instead of blocking the signals, change TASK_INTERRUPTIBLE to
TASK_NOLOAD.
In each case, no other functions called while signals are blocked
will sleep - just the one that has been fixed.
In one case, an extra 'interruptible' flag needs to be passed
down so the waiting decision can be made at the right place.

Signed-off-by: NeilBrown <neilb@suse.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
drivers/staging/lustre/include/linux/lnet/api.h
drivers/staging/lustre/lnet/lnet/api-ni.c
drivers/staging/lustre/lnet/lnet/lib-eq.c

index 31fcd33171b4dff73b7c11382a802cfe26352275..dae2e4f0056c1a55fb2c3c61da8f8e9191a20346 100644 (file)
@@ -169,6 +169,7 @@ int LNetEQFree(struct lnet_handle_eq eventq_in);
 int LNetEQPoll(struct lnet_handle_eq *eventqs_in,
               int               neq_in,
               int               timeout_ms,
+              int               interruptible,
               struct lnet_event *event_out,
               int              *which_eq_out);
 /** @} lnet_eq */
index 93e6274e9daceb489881a2b8b31eb00b943428b8..48d25ccadbb37fde07e63e21d0f4ded701042238 100644 (file)
@@ -961,19 +961,15 @@ static void
 lnet_ping_md_unlink(struct lnet_ping_info *pinfo,
                    struct lnet_handle_md *md_handle)
 {
-       sigset_t blocked = cfs_block_allsigs();
-
        LNetMDUnlink(*md_handle);
        LNetInvalidateMDHandle(md_handle);
 
        /* NB md could be busy; this just starts the unlink */
        while (pinfo->pi_features != LNET_PING_FEAT_INVAL) {
                CDEBUG(D_NET, "Still waiting for ping MD to unlink\n");
-               set_current_state(TASK_UNINTERRUPTIBLE);
+               set_current_state(TASK_NOLOAD);
                schedule_timeout(HZ);
        }
-
-       cfs_restore_sigs(blocked);
 }
 
 static void
@@ -2141,7 +2137,6 @@ static int lnet_ping(struct lnet_process_id id, int timeout_ms,
        int nob;
        int rc;
        int rc2;
-       sigset_t blocked;
 
        infosz = offsetof(struct lnet_ping_info, pi_ni[n_ids]);
 
@@ -2197,13 +2192,9 @@ static int lnet_ping(struct lnet_process_id id, int timeout_ms,
 
        do {
                /* MUST block for unlink to complete */
-               if (unlinked)
-                       blocked = cfs_block_allsigs();
-
-               rc2 = LNetEQPoll(&eqh, 1, timeout_ms, &event, &which);
 
-               if (unlinked)
-                       cfs_restore_sigs(blocked);
+               rc2 = LNetEQPoll(&eqh, 1, timeout_ms, !unlinked,
+                                &event, &which);
 
                CDEBUG(D_NET, "poll %d(%d %d)%s\n", rc2,
                       (rc2 <= 0) ? -1 : event.type,
index a173b69e2f922701aa572fba77b13f387b05cd97..ea53b5cb3f726df0c1ba05528e11339507bc5437 100644 (file)
@@ -308,7 +308,7 @@ lnet_eq_dequeue_event(struct lnet_eq *eq, struct lnet_event *ev)
  */
 
 static int
-lnet_eq_wait_locked(int *timeout_ms)
+lnet_eq_wait_locked(int *timeout_ms, long state)
 __must_hold(&the_lnet.ln_eq_wait_lock)
 {
        int tms = *timeout_ms;
@@ -320,7 +320,7 @@ __must_hold(&the_lnet.ln_eq_wait_lock)
                return -ENXIO; /* don't want to wait and no new event */
 
        init_waitqueue_entry(&wl, current);
-       set_current_state(TASK_INTERRUPTIBLE);
+       set_current_state(state);
        add_wait_queue(&the_lnet.ln_eq_waitq, &wl);
 
        lnet_eq_wait_unlock();
@@ -359,6 +359,7 @@ __must_hold(&the_lnet.ln_eq_wait_lock)
  * \param timeout_ms Time in milliseconds to wait for an event to occur on
  * one of the EQs. The constant LNET_TIME_FOREVER can be used to indicate an
  * infinite timeout.
+ * \param interruptible, if true, use TASK_INTERRUPTIBLE, else TASK_NOLOAD
  * \param event,which On successful return (1 or -EOVERFLOW), \a event will
  * hold the next event in the EQs, and \a which will contain the index of the
  * EQ from which the event was taken.
@@ -372,6 +373,7 @@ __must_hold(&the_lnet.ln_eq_wait_lock)
  */
 int
 LNetEQPoll(struct lnet_handle_eq *eventqs, int neq, int timeout_ms,
+          int interruptible,
           struct lnet_event *event, int *which)
 {
        int wait = 1;
@@ -412,7 +414,9 @@ LNetEQPoll(struct lnet_handle_eq *eventqs, int neq, int timeout_ms,
                 *  0 : don't want to wait anymore, but might have new event
                 *      so need to call dequeue again
                 */
-               wait = lnet_eq_wait_locked(&timeout_ms);
+               wait = lnet_eq_wait_locked(&timeout_ms,
+                                          interruptible ? TASK_INTERRUPTIBLE
+                                          : TASK_NOLOAD);
                if (wait < 0) /* no new event */
                        break;
        }