]> asedeno.scripts.mit.edu Git - linux.git/commitdiff
[SCSI] scsi_transport_fc: fix blocked bsg request when fc object deleted
authorJames Smart <james.smart@emulex.com>
Wed, 1 Sep 2010 02:27:31 +0000 (22:27 -0400)
committerJames Bottomley <James.Bottomley@suse.de>
Thu, 9 Sep 2010 20:37:40 +0000 (15:37 -0500)
When an rport is "blocked" and a bsg request is received, the bsg request gets
placed on the queue but the queue stalls. If the fc object is then deleted - the
bsg queue never restarts and keeps the reference on the object, and stops the
overall teardown.

This patch restarts the bsg queue on teardown and drains any pending requests,
allowing the teardown to succeed.

Signed-off-by: Carl Lajeunesse <carl.lajeunesse@emulex.com>
Signed-off-by: James Smart <james.smart@emulex.com>
Signed-off-by: James Bottomley <James.Bottomley@suse.de>
drivers/scsi/scsi_transport_fc.c

index 9f0f7d9c7422d7ec9f15fedfa8ffe8452c32bdfc..78486d5406525df31776f79f3ed7be3c9198817b 100644 (file)
@@ -4048,11 +4048,54 @@ fc_bsg_rportadd(struct Scsi_Host *shost, struct fc_rport *rport)
 /**
  * fc_bsg_remove - Deletes the bsg hooks on fchosts/rports
  * @q: the request_queue that is to be torn down.
+ *
+ * Notes:
+ *   Before unregistering the queue empty any requests that are blocked
+ *
+ *
  */
 static void
 fc_bsg_remove(struct request_queue *q)
 {
+       struct request *req; /* block request */
+       int counts; /* totals for request_list count and starved */
+
        if (q) {
+               /* Stop taking in new requests */
+               spin_lock_irq(q->queue_lock);
+               blk_stop_queue(q);
+
+               /* drain all requests in the queue */
+               while (1) {
+                       /* need the lock to fetch a request
+                        * this may fetch the same reqeust as the previous pass
+                        */
+                       req = blk_fetch_request(q);
+                       /* save requests in use and starved */
+                       counts = q->rq.count[0] + q->rq.count[1] +
+                               q->rq.starved[0] + q->rq.starved[1];
+                       spin_unlock_irq(q->queue_lock);
+                       /* any requests still outstanding? */
+                       if (counts == 0)
+                               break;
+
+                       /* This may be the same req as the previous iteration,
+                        * always send the blk_end_request_all after a prefetch.
+                        * It is not okay to not end the request because the
+                        * prefetch started the request.
+                        */
+                       if (req) {
+                               /* return -ENXIO to indicate that this queue is
+                                * going away
+                                */
+                               req->errors = -ENXIO;
+                               blk_end_request_all(req, -ENXIO);
+                       }
+
+                       msleep(200); /* allow bsg to possibly finish */
+                       spin_lock_irq(q->queue_lock);
+               }
+
                bsg_unregister_queue(q);
                blk_cleanup_queue(q);
        }