]> git.kernelconcepts.de Git - karo-tx-linux.git/commitdiff
hpsa: add ioaccel sg chaining for the ioaccel2 path
authorWebb Scales <webbnh@hp.com>
Thu, 23 Apr 2015 14:33:27 +0000 (09:33 -0500)
committerJames Bottomley <JBottomley@Odin.com>
Sun, 31 May 2015 18:32:26 +0000 (11:32 -0700)
Increase the request size for ioaccel2 path.

The error, if any, returned by hpsa_allocate_ioaccel2_sg_chain_blocks
to hpsa_alloc_ioaccel2_cmd_and_bft should be returned upstream rather
than assumed to be -ENOMEM.

This differs slightly from hpsa_alloc_ioaccel1_cmd_and_bft,
which does not call another hpsa_allocate function and only
has -ENOMEM to return from some kmalloc calls.

Reviewed-by: Scott Teel <scott.teel@pmcs.com>
Reviewed-by: Kevin Barnett <kevin.barnett@pmcs.com>
Signed-off-by: Robert Elliott <elliott@hp.com>
Signed-off-by: Don Brace <don.brace@pmcs.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: James Bottomley <JBottomley@Odin.com>
drivers/scsi/hpsa.c
drivers/scsi/hpsa.h

index 85585c2f92777c79773835d7cabb44e80785fd10..facb438061d55f70eff24c07b4384adce57375ee 100644 (file)
@@ -1699,6 +1699,46 @@ static void hpsa_slave_destroy(struct scsi_device *sdev)
        /* nothing to do. */
 }
 
+static void hpsa_free_ioaccel2_sg_chain_blocks(struct ctlr_info *h)
+{
+       int i;
+
+       if (!h->ioaccel2_cmd_sg_list)
+               return;
+       for (i = 0; i < h->nr_cmds; i++) {
+               kfree(h->ioaccel2_cmd_sg_list[i]);
+               h->ioaccel2_cmd_sg_list[i] = NULL;
+       }
+       kfree(h->ioaccel2_cmd_sg_list);
+       h->ioaccel2_cmd_sg_list = NULL;
+}
+
+static int hpsa_allocate_ioaccel2_sg_chain_blocks(struct ctlr_info *h)
+{
+       int i;
+
+       if (h->chainsize <= 0)
+               return 0;
+
+       h->ioaccel2_cmd_sg_list =
+               kzalloc(sizeof(*h->ioaccel2_cmd_sg_list) * h->nr_cmds,
+                                       GFP_KERNEL);
+       if (!h->ioaccel2_cmd_sg_list)
+               return -ENOMEM;
+       for (i = 0; i < h->nr_cmds; i++) {
+               h->ioaccel2_cmd_sg_list[i] =
+                       kmalloc(sizeof(*h->ioaccel2_cmd_sg_list[i]) *
+                                       h->maxsgentries, GFP_KERNEL);
+               if (!h->ioaccel2_cmd_sg_list[i])
+                       goto clean;
+       }
+       return 0;
+
+clean:
+       hpsa_free_ioaccel2_sg_chain_blocks(h);
+       return -ENOMEM;
+}
+
 static void hpsa_free_sg_chain_blocks(struct ctlr_info *h)
 {
        int i;
@@ -1741,6 +1781,39 @@ clean:
        return -ENOMEM;
 }
 
+static int hpsa_map_ioaccel2_sg_chain_block(struct ctlr_info *h,
+       struct io_accel2_cmd *cp, struct CommandList *c)
+{
+       struct ioaccel2_sg_element *chain_block;
+       u64 temp64;
+       u32 chain_size;
+
+       chain_block = h->ioaccel2_cmd_sg_list[c->cmdindex];
+       chain_size = le32_to_cpu(cp->data_len);
+       temp64 = pci_map_single(h->pdev, chain_block, chain_size,
+                               PCI_DMA_TODEVICE);
+       if (dma_mapping_error(&h->pdev->dev, temp64)) {
+               /* prevent subsequent unmapping */
+               cp->sg->address = 0;
+               return -1;
+       }
+       cp->sg->address = cpu_to_le64(temp64);
+       return 0;
+}
+
+static void hpsa_unmap_ioaccel2_sg_chain_block(struct ctlr_info *h,
+       struct io_accel2_cmd *cp)
+{
+       struct ioaccel2_sg_element *chain_sg;
+       u64 temp64;
+       u32 chain_size;
+
+       chain_sg = cp->sg;
+       temp64 = le64_to_cpu(chain_sg->address);
+       chain_size = le32_to_cpu(cp->data_len);
+       pci_unmap_single(h->pdev, temp64, chain_size, PCI_DMA_TODEVICE);
+}
+
 static int hpsa_map_sg_chain_block(struct ctlr_info *h,
        struct CommandList *c)
 {
@@ -1950,6 +2023,7 @@ static void complete_scsi_command(struct CommandList *cp)
        struct ctlr_info *h;
        struct ErrorInfo *ei;
        struct hpsa_scsi_dev_t *dev;
+       struct io_accel2_cmd *c2;
 
        u8 sense_key;
        u8 asc;      /* additional sense code */
@@ -1960,12 +2034,17 @@ static void complete_scsi_command(struct CommandList *cp)
        cmd = cp->scsi_cmd;
        h = cp->h;
        dev = cmd->device->hostdata;
+       c2 = &h->ioaccel2_cmd_pool[cp->cmdindex];
 
        scsi_dma_unmap(cmd); /* undo the DMA mappings */
        if ((cp->cmd_type == CMD_SCSI) &&
                (le16_to_cpu(cp->Header.SGTotal) > h->max_cmd_sg_entries))
                hpsa_unmap_sg_chain_block(h, cp);
 
+       if ((cp->cmd_type == CMD_IOACCEL2) &&
+               (c2->sg[0].chain_indicator == IOACCEL2_CHAIN))
+               hpsa_unmap_ioaccel2_sg_chain_block(h, c2);
+
        cmd->result = (DID_OK << 16);           /* host byte */
        cmd->result |= (COMMAND_COMPLETE << 8); /* msg byte */
 
@@ -3809,10 +3888,7 @@ static int hpsa_scsi_ioaccel2_queue_command(struct ctlr_info *h,
        u32 len;
        u32 total_len = 0;
 
-       if (scsi_sg_count(cmd) > h->ioaccel_maxsg) {
-               atomic_dec(&phys_disk->ioaccel_cmds_out);
-               return IO_ACCEL_INELIGIBLE;
-       }
+       BUG_ON(scsi_sg_count(cmd) > h->maxsgentries);
 
        if (fixup_ioaccel_cdb(cdb, &cdb_len)) {
                atomic_dec(&phys_disk->ioaccel_cmds_out);
@@ -3835,8 +3911,19 @@ static int hpsa_scsi_ioaccel2_queue_command(struct ctlr_info *h,
        }
 
        if (use_sg) {
-               BUG_ON(use_sg > IOACCEL2_MAXSGENTRIES);
                curr_sg = cp->sg;
+               if (use_sg > h->ioaccel_maxsg) {
+                       addr64 = le64_to_cpu(
+                               h->ioaccel2_cmd_sg_list[c->cmdindex]->address);
+                       curr_sg->address = cpu_to_le64(addr64);
+                       curr_sg->length = 0;
+                       curr_sg->reserved[0] = 0;
+                       curr_sg->reserved[1] = 0;
+                       curr_sg->reserved[2] = 0;
+                       curr_sg->chain_indicator = 0x80;
+
+                       curr_sg = h->ioaccel2_cmd_sg_list[c->cmdindex];
+               }
                scsi_for_each_sg(cmd, sg, use_sg, i) {
                        addr64 = (u64) sg_dma_address(sg);
                        len  = sg_dma_len(sg);
@@ -3881,14 +3968,22 @@ static int hpsa_scsi_ioaccel2_queue_command(struct ctlr_info *h,
        cp->Tag = cpu_to_le32(c->cmdindex << DIRECT_LOOKUP_SHIFT);
        memcpy(cp->cdb, cdb, sizeof(cp->cdb));
 
-       /* fill in sg elements */
-       cp->sg_count = (u8) use_sg;
-
        cp->data_len = cpu_to_le32(total_len);
        cp->err_ptr = cpu_to_le64(c->busaddr +
                        offsetof(struct io_accel2_cmd, error_data));
        cp->err_len = cpu_to_le32(sizeof(cp->error_data));
 
+       /* fill in sg elements */
+       if (use_sg > h->ioaccel_maxsg) {
+               cp->sg_count = 1;
+               if (hpsa_map_ioaccel2_sg_chain_block(h, cp, c)) {
+                       atomic_dec(&phys_disk->ioaccel_cmds_out);
+                       scsi_dma_unmap(cmd);
+                       return -1;
+               }
+       } else
+               cp->sg_count = (u8) use_sg;
+
        enqueue_cmd_and_start_io(h, c);
        return 0;
 }
@@ -7892,6 +7987,8 @@ clean_up:
 /* Free ioaccel2 mode command blocks and block fetch table */
 static void hpsa_free_ioaccel2_cmd_and_bft(struct ctlr_info *h)
 {
+       hpsa_free_ioaccel2_sg_chain_blocks(h);
+
        if (h->ioaccel2_cmd_pool)
                pci_free_consistent(h->pdev,
                        h->nr_cmds * sizeof(*h->ioaccel2_cmd_pool),
@@ -7903,6 +8000,8 @@ static void hpsa_free_ioaccel2_cmd_and_bft(struct ctlr_info *h)
 /* Allocate ioaccel2 mode command blocks and block fetch table */
 static int hpsa_alloc_ioaccel2_cmd_and_bft(struct ctlr_info *h)
 {
+       int rc;
+
        /* Allocate ioaccel2 mode command blocks and block fetch table */
 
        h->ioaccel_maxsg =
@@ -7922,7 +8021,13 @@ static int hpsa_alloc_ioaccel2_cmd_and_bft(struct ctlr_info *h)
                                sizeof(u32)), GFP_KERNEL);
 
        if ((h->ioaccel2_cmd_pool == NULL) ||
-               (h->ioaccel2_blockFetchTable == NULL))
+               (h->ioaccel2_blockFetchTable == NULL)) {
+               rc = -ENOMEM;
+               goto clean_up;
+       }
+
+       rc = hpsa_allocate_ioaccel2_sg_chain_blocks(h);
+       if (rc)
                goto clean_up;
 
        memset(h->ioaccel2_cmd_pool, 0,
@@ -7931,7 +8036,7 @@ static int hpsa_alloc_ioaccel2_cmd_and_bft(struct ctlr_info *h)
 
 clean_up:
        hpsa_free_ioaccel2_cmd_and_bft(h);
-       return 1;
+       return rc;
 }
 
 static void hpsa_put_ctlr_into_performant_mode(struct ctlr_info *h)
index 87a70b5fc5e262f3be7ec1cb6d7d792b86df471f..3acacf67bd22bd55fb664d53a5f7f20d39622adc 100644 (file)
@@ -162,6 +162,7 @@ struct ctlr_info {
        u8 max_cmd_sg_entries;
        int chainsize;
        struct SGDescriptor **cmd_sg_list;
+       struct ioaccel2_sg_element **ioaccel2_cmd_sg_list;
 
        /* pointers to command and error info pool */
        struct CommandList      *cmd_pool;