]> git.kernelconcepts.de Git - karo-tx-linux.git/commitdiff
Merge branch 'for-linus-core' of git://git.kernel.org/pub/scm/linux/kernel/git/nab...
authorLinus Torvalds <torvalds@linux-foundation.org>
Fri, 22 Jul 2011 23:01:14 +0000 (16:01 -0700)
committerLinus Torvalds <torvalds@linux-foundation.org>
Fri, 22 Jul 2011 23:01:14 +0000 (16:01 -0700)
* 'for-linus-core' of git://git.kernel.org/pub/scm/linux/kernel/git/nab/target-pending: (38 commits)
  target: Bump version to v4.1.0-rc1-ml
  target: remove custom hex2bin() implementation
  target: fix typo Assoication -> Association
  target: Update QUEUE ALGORITHM MODIFIER control page default
  target: ->map_task_SG conversion to ->map_control_SG and ->map_data_SG
  target: Follow up core updates from AGrover and HCH (round 4)
  target: Eliminate usage of struct se_mem
  target: Pass 2nd param of transport_split_cdb by value
  target: Enforce 1 page max for control cdb buffer sizes
  target: Make all control CDBs scatter-gather
  target: Implement Block Device Characteristics VPD page
  target: Fix reporting of supported VPD pages
  target: Allow for built-in target modules
  tcm_fc: Convert to wake_up_process and schedule_timeout_interruptible
  tcm_fc: Makefile cleanups
  loopback: Fix memory leak in tcm_loop_make_scsi_hba()
  loopback: Remove duplicate scsi/scsi_tcq.h include
  loopback: off by one in tcm_loop_make_naa_tpg()
  target/iblock: Remove unused iblock_dev members
  target/iblock: Use request_queue->nr_request for se_device defaults
  ...

37 files changed:
drivers/target/loopback/Kconfig
drivers/target/loopback/tcm_loop.c
drivers/target/loopback/tcm_loop.h
drivers/target/target_core_alua.c
drivers/target/target_core_cdb.c
drivers/target/target_core_configfs.c
drivers/target/target_core_device.c
drivers/target/target_core_fabric_configfs.c
drivers/target/target_core_fabric_lib.c
drivers/target/target_core_file.c
drivers/target/target_core_file.h
drivers/target/target_core_hba.c
drivers/target/target_core_iblock.c
drivers/target/target_core_iblock.h
drivers/target/target_core_pr.c
drivers/target/target_core_pr.h
drivers/target/target_core_pscsi.c
drivers/target/target_core_pscsi.h
drivers/target/target_core_rd.c
drivers/target/target_core_rd.h
drivers/target/target_core_scdb.c
drivers/target/target_core_scdb.h
drivers/target/target_core_stat.c
drivers/target/target_core_tmr.c
drivers/target/target_core_tpg.c
drivers/target/target_core_transport.c
drivers/target/target_core_ua.c
drivers/target/tcm_fc/Makefile
drivers/target/tcm_fc/tcm_fc.h
drivers/target/tcm_fc/tfc_cmd.c
drivers/target/tcm_fc/tfc_conf.c
drivers/target/tcm_fc/tfc_io.c
drivers/target/tcm_fc/tfc_sess.c
include/target/target_core_base.h
include/target/target_core_device.h
include/target/target_core_fabric_ops.h
include/target/target_core_transport.h

index 57dcbc2d711be591e4beb1f867b469e2c8e1d21a..abe8ecbcdf0648814100e1ecb809a523da7fa54d 100644 (file)
@@ -3,9 +3,3 @@ config LOOPBACK_TARGET
        help
          Say Y here to enable the TCM Virtual SAS target and Linux/SCSI LLD
          fabric loopback module.
-
-config LOOPBACK_TARGET_CDB_DEBUG
-       bool "TCM loopback fabric module CDB debug code"
-       depends on LOOPBACK_TARGET
-       help
-         Say Y here to enable the TCM loopback fabric module CDB debug code
index 70c2e7fa66643f7a7cdf040c62d74e977d7b9d1b..aa2d6799723569a490d566a71c0e7101a82fe407 100644 (file)
@@ -31,7 +31,6 @@
 #include <scsi/scsi_host.h>
 #include <scsi/scsi_device.h>
 #include <scsi/scsi_cmnd.h>
-#include <scsi/scsi_tcq.h>
 
 #include <target/target_core_base.h>
 #include <target/target_core_transport.h>
@@ -80,7 +79,7 @@ static struct se_cmd *tcm_loop_allocate_core_cmd(
 
        tl_cmd = kmem_cache_zalloc(tcm_loop_cmd_cache, GFP_ATOMIC);
        if (!tl_cmd) {
-               printk(KERN_ERR "Unable to allocate struct tcm_loop_cmd\n");
+               pr_err("Unable to allocate struct tcm_loop_cmd\n");
                set_host_byte(sc, DID_ERROR);
                return NULL;
        }
@@ -118,17 +117,16 @@ static struct se_cmd *tcm_loop_allocate_core_cmd(
         * Signal BIDI usage with T_TASK(cmd)->t_tasks_bidi
         */
        if (scsi_bidi_cmnd(sc))
-               T_TASK(se_cmd)->t_tasks_bidi = 1;
+               se_cmd->t_tasks_bidi = 1;
        /*
         * Locate the struct se_lun pointer and attach it to struct se_cmd
         */
-       if (transport_get_lun_for_cmd(se_cmd, NULL, tl_cmd->sc->device->lun) < 0) {
+       if (transport_lookup_cmd_lun(se_cmd, tl_cmd->sc->device->lun) < 0) {
                kmem_cache_free(tcm_loop_cmd_cache, tl_cmd);
                set_host_byte(sc, DID_NO_CONNECT);
                return NULL;
        }
 
-       transport_device_setup_cmd(se_cmd);
        return se_cmd;
 }
 
@@ -143,17 +141,17 @@ static int tcm_loop_new_cmd_map(struct se_cmd *se_cmd)
        struct tcm_loop_cmd *tl_cmd = container_of(se_cmd,
                                struct tcm_loop_cmd, tl_se_cmd);
        struct scsi_cmnd *sc = tl_cmd->sc;
-       void *mem_ptr, *mem_bidi_ptr = NULL;
-       u32 sg_no_bidi = 0;
+       struct scatterlist *sgl_bidi = NULL;
+       u32 sgl_bidi_count = 0;
        int ret;
        /*
         * Allocate the necessary tasks to complete the received CDB+data
         */
-       ret = transport_generic_allocate_tasks(se_cmd, tl_cmd->sc->cmnd);
-       if (ret == -1) {
+       ret = transport_generic_allocate_tasks(se_cmd, sc->cmnd);
+       if (ret == -ENOMEM) {
                /* Out of Resources */
                return PYX_TRANSPORT_LU_COMM_FAILURE;
-       } else if (ret == -2) {
+       } else if (ret == -EINVAL) {
                /*
                 * Handle case for SAM_STAT_RESERVATION_CONFLICT
                 */
@@ -165,35 +163,21 @@ static int tcm_loop_new_cmd_map(struct se_cmd *se_cmd)
                 */
                return PYX_TRANSPORT_USE_SENSE_REASON;
        }
+
        /*
-        * Setup the struct scatterlist memory from the received
-        * struct scsi_cmnd.
+        * For BIDI commands, pass in the extra READ buffer
+        * to transport_generic_map_mem_to_cmd() below..
         */
-       if (scsi_sg_count(sc)) {
-               se_cmd->se_cmd_flags |= SCF_PASSTHROUGH_SG_TO_MEM;
-               mem_ptr = (void *)scsi_sglist(sc);
-               /*
-                * For BIDI commands, pass in the extra READ buffer
-                * to transport_generic_map_mem_to_cmd() below..
-                */
-               if (T_TASK(se_cmd)->t_tasks_bidi) {
-                       struct scsi_data_buffer *sdb = scsi_in(sc);
+       if (se_cmd->t_tasks_bidi) {
+               struct scsi_data_buffer *sdb = scsi_in(sc);
 
-                       mem_bidi_ptr = (void *)sdb->table.sgl;
-                       sg_no_bidi = sdb->table.nents;
-               }
-       } else {
-               /*
-                * Used for DMA_NONE
-                */
-               mem_ptr = NULL;
+               sgl_bidi = sdb->table.sgl;
+               sgl_bidi_count = sdb->table.nents;
        }
-       /*
-        * Map the SG memory into struct se_mem->page linked list using the same
-        * physical memory at sg->page_link.
-        */
-       ret = transport_generic_map_mem_to_cmd(se_cmd, mem_ptr,
-                       scsi_sg_count(sc), mem_bidi_ptr, sg_no_bidi);
+
+       /* Tell the core about our preallocated memory */
+       ret = transport_generic_map_mem_to_cmd(se_cmd, scsi_sglist(sc),
+                       scsi_sg_count(sc), sgl_bidi, sgl_bidi_count);
        if (ret < 0)
                return PYX_TRANSPORT_LU_COMM_FAILURE;
 
@@ -216,13 +200,10 @@ static void tcm_loop_check_stop_free(struct se_cmd *se_cmd)
         * Release the struct se_cmd, which will make a callback to release
         * struct tcm_loop_cmd * in tcm_loop_deallocate_core_cmd()
         */
-       transport_generic_free_cmd(se_cmd, 0, 1, 0);
+       transport_generic_free_cmd(se_cmd, 0, 0);
 }
 
-/*
- * Called from struct target_core_fabric_ops->release_cmd_to_pool()
- */
-static void tcm_loop_deallocate_core_cmd(struct se_cmd *se_cmd)
+static void tcm_loop_release_cmd(struct se_cmd *se_cmd)
 {
        struct tcm_loop_cmd *tl_cmd = container_of(se_cmd,
                                struct tcm_loop_cmd, tl_se_cmd);
@@ -300,7 +281,7 @@ static int tcm_loop_queuecommand(
        struct tcm_loop_hba *tl_hba;
        struct tcm_loop_tpg *tl_tpg;
 
-       TL_CDB_DEBUG("tcm_loop_queuecommand() %d:%d:%d:%d got CDB: 0x%02x"
+       pr_debug("tcm_loop_queuecommand() %d:%d:%d:%d got CDB: 0x%02x"
                " scsi_buf_len: %u\n", sc->device->host->host_no,
                sc->device->id, sc->device->channel, sc->device->lun,
                sc->cmnd[0], scsi_bufflen(sc));
@@ -350,7 +331,7 @@ static int tcm_loop_device_reset(struct scsi_cmnd *sc)
         */
        tl_nexus = tl_hba->tl_nexus;
        if (!tl_nexus) {
-               printk(KERN_ERR "Unable to perform device reset without"
+               pr_err("Unable to perform device reset without"
                                " active I_T Nexus\n");
                return FAILED;
        }
@@ -363,13 +344,13 @@ static int tcm_loop_device_reset(struct scsi_cmnd *sc)
 
        tl_cmd = kmem_cache_zalloc(tcm_loop_cmd_cache, GFP_KERNEL);
        if (!tl_cmd) {
-               printk(KERN_ERR "Unable to allocate memory for tl_cmd\n");
+               pr_err("Unable to allocate memory for tl_cmd\n");
                return FAILED;
        }
 
        tl_tmr = kzalloc(sizeof(struct tcm_loop_tmr), GFP_KERNEL);
        if (!tl_tmr) {
-               printk(KERN_ERR "Unable to allocate memory for tl_tmr\n");
+               pr_err("Unable to allocate memory for tl_tmr\n");
                goto release;
        }
        init_waitqueue_head(&tl_tmr->tl_tmr_wait);
@@ -384,14 +365,14 @@ static int tcm_loop_device_reset(struct scsi_cmnd *sc)
        /*
         * Allocate the LUN_RESET TMR
         */
-       se_cmd->se_tmr_req = core_tmr_alloc_req(se_cmd, (void *)tl_tmr,
+       se_cmd->se_tmr_req = core_tmr_alloc_req(se_cmd, tl_tmr,
                                TMR_LUN_RESET);
        if (IS_ERR(se_cmd->se_tmr_req))
                goto release;
        /*
         * Locate the underlying TCM struct se_lun from sc->device->lun
         */
-       if (transport_get_lun_for_tmr(se_cmd, sc->device->lun) < 0)
+       if (transport_lookup_tmr_lun(se_cmd, sc->device->lun) < 0)
                goto release;
        /*
         * Queue the TMR to TCM Core and sleep waiting for tcm_loop_queue_tm_rsp()
@@ -407,7 +388,7 @@ static int tcm_loop_device_reset(struct scsi_cmnd *sc)
                SUCCESS : FAILED;
 release:
        if (se_cmd)
-               transport_generic_free_cmd(se_cmd, 1, 1, 0);
+               transport_generic_free_cmd(se_cmd, 1, 0);
        else
                kmem_cache_free(tcm_loop_cmd_cache, tl_cmd);
        kfree(tl_tmr);
@@ -454,7 +435,7 @@ static int tcm_loop_driver_probe(struct device *dev)
        sh = scsi_host_alloc(&tcm_loop_driver_template,
                        sizeof(struct tcm_loop_hba));
        if (!sh) {
-               printk(KERN_ERR "Unable to allocate struct scsi_host\n");
+               pr_err("Unable to allocate struct scsi_host\n");
                return -ENODEV;
        }
        tl_hba->sh = sh;
@@ -473,7 +454,7 @@ static int tcm_loop_driver_probe(struct device *dev)
 
        error = scsi_add_host(sh, &tl_hba->dev);
        if (error) {
-               printk(KERN_ERR "%s: scsi_add_host failed\n", __func__);
+               pr_err("%s: scsi_add_host failed\n", __func__);
                scsi_host_put(sh);
                return -ENODEV;
        }
@@ -514,7 +495,7 @@ static int tcm_loop_setup_hba_bus(struct tcm_loop_hba *tl_hba, int tcm_loop_host
 
        ret = device_register(&tl_hba->dev);
        if (ret) {
-               printk(KERN_ERR "device_register() failed for"
+               pr_err("device_register() failed for"
                                " tl_hba->dev: %d\n", ret);
                return -ENODEV;
        }
@@ -532,24 +513,24 @@ static int tcm_loop_alloc_core_bus(void)
 
        tcm_loop_primary = root_device_register("tcm_loop_0");
        if (IS_ERR(tcm_loop_primary)) {
-               printk(KERN_ERR "Unable to allocate tcm_loop_primary\n");
+               pr_err("Unable to allocate tcm_loop_primary\n");
                return PTR_ERR(tcm_loop_primary);
        }
 
        ret = bus_register(&tcm_loop_lld_bus);
        if (ret) {
-               printk(KERN_ERR "bus_register() failed for tcm_loop_lld_bus\n");
+               pr_err("bus_register() failed for tcm_loop_lld_bus\n");
                goto dev_unreg;
        }
 
        ret = driver_register(&tcm_loop_driverfs);
        if (ret) {
-               printk(KERN_ERR "driver_register() failed for"
+               pr_err("driver_register() failed for"
                                "tcm_loop_driverfs\n");
                goto bus_unreg;
        }
 
-       printk(KERN_INFO "Initialized TCM Loop Core Bus\n");
+       pr_debug("Initialized TCM Loop Core Bus\n");
        return ret;
 
 bus_unreg:
@@ -565,7 +546,7 @@ static void tcm_loop_release_core_bus(void)
        bus_unregister(&tcm_loop_lld_bus);
        root_device_unregister(tcm_loop_primary);
 
-       printk(KERN_INFO "Releasing TCM Loop Core BUS\n");
+       pr_debug("Releasing TCM Loop Core BUS\n");
 }
 
 static char *tcm_loop_get_fabric_name(void)
@@ -593,7 +574,7 @@ static u8 tcm_loop_get_fabric_proto_ident(struct se_portal_group *se_tpg)
        case SCSI_PROTOCOL_ISCSI:
                return iscsi_get_fabric_proto_ident(se_tpg);
        default:
-               printk(KERN_ERR "Unknown tl_proto_id: 0x%02x, using"
+               pr_err("Unknown tl_proto_id: 0x%02x, using"
                        " SAS emulation\n", tl_hba->tl_proto_id);
                break;
        }
@@ -649,7 +630,7 @@ static u32 tcm_loop_get_pr_transport_id(
                return iscsi_get_pr_transport_id(se_tpg, se_nacl, pr_reg,
                                        format_code, buf);
        default:
-               printk(KERN_ERR "Unknown tl_proto_id: 0x%02x, using"
+               pr_err("Unknown tl_proto_id: 0x%02x, using"
                        " SAS emulation\n", tl_hba->tl_proto_id);
                break;
        }
@@ -679,7 +660,7 @@ static u32 tcm_loop_get_pr_transport_id_len(
                return iscsi_get_pr_transport_id_len(se_tpg, se_nacl, pr_reg,
                                        format_code);
        default:
-               printk(KERN_ERR "Unknown tl_proto_id: 0x%02x, using"
+               pr_err("Unknown tl_proto_id: 0x%02x, using"
                        " SAS emulation\n", tl_hba->tl_proto_id);
                break;
        }
@@ -713,7 +694,7 @@ static char *tcm_loop_parse_pr_out_transport_id(
                return iscsi_parse_pr_out_transport_id(se_tpg, buf, out_tid_len,
                                        port_nexus_ptr);
        default:
-               printk(KERN_ERR "Unknown tl_proto_id: 0x%02x, using"
+               pr_err("Unknown tl_proto_id: 0x%02x, using"
                        " SAS emulation\n", tl_hba->tl_proto_id);
                break;
        }
@@ -762,7 +743,7 @@ static struct se_node_acl *tcm_loop_tpg_alloc_fabric_acl(
 
        tl_nacl = kzalloc(sizeof(struct tcm_loop_nacl), GFP_KERNEL);
        if (!tl_nacl) {
-               printk(KERN_ERR "Unable to allocate struct tcm_loop_nacl\n");
+               pr_err("Unable to allocate struct tcm_loop_nacl\n");
                return NULL;
        }
 
@@ -784,16 +765,6 @@ static u32 tcm_loop_get_inst_index(struct se_portal_group *se_tpg)
        return 1;
 }
 
-static void tcm_loop_new_cmd_failure(struct se_cmd *se_cmd)
-{
-       /*
-        * Since TCM_loop is already passing struct scatterlist data from
-        * struct scsi_cmnd, no more Linux/SCSI failure dependent state need
-        * to be handled here.
-        */
-       return;
-}
-
 static int tcm_loop_is_state_remove(struct se_cmd *se_cmd)
 {
        /*
@@ -882,7 +853,7 @@ static int tcm_loop_queue_data_in(struct se_cmd *se_cmd)
                                struct tcm_loop_cmd, tl_se_cmd);
        struct scsi_cmnd *sc = tl_cmd->sc;
 
-       TL_CDB_DEBUG("tcm_loop_queue_data_in() called for scsi_cmnd: %p"
+       pr_debug("tcm_loop_queue_data_in() called for scsi_cmnd: %p"
                     " cdb: 0x%02x\n", sc, sc->cmnd[0]);
 
        sc->result = SAM_STAT_GOOD;
@@ -897,14 +868,14 @@ static int tcm_loop_queue_status(struct se_cmd *se_cmd)
                                struct tcm_loop_cmd, tl_se_cmd);
        struct scsi_cmnd *sc = tl_cmd->sc;
 
-       TL_CDB_DEBUG("tcm_loop_queue_status() called for scsi_cmnd: %p"
+       pr_debug("tcm_loop_queue_status() called for scsi_cmnd: %p"
                        " cdb: 0x%02x\n", sc, sc->cmnd[0]);
 
        if (se_cmd->sense_buffer &&
           ((se_cmd->se_cmd_flags & SCF_TRANSPORT_TASK_SENSE) ||
            (se_cmd->se_cmd_flags & SCF_EMULATED_TASK_SENSE))) {
 
-               memcpy((void *)sc->sense_buffer, (void *)se_cmd->sense_buffer,
+               memcpy(sc->sense_buffer, se_cmd->sense_buffer,
                                SCSI_SENSE_BUFFERSIZE);
                sc->result = SAM_STAT_CHECK_CONDITION;
                set_driver_byte(sc, DRIVER_SENSE);
@@ -972,7 +943,7 @@ static int tcm_loop_port_link(
         */
        scsi_add_device(tl_hba->sh, 0, tl_tpg->tl_tpgt, lun->unpacked_lun);
 
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Port Link Successful\n");
+       pr_debug("TCM_Loop_ConfigFS: Port Link Successful\n");
        return 0;
 }
 
@@ -990,7 +961,7 @@ static void tcm_loop_port_unlink(
        sd = scsi_device_lookup(tl_hba->sh, 0, tl_tpg->tl_tpgt,
                                se_lun->unpacked_lun);
        if (!sd) {
-               printk(KERN_ERR "Unable to locate struct scsi_device for %d:%d:"
+               pr_err("Unable to locate struct scsi_device for %d:%d:"
                        "%d\n", 0, tl_tpg->tl_tpgt, se_lun->unpacked_lun);
                return;
        }
@@ -1003,7 +974,7 @@ static void tcm_loop_port_unlink(
        atomic_dec(&tl_tpg->tl_tpg_port_count);
        smp_mb__after_atomic_dec();
 
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Port Unlink Successful\n");
+       pr_debug("TCM_Loop_ConfigFS: Port Unlink Successful\n");
 }
 
 /* End items for tcm_loop_port_cit */
@@ -1020,14 +991,14 @@ static int tcm_loop_make_nexus(
        int ret = -ENOMEM;
 
        if (tl_tpg->tl_hba->tl_nexus) {
-               printk(KERN_INFO "tl_tpg->tl_hba->tl_nexus already exists\n");
+               pr_debug("tl_tpg->tl_hba->tl_nexus already exists\n");
                return -EEXIST;
        }
        se_tpg = &tl_tpg->tl_se_tpg;
 
        tl_nexus = kzalloc(sizeof(struct tcm_loop_nexus), GFP_KERNEL);
        if (!tl_nexus) {
-               printk(KERN_ERR "Unable to allocate struct tcm_loop_nexus\n");
+               pr_err("Unable to allocate struct tcm_loop_nexus\n");
                return -ENOMEM;
        }
        /*
@@ -1054,9 +1025,9 @@ static int tcm_loop_make_nexus(
         * transport_register_session()
         */
        __transport_register_session(se_tpg, tl_nexus->se_sess->se_node_acl,
-                       tl_nexus->se_sess, (void *)tl_nexus);
+                       tl_nexus->se_sess, tl_nexus);
        tl_tpg->tl_hba->tl_nexus = tl_nexus;
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Established I_T Nexus to emulated"
+       pr_debug("TCM_Loop_ConfigFS: Established I_T Nexus to emulated"
                " %s Initiator Port: %s\n", tcm_loop_dump_proto_id(tl_hba),
                name);
        return 0;
@@ -1082,13 +1053,13 @@ static int tcm_loop_drop_nexus(
                return -ENODEV;
 
        if (atomic_read(&tpg->tl_tpg_port_count)) {
-               printk(KERN_ERR "Unable to remove TCM_Loop I_T Nexus with"
+               pr_err("Unable to remove TCM_Loop I_T Nexus with"
                        " active TPG port count: %d\n",
                        atomic_read(&tpg->tl_tpg_port_count));
                return -EPERM;
        }
 
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Removing I_T Nexus to emulated"
+       pr_debug("TCM_Loop_ConfigFS: Removing I_T Nexus to emulated"
                " %s Initiator Port: %s\n", tcm_loop_dump_proto_id(tl_hba),
                tl_nexus->se_sess->se_node_acl->initiatorname);
        /*
@@ -1144,7 +1115,7 @@ static ssize_t tcm_loop_tpg_store_nexus(
         * tcm_loop_make_nexus()
         */
        if (strlen(page) >= TL_WWN_ADDR_LEN) {
-               printk(KERN_ERR "Emulated NAA Sas Address: %s, exceeds"
+               pr_err("Emulated NAA Sas Address: %s, exceeds"
                                " max: %d\n", page, TL_WWN_ADDR_LEN);
                return -EINVAL;
        }
@@ -1153,7 +1124,7 @@ static ssize_t tcm_loop_tpg_store_nexus(
        ptr = strstr(i_port, "naa.");
        if (ptr) {
                if (tl_hba->tl_proto_id != SCSI_PROTOCOL_SAS) {
-                       printk(KERN_ERR "Passed SAS Initiator Port %s does not"
+                       pr_err("Passed SAS Initiator Port %s does not"
                                " match target port protoid: %s\n", i_port,
                                tcm_loop_dump_proto_id(tl_hba));
                        return -EINVAL;
@@ -1164,7 +1135,7 @@ static ssize_t tcm_loop_tpg_store_nexus(
        ptr = strstr(i_port, "fc.");
        if (ptr) {
                if (tl_hba->tl_proto_id != SCSI_PROTOCOL_FCP) {
-                       printk(KERN_ERR "Passed FCP Initiator Port %s does not"
+                       pr_err("Passed FCP Initiator Port %s does not"
                                " match target port protoid: %s\n", i_port,
                                tcm_loop_dump_proto_id(tl_hba));
                        return -EINVAL;
@@ -1175,7 +1146,7 @@ static ssize_t tcm_loop_tpg_store_nexus(
        ptr = strstr(i_port, "iqn.");
        if (ptr) {
                if (tl_hba->tl_proto_id != SCSI_PROTOCOL_ISCSI) {
-                       printk(KERN_ERR "Passed iSCSI Initiator Port %s does not"
+                       pr_err("Passed iSCSI Initiator Port %s does not"
                                " match target port protoid: %s\n", i_port,
                                tcm_loop_dump_proto_id(tl_hba));
                        return -EINVAL;
@@ -1183,7 +1154,7 @@ static ssize_t tcm_loop_tpg_store_nexus(
                port_ptr = &i_port[0];
                goto check_newline;
        }
-       printk(KERN_ERR "Unable to locate prefix for emulated Initiator Port:"
+       pr_err("Unable to locate prefix for emulated Initiator Port:"
                        " %s\n", i_port);
        return -EINVAL;
        /*
@@ -1223,15 +1194,15 @@ struct se_portal_group *tcm_loop_make_naa_tpg(
 
        tpgt_str = strstr(name, "tpgt_");
        if (!tpgt_str) {
-               printk(KERN_ERR "Unable to locate \"tpgt_#\" directory"
+               pr_err("Unable to locate \"tpgt_#\" directory"
                                " group\n");
                return ERR_PTR(-EINVAL);
        }
        tpgt_str += 5; /* Skip ahead of "tpgt_" */
        tpgt = (unsigned short int) simple_strtoul(tpgt_str, &end_ptr, 0);
 
-       if (tpgt > TL_TPGS_PER_HBA) {
-               printk(KERN_ERR "Passed tpgt: %hu exceeds TL_TPGS_PER_HBA:"
+       if (tpgt >= TL_TPGS_PER_HBA) {
+               pr_err("Passed tpgt: %hu exceeds TL_TPGS_PER_HBA:"
                                " %u\n", tpgt, TL_TPGS_PER_HBA);
                return ERR_PTR(-EINVAL);
        }
@@ -1242,12 +1213,12 @@ struct se_portal_group *tcm_loop_make_naa_tpg(
         * Register the tl_tpg as a emulated SAS TCM Target Endpoint
         */
        ret = core_tpg_register(&tcm_loop_fabric_configfs->tf_ops,
-                       wwn, &tl_tpg->tl_se_tpg, (void *)tl_tpg,
+                       wwn, &tl_tpg->tl_se_tpg, tl_tpg,
                        TRANSPORT_TPG_TYPE_NORMAL);
        if (ret < 0)
                return ERR_PTR(-ENOMEM);
 
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Allocated Emulated %s"
+       pr_debug("TCM_Loop_ConfigFS: Allocated Emulated %s"
                " Target Port %s,t,0x%04x\n", tcm_loop_dump_proto_id(tl_hba),
                config_item_name(&wwn->wwn_group.cg_item), tpgt);
 
@@ -1274,7 +1245,7 @@ void tcm_loop_drop_naa_tpg(
         */
        core_tpg_deregister(se_tpg);
 
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Deallocated Emulated %s"
+       pr_debug("TCM_Loop_ConfigFS: Deallocated Emulated %s"
                " Target Port %s,t,0x%04x\n", tcm_loop_dump_proto_id(tl_hba),
                config_item_name(&wwn->wwn_group.cg_item), tpgt);
 }
@@ -1295,7 +1266,7 @@ struct se_wwn *tcm_loop_make_scsi_hba(
 
        tl_hba = kzalloc(sizeof(struct tcm_loop_hba), GFP_KERNEL);
        if (!tl_hba) {
-               printk(KERN_ERR "Unable to allocate struct tcm_loop_hba\n");
+               pr_err("Unable to allocate struct tcm_loop_hba\n");
                return ERR_PTR(-ENOMEM);
        }
        /*
@@ -1314,22 +1285,21 @@ struct se_wwn *tcm_loop_make_scsi_hba(
                goto check_len;
        }
        ptr = strstr(name, "iqn.");
-       if (ptr) {
-               tl_hba->tl_proto_id = SCSI_PROTOCOL_ISCSI;
-               goto check_len;
+       if (!ptr) {
+               pr_err("Unable to locate prefix for emulated Target "
+                               "Port: %s\n", name);
+               ret = -EINVAL;
+               goto out;
        }
-
-       printk(KERN_ERR "Unable to locate prefix for emulated Target Port:"
-                       " %s\n", name);
-       return ERR_PTR(-EINVAL);
+       tl_hba->tl_proto_id = SCSI_PROTOCOL_ISCSI;
 
 check_len:
        if (strlen(name) >= TL_WWN_ADDR_LEN) {
-               printk(KERN_ERR "Emulated NAA %s Address: %s, exceeds"
+               pr_err("Emulated NAA %s Address: %s, exceeds"
                        " max: %d\n", name, tcm_loop_dump_proto_id(tl_hba),
                        TL_WWN_ADDR_LEN);
-               kfree(tl_hba);
-               return ERR_PTR(-EINVAL);
+               ret = -EINVAL;
+               goto out;
        }
        snprintf(&tl_hba->tl_wwn_address[0], TL_WWN_ADDR_LEN, "%s", &name[off]);
 
@@ -1344,7 +1314,7 @@ check_len:
 
        sh = tl_hba->sh;
        tcm_loop_hba_no_cnt++;
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Allocated emulated Target"
+       pr_debug("TCM_Loop_ConfigFS: Allocated emulated Target"
                " %s Address: %s at Linux/SCSI Host ID: %d\n",
                tcm_loop_dump_proto_id(tl_hba), name, sh->host_no);
 
@@ -1367,7 +1337,7 @@ void tcm_loop_drop_scsi_hba(
         */
        device_unregister(&tl_hba->dev);
 
-       printk(KERN_INFO "TCM_Loop_ConfigFS: Deallocated emulated Target"
+       pr_debug("TCM_Loop_ConfigFS: Deallocated emulated Target"
                " SAS Address: %s at Linux/SCSI Host ID: %d\n",
                config_item_name(&wwn->wwn_group.cg_item), host_no);
 }
@@ -1402,9 +1372,9 @@ static int tcm_loop_register_configfs(void)
         * Register the top level struct config_item_type with TCM core
         */
        fabric = target_fabric_configfs_init(THIS_MODULE, "loopback");
-       if (!fabric) {
-               printk(KERN_ERR "tcm_loop_register_configfs() failed!\n");
-               return -1;
+       if (IS_ERR(fabric)) {
+               pr_err("tcm_loop_register_configfs() failed!\n");
+               return PTR_ERR(fabric);
        }
        /*
         * Setup the fabric API of function pointers used by target_core_mod
@@ -1435,20 +1405,12 @@ static int tcm_loop_register_configfs(void)
        fabric->tf_ops.tpg_release_fabric_acl =
                                        &tcm_loop_tpg_release_fabric_acl;
        fabric->tf_ops.tpg_get_inst_index = &tcm_loop_get_inst_index;
-       /*
-        * Since tcm_loop is mapping physical memory from Linux/SCSI
-        * struct scatterlist arrays for each struct scsi_cmnd I/O,
-        * we do not need TCM to allocate a iovec array for
-        * virtual memory address mappings
-        */
-       fabric->tf_ops.alloc_cmd_iovecs = NULL;
        /*
         * Used for setting up remaining TCM resources in process context
         */
        fabric->tf_ops.new_cmd_map = &tcm_loop_new_cmd_map;
        fabric->tf_ops.check_stop_free = &tcm_loop_check_stop_free;
-       fabric->tf_ops.release_cmd_to_pool = &tcm_loop_deallocate_core_cmd;
-       fabric->tf_ops.release_cmd_direct = &tcm_loop_deallocate_core_cmd;
+       fabric->tf_ops.release_cmd = &tcm_loop_release_cmd;
        fabric->tf_ops.shutdown_session = &tcm_loop_shutdown_session;
        fabric->tf_ops.close_session = &tcm_loop_close_session;
        fabric->tf_ops.stop_session = &tcm_loop_stop_session;
@@ -1465,7 +1427,6 @@ static int tcm_loop_register_configfs(void)
                                        &tcm_loop_set_default_node_attributes;
        fabric->tf_ops.get_task_tag = &tcm_loop_get_task_tag;
        fabric->tf_ops.get_cmd_state = &tcm_loop_get_cmd_state;
-       fabric->tf_ops.new_cmd_failure = &tcm_loop_new_cmd_failure;
        fabric->tf_ops.queue_data_in = &tcm_loop_queue_data_in;
        fabric->tf_ops.queue_status = &tcm_loop_queue_status;
        fabric->tf_ops.queue_tm_rsp = &tcm_loop_queue_tm_rsp;
@@ -1503,7 +1464,7 @@ static int tcm_loop_register_configfs(void)
         */
        ret = target_fabric_configfs_register(fabric);
        if (ret < 0) {
-               printk(KERN_ERR "target_fabric_configfs_register() for"
+               pr_err("target_fabric_configfs_register() for"
                                " TCM_Loop failed!\n");
                target_fabric_configfs_free(fabric);
                return -1;
@@ -1512,7 +1473,7 @@ static int tcm_loop_register_configfs(void)
         * Setup our local pointer to *fabric.
         */
        tcm_loop_fabric_configfs = fabric;
-       printk(KERN_INFO "TCM_LOOP[0] - Set fabric ->"
+       pr_debug("TCM_LOOP[0] - Set fabric ->"
                        " tcm_loop_fabric_configfs\n");
        return 0;
 }
@@ -1524,7 +1485,7 @@ static void tcm_loop_deregister_configfs(void)
 
        target_fabric_configfs_deregister(tcm_loop_fabric_configfs);
        tcm_loop_fabric_configfs = NULL;
-       printk(KERN_INFO "TCM_LOOP[0] - Cleared"
+       pr_debug("TCM_LOOP[0] - Cleared"
                                " tcm_loop_fabric_configfs\n");
 }
 
@@ -1537,7 +1498,7 @@ static int __init tcm_loop_fabric_init(void)
                                __alignof__(struct tcm_loop_cmd),
                                0, NULL);
        if (!tcm_loop_cmd_cache) {
-               printk(KERN_ERR "kmem_cache_create() for"
+               pr_debug("kmem_cache_create() for"
                        " tcm_loop_cmd_cache failed\n");
                return -ENOMEM;
        }
index 7e9f7ab455487c1734377d7126670d4d6dd53570..6b76c7a22bb062fa692a89e89d903b747e67e157 100644 (file)
  */
 #define TL_SCSI_MAX_CMD_LEN            32
 
-#ifdef CONFIG_LOOPBACK_TARGET_CDB_DEBUG
-# define TL_CDB_DEBUG(x...)            printk(KERN_INFO x)
-#else
-# define TL_CDB_DEBUG(x...)
-#endif
-
 struct tcm_loop_cmd {
        /* State of Linux/SCSI CDB+Data descriptor */
        u32 sc_cmd_state;
index 47abb42d9c36b04abf0f2e3b735465bf96e0fb88..98c98a3a0250c9c9b74a373ea346e58a15b32844 100644 (file)
@@ -46,6 +46,14 @@ static int core_alua_set_tg_pt_secondary_state(
                struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem,
                struct se_port *port, int explict, int offline);
 
+static u16 alua_lu_gps_counter;
+static u32 alua_lu_gps_count;
+
+static DEFINE_SPINLOCK(lu_gps_lock);
+static LIST_HEAD(lu_gps_list);
+
+struct t10_alua_lu_gp *default_lu_gp;
+
 /*
  * REPORT_TARGET_PORT_GROUPS
  *
@@ -53,16 +61,18 @@ static int core_alua_set_tg_pt_secondary_state(
  */
 int core_emulate_report_target_port_groups(struct se_cmd *cmd)
 {
-       struct se_subsystem_dev *su_dev = SE_DEV(cmd)->se_sub_dev;
+       struct se_subsystem_dev *su_dev = cmd->se_dev->se_sub_dev;
        struct se_port *port;
        struct t10_alua_tg_pt_gp *tg_pt_gp;
        struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       unsigned char *buf;
        u32 rd_len = 0, off = 4; /* Skip over RESERVED area to first
                                    Target port group descriptor */
 
-       spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
-       list_for_each_entry(tg_pt_gp, &T10_ALUA(su_dev)->tg_pt_gps_list,
+       buf = transport_kmap_first_data_page(cmd);
+
+       spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
+       list_for_each_entry(tg_pt_gp, &su_dev->t10_alua.tg_pt_gps_list,
                        tg_pt_gp_list) {
                /*
                 * PREF: Preferred target port bit, determine if this
@@ -124,7 +134,7 @@ int core_emulate_report_target_port_groups(struct se_cmd *cmd)
                }
                spin_unlock(&tg_pt_gp->tg_pt_gp_lock);
        }
-       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
        /*
         * Set the RETURN DATA LENGTH set in the header of the DataIN Payload
         */
@@ -133,6 +143,8 @@ int core_emulate_report_target_port_groups(struct se_cmd *cmd)
        buf[2] = ((rd_len >> 8) & 0xff);
        buf[3] = (rd_len & 0xff);
 
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
@@ -143,45 +155,53 @@ int core_emulate_report_target_port_groups(struct se_cmd *cmd)
  */
 int core_emulate_set_target_port_groups(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
-       struct se_subsystem_dev *su_dev = SE_DEV(cmd)->se_sub_dev;
-       struct se_port *port, *l_port = SE_LUN(cmd)->lun_sep;
-       struct se_node_acl *nacl = SE_SESS(cmd)->se_node_acl;
+       struct se_device *dev = cmd->se_dev;
+       struct se_subsystem_dev *su_dev = dev->se_sub_dev;
+       struct se_port *port, *l_port = cmd->se_lun->lun_sep;
+       struct se_node_acl *nacl = cmd->se_sess->se_node_acl;
        struct t10_alua_tg_pt_gp *tg_pt_gp = NULL, *l_tg_pt_gp;
        struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem, *l_tg_pt_gp_mem;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
-       unsigned char *ptr = &buf[4]; /* Skip over RESERVED area in header */
+       unsigned char *buf;
+       unsigned char *ptr;
        u32 len = 4; /* Skip over RESERVED area in header */
        int alua_access_state, primary = 0, rc;
        u16 tg_pt_id, rtpi;
 
-       if (!(l_port))
+       if (!l_port)
                return PYX_TRANSPORT_LU_COMM_FAILURE;
+
+       buf = transport_kmap_first_data_page(cmd);
+
        /*
         * Determine if explict ALUA via SET_TARGET_PORT_GROUPS is allowed
         * for the local tg_pt_gp.
         */
        l_tg_pt_gp_mem = l_port->sep_alua_tg_pt_gp_mem;
-       if (!(l_tg_pt_gp_mem)) {
-               printk(KERN_ERR "Unable to access l_port->sep_alua_tg_pt_gp_mem\n");
-               return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+       if (!l_tg_pt_gp_mem) {
+               pr_err("Unable to access l_port->sep_alua_tg_pt_gp_mem\n");
+               rc = PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+               goto out;
        }
        spin_lock(&l_tg_pt_gp_mem->tg_pt_gp_mem_lock);
        l_tg_pt_gp = l_tg_pt_gp_mem->tg_pt_gp;
-       if (!(l_tg_pt_gp)) {
+       if (!l_tg_pt_gp) {
                spin_unlock(&l_tg_pt_gp_mem->tg_pt_gp_mem_lock);
-               printk(KERN_ERR "Unable to access *l_tg_pt_gp_mem->tg_pt_gp\n");
-               return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+               pr_err("Unable to access *l_tg_pt_gp_mem->tg_pt_gp\n");
+               rc = PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+               goto out;
        }
        rc = (l_tg_pt_gp->tg_pt_gp_alua_access_type & TPGS_EXPLICT_ALUA);
        spin_unlock(&l_tg_pt_gp_mem->tg_pt_gp_mem_lock);
 
-       if (!(rc)) {
-               printk(KERN_INFO "Unable to process SET_TARGET_PORT_GROUPS"
+       if (!rc) {
+               pr_debug("Unable to process SET_TARGET_PORT_GROUPS"
                                " while TPGS_EXPLICT_ALUA is disabled\n");
-               return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+               rc = PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
+               goto out;
        }
 
+       ptr = &buf[4]; /* Skip over RESERVED area in header */
+
        while (len < cmd->data_length) {
                alua_access_state = (ptr[0] & 0x0f);
                /*
@@ -201,7 +221,8 @@ int core_emulate_set_target_port_groups(struct se_cmd *cmd)
                         * REQUEST, and the additional sense code set to INVALID
                         * FIELD IN PARAMETER LIST.
                         */
-                       return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
+                       rc = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
+                       goto out;
                }
                rc = -1;
                /*
@@ -224,11 +245,11 @@ int core_emulate_set_target_port_groups(struct se_cmd *cmd)
                         * Locate the matching target port group ID from
                         * the global tg_pt_gp list
                         */
-                       spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                       spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
                        list_for_each_entry(tg_pt_gp,
-                                       &T10_ALUA(su_dev)->tg_pt_gps_list,
+                                       &su_dev->t10_alua.tg_pt_gps_list,
                                        tg_pt_gp_list) {
-                               if (!(tg_pt_gp->tg_pt_gp_valid_id))
+                               if (!tg_pt_gp->tg_pt_gp_valid_id)
                                        continue;
 
                                if (tg_pt_id != tg_pt_gp->tg_pt_gp_id)
@@ -236,24 +257,26 @@ int core_emulate_set_target_port_groups(struct se_cmd *cmd)
 
                                atomic_inc(&tg_pt_gp->tg_pt_gp_ref_cnt);
                                smp_mb__after_atomic_inc();
-                               spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                               spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
 
                                rc = core_alua_do_port_transition(tg_pt_gp,
                                                dev, l_port, nacl,
                                                alua_access_state, 1);
 
-                               spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                               spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
                                atomic_dec(&tg_pt_gp->tg_pt_gp_ref_cnt);
                                smp_mb__after_atomic_dec();
                                break;
                        }
-                       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
                        /*
                         * If not matching target port group ID can be located
                         * throw an exception with ASCQ: INVALID_PARAMETER_LIST
                         */
-                       if (rc != 0)
-                               return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
+                       if (rc != 0) {
+                               rc = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
+                               goto out;
+                       }
                } else {
                        /*
                         * Extact the RELATIVE TARGET PORT IDENTIFIER to identify
@@ -287,14 +310,19 @@ int core_emulate_set_target_port_groups(struct se_cmd *cmd)
                         * be located, throw an exception with ASCQ:
                         * INVALID_PARAMETER_LIST
                         */
-                       if (rc != 0)
-                               return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
+                       if (rc != 0) {
+                               rc = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
+                               goto out;
+                       }
                }
 
                ptr += 4;
                len += 4;
        }
 
+out:
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
@@ -464,13 +492,13 @@ static int core_alua_state_check(
        unsigned char *cdb,
        u8 *alua_ascq)
 {
-       struct se_lun *lun = SE_LUN(cmd);
+       struct se_lun *lun = cmd->se_lun;
        struct se_port *port = lun->lun_sep;
        struct t10_alua_tg_pt_gp *tg_pt_gp;
        struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
        int out_alua_state, nonop_delay_msecs;
 
-       if (!(port))
+       if (!port)
                return 0;
        /*
         * First, check for a struct se_port specific secondary ALUA target port
@@ -478,7 +506,7 @@ static int core_alua_state_check(
         */
        if (atomic_read(&port->sep_tg_pt_secondary_offline)) {
                *alua_ascq = ASCQ_04H_ALUA_OFFLINE;
-               printk(KERN_INFO "ALUA: Got secondary offline status for local"
+               pr_debug("ALUA: Got secondary offline status for local"
                                " target port\n");
                *alua_ascq = ASCQ_04H_ALUA_OFFLINE;
                return 1;
@@ -520,9 +548,9 @@ static int core_alua_state_check(
         */
        case ALUA_ACCESS_STATE_OFFLINE:
        default:
-               printk(KERN_ERR "Unknown ALUA access state: 0x%02x\n",
+               pr_err("Unknown ALUA access state: 0x%02x\n",
                                out_alua_state);
-               return -1;
+               return -EINVAL;
        }
 
        return 0;
@@ -552,8 +580,8 @@ static int core_alua_check_transition(int state, int *primary)
                *primary = 0;
                break;
        default:
-               printk(KERN_ERR "Unknown ALUA access state: 0x%02x\n", state);
-               return -1;
+               pr_err("Unknown ALUA access state: 0x%02x\n", state);
+               return -EINVAL;
        }
 
        return 0;
@@ -610,7 +638,7 @@ int core_alua_check_nonop_delay(
         * The ALUA Active/NonOptimized access state delay can be disabled
         * in via configfs with a value of zero
         */
-       if (!(cmd->alua_nonop_delay))
+       if (!cmd->alua_nonop_delay)
                return 0;
        /*
         * struct se_cmd->alua_nonop_delay gets set by a target port group
@@ -639,7 +667,7 @@ static int core_alua_write_tpg_metadata(
 
        file = filp_open(path, flags, 0600);
        if (IS_ERR(file) || !file || !file->f_dentry) {
-               printk(KERN_ERR "filp_open(%s) for ALUA metadata failed\n",
+               pr_err("filp_open(%s) for ALUA metadata failed\n",
                        path);
                return -ENODEV;
        }
@@ -653,7 +681,7 @@ static int core_alua_write_tpg_metadata(
        set_fs(old_fs);
 
        if (ret < 0) {
-               printk(KERN_ERR "Error writing ALUA metadata file: %s\n", path);
+               pr_err("Error writing ALUA metadata file: %s\n", path);
                filp_close(file, NULL);
                return -EIO;
        }
@@ -750,7 +778,7 @@ static int core_alua_do_transition_tg_pt(
                         * se_deve->se_lun_acl pointer may be NULL for a
                         * entry created without explict Node+MappedLUN ACLs
                         */
-                       if (!(lacl))
+                       if (!lacl)
                                continue;
 
                        if (explict &&
@@ -792,7 +820,7 @@ static int core_alua_do_transition_tg_pt(
         */
        atomic_set(&tg_pt_gp->tg_pt_gp_alua_access_state, new_state);
 
-       printk(KERN_INFO "Successful %s ALUA transition TG PT Group: %s ID: %hu"
+       pr_debug("Successful %s ALUA transition TG PT Group: %s ID: %hu"
                " from primary access state %s to %s\n", (explict) ? "explict" :
                "implict", config_item_name(&tg_pt_gp->tg_pt_gp_group.cg_item),
                tg_pt_gp->tg_pt_gp_id, core_alua_dump_state(old_state),
@@ -823,8 +851,8 @@ int core_alua_do_port_transition(
                return -EINVAL;
 
        md_buf = kzalloc(l_tg_pt_gp->tg_pt_gp_md_buf_len, GFP_KERNEL);
-       if (!(md_buf)) {
-               printk("Unable to allocate buf for ALUA metadata\n");
+       if (!md_buf) {
+               pr_err("Unable to allocate buf for ALUA metadata\n");
                return -ENOMEM;
        }
 
@@ -839,7 +867,7 @@ int core_alua_do_port_transition(
         * we only do transition on the passed *l_tp_pt_gp, and not
         * on all of the matching target port groups IDs in default_lu_gp.
         */
-       if (!(lu_gp->lu_gp_id)) {
+       if (!lu_gp->lu_gp_id) {
                /*
                 * core_alua_do_transition_tg_pt() will always return
                 * success.
@@ -866,12 +894,12 @@ int core_alua_do_port_transition(
                smp_mb__after_atomic_inc();
                spin_unlock(&lu_gp->lu_gp_lock);
 
-               spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+               spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
                list_for_each_entry(tg_pt_gp,
-                               &T10_ALUA(su_dev)->tg_pt_gps_list,
+                               &su_dev->t10_alua.tg_pt_gps_list,
                                tg_pt_gp_list) {
 
-                       if (!(tg_pt_gp->tg_pt_gp_valid_id))
+                       if (!tg_pt_gp->tg_pt_gp_valid_id)
                                continue;
                        /*
                         * If the target behavior port asymmetric access state
@@ -893,7 +921,7 @@ int core_alua_do_port_transition(
                        }
                        atomic_inc(&tg_pt_gp->tg_pt_gp_ref_cnt);
                        smp_mb__after_atomic_inc();
-                       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
                        /*
                         * core_alua_do_transition_tg_pt() will always return
                         * success.
@@ -901,11 +929,11 @@ int core_alua_do_port_transition(
                        core_alua_do_transition_tg_pt(tg_pt_gp, port,
                                        nacl, md_buf, new_state, explict);
 
-                       spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                       spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
                        atomic_dec(&tg_pt_gp->tg_pt_gp_ref_cnt);
                        smp_mb__after_atomic_dec();
                }
-               spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+               spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
 
                spin_lock(&lu_gp->lu_gp_lock);
                atomic_dec(&lu_gp_mem->lu_gp_mem_ref_cnt);
@@ -913,7 +941,7 @@ int core_alua_do_port_transition(
        }
        spin_unlock(&lu_gp->lu_gp_lock);
 
-       printk(KERN_INFO "Successfully processed LU Group: %s all ALUA TG PT"
+       pr_debug("Successfully processed LU Group: %s all ALUA TG PT"
                " Group IDs: %hu %s transition to primary state: %s\n",
                config_item_name(&lu_gp->lu_gp_group.cg_item),
                l_tg_pt_gp->tg_pt_gp_id, (explict) ? "explict" : "implict",
@@ -942,11 +970,11 @@ static int core_alua_update_tpg_secondary_metadata(
        memset(wwn, 0, ALUA_SECONDARY_METADATA_WWN_LEN);
 
        len = snprintf(wwn, ALUA_SECONDARY_METADATA_WWN_LEN, "%s",
-                       TPG_TFO(se_tpg)->tpg_get_wwn(se_tpg));
+                       se_tpg->se_tpg_tfo->tpg_get_wwn(se_tpg));
 
-       if (TPG_TFO(se_tpg)->tpg_get_tag != NULL)
+       if (se_tpg->se_tpg_tfo->tpg_get_tag != NULL)
                snprintf(wwn+len, ALUA_SECONDARY_METADATA_WWN_LEN-len, "+%hu",
-                               TPG_TFO(se_tpg)->tpg_get_tag(se_tpg));
+                               se_tpg->se_tpg_tfo->tpg_get_tag(se_tpg));
 
        len = snprintf(md_buf, md_buf_len, "alua_tg_pt_offline=%d\n"
                        "alua_tg_pt_status=0x%02x\n",
@@ -954,7 +982,7 @@ static int core_alua_update_tpg_secondary_metadata(
                        port->sep_tg_pt_secondary_stat);
 
        snprintf(path, ALUA_METADATA_PATH_LEN, "/var/target/alua/%s/%s/lun_%u",
-                       TPG_TFO(se_tpg)->get_fabric_name(), wwn,
+                       se_tpg->se_tpg_tfo->get_fabric_name(), wwn,
                        port->sep_lun->unpacked_lun);
 
        return core_alua_write_tpg_metadata(path, md_buf, len);
@@ -973,11 +1001,11 @@ static int core_alua_set_tg_pt_secondary_state(
 
        spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
        tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
-       if (!(tg_pt_gp)) {
+       if (!tg_pt_gp) {
                spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
-               printk(KERN_ERR "Unable to complete secondary state"
+               pr_err("Unable to complete secondary state"
                                " transition\n");
-               return -1;
+               return -EINVAL;
        }
        trans_delay_msecs = tg_pt_gp->tg_pt_gp_trans_delay_msecs;
        /*
@@ -994,7 +1022,7 @@ static int core_alua_set_tg_pt_secondary_state(
                        ALUA_STATUS_ALTERED_BY_EXPLICT_STPG :
                        ALUA_STATUS_ALTERED_BY_IMPLICT_ALUA;
 
-       printk(KERN_INFO "Successful %s ALUA transition TG PT Group: %s ID: %hu"
+       pr_debug("Successful %s ALUA transition TG PT Group: %s ID: %hu"
                " to secondary access state: %s\n", (explict) ? "explict" :
                "implict", config_item_name(&tg_pt_gp->tg_pt_gp_group.cg_item),
                tg_pt_gp->tg_pt_gp_id, (offline) ? "OFFLINE" : "ONLINE");
@@ -1012,10 +1040,10 @@ static int core_alua_set_tg_pt_secondary_state(
         */
        if (port->sep_tg_pt_secondary_write_md) {
                md_buf = kzalloc(md_buf_len, GFP_KERNEL);
-               if (!(md_buf)) {
-                       printk(KERN_ERR "Unable to allocate md_buf for"
+               if (!md_buf) {
+                       pr_err("Unable to allocate md_buf for"
                                " secondary ALUA access metadata\n");
-                       return -1;
+                       return -ENOMEM;
                }
                mutex_lock(&port->sep_tg_pt_md_mutex);
                core_alua_update_tpg_secondary_metadata(tg_pt_gp_mem, port,
@@ -1034,19 +1062,19 @@ core_alua_allocate_lu_gp(const char *name, int def_group)
        struct t10_alua_lu_gp *lu_gp;
 
        lu_gp = kmem_cache_zalloc(t10_alua_lu_gp_cache, GFP_KERNEL);
-       if (!(lu_gp)) {
-               printk(KERN_ERR "Unable to allocate struct t10_alua_lu_gp\n");
+       if (!lu_gp) {
+               pr_err("Unable to allocate struct t10_alua_lu_gp\n");
                return ERR_PTR(-ENOMEM);
        }
-       INIT_LIST_HEAD(&lu_gp->lu_gp_list);
+       INIT_LIST_HEAD(&lu_gp->lu_gp_node);
        INIT_LIST_HEAD(&lu_gp->lu_gp_mem_list);
        spin_lock_init(&lu_gp->lu_gp_lock);
        atomic_set(&lu_gp->lu_gp_ref_cnt, 0);
 
        if (def_group) {
-               lu_gp->lu_gp_id = se_global->alua_lu_gps_counter++;
+               lu_gp->lu_gp_id = alua_lu_gps_counter++;
                lu_gp->lu_gp_valid_id = 1;
-               se_global->alua_lu_gps_count++;
+               alua_lu_gps_count++;
        }
 
        return lu_gp;
@@ -1060,41 +1088,41 @@ int core_alua_set_lu_gp_id(struct t10_alua_lu_gp *lu_gp, u16 lu_gp_id)
         * The lu_gp->lu_gp_id may only be set once..
         */
        if (lu_gp->lu_gp_valid_id) {
-               printk(KERN_WARNING "ALUA LU Group already has a valid ID,"
+               pr_warn("ALUA LU Group already has a valid ID,"
                        " ignoring request\n");
-               return -1;
+               return -EINVAL;
        }
 
-       spin_lock(&se_global->lu_gps_lock);
-       if (se_global->alua_lu_gps_count == 0x0000ffff) {
-               printk(KERN_ERR "Maximum ALUA se_global->alua_lu_gps_count:"
+       spin_lock(&lu_gps_lock);
+       if (alua_lu_gps_count == 0x0000ffff) {
+               pr_err("Maximum ALUA alua_lu_gps_count:"
                                " 0x0000ffff reached\n");
-               spin_unlock(&se_global->lu_gps_lock);
+               spin_unlock(&lu_gps_lock);
                kmem_cache_free(t10_alua_lu_gp_cache, lu_gp);
-               return -1;
+               return -ENOSPC;
        }
 again:
        lu_gp_id_tmp = (lu_gp_id != 0) ? lu_gp_id :
-                               se_global->alua_lu_gps_counter++;
+                               alua_lu_gps_counter++;
 
-       list_for_each_entry(lu_gp_tmp, &se_global->g_lu_gps_list, lu_gp_list) {
+       list_for_each_entry(lu_gp_tmp, &lu_gps_list, lu_gp_node) {
                if (lu_gp_tmp->lu_gp_id == lu_gp_id_tmp) {
-                       if (!(lu_gp_id))
+                       if (!lu_gp_id)
                                goto again;
 
-                       printk(KERN_WARNING "ALUA Logical Unit Group ID: %hu"
+                       pr_warn("ALUA Logical Unit Group ID: %hu"
                                " already exists, ignoring request\n",
                                lu_gp_id);
-                       spin_unlock(&se_global->lu_gps_lock);
-                       return -1;
+                       spin_unlock(&lu_gps_lock);
+                       return -EINVAL;
                }
        }
 
        lu_gp->lu_gp_id = lu_gp_id_tmp;
        lu_gp->lu_gp_valid_id = 1;
-       list_add_tail(&lu_gp->lu_gp_list, &se_global->g_lu_gps_list);
-       se_global->alua_lu_gps_count++;
-       spin_unlock(&se_global->lu_gps_lock);
+       list_add_tail(&lu_gp->lu_gp_node, &lu_gps_list);
+       alua_lu_gps_count++;
+       spin_unlock(&lu_gps_lock);
 
        return 0;
 }
@@ -1105,8 +1133,8 @@ core_alua_allocate_lu_gp_mem(struct se_device *dev)
        struct t10_alua_lu_gp_member *lu_gp_mem;
 
        lu_gp_mem = kmem_cache_zalloc(t10_alua_lu_gp_mem_cache, GFP_KERNEL);
-       if (!(lu_gp_mem)) {
-               printk(KERN_ERR "Unable to allocate struct t10_alua_lu_gp_member\n");
+       if (!lu_gp_mem) {
+               pr_err("Unable to allocate struct t10_alua_lu_gp_member\n");
                return ERR_PTR(-ENOMEM);
        }
        INIT_LIST_HEAD(&lu_gp_mem->lu_gp_mem_list);
@@ -1130,11 +1158,11 @@ void core_alua_free_lu_gp(struct t10_alua_lu_gp *lu_gp)
         * no associations can be made while we are releasing
         * struct t10_alua_lu_gp.
         */
-       spin_lock(&se_global->lu_gps_lock);
+       spin_lock(&lu_gps_lock);
        atomic_set(&lu_gp->lu_gp_shutdown, 1);
-       list_del(&lu_gp->lu_gp_list);
-       se_global->alua_lu_gps_count--;
-       spin_unlock(&se_global->lu_gps_lock);
+       list_del(&lu_gp->lu_gp_node);
+       alua_lu_gps_count--;
+       spin_unlock(&lu_gps_lock);
        /*
         * Allow struct t10_alua_lu_gp * referenced by core_alua_get_lu_gp_by_name()
         * in target_core_configfs.c:target_core_store_alua_lu_gp() to be
@@ -1165,9 +1193,9 @@ void core_alua_free_lu_gp(struct t10_alua_lu_gp *lu_gp)
                 * we want to re-assocate a given lu_gp_mem with default_lu_gp.
                 */
                spin_lock(&lu_gp_mem->lu_gp_mem_lock);
-               if (lu_gp != se_global->default_lu_gp)
+               if (lu_gp != default_lu_gp)
                        __core_alua_attach_lu_gp_mem(lu_gp_mem,
-                                       se_global->default_lu_gp);
+                                       default_lu_gp);
                else
                        lu_gp_mem->lu_gp = NULL;
                spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
@@ -1182,7 +1210,7 @@ void core_alua_free_lu_gp(struct t10_alua_lu_gp *lu_gp)
 void core_alua_free_lu_gp_mem(struct se_device *dev)
 {
        struct se_subsystem_dev *su_dev = dev->se_sub_dev;
-       struct t10_alua *alua = T10_ALUA(su_dev);
+       struct t10_alua *alua = &su_dev->t10_alua;
        struct t10_alua_lu_gp *lu_gp;
        struct t10_alua_lu_gp_member *lu_gp_mem;
 
@@ -1190,7 +1218,7 @@ void core_alua_free_lu_gp_mem(struct se_device *dev)
                return;
 
        lu_gp_mem = dev->dev_alua_lu_gp_mem;
-       if (!(lu_gp_mem))
+       if (!lu_gp_mem)
                return;
 
        while (atomic_read(&lu_gp_mem->lu_gp_mem_ref_cnt))
@@ -1198,7 +1226,7 @@ void core_alua_free_lu_gp_mem(struct se_device *dev)
 
        spin_lock(&lu_gp_mem->lu_gp_mem_lock);
        lu_gp = lu_gp_mem->lu_gp;
-       if ((lu_gp)) {
+       if (lu_gp) {
                spin_lock(&lu_gp->lu_gp_lock);
                if (lu_gp_mem->lu_gp_assoc) {
                        list_del(&lu_gp_mem->lu_gp_mem_list);
@@ -1218,27 +1246,27 @@ struct t10_alua_lu_gp *core_alua_get_lu_gp_by_name(const char *name)
        struct t10_alua_lu_gp *lu_gp;
        struct config_item *ci;
 
-       spin_lock(&se_global->lu_gps_lock);
-       list_for_each_entry(lu_gp, &se_global->g_lu_gps_list, lu_gp_list) {
-               if (!(lu_gp->lu_gp_valid_id))
+       spin_lock(&lu_gps_lock);
+       list_for_each_entry(lu_gp, &lu_gps_list, lu_gp_node) {
+               if (!lu_gp->lu_gp_valid_id)
                        continue;
                ci = &lu_gp->lu_gp_group.cg_item;
-               if (!(strcmp(config_item_name(ci), name))) {
+               if (!strcmp(config_item_name(ci), name)) {
                        atomic_inc(&lu_gp->lu_gp_ref_cnt);
-                       spin_unlock(&se_global->lu_gps_lock);
+                       spin_unlock(&lu_gps_lock);
                        return lu_gp;
                }
        }
-       spin_unlock(&se_global->lu_gps_lock);
+       spin_unlock(&lu_gps_lock);
 
        return NULL;
 }
 
 void core_alua_put_lu_gp_from_name(struct t10_alua_lu_gp *lu_gp)
 {
-       spin_lock(&se_global->lu_gps_lock);
+       spin_lock(&lu_gps_lock);
        atomic_dec(&lu_gp->lu_gp_ref_cnt);
-       spin_unlock(&se_global->lu_gps_lock);
+       spin_unlock(&lu_gps_lock);
 }
 
 /*
@@ -1279,8 +1307,8 @@ struct t10_alua_tg_pt_gp *core_alua_allocate_tg_pt_gp(
        struct t10_alua_tg_pt_gp *tg_pt_gp;
 
        tg_pt_gp = kmem_cache_zalloc(t10_alua_tg_pt_gp_cache, GFP_KERNEL);
-       if (!(tg_pt_gp)) {
-               printk(KERN_ERR "Unable to allocate struct t10_alua_tg_pt_gp\n");
+       if (!tg_pt_gp) {
+               pr_err("Unable to allocate struct t10_alua_tg_pt_gp\n");
                return NULL;
        }
        INIT_LIST_HEAD(&tg_pt_gp->tg_pt_gp_list);
@@ -1304,14 +1332,14 @@ struct t10_alua_tg_pt_gp *core_alua_allocate_tg_pt_gp(
        tg_pt_gp->tg_pt_gp_trans_delay_msecs = ALUA_DEFAULT_TRANS_DELAY_MSECS;
 
        if (def_group) {
-               spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+               spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
                tg_pt_gp->tg_pt_gp_id =
-                               T10_ALUA(su_dev)->alua_tg_pt_gps_counter++;
+                               su_dev->t10_alua.alua_tg_pt_gps_counter++;
                tg_pt_gp->tg_pt_gp_valid_id = 1;
-               T10_ALUA(su_dev)->alua_tg_pt_gps_count++;
+               su_dev->t10_alua.alua_tg_pt_gps_count++;
                list_add_tail(&tg_pt_gp->tg_pt_gp_list,
-                             &T10_ALUA(su_dev)->tg_pt_gps_list);
-               spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                             &su_dev->t10_alua.tg_pt_gps_list);
+               spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
        }
 
        return tg_pt_gp;
@@ -1328,42 +1356,42 @@ int core_alua_set_tg_pt_gp_id(
         * The tg_pt_gp->tg_pt_gp_id may only be set once..
         */
        if (tg_pt_gp->tg_pt_gp_valid_id) {
-               printk(KERN_WARNING "ALUA TG PT Group already has a valid ID,"
+               pr_warn("ALUA TG PT Group already has a valid ID,"
                        " ignoring request\n");
-               return -1;
+               return -EINVAL;
        }
 
-       spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
-       if (T10_ALUA(su_dev)->alua_tg_pt_gps_count == 0x0000ffff) {
-               printk(KERN_ERR "Maximum ALUA alua_tg_pt_gps_count:"
+       spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
+       if (su_dev->t10_alua.alua_tg_pt_gps_count == 0x0000ffff) {
+               pr_err("Maximum ALUA alua_tg_pt_gps_count:"
                        " 0x0000ffff reached\n");
-               spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+               spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
                kmem_cache_free(t10_alua_tg_pt_gp_cache, tg_pt_gp);
-               return -1;
+               return -ENOSPC;
        }
 again:
        tg_pt_gp_id_tmp = (tg_pt_gp_id != 0) ? tg_pt_gp_id :
-                       T10_ALUA(su_dev)->alua_tg_pt_gps_counter++;
+                       su_dev->t10_alua.alua_tg_pt_gps_counter++;
 
-       list_for_each_entry(tg_pt_gp_tmp, &T10_ALUA(su_dev)->tg_pt_gps_list,
+       list_for_each_entry(tg_pt_gp_tmp, &su_dev->t10_alua.tg_pt_gps_list,
                        tg_pt_gp_list) {
                if (tg_pt_gp_tmp->tg_pt_gp_id == tg_pt_gp_id_tmp) {
-                       if (!(tg_pt_gp_id))
+                       if (!tg_pt_gp_id)
                                goto again;
 
-                       printk(KERN_ERR "ALUA Target Port Group ID: %hu already"
+                       pr_err("ALUA Target Port Group ID: %hu already"
                                " exists, ignoring request\n", tg_pt_gp_id);
-                       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
-                       return -1;
+                       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
+                       return -EINVAL;
                }
        }
 
        tg_pt_gp->tg_pt_gp_id = tg_pt_gp_id_tmp;
        tg_pt_gp->tg_pt_gp_valid_id = 1;
        list_add_tail(&tg_pt_gp->tg_pt_gp_list,
-                       &T10_ALUA(su_dev)->tg_pt_gps_list);
-       T10_ALUA(su_dev)->alua_tg_pt_gps_count++;
-       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                       &su_dev->t10_alua.tg_pt_gps_list);
+       su_dev->t10_alua.alua_tg_pt_gps_count++;
+       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
 
        return 0;
 }
@@ -1375,8 +1403,8 @@ struct t10_alua_tg_pt_gp_member *core_alua_allocate_tg_pt_gp_mem(
 
        tg_pt_gp_mem = kmem_cache_zalloc(t10_alua_tg_pt_gp_mem_cache,
                                GFP_KERNEL);
-       if (!(tg_pt_gp_mem)) {
-               printk(KERN_ERR "Unable to allocate struct t10_alua_tg_pt_gp_member\n");
+       if (!tg_pt_gp_mem) {
+               pr_err("Unable to allocate struct t10_alua_tg_pt_gp_member\n");
                return ERR_PTR(-ENOMEM);
        }
        INIT_LIST_HEAD(&tg_pt_gp_mem->tg_pt_gp_mem_list);
@@ -1403,10 +1431,10 @@ void core_alua_free_tg_pt_gp(
         * no assications *OR* explict ALUA via SET_TARGET_PORT_GROUPS
         * can be made while we are releasing struct t10_alua_tg_pt_gp.
         */
-       spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+       spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
        list_del(&tg_pt_gp->tg_pt_gp_list);
-       T10_ALUA(su_dev)->alua_tg_pt_gps_counter--;
-       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+       su_dev->t10_alua.alua_tg_pt_gps_counter--;
+       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
        /*
         * Allow a struct t10_alua_tg_pt_gp_member * referenced by
         * core_alua_get_tg_pt_gp_by_name() in
@@ -1438,9 +1466,9 @@ void core_alua_free_tg_pt_gp(
                 * default_tg_pt_gp.
                 */
                spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
-               if (tg_pt_gp != T10_ALUA(su_dev)->default_tg_pt_gp) {
+               if (tg_pt_gp != su_dev->t10_alua.default_tg_pt_gp) {
                        __core_alua_attach_tg_pt_gp_mem(tg_pt_gp_mem,
-                                       T10_ALUA(su_dev)->default_tg_pt_gp);
+                                       su_dev->t10_alua.default_tg_pt_gp);
                } else
                        tg_pt_gp_mem->tg_pt_gp = NULL;
                spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
@@ -1455,7 +1483,7 @@ void core_alua_free_tg_pt_gp(
 void core_alua_free_tg_pt_gp_mem(struct se_port *port)
 {
        struct se_subsystem_dev *su_dev = port->sep_lun->lun_se_dev->se_sub_dev;
-       struct t10_alua *alua = T10_ALUA(su_dev);
+       struct t10_alua *alua = &su_dev->t10_alua;
        struct t10_alua_tg_pt_gp *tg_pt_gp;
        struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
 
@@ -1463,7 +1491,7 @@ void core_alua_free_tg_pt_gp_mem(struct se_port *port)
                return;
 
        tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
-       if (!(tg_pt_gp_mem))
+       if (!tg_pt_gp_mem)
                return;
 
        while (atomic_read(&tg_pt_gp_mem->tg_pt_gp_mem_ref_cnt))
@@ -1471,7 +1499,7 @@ void core_alua_free_tg_pt_gp_mem(struct se_port *port)
 
        spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
        tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
-       if ((tg_pt_gp)) {
+       if (tg_pt_gp) {
                spin_lock(&tg_pt_gp->tg_pt_gp_lock);
                if (tg_pt_gp_mem->tg_pt_gp_assoc) {
                        list_del(&tg_pt_gp_mem->tg_pt_gp_mem_list);
@@ -1493,19 +1521,19 @@ static struct t10_alua_tg_pt_gp *core_alua_get_tg_pt_gp_by_name(
        struct t10_alua_tg_pt_gp *tg_pt_gp;
        struct config_item *ci;
 
-       spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
-       list_for_each_entry(tg_pt_gp, &T10_ALUA(su_dev)->tg_pt_gps_list,
+       spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
+       list_for_each_entry(tg_pt_gp, &su_dev->t10_alua.tg_pt_gps_list,
                        tg_pt_gp_list) {
-               if (!(tg_pt_gp->tg_pt_gp_valid_id))
+               if (!tg_pt_gp->tg_pt_gp_valid_id)
                        continue;
                ci = &tg_pt_gp->tg_pt_gp_group.cg_item;
-               if (!(strcmp(config_item_name(ci), name))) {
+               if (!strcmp(config_item_name(ci), name)) {
                        atomic_inc(&tg_pt_gp->tg_pt_gp_ref_cnt);
-                       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+                       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
                        return tg_pt_gp;
                }
        }
-       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
 
        return NULL;
 }
@@ -1515,9 +1543,9 @@ static void core_alua_put_tg_pt_gp_from_name(
 {
        struct se_subsystem_dev *su_dev = tg_pt_gp->tg_pt_gp_su_dev;
 
-       spin_lock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+       spin_lock(&su_dev->t10_alua.tg_pt_gps_lock);
        atomic_dec(&tg_pt_gp->tg_pt_gp_ref_cnt);
-       spin_unlock(&T10_ALUA(su_dev)->tg_pt_gps_lock);
+       spin_unlock(&su_dev->t10_alua.tg_pt_gps_lock);
 }
 
 /*
@@ -1555,7 +1583,7 @@ ssize_t core_alua_show_tg_pt_gp_info(struct se_port *port, char *page)
 {
        struct se_subsystem_dev *su_dev = port->sep_lun->lun_se_dev->se_sub_dev;
        struct config_item *tg_pt_ci;
-       struct t10_alua *alua = T10_ALUA(su_dev);
+       struct t10_alua *alua = &su_dev->t10_alua;
        struct t10_alua_tg_pt_gp *tg_pt_gp;
        struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
        ssize_t len = 0;
@@ -1564,12 +1592,12 @@ ssize_t core_alua_show_tg_pt_gp_info(struct se_port *port, char *page)
                return len;
 
        tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
-       if (!(tg_pt_gp_mem))
+       if (!tg_pt_gp_mem)
                return len;
 
        spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
        tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
-       if ((tg_pt_gp)) {
+       if (tg_pt_gp) {
                tg_pt_ci = &tg_pt_gp->tg_pt_gp_group.cg_item;
                len += sprintf(page, "TG Port Alias: %s\nTG Port Group ID:"
                        " %hu\nTG Port Primary Access State: %s\nTG Port "
@@ -1605,16 +1633,16 @@ ssize_t core_alua_store_tg_pt_gp_info(
        tpg = port->sep_tpg;
        lun = port->sep_lun;
 
-       if (T10_ALUA(su_dev)->alua_type != SPC3_ALUA_EMULATED) {
-               printk(KERN_WARNING "SPC3_ALUA_EMULATED not enabled for"
-                       " %s/tpgt_%hu/%s\n", TPG_TFO(tpg)->tpg_get_wwn(tpg),
-                       TPG_TFO(tpg)->tpg_get_tag(tpg),
+       if (su_dev->t10_alua.alua_type != SPC3_ALUA_EMULATED) {
+               pr_warn("SPC3_ALUA_EMULATED not enabled for"
+                       " %s/tpgt_%hu/%s\n", tpg->se_tpg_tfo->tpg_get_wwn(tpg),
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg),
                        config_item_name(&lun->lun_group.cg_item));
                return -EINVAL;
        }
 
        if (count > TG_PT_GROUP_NAME_BUF) {
-               printk(KERN_ERR "ALUA Target Port Group alias too large!\n");
+               pr_err("ALUA Target Port Group alias too large!\n");
                return -EINVAL;
        }
        memset(buf, 0, TG_PT_GROUP_NAME_BUF);
@@ -1631,31 +1659,31 @@ ssize_t core_alua_store_tg_pt_gp_info(
                 */
                tg_pt_gp_new = core_alua_get_tg_pt_gp_by_name(su_dev,
                                        strstrip(buf));
-               if (!(tg_pt_gp_new))
+               if (!tg_pt_gp_new)
                        return -ENODEV;
        }
        tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
-       if (!(tg_pt_gp_mem)) {
+       if (!tg_pt_gp_mem) {
                if (tg_pt_gp_new)
                        core_alua_put_tg_pt_gp_from_name(tg_pt_gp_new);
-               printk(KERN_ERR "NULL struct se_port->sep_alua_tg_pt_gp_mem pointer\n");
+               pr_err("NULL struct se_port->sep_alua_tg_pt_gp_mem pointer\n");
                return -EINVAL;
        }
 
        spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
        tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
-       if ((tg_pt_gp)) {
+       if (tg_pt_gp) {
                /*
                 * Clearing an existing tg_pt_gp association, and replacing
                 * with the default_tg_pt_gp.
                 */
-               if (!(tg_pt_gp_new)) {
-                       printk(KERN_INFO "Target_Core_ConfigFS: Moving"
+               if (!tg_pt_gp_new) {
+                       pr_debug("Target_Core_ConfigFS: Moving"
                                " %s/tpgt_%hu/%s from ALUA Target Port Group:"
                                " alua/%s, ID: %hu back to"
                                " default_tg_pt_gp\n",
-                               TPG_TFO(tpg)->tpg_get_wwn(tpg),
-                               TPG_TFO(tpg)->tpg_get_tag(tpg),
+                               tpg->se_tpg_tfo->tpg_get_wwn(tpg),
+                               tpg->se_tpg_tfo->tpg_get_tag(tpg),
                                config_item_name(&lun->lun_group.cg_item),
                                config_item_name(
                                        &tg_pt_gp->tg_pt_gp_group.cg_item),
@@ -1663,7 +1691,7 @@ ssize_t core_alua_store_tg_pt_gp_info(
 
                        __core_alua_drop_tg_pt_gp_mem(tg_pt_gp_mem, tg_pt_gp);
                        __core_alua_attach_tg_pt_gp_mem(tg_pt_gp_mem,
-                                       T10_ALUA(su_dev)->default_tg_pt_gp);
+                                       su_dev->t10_alua.default_tg_pt_gp);
                        spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
 
                        return count;
@@ -1679,10 +1707,10 @@ ssize_t core_alua_store_tg_pt_gp_info(
         */
        __core_alua_attach_tg_pt_gp_mem(tg_pt_gp_mem, tg_pt_gp_new);
        spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
-       printk(KERN_INFO "Target_Core_ConfigFS: %s %s/tpgt_%hu/%s to ALUA"
+       pr_debug("Target_Core_ConfigFS: %s %s/tpgt_%hu/%s to ALUA"
                " Target Port Group: alua/%s, ID: %hu\n", (move) ?
-               "Moving" : "Adding", TPG_TFO(tpg)->tpg_get_wwn(tpg),
-               TPG_TFO(tpg)->tpg_get_tag(tpg),
+               "Moving" : "Adding", tpg->se_tpg_tfo->tpg_get_wwn(tpg),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg),
                config_item_name(&lun->lun_group.cg_item),
                config_item_name(&tg_pt_gp_new->tg_pt_gp_group.cg_item),
                tg_pt_gp_new->tg_pt_gp_id);
@@ -1716,11 +1744,11 @@ ssize_t core_alua_store_access_type(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract alua_access_type\n");
+               pr_err("Unable to extract alua_access_type\n");
                return -EINVAL;
        }
        if ((tmp != 0) && (tmp != 1) && (tmp != 2) && (tmp != 3)) {
-               printk(KERN_ERR "Illegal value for alua_access_type:"
+               pr_err("Illegal value for alua_access_type:"
                                " %lu\n", tmp);
                return -EINVAL;
        }
@@ -1754,11 +1782,11 @@ ssize_t core_alua_store_nonop_delay_msecs(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract nonop_delay_msecs\n");
+               pr_err("Unable to extract nonop_delay_msecs\n");
                return -EINVAL;
        }
        if (tmp > ALUA_MAX_NONOP_DELAY_MSECS) {
-               printk(KERN_ERR "Passed nonop_delay_msecs: %lu, exceeds"
+               pr_err("Passed nonop_delay_msecs: %lu, exceeds"
                        " ALUA_MAX_NONOP_DELAY_MSECS: %d\n", tmp,
                        ALUA_MAX_NONOP_DELAY_MSECS);
                return -EINVAL;
@@ -1785,11 +1813,11 @@ ssize_t core_alua_store_trans_delay_msecs(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract trans_delay_msecs\n");
+               pr_err("Unable to extract trans_delay_msecs\n");
                return -EINVAL;
        }
        if (tmp > ALUA_MAX_TRANS_DELAY_MSECS) {
-               printk(KERN_ERR "Passed trans_delay_msecs: %lu, exceeds"
+               pr_err("Passed trans_delay_msecs: %lu, exceeds"
                        " ALUA_MAX_TRANS_DELAY_MSECS: %d\n", tmp,
                        ALUA_MAX_TRANS_DELAY_MSECS);
                return -EINVAL;
@@ -1816,11 +1844,11 @@ ssize_t core_alua_store_preferred_bit(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract preferred ALUA value\n");
+               pr_err("Unable to extract preferred ALUA value\n");
                return -EINVAL;
        }
        if ((tmp != 0) && (tmp != 1)) {
-               printk(KERN_ERR "Illegal value for preferred ALUA: %lu\n", tmp);
+               pr_err("Illegal value for preferred ALUA: %lu\n", tmp);
                return -EINVAL;
        }
        tg_pt_gp->tg_pt_gp_pref = (int)tmp;
@@ -1830,7 +1858,7 @@ ssize_t core_alua_store_preferred_bit(
 
 ssize_t core_alua_show_offline_bit(struct se_lun *lun, char *page)
 {
-       if (!(lun->lun_sep))
+       if (!lun->lun_sep)
                return -ENODEV;
 
        return sprintf(page, "%d\n",
@@ -1846,22 +1874,22 @@ ssize_t core_alua_store_offline_bit(
        unsigned long tmp;
        int ret;
 
-       if (!(lun->lun_sep))
+       if (!lun->lun_sep)
                return -ENODEV;
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract alua_tg_pt_offline value\n");
+               pr_err("Unable to extract alua_tg_pt_offline value\n");
                return -EINVAL;
        }
        if ((tmp != 0) && (tmp != 1)) {
-               printk(KERN_ERR "Illegal value for alua_tg_pt_offline: %lu\n",
+               pr_err("Illegal value for alua_tg_pt_offline: %lu\n",
                                tmp);
                return -EINVAL;
        }
        tg_pt_gp_mem = lun->lun_sep->sep_alua_tg_pt_gp_mem;
-       if (!(tg_pt_gp_mem)) {
-               printk(KERN_ERR "Unable to locate *tg_pt_gp_mem\n");
+       if (!tg_pt_gp_mem) {
+               pr_err("Unable to locate *tg_pt_gp_mem\n");
                return -EINVAL;
        }
 
@@ -1890,13 +1918,13 @@ ssize_t core_alua_store_secondary_status(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract alua_tg_pt_status\n");
+               pr_err("Unable to extract alua_tg_pt_status\n");
                return -EINVAL;
        }
        if ((tmp != ALUA_STATUS_NONE) &&
            (tmp != ALUA_STATUS_ALTERED_BY_EXPLICT_STPG) &&
            (tmp != ALUA_STATUS_ALTERED_BY_IMPLICT_ALUA)) {
-               printk(KERN_ERR "Illegal value for alua_tg_pt_status: %lu\n",
+               pr_err("Illegal value for alua_tg_pt_status: %lu\n",
                                tmp);
                return -EINVAL;
        }
@@ -1923,11 +1951,11 @@ ssize_t core_alua_store_secondary_write_metadata(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract alua_tg_pt_write_md\n");
+               pr_err("Unable to extract alua_tg_pt_write_md\n");
                return -EINVAL;
        }
        if ((tmp != 0) && (tmp != 1)) {
-               printk(KERN_ERR "Illegal value for alua_tg_pt_write_md:"
+               pr_err("Illegal value for alua_tg_pt_write_md:"
                                " %lu\n", tmp);
                return -EINVAL;
        }
@@ -1939,7 +1967,7 @@ ssize_t core_alua_store_secondary_write_metadata(
 int core_setup_alua(struct se_device *dev, int force_pt)
 {
        struct se_subsystem_dev *su_dev = dev->se_sub_dev;
-       struct t10_alua *alua = T10_ALUA(su_dev);
+       struct t10_alua *alua = &su_dev->t10_alua;
        struct t10_alua_lu_gp_member *lu_gp_mem;
        /*
         * If this device is from Target_Core_Mod/pSCSI, use the ALUA logic
@@ -1947,44 +1975,44 @@ int core_setup_alua(struct se_device *dev, int force_pt)
         * cause a problem because libata and some SATA RAID HBAs appear
         * under Linux/SCSI, but emulate SCSI logic themselves.
         */
-       if (((TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) &&
-           !(DEV_ATTRIB(dev)->emulate_alua)) || force_pt) {
+       if (((dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) &&
+           !(dev->se_sub_dev->se_dev_attrib.emulate_alua)) || force_pt) {
                alua->alua_type = SPC_ALUA_PASSTHROUGH;
                alua->alua_state_check = &core_alua_state_check_nop;
-               printk(KERN_INFO "%s: Using SPC_ALUA_PASSTHROUGH, no ALUA"
-                       " emulation\n", TRANSPORT(dev)->name);
+               pr_debug("%s: Using SPC_ALUA_PASSTHROUGH, no ALUA"
+                       " emulation\n", dev->transport->name);
                return 0;
        }
        /*
         * If SPC-3 or above is reported by real or emulated struct se_device,
         * use emulated ALUA.
         */
-       if (TRANSPORT(dev)->get_device_rev(dev) >= SCSI_3) {
-               printk(KERN_INFO "%s: Enabling ALUA Emulation for SPC-3"
-                       " device\n", TRANSPORT(dev)->name);
+       if (dev->transport->get_device_rev(dev) >= SCSI_3) {
+               pr_debug("%s: Enabling ALUA Emulation for SPC-3"
+                       " device\n", dev->transport->name);
                /*
                 * Associate this struct se_device with the default ALUA
                 * LUN Group.
                 */
                lu_gp_mem = core_alua_allocate_lu_gp_mem(dev);
-               if (IS_ERR(lu_gp_mem) || !lu_gp_mem)
-                       return -1;
+               if (IS_ERR(lu_gp_mem))
+                       return PTR_ERR(lu_gp_mem);
 
                alua->alua_type = SPC3_ALUA_EMULATED;
                alua->alua_state_check = &core_alua_state_check;
                spin_lock(&lu_gp_mem->lu_gp_mem_lock);
                __core_alua_attach_lu_gp_mem(lu_gp_mem,
-                               se_global->default_lu_gp);
+                               default_lu_gp);
                spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
 
-               printk(KERN_INFO "%s: Adding to default ALUA LU Group:"
+               pr_debug("%s: Adding to default ALUA LU Group:"
                        " core/alua/lu_gps/default_lu_gp\n",
-                       TRANSPORT(dev)->name);
+                       dev->transport->name);
        } else {
                alua->alua_type = SPC2_ALUA_DISABLED;
                alua->alua_state_check = &core_alua_state_check_nop;
-               printk(KERN_INFO "%s: Disabling ALUA Emulation for SPC-2"
-                       " device\n", TRANSPORT(dev)->name);
+               pr_debug("%s: Disabling ALUA Emulation for SPC-2"
+                       " device\n", dev->transport->name);
        }
 
        return 0;
index 7f19c8b7b84c704c0ab74ffdef7a19dca63737e2..8ae09a1bdf74a8744b4884df7cddb792af6f8f0f 100644 (file)
@@ -23,6 +23,7 @@
  * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
  */
 
+#include <linux/kernel.h>
 #include <asm/unaligned.h>
 #include <scsi/scsi.h>
 
@@ -64,20 +65,22 @@ target_fill_alua_data(struct se_port *port, unsigned char *buf)
 static int
 target_emulate_inquiry_std(struct se_cmd *cmd)
 {
-       struct se_lun *lun = SE_LUN(cmd);
-       struct se_device *dev = SE_DEV(cmd);
-       unsigned char *buf = cmd->t_task->t_task_buf;
+       struct se_lun *lun = cmd->se_lun;
+       struct se_device *dev = cmd->se_dev;
+       unsigned char *buf;
 
        /*
         * Make sure we at least have 6 bytes of INQUIRY response
         * payload going back for EVPD=0
         */
        if (cmd->data_length < 6) {
-               printk(KERN_ERR "SCSI Inquiry payload length: %u"
+               pr_err("SCSI Inquiry payload length: %u"
                        " too small for EVPD=0\n", cmd->data_length);
-               return -1;
+               return -EINVAL;
        }
 
+       buf = transport_kmap_first_data_page(cmd);
+
        buf[0] = dev->transport->get_device_type(dev);
        if (buf[0] == TYPE_TAPE)
                buf[1] = 0x80;
@@ -86,12 +89,12 @@ target_emulate_inquiry_std(struct se_cmd *cmd)
        /*
         * Enable SCCS and TPGS fields for Emulated ALUA
         */
-       if (T10_ALUA(dev->se_sub_dev)->alua_type == SPC3_ALUA_EMULATED)
+       if (dev->se_sub_dev->t10_alua.alua_type == SPC3_ALUA_EMULATED)
                target_fill_alua_data(lun->lun_sep, buf);
 
        if (cmd->data_length < 8) {
                buf[4] = 1; /* Set additional length to 1 */
-               return 0;
+               goto out;
        }
 
        buf[7] = 0x32; /* Sync=1 and CmdQue=1 */
@@ -102,40 +105,18 @@ target_emulate_inquiry_std(struct se_cmd *cmd)
         */
        if (cmd->data_length < 36) {
                buf[4] = 3; /* Set additional length to 3 */
-               return 0;
+               goto out;
        }
 
        snprintf((unsigned char *)&buf[8], 8, "LIO-ORG");
        snprintf((unsigned char *)&buf[16], 16, "%s",
-                &DEV_T10_WWN(dev)->model[0]);
+                &dev->se_sub_dev->t10_wwn.model[0]);
        snprintf((unsigned char *)&buf[32], 4, "%s",
-                &DEV_T10_WWN(dev)->revision[0]);
+                &dev->se_sub_dev->t10_wwn.revision[0]);
        buf[4] = 31; /* Set additional length to 31 */
-       return 0;
-}
-
-/* supported vital product data pages */
-static int
-target_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf)
-{
-       buf[1] = 0x00;
-       if (cmd->data_length < 8)
-               return 0;
-
-       buf[4] = 0x0;
-       /*
-        * Only report the INQUIRY EVPD=1 pages after a valid NAA
-        * Registered Extended LUN WWN has been set via ConfigFS
-        * during device creation/restart.
-        */
-       if (SE_DEV(cmd)->se_sub_dev->su_dev_flags &
-                       SDF_EMULATED_VPD_UNIT_SERIAL) {
-               buf[3] = 3;
-               buf[5] = 0x80;
-               buf[6] = 0x83;
-               buf[7] = 0x86;
-       }
 
+out:
+       transport_kunmap_first_data_page(cmd);
        return 0;
 }
 
@@ -143,16 +124,15 @@ target_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf)
 static int
 target_emulate_evpd_80(struct se_cmd *cmd, unsigned char *buf)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        u16 len = 0;
 
-       buf[1] = 0x80;
        if (dev->se_sub_dev->su_dev_flags &
                        SDF_EMULATED_VPD_UNIT_SERIAL) {
                u32 unit_serial_len;
 
                unit_serial_len =
-                       strlen(&DEV_T10_WWN(dev)->unit_serial[0]);
+                       strlen(&dev->se_sub_dev->t10_wwn.unit_serial[0]);
                unit_serial_len++; /* For NULL Terminator */
 
                if (((len + 4) + unit_serial_len) > cmd->data_length) {
@@ -162,7 +142,7 @@ target_emulate_evpd_80(struct se_cmd *cmd, unsigned char *buf)
                        return 0;
                }
                len += sprintf((unsigned char *)&buf[4], "%s",
-                       &DEV_T10_WWN(dev)->unit_serial[0]);
+                       &dev->se_sub_dev->t10_wwn.unit_serial[0]);
                len++; /* Extra Byte for NULL Terminator */
                buf[3] = len;
        }
@@ -176,21 +156,18 @@ target_emulate_evpd_80(struct se_cmd *cmd, unsigned char *buf)
 static int
 target_emulate_evpd_83(struct se_cmd *cmd, unsigned char *buf)
 {
-       struct se_device *dev = SE_DEV(cmd);
-       struct se_lun *lun = SE_LUN(cmd);
+       struct se_device *dev = cmd->se_dev;
+       struct se_lun *lun = cmd->se_lun;
        struct se_port *port = NULL;
        struct se_portal_group *tpg = NULL;
        struct t10_alua_lu_gp_member *lu_gp_mem;
        struct t10_alua_tg_pt_gp *tg_pt_gp;
        struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
-       unsigned char binary, binary_new;
-       unsigned char *prod = &DEV_T10_WWN(dev)->model[0];
+       unsigned char *prod = &dev->se_sub_dev->t10_wwn.model[0];
        u32 prod_len;
        u32 unit_serial_len, off = 0;
-       int i;
        u16 len = 0, id_len;
 
-       buf[1] = 0x83;
        off = 4;
 
        /*
@@ -210,11 +187,11 @@ target_emulate_evpd_83(struct se_cmd *cmd, unsigned char *buf)
        /* CODE SET == Binary */
        buf[off++] = 0x1;
 
-       /* Set ASSOICATION == addressed logical unit: 0)b */
+       /* Set ASSOCIATION == addressed logical unit: 0)b */
        buf[off] = 0x00;
 
        /* Identifier/Designator type == NAA identifier */
-       buf[off++] = 0x3;
+       buf[off++] |= 0x3;
        off++;
 
        /* Identifier/Designator length */
@@ -237,16 +214,9 @@ target_emulate_evpd_83(struct se_cmd *cmd, unsigned char *buf)
         * VENDOR_SPECIFIC_IDENTIFIER and
         * VENDOR_SPECIFIC_IDENTIFIER_EXTENTION
         */
-       binary = transport_asciihex_to_binaryhex(
-                               &DEV_T10_WWN(dev)->unit_serial[0]);
-       buf[off++] |= (binary & 0xf0) >> 4;
-       for (i = 0; i < 24; i += 2) {
-               binary_new = transport_asciihex_to_binaryhex(
-                       &DEV_T10_WWN(dev)->unit_serial[i+2]);
-               buf[off] = (binary & 0x0f) << 4;
-               buf[off++] |= (binary_new & 0xf0) >> 4;
-               binary = binary_new;
-       }
+       buf[off++] |= hex_to_bin(dev->se_sub_dev->t10_wwn.unit_serial[0]);
+       hex2bin(&buf[off], &dev->se_sub_dev->t10_wwn.unit_serial[1], 12);
+
        len = 20;
        off = (len + 4);
 
@@ -263,7 +233,7 @@ check_t10_vend_desc:
        if (dev->se_sub_dev->su_dev_flags &
                        SDF_EMULATED_VPD_UNIT_SERIAL) {
                unit_serial_len =
-                       strlen(&DEV_T10_WWN(dev)->unit_serial[0]);
+                       strlen(&dev->se_sub_dev->t10_wwn.unit_serial[0]);
                unit_serial_len++; /* For NULL Terminator */
 
                if ((len + (id_len + 4) +
@@ -274,7 +244,7 @@ check_t10_vend_desc:
                }
                id_len += sprintf((unsigned char *)&buf[off+12],
                                "%s:%s", prod,
-                               &DEV_T10_WWN(dev)->unit_serial[0]);
+                               &dev->se_sub_dev->t10_wwn.unit_serial[0]);
        }
        buf[off] = 0x2; /* ASCII */
        buf[off+1] = 0x1; /* T10 Vendor ID */
@@ -312,10 +282,10 @@ check_port:
                        goto check_tpgi;
                }
                buf[off] =
-                       (TPG_TFO(tpg)->get_fabric_proto_ident(tpg) << 4);
+                       (tpg->se_tpg_tfo->get_fabric_proto_ident(tpg) << 4);
                buf[off++] |= 0x1; /* CODE SET == Binary */
                buf[off] = 0x80; /* Set PIV=1 */
-               /* Set ASSOICATION == target port: 01b */
+               /* Set ASSOCIATION == target port: 01b */
                buf[off] |= 0x10;
                /* DESIGNATOR TYPE == Relative target port identifer */
                buf[off++] |= 0x4;
@@ -335,7 +305,7 @@ check_port:
                 * section 7.5.1 Table 362
                 */
 check_tpgi:
-               if (T10_ALUA(dev->se_sub_dev)->alua_type !=
+               if (dev->se_sub_dev->t10_alua.alua_type !=
                                SPC3_ALUA_EMULATED)
                        goto check_scsi_name;
 
@@ -349,7 +319,7 @@ check_tpgi:
 
                spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
                tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
-               if (!(tg_pt_gp)) {
+               if (!tg_pt_gp) {
                        spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
                        goto check_lu_gp;
                }
@@ -357,10 +327,10 @@ check_tpgi:
                spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
 
                buf[off] =
-                       (TPG_TFO(tpg)->get_fabric_proto_ident(tpg) << 4);
+                       (tpg->se_tpg_tfo->get_fabric_proto_ident(tpg) << 4);
                buf[off++] |= 0x1; /* CODE SET == Binary */
                buf[off] = 0x80; /* Set PIV=1 */
-               /* Set ASSOICATION == target port: 01b */
+               /* Set ASSOCIATION == target port: 01b */
                buf[off] |= 0x10;
                /* DESIGNATOR TYPE == Target port group identifier */
                buf[off++] |= 0x5;
@@ -380,12 +350,12 @@ check_lu_gp:
                        goto check_scsi_name;
                }
                lu_gp_mem = dev->dev_alua_lu_gp_mem;
-               if (!(lu_gp_mem))
+               if (!lu_gp_mem)
                        goto check_scsi_name;
 
                spin_lock(&lu_gp_mem->lu_gp_mem_lock);
                lu_gp = lu_gp_mem->lu_gp;
-               if (!(lu_gp)) {
+               if (!lu_gp) {
                        spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
                        goto check_scsi_name;
                }
@@ -409,7 +379,7 @@ check_lu_gp:
                 * section 7.5.1 Table 362
                 */
 check_scsi_name:
-               scsi_name_len = strlen(TPG_TFO(tpg)->tpg_get_wwn(tpg));
+               scsi_name_len = strlen(tpg->se_tpg_tfo->tpg_get_wwn(tpg));
                /* UTF-8 ",t,0x<16-bit TPGT>" + NULL Terminator */
                scsi_name_len += 10;
                /* Check for 4-byte padding */
@@ -424,10 +394,10 @@ check_scsi_name:
                        goto set_len;
                }
                buf[off] =
-                       (TPG_TFO(tpg)->get_fabric_proto_ident(tpg) << 4);
+                       (tpg->se_tpg_tfo->get_fabric_proto_ident(tpg) << 4);
                buf[off++] |= 0x3; /* CODE SET == UTF-8 */
                buf[off] = 0x80; /* Set PIV=1 */
-               /* Set ASSOICATION == target port: 01b */
+               /* Set ASSOCIATION == target port: 01b */
                buf[off] |= 0x10;
                /* DESIGNATOR TYPE == SCSI name string */
                buf[off++] |= 0x8;
@@ -438,9 +408,9 @@ check_scsi_name:
                 * Target Port, this means "<iSCSI name>,t,0x<TPGT> in
                 * UTF-8 encoding.
                 */
-               tpgt = TPG_TFO(tpg)->tpg_get_tag(tpg);
+               tpgt = tpg->se_tpg_tfo->tpg_get_tag(tpg);
                scsi_name_len = sprintf(&buf[off], "%s,t,0x%04x",
-                                       TPG_TFO(tpg)->tpg_get_wwn(tpg), tpgt);
+                                       tpg->se_tpg_tfo->tpg_get_wwn(tpg), tpgt);
                scsi_name_len += 1 /* Include  NULL terminator */;
                /*
                 * The null-terminated, null-padded (see 4.4.2) SCSI
@@ -471,13 +441,12 @@ target_emulate_evpd_86(struct se_cmd *cmd, unsigned char *buf)
        if (cmd->data_length < 60)
                return 0;
 
-       buf[1] = 0x86;
        buf[2] = 0x3c;
        /* Set HEADSUP, ORDSUP, SIMPSUP */
        buf[5] = 0x07;
 
        /* If WriteCache emulation is enabled, set V_SUP */
-       if (DEV_ATTRIB(SE_DEV(cmd))->emulate_write_cache > 0)
+       if (cmd->se_dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0)
                buf[6] = 0x01;
        return 0;
 }
@@ -486,7 +455,7 @@ target_emulate_evpd_86(struct se_cmd *cmd, unsigned char *buf)
 static int
 target_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        int have_tp = 0;
 
        /*
@@ -494,27 +463,29 @@ target_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
         * emulate_tpu=1 or emulate_tpws=1 we will be expect a
         * different page length for Thin Provisioning.
         */
-       if (DEV_ATTRIB(dev)->emulate_tpu || DEV_ATTRIB(dev)->emulate_tpws)
+       if (dev->se_sub_dev->se_dev_attrib.emulate_tpu || dev->se_sub_dev->se_dev_attrib.emulate_tpws)
                have_tp = 1;
 
        if (cmd->data_length < (0x10 + 4)) {
-               printk(KERN_INFO "Received data_length: %u"
+               pr_debug("Received data_length: %u"
                        " too small for EVPD 0xb0\n",
                        cmd->data_length);
-               return -1;
+               return -EINVAL;
        }
 
        if (have_tp && cmd->data_length < (0x3c + 4)) {
-               printk(KERN_INFO "Received data_length: %u"
+               pr_debug("Received data_length: %u"
                        " too small for TPE=1 EVPD 0xb0\n",
                        cmd->data_length);
                have_tp = 0;
        }
 
        buf[0] = dev->transport->get_device_type(dev);
-       buf[1] = 0xb0;
        buf[3] = have_tp ? 0x3c : 0x10;
 
+       /* Set WSNZ to 1 */
+       buf[4] = 0x01;
+
        /*
         * Set OPTIMAL TRANSFER LENGTH GRANULARITY
         */
@@ -523,12 +494,12 @@ target_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
        /*
         * Set MAXIMUM TRANSFER LENGTH
         */
-       put_unaligned_be32(DEV_ATTRIB(dev)->max_sectors, &buf[8]);
+       put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.max_sectors, &buf[8]);
 
        /*
         * Set OPTIMAL TRANSFER LENGTH
         */
-       put_unaligned_be32(DEV_ATTRIB(dev)->optimal_sectors, &buf[12]);
+       put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.optimal_sectors, &buf[12]);
 
        /*
         * Exit now if we don't support TP or the initiator sent a too
@@ -540,35 +511,51 @@ target_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
        /*
         * Set MAXIMUM UNMAP LBA COUNT
         */
-       put_unaligned_be32(DEV_ATTRIB(dev)->max_unmap_lba_count, &buf[20]);
+       put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.max_unmap_lba_count, &buf[20]);
 
        /*
         * Set MAXIMUM UNMAP BLOCK DESCRIPTOR COUNT
         */
-       put_unaligned_be32(DEV_ATTRIB(dev)->max_unmap_block_desc_count,
+       put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count,
                           &buf[24]);
 
        /*
         * Set OPTIMAL UNMAP GRANULARITY
         */
-       put_unaligned_be32(DEV_ATTRIB(dev)->unmap_granularity, &buf[28]);
+       put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.unmap_granularity, &buf[28]);
 
        /*
         * UNMAP GRANULARITY ALIGNMENT
         */
-       put_unaligned_be32(DEV_ATTRIB(dev)->unmap_granularity_alignment,
+       put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment,
                           &buf[32]);
-       if (DEV_ATTRIB(dev)->unmap_granularity_alignment != 0)
+       if (dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment != 0)
                buf[32] |= 0x80; /* Set the UGAVALID bit */
 
        return 0;
 }
 
+/* Block Device Characteristics VPD page */
+static int
+target_emulate_evpd_b1(struct se_cmd *cmd, unsigned char *buf)
+{
+       struct se_device *dev = cmd->se_dev;
+
+       buf[0] = dev->transport->get_device_type(dev);
+       buf[3] = 0x3c;
+
+       if (cmd->data_length >= 5 &&
+           dev->se_sub_dev->se_dev_attrib.is_nonrot)
+               buf[5] = 1;
+
+       return 0;
+}
+
 /* Thin Provisioning VPD */
 static int
 target_emulate_evpd_b2(struct se_cmd *cmd, unsigned char *buf)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
 
        /*
         * From sbc3r22 section 6.5.4 Thin Provisioning VPD page:
@@ -579,7 +566,6 @@ target_emulate_evpd_b2(struct se_cmd *cmd, unsigned char *buf)
         * defined in table 162.
         */
        buf[0] = dev->transport->get_device_type(dev);
-       buf[1] = 0xb2;
 
        /*
         * Set Hardcoded length mentioned above for DP=0
@@ -602,7 +588,7 @@ target_emulate_evpd_b2(struct se_cmd *cmd, unsigned char *buf)
         * the UNMAP command (see 5.25). A TPU bit set to zero indicates
         * that the device server does not support the UNMAP command.
         */
-       if (DEV_ATTRIB(dev)->emulate_tpu != 0)
+       if (dev->se_sub_dev->se_dev_attrib.emulate_tpu != 0)
                buf[5] = 0x80;
 
        /*
@@ -611,18 +597,59 @@ target_emulate_evpd_b2(struct se_cmd *cmd, unsigned char *buf)
         * A TPWS bit set to zero indicates that the device server does not
         * support the use of the WRITE SAME (16) command to unmap LBAs.
         */
-       if (DEV_ATTRIB(dev)->emulate_tpws != 0)
+       if (dev->se_sub_dev->se_dev_attrib.emulate_tpws != 0)
                buf[5] |= 0x40;
 
        return 0;
 }
 
+static int
+target_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf);
+
+static struct {
+       uint8_t         page;
+       int             (*emulate)(struct se_cmd *, unsigned char *);
+} evpd_handlers[] = {
+       { .page = 0x00, .emulate = target_emulate_evpd_00 },
+       { .page = 0x80, .emulate = target_emulate_evpd_80 },
+       { .page = 0x83, .emulate = target_emulate_evpd_83 },
+       { .page = 0x86, .emulate = target_emulate_evpd_86 },
+       { .page = 0xb0, .emulate = target_emulate_evpd_b0 },
+       { .page = 0xb1, .emulate = target_emulate_evpd_b1 },
+       { .page = 0xb2, .emulate = target_emulate_evpd_b2 },
+};
+
+/* supported vital product data pages */
+static int
+target_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf)
+{
+       int p;
+
+       if (cmd->data_length < 8)
+               return 0;
+       /*
+        * Only report the INQUIRY EVPD=1 pages after a valid NAA
+        * Registered Extended LUN WWN has been set via ConfigFS
+        * during device creation/restart.
+        */
+       if (cmd->se_dev->se_sub_dev->su_dev_flags &
+                       SDF_EMULATED_VPD_UNIT_SERIAL) {
+               buf[3] = ARRAY_SIZE(evpd_handlers);
+               for (p = 0; p < min_t(int, ARRAY_SIZE(evpd_handlers),
+                                     cmd->data_length - 4); ++p)
+                       buf[p + 4] = evpd_handlers[p].page;
+       }
+
+       return 0;
+}
+
 static int
 target_emulate_inquiry(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
-       unsigned char *buf = cmd->t_task->t_task_buf;
-       unsigned char *cdb = cmd->t_task->t_task_cdb;
+       struct se_device *dev = cmd->se_dev;
+       unsigned char *buf;
+       unsigned char *cdb = cmd->t_task_cdb;
+       int p, ret;
 
        if (!(cdb[1] & 0x1))
                return target_emulate_inquiry_std(cmd);
@@ -635,38 +662,33 @@ target_emulate_inquiry(struct se_cmd *cmd)
         * payload length left for the next outgoing EVPD metadata
         */
        if (cmd->data_length < 4) {
-               printk(KERN_ERR "SCSI Inquiry payload length: %u"
+               pr_err("SCSI Inquiry payload length: %u"
                        " too small for EVPD=1\n", cmd->data_length);
-               return -1;
+               return -EINVAL;
        }
+
+       buf = transport_kmap_first_data_page(cmd);
+
        buf[0] = dev->transport->get_device_type(dev);
 
-       switch (cdb[2]) {
-       case 0x00:
-               return target_emulate_evpd_00(cmd, buf);
-       case 0x80:
-               return target_emulate_evpd_80(cmd, buf);
-       case 0x83:
-               return target_emulate_evpd_83(cmd, buf);
-       case 0x86:
-               return target_emulate_evpd_86(cmd, buf);
-       case 0xb0:
-               return target_emulate_evpd_b0(cmd, buf);
-       case 0xb2:
-               return target_emulate_evpd_b2(cmd, buf);
-       default:
-               printk(KERN_ERR "Unknown VPD Code: 0x%02x\n", cdb[2]);
-               return -1;
-       }
+       for (p = 0; p < ARRAY_SIZE(evpd_handlers); ++p)
+               if (cdb[2] == evpd_handlers[p].page) {
+                       buf[1] = cdb[2];
+                       ret = evpd_handlers[p].emulate(cmd, buf);
+                       transport_kunmap_first_data_page(cmd);
+                       return ret;
+               }
 
-       return 0;
+       transport_kunmap_first_data_page(cmd);
+       pr_err("Unknown VPD Code: 0x%02x\n", cdb[2]);
+       return -EINVAL;
 }
 
 static int
 target_emulate_readcapacity(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
-       unsigned char *buf = cmd->t_task->t_task_buf;
+       struct se_device *dev = cmd->se_dev;
+       unsigned char *buf;
        unsigned long long blocks_long = dev->transport->get_blocks(dev);
        u32 blocks;
 
@@ -675,30 +697,36 @@ target_emulate_readcapacity(struct se_cmd *cmd)
        else
                blocks = (u32)blocks_long;
 
+       buf = transport_kmap_first_data_page(cmd);
+
        buf[0] = (blocks >> 24) & 0xff;
        buf[1] = (blocks >> 16) & 0xff;
        buf[2] = (blocks >> 8) & 0xff;
        buf[3] = blocks & 0xff;
-       buf[4] = (DEV_ATTRIB(dev)->block_size >> 24) & 0xff;
-       buf[5] = (DEV_ATTRIB(dev)->block_size >> 16) & 0xff;
-       buf[6] = (DEV_ATTRIB(dev)->block_size >> 8) & 0xff;
-       buf[7] = DEV_ATTRIB(dev)->block_size & 0xff;
+       buf[4] = (dev->se_sub_dev->se_dev_attrib.block_size >> 24) & 0xff;
+       buf[5] = (dev->se_sub_dev->se_dev_attrib.block_size >> 16) & 0xff;
+       buf[6] = (dev->se_sub_dev->se_dev_attrib.block_size >> 8) & 0xff;
+       buf[7] = dev->se_sub_dev->se_dev_attrib.block_size & 0xff;
        /*
         * Set max 32-bit blocks to signal SERVICE ACTION READ_CAPACITY_16
        */
-       if (DEV_ATTRIB(dev)->emulate_tpu || DEV_ATTRIB(dev)->emulate_tpws)
+       if (dev->se_sub_dev->se_dev_attrib.emulate_tpu || dev->se_sub_dev->se_dev_attrib.emulate_tpws)
                put_unaligned_be32(0xFFFFFFFF, &buf[0]);
 
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
 static int
 target_emulate_readcapacity_16(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
-       unsigned char *buf = cmd->t_task->t_task_buf;
+       struct se_device *dev = cmd->se_dev;
+       unsigned char *buf;
        unsigned long long blocks = dev->transport->get_blocks(dev);
 
+       buf = transport_kmap_first_data_page(cmd);
+
        buf[0] = (blocks >> 56) & 0xff;
        buf[1] = (blocks >> 48) & 0xff;
        buf[2] = (blocks >> 40) & 0xff;
@@ -707,17 +735,19 @@ target_emulate_readcapacity_16(struct se_cmd *cmd)
        buf[5] = (blocks >> 16) & 0xff;
        buf[6] = (blocks >> 8) & 0xff;
        buf[7] = blocks & 0xff;
-       buf[8] = (DEV_ATTRIB(dev)->block_size >> 24) & 0xff;
-       buf[9] = (DEV_ATTRIB(dev)->block_size >> 16) & 0xff;
-       buf[10] = (DEV_ATTRIB(dev)->block_size >> 8) & 0xff;
-       buf[11] = DEV_ATTRIB(dev)->block_size & 0xff;
+       buf[8] = (dev->se_sub_dev->se_dev_attrib.block_size >> 24) & 0xff;
+       buf[9] = (dev->se_sub_dev->se_dev_attrib.block_size >> 16) & 0xff;
+       buf[10] = (dev->se_sub_dev->se_dev_attrib.block_size >> 8) & 0xff;
+       buf[11] = dev->se_sub_dev->se_dev_attrib.block_size & 0xff;
        /*
         * Set Thin Provisioning Enable bit following sbc3r22 in section
         * READ CAPACITY (16) byte 14 if emulate_tpu or emulate_tpws is enabled.
         */
-       if (DEV_ATTRIB(dev)->emulate_tpu || DEV_ATTRIB(dev)->emulate_tpws)
+       if (dev->se_sub_dev->se_dev_attrib.emulate_tpu || dev->se_sub_dev->se_dev_attrib.emulate_tpws)
                buf[14] = 0x80;
 
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
@@ -736,6 +766,35 @@ target_modesense_control(struct se_device *dev, unsigned char *p)
        p[0] = 0x0a;
        p[1] = 0x0a;
        p[2] = 2;
+       /*
+        * From spc4r23, 7.4.7 Control mode page
+        *
+        * The QUEUE ALGORITHM MODIFIER field (see table 368) specifies
+        * restrictions on the algorithm used for reordering commands
+        * having the SIMPLE task attribute (see SAM-4).
+        *
+        *                    Table 368 -- QUEUE ALGORITHM MODIFIER field
+        *                         Code      Description
+        *                          0h       Restricted reordering
+        *                          1h       Unrestricted reordering allowed
+        *                          2h to 7h    Reserved
+        *                          8h to Fh    Vendor specific
+        *
+        * A value of zero in the QUEUE ALGORITHM MODIFIER field specifies that
+        * the device server shall order the processing sequence of commands
+        * having the SIMPLE task attribute such that data integrity is maintained
+        * for that I_T nexus (i.e., if the transmission of new SCSI transport protocol
+        * requests is halted at any time, the final value of all data observable
+        * on the medium shall be the same as if all the commands had been processed
+        * with the ORDERED task attribute).
+        *
+        * A value of one in the QUEUE ALGORITHM MODIFIER field specifies that the
+        * device server may reorder the processing sequence of commands having the
+        * SIMPLE task attribute in any manner. Any data integrity exposures related to
+        * command sequence order shall be explicitly handled by the application client
+        * through the selection of appropriate ommands and task attributes.
+        */
+       p[3] = (dev->se_sub_dev->se_dev_attrib.emulate_rest_reord == 1) ? 0x00 : 0x10;
        /*
         * From spc4r17, section 7.4.6 Control mode Page
         *
@@ -765,8 +824,8 @@ target_modesense_control(struct se_device *dev, unsigned char *p)
         * for a BUSY, TASK SET FULL, or RESERVATION CONFLICT status regardless
         * to the number of commands completed with one of those status codes.
         */
-       p[4] = (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl == 2) ? 0x30 :
-              (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl == 1) ? 0x20 : 0x00;
+       p[4] = (dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl == 2) ? 0x30 :
+              (dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl == 1) ? 0x20 : 0x00;
        /*
         * From spc4r17, section 7.4.6 Control mode Page
         *
@@ -779,7 +838,7 @@ target_modesense_control(struct se_device *dev, unsigned char *p)
         * which the command was received shall be completed with TASK ABORTED
         * status (see SAM-4).
         */
-       p[5] = (DEV_ATTRIB(dev)->emulate_tas) ? 0x40 : 0x00;
+       p[5] = (dev->se_sub_dev->se_dev_attrib.emulate_tas) ? 0x40 : 0x00;
        p[8] = 0xff;
        p[9] = 0xff;
        p[11] = 30;
@@ -792,7 +851,7 @@ target_modesense_caching(struct se_device *dev, unsigned char *p)
 {
        p[0] = 0x08;
        p[1] = 0x12;
-       if (DEV_ATTRIB(dev)->emulate_write_cache > 0)
+       if (dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0)
                p[2] = 0x04; /* Write Cache Enable */
        p[12] = 0x20; /* Disabled Read Ahead */
 
@@ -830,9 +889,9 @@ target_modesense_dpofua(unsigned char *buf, int type)
 static int
 target_emulate_modesense(struct se_cmd *cmd, int ten)
 {
-       struct se_device *dev = SE_DEV(cmd);
-       char *cdb = cmd->t_task->t_task_cdb;
-       unsigned char *rbuf = cmd->t_task->t_task_buf;
+       struct se_device *dev = cmd->se_dev;
+       char *cdb = cmd->t_task_cdb;
+       unsigned char *rbuf;
        int type = dev->transport->get_device_type(dev);
        int offset = (ten) ? 8 : 4;
        int length = 0;
@@ -856,7 +915,7 @@ target_emulate_modesense(struct se_cmd *cmd, int ten)
                length += target_modesense_control(dev, &buf[offset+length]);
                break;
        default:
-               printk(KERN_ERR "Got Unknown Mode Page: 0x%02x\n",
+               pr_err("Got Unknown Mode Page: 0x%02x\n",
                                cdb[2] & 0x3f);
                return PYX_TRANSPORT_UNKNOWN_MODE_PAGE;
        }
@@ -867,13 +926,13 @@ target_emulate_modesense(struct se_cmd *cmd, int ten)
                buf[0] = (offset >> 8) & 0xff;
                buf[1] = offset & 0xff;
 
-               if ((SE_LUN(cmd)->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
+               if ((cmd->se_lun->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
                    (cmd->se_deve &&
                    (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
                        target_modesense_write_protect(&buf[3], type);
 
-               if ((DEV_ATTRIB(dev)->emulate_write_cache > 0) &&
-                   (DEV_ATTRIB(dev)->emulate_fua_write > 0))
+               if ((dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0) &&
+                   (dev->se_sub_dev->se_dev_attrib.emulate_fua_write > 0))
                        target_modesense_dpofua(&buf[3], type);
 
                if ((offset + 2) > cmd->data_length)
@@ -883,19 +942,22 @@ target_emulate_modesense(struct se_cmd *cmd, int ten)
                offset -= 1;
                buf[0] = offset & 0xff;
 
-               if ((SE_LUN(cmd)->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
+               if ((cmd->se_lun->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
                    (cmd->se_deve &&
                    (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
                        target_modesense_write_protect(&buf[2], type);
 
-               if ((DEV_ATTRIB(dev)->emulate_write_cache > 0) &&
-                   (DEV_ATTRIB(dev)->emulate_fua_write > 0))
+               if ((dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0) &&
+                   (dev->se_sub_dev->se_dev_attrib.emulate_fua_write > 0))
                        target_modesense_dpofua(&buf[2], type);
 
                if ((offset + 1) > cmd->data_length)
                        offset = cmd->data_length;
        }
+
+       rbuf = transport_kmap_first_data_page(cmd);
        memcpy(rbuf, buf, offset);
+       transport_kunmap_first_data_page(cmd);
 
        return 0;
 }
@@ -903,16 +965,20 @@ target_emulate_modesense(struct se_cmd *cmd, int ten)
 static int
 target_emulate_request_sense(struct se_cmd *cmd)
 {
-       unsigned char *cdb = cmd->t_task->t_task_cdb;
-       unsigned char *buf = cmd->t_task->t_task_buf;
+       unsigned char *cdb = cmd->t_task_cdb;
+       unsigned char *buf;
        u8 ua_asc = 0, ua_ascq = 0;
+       int err = 0;
 
        if (cdb[1] & 0x01) {
-               printk(KERN_ERR "REQUEST_SENSE description emulation not"
+               pr_err("REQUEST_SENSE description emulation not"
                        " supported\n");
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
-       if (!(core_scsi3_ua_clear_for_request_sense(cmd, &ua_asc, &ua_ascq))) {
+
+       buf = transport_kmap_first_data_page(cmd);
+
+       if (!core_scsi3_ua_clear_for_request_sense(cmd, &ua_asc, &ua_ascq)) {
                /*
                 * CURRENT ERROR, UNIT ATTENTION
                 */
@@ -924,7 +990,8 @@ target_emulate_request_sense(struct se_cmd *cmd)
                 */
                if (cmd->data_length <= 18) {
                        buf[7] = 0x00;
-                       return 0;
+                       err = -EINVAL;
+                       goto end;
                }
                /*
                 * The Additional Sense Code (ASC) from the UNIT ATTENTION
@@ -944,7 +1011,8 @@ target_emulate_request_sense(struct se_cmd *cmd)
                 */
                if (cmd->data_length <= 18) {
                        buf[7] = 0x00;
-                       return 0;
+                       err = -EINVAL;
+                       goto end;
                }
                /*
                 * NO ADDITIONAL SENSE INFORMATION
@@ -953,6 +1021,9 @@ target_emulate_request_sense(struct se_cmd *cmd)
                buf[7] = 0x0A;
        }
 
+end:
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
@@ -963,13 +1034,13 @@ target_emulate_request_sense(struct se_cmd *cmd)
 static int
 target_emulate_unmap(struct se_task *task)
 {
-       struct se_cmd *cmd = TASK_CMD(task);
-       struct se_device *dev = SE_DEV(cmd);
-       unsigned char *buf = cmd->t_task->t_task_buf, *ptr = NULL;
-       unsigned char *cdb = &cmd->t_task->t_task_cdb[0];
+       struct se_cmd *cmd = task->task_se_cmd;
+       struct se_device *dev = cmd->se_dev;
+       unsigned char *buf, *ptr = NULL;
+       unsigned char *cdb = &cmd->t_task_cdb[0];
        sector_t lba;
        unsigned int size = cmd->data_length, range;
-       int ret, offset;
+       int ret = 0, offset;
        unsigned short dl, bd_dl;
 
        /* First UNMAP block descriptor starts at 8 byte offset */
@@ -977,21 +1048,24 @@ target_emulate_unmap(struct se_task *task)
        size -= 8;
        dl = get_unaligned_be16(&cdb[0]);
        bd_dl = get_unaligned_be16(&cdb[2]);
+
+       buf = transport_kmap_first_data_page(cmd);
+
        ptr = &buf[offset];
-       printk(KERN_INFO "UNMAP: Sub: %s Using dl: %hu bd_dl: %hu size: %hu"
+       pr_debug("UNMAP: Sub: %s Using dl: %hu bd_dl: %hu size: %hu"
                " ptr: %p\n", dev->transport->name, dl, bd_dl, size, ptr);
 
        while (size) {
                lba = get_unaligned_be64(&ptr[0]);
                range = get_unaligned_be32(&ptr[8]);
-               printk(KERN_INFO "UNMAP: Using lba: %llu and range: %u\n",
+               pr_debug("UNMAP: Using lba: %llu and range: %u\n",
                                 (unsigned long long)lba, range);
 
                ret = dev->transport->do_discard(dev, lba, range);
                if (ret < 0) {
-                       printk(KERN_ERR "blkdev_issue_discard() failed: %d\n",
+                       pr_err("blkdev_issue_discard() failed: %d\n",
                                        ret);
-                       return -1;
+                       goto err;
                }
 
                ptr += 16;
@@ -1000,7 +1074,10 @@ target_emulate_unmap(struct se_task *task)
 
        task->task_scsi_status = GOOD;
        transport_complete_task(task, 1);
-       return 0;
+err:
+       transport_kunmap_first_data_page(cmd);
+
+       return ret;
 }
 
 /*
@@ -1008,23 +1085,36 @@ target_emulate_unmap(struct se_task *task)
  * Note this is not used for TCM/pSCSI passthrough
  */
 static int
-target_emulate_write_same(struct se_task *task)
+target_emulate_write_same(struct se_task *task, int write_same32)
 {
-       struct se_cmd *cmd = TASK_CMD(task);
-       struct se_device *dev = SE_DEV(cmd);
-       sector_t lba = cmd->t_task->t_task_lba;
-       unsigned int range;
+       struct se_cmd *cmd = task->task_se_cmd;
+       struct se_device *dev = cmd->se_dev;
+       sector_t range;
+       sector_t lba = cmd->t_task_lba;
+       unsigned int num_blocks;
        int ret;
+       /*
+        * Extract num_blocks from the WRITE_SAME_* CDB.  Then use the explict
+        * range when non zero is supplied, otherwise calculate the remaining
+        * range based on ->get_blocks() - starting LBA.
+        */
+       if (write_same32)
+               num_blocks = get_unaligned_be32(&cmd->t_task_cdb[28]);
+       else
+               num_blocks = get_unaligned_be32(&cmd->t_task_cdb[10]);
 
-       range = (cmd->data_length / DEV_ATTRIB(dev)->block_size);
+       if (num_blocks != 0)
+               range = num_blocks;
+       else
+               range = (dev->transport->get_blocks(dev) - lba);
 
-       printk(KERN_INFO "WRITE_SAME UNMAP: LBA: %llu Range: %u\n",
-                        (unsigned long long)lba, range);
+       pr_debug("WRITE_SAME UNMAP: LBA: %llu Range: %llu\n",
+                (unsigned long long)lba, (unsigned long long)range);
 
        ret = dev->transport->do_discard(dev, lba, range);
        if (ret < 0) {
-               printk(KERN_INFO "blkdev_issue_discard() failed for WRITE_SAME\n");
-               return -1;
+               pr_debug("blkdev_issue_discard() failed for WRITE_SAME\n");
+               return ret;
        }
 
        task->task_scsi_status = GOOD;
@@ -1035,12 +1125,12 @@ target_emulate_write_same(struct se_task *task)
 int
 transport_emulate_control_cdb(struct se_task *task)
 {
-       struct se_cmd *cmd = TASK_CMD(task);
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_cmd *cmd = task->task_se_cmd;
+       struct se_device *dev = cmd->se_dev;
        unsigned short service_action;
        int ret = 0;
 
-       switch (cmd->t_task->t_task_cdb[0]) {
+       switch (cmd->t_task_cdb[0]) {
        case INQUIRY:
                ret = target_emulate_inquiry(cmd);
                break;
@@ -1054,13 +1144,13 @@ transport_emulate_control_cdb(struct se_task *task)
                ret = target_emulate_modesense(cmd, 1);
                break;
        case SERVICE_ACTION_IN:
-               switch (cmd->t_task->t_task_cdb[1] & 0x1f) {
+               switch (cmd->t_task_cdb[1] & 0x1f) {
                case SAI_READ_CAPACITY_16:
                        ret = target_emulate_readcapacity_16(cmd);
                        break;
                default:
-                       printk(KERN_ERR "Unsupported SA: 0x%02x\n",
-                               cmd->t_task->t_task_cdb[1] & 0x1f);
+                       pr_err("Unsupported SA: 0x%02x\n",
+                               cmd->t_task_cdb[1] & 0x1f);
                        return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
                }
                break;
@@ -1069,7 +1159,7 @@ transport_emulate_control_cdb(struct se_task *task)
                break;
        case UNMAP:
                if (!dev->transport->do_discard) {
-                       printk(KERN_ERR "UNMAP emulation not supported for: %s\n",
+                       pr_err("UNMAP emulation not supported for: %s\n",
                                        dev->transport->name);
                        return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
                }
@@ -1077,27 +1167,27 @@ transport_emulate_control_cdb(struct se_task *task)
                break;
        case WRITE_SAME_16:
                if (!dev->transport->do_discard) {
-                       printk(KERN_ERR "WRITE_SAME_16 emulation not supported"
+                       pr_err("WRITE_SAME_16 emulation not supported"
                                        " for: %s\n", dev->transport->name);
                        return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
                }
-               ret = target_emulate_write_same(task);
+               ret = target_emulate_write_same(task, 0);
                break;
        case VARIABLE_LENGTH_CMD:
                service_action =
-                       get_unaligned_be16(&cmd->t_task->t_task_cdb[8]);
+                       get_unaligned_be16(&cmd->t_task_cdb[8]);
                switch (service_action) {
                case WRITE_SAME_32:
                        if (!dev->transport->do_discard) {
-                               printk(KERN_ERR "WRITE_SAME_32 SA emulation not"
+                               pr_err("WRITE_SAME_32 SA emulation not"
                                        " supported for: %s\n",
                                        dev->transport->name);
                                return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
                        }
-                       ret = target_emulate_write_same(task);
+                       ret = target_emulate_write_same(task, 1);
                        break;
                default:
-                       printk(KERN_ERR "Unsupported VARIABLE_LENGTH_CMD SA:"
+                       pr_err("Unsupported VARIABLE_LENGTH_CMD SA:"
                                        " 0x%02x\n", service_action);
                        break;
                }
@@ -1105,8 +1195,7 @@ transport_emulate_control_cdb(struct se_task *task)
        case SYNCHRONIZE_CACHE:
        case 0x91: /* SYNCHRONIZE_CACHE_16: */
                if (!dev->transport->do_sync_cache) {
-                       printk(KERN_ERR
-                               "SYNCHRONIZE_CACHE emulation not supported"
+                       pr_err("SYNCHRONIZE_CACHE emulation not supported"
                                " for: %s\n", dev->transport->name);
                        return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
                }
@@ -1123,8 +1212,8 @@ transport_emulate_control_cdb(struct se_task *task)
        case WRITE_FILEMARKS:
                break;
        default:
-               printk(KERN_ERR "Unsupported SCSI Opcode: 0x%02x for %s\n",
-                       cmd->t_task->t_task_cdb[0], dev->transport->name);
+               pr_err("Unsupported SCSI Opcode: 0x%02x for %s\n",
+                       cmd->t_task_cdb[0], dev->transport->name);
                return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
        }
 
index 25c1f49a7d8bae7ee61f20df5830ba5a782277c7..b2575d8568cc174ceadb7648c6fa9deb0a438a3f 100644 (file)
@@ -37,6 +37,7 @@
 #include <linux/parser.h>
 #include <linux/syscalls.h>
 #include <linux/configfs.h>
+#include <linux/spinlock.h>
 
 #include <target/target_core_base.h>
 #include <target/target_core_device.h>
@@ -52,6 +53,8 @@
 #include "target_core_rd.h"
 #include "target_core_stat.h"
 
+extern struct t10_alua_lu_gp *default_lu_gp;
+
 static struct list_head g_tf_list;
 static struct mutex g_tf_lock;
 
@@ -61,6 +64,13 @@ struct target_core_configfs_attribute {
        ssize_t (*store)(void *, const char *, size_t);
 };
 
+static struct config_group target_core_hbagroup;
+static struct config_group alua_group;
+static struct config_group alua_lu_gps_group;
+
+static DEFINE_SPINLOCK(se_device_lock);
+static LIST_HEAD(se_dev_list);
+
 static inline struct se_hba *
 item_to_hba(struct config_item *item)
 {
@@ -94,12 +104,12 @@ static struct target_fabric_configfs *target_core_get_fabric(
 {
        struct target_fabric_configfs *tf;
 
-       if (!(name))
+       if (!name)
                return NULL;
 
        mutex_lock(&g_tf_lock);
        list_for_each_entry(tf, &g_tf_list, tf_list) {
-               if (!(strcmp(tf->tf_name, name))) {
+               if (!strcmp(tf->tf_name, name)) {
                        atomic_inc(&tf->tf_access_cnt);
                        mutex_unlock(&g_tf_lock);
                        return tf;
@@ -120,7 +130,7 @@ static struct config_group *target_core_register_fabric(
        struct target_fabric_configfs *tf;
        int ret;
 
-       printk(KERN_INFO "Target_Core_ConfigFS: REGISTER -> group: %p name:"
+       pr_debug("Target_Core_ConfigFS: REGISTER -> group: %p name:"
                        " %s\n", group, name);
        /*
         * Ensure that TCM subsystem plugins are loaded at this point for
@@ -140,7 +150,7 @@ static struct config_group *target_core_register_fabric(
         * registered, but simply provids auto loading logic for modules with
         * mkdir(2) system calls with known TCM fabric modules.
         */
-       if (!(strncmp(name, "iscsi", 5))) {
+       if (!strncmp(name, "iscsi", 5)) {
                /*
                 * Automatically load the LIO Target fabric module when the
                 * following is called:
@@ -149,11 +159,11 @@ static struct config_group *target_core_register_fabric(
                 */
                ret = request_module("iscsi_target_mod");
                if (ret < 0) {
-                       printk(KERN_ERR "request_module() failed for"
+                       pr_err("request_module() failed for"
                                " iscsi_target_mod.ko: %d\n", ret);
                        return ERR_PTR(-EINVAL);
                }
-       } else if (!(strncmp(name, "loopback", 8))) {
+       } else if (!strncmp(name, "loopback", 8)) {
                /*
                 * Automatically load the tcm_loop fabric module when the
                 * following is called:
@@ -162,25 +172,25 @@ static struct config_group *target_core_register_fabric(
                 */
                ret = request_module("tcm_loop");
                if (ret < 0) {
-                       printk(KERN_ERR "request_module() failed for"
+                       pr_err("request_module() failed for"
                                " tcm_loop.ko: %d\n", ret);
                        return ERR_PTR(-EINVAL);
                }
        }
 
        tf = target_core_get_fabric(name);
-       if (!(tf)) {
-               printk(KERN_ERR "target_core_get_fabric() failed for %s\n",
+       if (!tf) {
+               pr_err("target_core_get_fabric() failed for %s\n",
                        name);
                return ERR_PTR(-EINVAL);
        }
-       printk(KERN_INFO "Target_Core_ConfigFS: REGISTER -> Located fabric:"
+       pr_debug("Target_Core_ConfigFS: REGISTER -> Located fabric:"
                        " %s\n", tf->tf_name);
        /*
         * On a successful target_core_get_fabric() look, the returned
         * struct target_fabric_configfs *tf will contain a usage reference.
         */
-       printk(KERN_INFO "Target_Core_ConfigFS: REGISTER tfc_wwn_cit -> %p\n",
+       pr_debug("Target_Core_ConfigFS: REGISTER tfc_wwn_cit -> %p\n",
                        &TF_CIT_TMPL(tf)->tfc_wwn_cit);
 
        tf->tf_group.default_groups = tf->tf_default_groups;
@@ -192,14 +202,14 @@ static struct config_group *target_core_register_fabric(
        config_group_init_type_name(&tf->tf_disc_group, "discovery_auth",
                        &TF_CIT_TMPL(tf)->tfc_discovery_cit);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: REGISTER -> Allocated Fabric:"
+       pr_debug("Target_Core_ConfigFS: REGISTER -> Allocated Fabric:"
                        " %s\n", tf->tf_group.cg_item.ci_name);
        /*
         * Setup tf_ops.tf_subsys pointer for usage with configfs_depend_item()
         */
        tf->tf_ops.tf_subsys = tf->tf_subsys;
        tf->tf_fabric = &tf->tf_group.cg_item;
-       printk(KERN_INFO "Target_Core_ConfigFS: REGISTER -> Set tf->tf_fabric"
+       pr_debug("Target_Core_ConfigFS: REGISTER -> Set tf->tf_fabric"
                        " for %s\n", name);
 
        return &tf->tf_group;
@@ -218,18 +228,18 @@ static void target_core_deregister_fabric(
        struct config_item *df_item;
        int i;
 
-       printk(KERN_INFO "Target_Core_ConfigFS: DEREGISTER -> Looking up %s in"
+       pr_debug("Target_Core_ConfigFS: DEREGISTER -> Looking up %s in"
                " tf list\n", config_item_name(item));
 
-       printk(KERN_INFO "Target_Core_ConfigFS: DEREGISTER -> located fabric:"
+       pr_debug("Target_Core_ConfigFS: DEREGISTER -> located fabric:"
                        " %s\n", tf->tf_name);
        atomic_dec(&tf->tf_access_cnt);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: DEREGISTER -> Releasing"
+       pr_debug("Target_Core_ConfigFS: DEREGISTER -> Releasing"
                        " tf->tf_fabric for %s\n", tf->tf_name);
        tf->tf_fabric = NULL;
 
-       printk(KERN_INFO "Target_Core_ConfigFS: DEREGISTER -> Releasing ci"
+       pr_debug("Target_Core_ConfigFS: DEREGISTER -> Releasing ci"
                        " %s\n", config_item_name(item));
 
        tf_group = &tf->tf_group;
@@ -296,23 +306,19 @@ struct target_fabric_configfs *target_fabric_configfs_init(
 {
        struct target_fabric_configfs *tf;
 
-       if (!(fabric_mod)) {
-               printk(KERN_ERR "Missing struct module *fabric_mod pointer\n");
-               return NULL;
-       }
        if (!(name)) {
-               printk(KERN_ERR "Unable to locate passed fabric name\n");
-               return NULL;
+               pr_err("Unable to locate passed fabric name\n");
+               return ERR_PTR(-EINVAL);
        }
        if (strlen(name) >= TARGET_FABRIC_NAME_SIZE) {
-               printk(KERN_ERR "Passed name: %s exceeds TARGET_FABRIC"
+               pr_err("Passed name: %s exceeds TARGET_FABRIC"
                        "_NAME_SIZE\n", name);
-               return NULL;
+               return ERR_PTR(-EINVAL);
        }
 
        tf = kzalloc(sizeof(struct target_fabric_configfs), GFP_KERNEL);
-       if (!(tf))
-               return NULL;
+       if (!tf)
+               return ERR_PTR(-ENOMEM);
 
        INIT_LIST_HEAD(&tf->tf_list);
        atomic_set(&tf->tf_access_cnt, 0);
@@ -330,9 +336,9 @@ struct target_fabric_configfs *target_fabric_configfs_init(
        list_add_tail(&tf->tf_list, &g_tf_list);
        mutex_unlock(&g_tf_lock);
 
-       printk(KERN_INFO "<<<<<<<<<<<<<<<<<<<<<< BEGIN FABRIC API >>>>>>>>"
+       pr_debug("<<<<<<<<<<<<<<<<<<<<<< BEGIN FABRIC API >>>>>>>>"
                        ">>>>>>>>>>>>>>\n");
-       printk(KERN_INFO "Initialized struct target_fabric_configfs: %p for"
+       pr_debug("Initialized struct target_fabric_configfs: %p for"
                        " %s\n", tf, tf->tf_name);
        return tf;
 }
@@ -361,140 +367,132 @@ static int target_fabric_tf_ops_check(
 {
        struct target_core_fabric_ops *tfo = &tf->tf_ops;
 
-       if (!(tfo->get_fabric_name)) {
-               printk(KERN_ERR "Missing tfo->get_fabric_name()\n");
-               return -EINVAL;
-       }
-       if (!(tfo->get_fabric_proto_ident)) {
-               printk(KERN_ERR "Missing tfo->get_fabric_proto_ident()\n");
-               return -EINVAL;
-       }
-       if (!(tfo->tpg_get_wwn)) {
-               printk(KERN_ERR "Missing tfo->tpg_get_wwn()\n");
+       if (!tfo->get_fabric_name) {
+               pr_err("Missing tfo->get_fabric_name()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_get_tag)) {
-               printk(KERN_ERR "Missing tfo->tpg_get_tag()\n");
+       if (!tfo->get_fabric_proto_ident) {
+               pr_err("Missing tfo->get_fabric_proto_ident()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_get_default_depth)) {
-               printk(KERN_ERR "Missing tfo->tpg_get_default_depth()\n");
+       if (!tfo->tpg_get_wwn) {
+               pr_err("Missing tfo->tpg_get_wwn()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_get_pr_transport_id)) {
-               printk(KERN_ERR "Missing tfo->tpg_get_pr_transport_id()\n");
+       if (!tfo->tpg_get_tag) {
+               pr_err("Missing tfo->tpg_get_tag()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_get_pr_transport_id_len)) {
-               printk(KERN_ERR "Missing tfo->tpg_get_pr_transport_id_len()\n");
+       if (!tfo->tpg_get_default_depth) {
+               pr_err("Missing tfo->tpg_get_default_depth()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_check_demo_mode)) {
-               printk(KERN_ERR "Missing tfo->tpg_check_demo_mode()\n");
+       if (!tfo->tpg_get_pr_transport_id) {
+               pr_err("Missing tfo->tpg_get_pr_transport_id()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_check_demo_mode_cache)) {
-               printk(KERN_ERR "Missing tfo->tpg_check_demo_mode_cache()\n");
+       if (!tfo->tpg_get_pr_transport_id_len) {
+               pr_err("Missing tfo->tpg_get_pr_transport_id_len()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_check_demo_mode_write_protect)) {
-               printk(KERN_ERR "Missing tfo->tpg_check_demo_mode_write_protect()\n");
+       if (!tfo->tpg_check_demo_mode) {
+               pr_err("Missing tfo->tpg_check_demo_mode()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_check_prod_mode_write_protect)) {
-               printk(KERN_ERR "Missing tfo->tpg_check_prod_mode_write_protect()\n");
+       if (!tfo->tpg_check_demo_mode_cache) {
+               pr_err("Missing tfo->tpg_check_demo_mode_cache()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_alloc_fabric_acl)) {
-               printk(KERN_ERR "Missing tfo->tpg_alloc_fabric_acl()\n");
+       if (!tfo->tpg_check_demo_mode_write_protect) {
+               pr_err("Missing tfo->tpg_check_demo_mode_write_protect()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_release_fabric_acl)) {
-               printk(KERN_ERR "Missing tfo->tpg_release_fabric_acl()\n");
+       if (!tfo->tpg_check_prod_mode_write_protect) {
+               pr_err("Missing tfo->tpg_check_prod_mode_write_protect()\n");
                return -EINVAL;
        }
-       if (!(tfo->tpg_get_inst_index)) {
-               printk(KERN_ERR "Missing tfo->tpg_get_inst_index()\n");
+       if (!tfo->tpg_alloc_fabric_acl) {
+               pr_err("Missing tfo->tpg_alloc_fabric_acl()\n");
                return -EINVAL;
        }
-       if (!(tfo->release_cmd_to_pool)) {
-               printk(KERN_ERR "Missing tfo->release_cmd_to_pool()\n");
+       if (!tfo->tpg_release_fabric_acl) {
+               pr_err("Missing tfo->tpg_release_fabric_acl()\n");
                return -EINVAL;
        }
-       if (!(tfo->release_cmd_direct)) {
-               printk(KERN_ERR "Missing tfo->release_cmd_direct()\n");
+       if (!tfo->tpg_get_inst_index) {
+               pr_err("Missing tfo->tpg_get_inst_index()\n");
                return -EINVAL;
        }
-       if (!(tfo->shutdown_session)) {
-               printk(KERN_ERR "Missing tfo->shutdown_session()\n");
+       if (!tfo->release_cmd) {
+               pr_err("Missing tfo->release_cmd()\n");
                return -EINVAL;
        }
-       if (!(tfo->close_session)) {
-               printk(KERN_ERR "Missing tfo->close_session()\n");
+       if (!tfo->shutdown_session) {
+               pr_err("Missing tfo->shutdown_session()\n");
                return -EINVAL;
        }
-       if (!(tfo->stop_session)) {
-               printk(KERN_ERR "Missing tfo->stop_session()\n");
+       if (!tfo->close_session) {
+               pr_err("Missing tfo->close_session()\n");
                return -EINVAL;
        }
-       if (!(tfo->fall_back_to_erl0)) {
-               printk(KERN_ERR "Missing tfo->fall_back_to_erl0()\n");
+       if (!tfo->stop_session) {
+               pr_err("Missing tfo->stop_session()\n");
                return -EINVAL;
        }
-       if (!(tfo->sess_logged_in)) {
-               printk(KERN_ERR "Missing tfo->sess_logged_in()\n");
+       if (!tfo->fall_back_to_erl0) {
+               pr_err("Missing tfo->fall_back_to_erl0()\n");
                return -EINVAL;
        }
-       if (!(tfo->sess_get_index)) {
-               printk(KERN_ERR "Missing tfo->sess_get_index()\n");
+       if (!tfo->sess_logged_in) {
+               pr_err("Missing tfo->sess_logged_in()\n");
                return -EINVAL;
        }
-       if (!(tfo->write_pending)) {
-               printk(KERN_ERR "Missing tfo->write_pending()\n");
+       if (!tfo->sess_get_index) {
+               pr_err("Missing tfo->sess_get_index()\n");
                return -EINVAL;
        }
-       if (!(tfo->write_pending_status)) {
-               printk(KERN_ERR "Missing tfo->write_pending_status()\n");
+       if (!tfo->write_pending) {
+               pr_err("Missing tfo->write_pending()\n");
                return -EINVAL;
        }
-       if (!(tfo->set_default_node_attributes)) {
-               printk(KERN_ERR "Missing tfo->set_default_node_attributes()\n");
+       if (!tfo->write_pending_status) {
+               pr_err("Missing tfo->write_pending_status()\n");
                return -EINVAL;
        }
-       if (!(tfo->get_task_tag)) {
-               printk(KERN_ERR "Missing tfo->get_task_tag()\n");
+       if (!tfo->set_default_node_attributes) {
+               pr_err("Missing tfo->set_default_node_attributes()\n");
                return -EINVAL;
        }
-       if (!(tfo->get_cmd_state)) {
-               printk(KERN_ERR "Missing tfo->get_cmd_state()\n");
+       if (!tfo->get_task_tag) {
+               pr_err("Missing tfo->get_task_tag()\n");
                return -EINVAL;
        }
-       if (!(tfo->new_cmd_failure)) {
-               printk(KERN_ERR "Missing tfo->new_cmd_failure()\n");
+       if (!tfo->get_cmd_state) {
+               pr_err("Missing tfo->get_cmd_state()\n");
                return -EINVAL;
        }
-       if (!(tfo->queue_data_in)) {
-               printk(KERN_ERR "Missing tfo->queue_data_in()\n");
+       if (!tfo->queue_data_in) {
+               pr_err("Missing tfo->queue_data_in()\n");
                return -EINVAL;
        }
-       if (!(tfo->queue_status)) {
-               printk(KERN_ERR "Missing tfo->queue_status()\n");
+       if (!tfo->queue_status) {
+               pr_err("Missing tfo->queue_status()\n");
                return -EINVAL;
        }
-       if (!(tfo->queue_tm_rsp)) {
-               printk(KERN_ERR "Missing tfo->queue_tm_rsp()\n");
+       if (!tfo->queue_tm_rsp) {
+               pr_err("Missing tfo->queue_tm_rsp()\n");
                return -EINVAL;
        }
-       if (!(tfo->set_fabric_sense_len)) {
-               printk(KERN_ERR "Missing tfo->set_fabric_sense_len()\n");
+       if (!tfo->set_fabric_sense_len) {
+               pr_err("Missing tfo->set_fabric_sense_len()\n");
                return -EINVAL;
        }
-       if (!(tfo->get_fabric_sense_len)) {
-               printk(KERN_ERR "Missing tfo->get_fabric_sense_len()\n");
+       if (!tfo->get_fabric_sense_len) {
+               pr_err("Missing tfo->get_fabric_sense_len()\n");
                return -EINVAL;
        }
-       if (!(tfo->is_state_remove)) {
-               printk(KERN_ERR "Missing tfo->is_state_remove()\n");
+       if (!tfo->is_state_remove) {
+               pr_err("Missing tfo->is_state_remove()\n");
                return -EINVAL;
        }
        /*
@@ -502,20 +500,20 @@ static int target_fabric_tf_ops_check(
         * tfo->fabric_make_tpg() and tfo->fabric_drop_tpg() in
         * target_core_fabric_configfs.c WWN+TPG group context code.
         */
-       if (!(tfo->fabric_make_wwn)) {
-               printk(KERN_ERR "Missing tfo->fabric_make_wwn()\n");
+       if (!tfo->fabric_make_wwn) {
+               pr_err("Missing tfo->fabric_make_wwn()\n");
                return -EINVAL;
        }
-       if (!(tfo->fabric_drop_wwn)) {
-               printk(KERN_ERR "Missing tfo->fabric_drop_wwn()\n");
+       if (!tfo->fabric_drop_wwn) {
+               pr_err("Missing tfo->fabric_drop_wwn()\n");
                return -EINVAL;
        }
-       if (!(tfo->fabric_make_tpg)) {
-               printk(KERN_ERR "Missing tfo->fabric_make_tpg()\n");
+       if (!tfo->fabric_make_tpg) {
+               pr_err("Missing tfo->fabric_make_tpg()\n");
                return -EINVAL;
        }
-       if (!(tfo->fabric_drop_tpg)) {
-               printk(KERN_ERR "Missing tfo->fabric_drop_tpg()\n");
+       if (!tfo->fabric_drop_tpg) {
+               pr_err("Missing tfo->fabric_drop_tpg()\n");
                return -EINVAL;
        }
 
@@ -533,22 +531,15 @@ static int target_fabric_tf_ops_check(
 int target_fabric_configfs_register(
        struct target_fabric_configfs *tf)
 {
-       struct config_group *su_group;
        int ret;
 
-       if (!(tf)) {
-               printk(KERN_ERR "Unable to locate target_fabric_configfs"
+       if (!tf) {
+               pr_err("Unable to locate target_fabric_configfs"
                        " pointer\n");
                return -EINVAL;
        }
-       if (!(tf->tf_subsys)) {
-               printk(KERN_ERR "Unable to target struct config_subsystem"
-                       " pointer\n");
-               return -EINVAL;
-       }
-       su_group = &tf->tf_subsys->su_group;
-       if (!(su_group)) {
-               printk(KERN_ERR "Unable to locate target struct config_group"
+       if (!tf->tf_subsys) {
+               pr_err("Unable to target struct config_subsystem"
                        " pointer\n");
                return -EINVAL;
        }
@@ -556,7 +547,7 @@ int target_fabric_configfs_register(
        if (ret < 0)
                return ret;
 
-       printk(KERN_INFO "<<<<<<<<<<<<<<<<<<<<<< END FABRIC API >>>>>>>>>>>>"
+       pr_debug("<<<<<<<<<<<<<<<<<<<<<< END FABRIC API >>>>>>>>>>>>"
                ">>>>>>>>>>\n");
        return 0;
 }
@@ -565,48 +556,39 @@ EXPORT_SYMBOL(target_fabric_configfs_register);
 void target_fabric_configfs_deregister(
        struct target_fabric_configfs *tf)
 {
-       struct config_group *su_group;
        struct configfs_subsystem *su;
 
-       if (!(tf)) {
-               printk(KERN_ERR "Unable to locate passed target_fabric_"
+       if (!tf) {
+               pr_err("Unable to locate passed target_fabric_"
                        "configfs\n");
                return;
        }
        su = tf->tf_subsys;
-       if (!(su)) {
-               printk(KERN_ERR "Unable to locate passed tf->tf_subsys"
+       if (!su) {
+               pr_err("Unable to locate passed tf->tf_subsys"
                        " pointer\n");
                return;
        }
-       su_group = &tf->tf_subsys->su_group;
-       if (!(su_group)) {
-               printk(KERN_ERR "Unable to locate target struct config_group"
-                       " pointer\n");
-               return;
-       }
-
-       printk(KERN_INFO "<<<<<<<<<<<<<<<<<<<<<< BEGIN FABRIC API >>>>>>>>>>"
+       pr_debug("<<<<<<<<<<<<<<<<<<<<<< BEGIN FABRIC API >>>>>>>>>>"
                        ">>>>>>>>>>>>\n");
        mutex_lock(&g_tf_lock);
        if (atomic_read(&tf->tf_access_cnt)) {
                mutex_unlock(&g_tf_lock);
-               printk(KERN_ERR "Non zero tf->tf_access_cnt for fabric %s\n",
+               pr_err("Non zero tf->tf_access_cnt for fabric %s\n",
                        tf->tf_name);
                BUG();
        }
        list_del(&tf->tf_list);
        mutex_unlock(&g_tf_lock);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: DEREGISTER -> Releasing tf:"
+       pr_debug("Target_Core_ConfigFS: DEREGISTER -> Releasing tf:"
                        " %s\n", tf->tf_name);
        tf->tf_module = NULL;
        tf->tf_subsys = NULL;
        kfree(tf);
 
-       printk("<<<<<<<<<<<<<<<<<<<<<< END FABRIC API >>>>>>>>>>>>>>>>>"
+       pr_debug("<<<<<<<<<<<<<<<<<<<<<< END FABRIC API >>>>>>>>>>>>>>>>>"
                        ">>>>>\n");
-       return;
 }
 EXPORT_SYMBOL(target_fabric_configfs_deregister);
 
@@ -627,11 +609,12 @@ static ssize_t target_core_dev_show_attr_##_name(                 \
                                                                        \
        spin_lock(&se_dev->se_dev_lock);                                \
        dev = se_dev->se_dev_ptr;                                       \
-       if (!(dev)) {                                                   \
+       if (!dev) {                                                     \
                spin_unlock(&se_dev->se_dev_lock);                      \
                return -ENODEV;                                         \
        }                                                               \
-       rb = snprintf(page, PAGE_SIZE, "%u\n", (u32)DEV_ATTRIB(dev)->_name); \
+       rb = snprintf(page, PAGE_SIZE, "%u\n",                          \
+               (u32)dev->se_sub_dev->se_dev_attrib._name);             \
        spin_unlock(&se_dev->se_dev_lock);                              \
                                                                        \
        return rb;                                                      \
@@ -650,14 +633,14 @@ static ssize_t target_core_dev_store_attr_##_name(                        \
                                                                        \
        spin_lock(&se_dev->se_dev_lock);                                \
        dev = se_dev->se_dev_ptr;                                       \
-       if (!(dev)) {                                                   \
+       if (!dev) {                                                     \
                spin_unlock(&se_dev->se_dev_lock);                      \
                return -ENODEV;                                         \
        }                                                               \
        ret = strict_strtoul(page, 0, &val);                            \
        if (ret < 0) {                                                  \
                spin_unlock(&se_dev->se_dev_lock);                      \
-               printk(KERN_ERR "strict_strtoul() failed with"          \
+               pr_err("strict_strtoul() failed with"           \
                        " ret: %d\n", ret);                             \
                return -EINVAL;                                         \
        }                                                               \
@@ -715,6 +698,12 @@ SE_DEV_ATTR(emulate_tpws, S_IRUGO | S_IWUSR);
 DEF_DEV_ATTRIB(enforce_pr_isids);
 SE_DEV_ATTR(enforce_pr_isids, S_IRUGO | S_IWUSR);
 
+DEF_DEV_ATTRIB(is_nonrot);
+SE_DEV_ATTR(is_nonrot, S_IRUGO | S_IWUSR);
+
+DEF_DEV_ATTRIB(emulate_rest_reord);
+SE_DEV_ATTR(emulate_rest_reord, S_IRUGO | S_IWUSR);
+
 DEF_DEV_ATTRIB_RO(hw_block_size);
 SE_DEV_ATTR_RO(hw_block_size);
 
@@ -763,6 +752,8 @@ static struct configfs_attribute *target_core_dev_attrib_attrs[] = {
        &target_core_dev_attrib_emulate_tpu.attr,
        &target_core_dev_attrib_emulate_tpws.attr,
        &target_core_dev_attrib_enforce_pr_isids.attr,
+       &target_core_dev_attrib_is_nonrot.attr,
+       &target_core_dev_attrib_emulate_rest_reord.attr,
        &target_core_dev_attrib_hw_block_size.attr,
        &target_core_dev_attrib_block_size.attr,
        &target_core_dev_attrib_hw_max_sectors.attr,
@@ -819,7 +810,7 @@ static ssize_t target_core_dev_wwn_show_attr_vpd_unit_serial(
        struct se_device *dev;
 
        dev = se_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
        return sprintf(page, "T10 VPD Unit Serial Number: %s\n",
@@ -846,13 +837,13 @@ static ssize_t target_core_dev_wwn_store_attr_vpd_unit_serial(
         * VPD Unit Serial Number that OS dependent multipath can depend on.
         */
        if (su_dev->su_dev_flags & SDF_FIRMWARE_VPD_UNIT_SERIAL) {
-               printk(KERN_ERR "Underlying SCSI device firmware provided VPD"
+               pr_err("Underlying SCSI device firmware provided VPD"
                        " Unit Serial, ignoring request\n");
                return -EOPNOTSUPP;
        }
 
        if (strlen(page) >= INQUIRY_VPD_SERIAL_LEN) {
-               printk(KERN_ERR "Emulated VPD Unit Serial exceeds"
+               pr_err("Emulated VPD Unit Serial exceeds"
                " INQUIRY_VPD_SERIAL_LEN: %d\n", INQUIRY_VPD_SERIAL_LEN);
                return -EOVERFLOW;
        }
@@ -863,9 +854,9 @@ static ssize_t target_core_dev_wwn_store_attr_vpd_unit_serial(
         * could cause negative effects.
         */
        dev = su_dev->se_dev_ptr;
-       if ((dev)) {
+       if (dev) {
                if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-                       printk(KERN_ERR "Unable to set VPD Unit Serial while"
+                       pr_err("Unable to set VPD Unit Serial while"
                                " active %d $FABRIC_MOD exports exist\n",
                                atomic_read(&dev->dev_export_obj.obj_access_count));
                        return -EINVAL;
@@ -883,7 +874,7 @@ static ssize_t target_core_dev_wwn_store_attr_vpd_unit_serial(
                        "%s", strstrip(buf));
        su_dev->su_dev_flags |= SDF_EMULATED_VPD_UNIT_SERIAL;
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Set emulated VPD Unit Serial:"
+       pr_debug("Target_Core_ConfigFS: Set emulated VPD Unit Serial:"
                        " %s\n", su_dev->t10_wwn.unit_serial);
 
        return count;
@@ -905,19 +896,19 @@ static ssize_t target_core_dev_wwn_show_attr_vpd_protocol_identifier(
        ssize_t len = 0;
 
        dev = se_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
        memset(buf, 0, VPD_TMP_BUF_SIZE);
 
        spin_lock(&t10_wwn->t10_vpd_lock);
        list_for_each_entry(vpd, &t10_wwn->t10_vpd_list, vpd_list) {
-               if (!(vpd->protocol_identifier_set))
+               if (!vpd->protocol_identifier_set)
                        continue;
 
                transport_dump_vpd_proto_id(vpd, buf, VPD_TMP_BUF_SIZE);
 
-               if ((len + strlen(buf) >= PAGE_SIZE))
+               if (len + strlen(buf) >= PAGE_SIZE)
                        break;
 
                len += sprintf(page+len, "%s", buf);
@@ -952,7 +943,7 @@ static ssize_t target_core_dev_wwn_show_attr_##_name(                       \
        ssize_t len = 0;                                                \
                                                                        \
        dev = se_dev->se_dev_ptr;                                       \
-       if (!(dev))                                                     \
+       if (!dev)                                                       \
                return -ENODEV;                                         \
                                                                        \
        spin_lock(&t10_wwn->t10_vpd_lock);                              \
@@ -962,19 +953,19 @@ static ssize_t target_core_dev_wwn_show_attr_##_name(                     \
                                                                        \
                memset(buf, 0, VPD_TMP_BUF_SIZE);                       \
                transport_dump_vpd_assoc(vpd, buf, VPD_TMP_BUF_SIZE);   \
-               if ((len + strlen(buf) >= PAGE_SIZE))                   \
+               if (len + strlen(buf) >= PAGE_SIZE)                     \
                        break;                                          \
                len += sprintf(page+len, "%s", buf);                    \
                                                                        \
                memset(buf, 0, VPD_TMP_BUF_SIZE);                       \
                transport_dump_vpd_ident_type(vpd, buf, VPD_TMP_BUF_SIZE); \
-               if ((len + strlen(buf) >= PAGE_SIZE))                   \
+               if (len + strlen(buf) >= PAGE_SIZE)                     \
                        break;                                          \
                len += sprintf(page+len, "%s", buf);                    \
                                                                        \
                memset(buf, 0, VPD_TMP_BUF_SIZE);                       \
                transport_dump_vpd_ident(vpd, buf, VPD_TMP_BUF_SIZE); \
-               if ((len + strlen(buf) >= PAGE_SIZE))                   \
+               if (len + strlen(buf) >= PAGE_SIZE)                     \
                        break;                                          \
                len += sprintf(page+len, "%s", buf);                    \
        }                                                               \
@@ -984,7 +975,7 @@ static ssize_t target_core_dev_wwn_show_attr_##_name(                       \
 }
 
 /*
- * VPD page 0x83 Assoication: Logical Unit
+ * VPD page 0x83 Association: Logical Unit
  */
 DEF_DEV_WWN_ASSOC_SHOW(vpd_assoc_logical_unit, 0x00);
 
@@ -1083,7 +1074,7 @@ static ssize_t target_core_dev_pr_show_spc3_res(
 
        spin_lock(&dev->dev_reservation_lock);
        pr_reg = dev->dev_pr_res_holder;
-       if (!(pr_reg)) {
+       if (!pr_reg) {
                *len += sprintf(page + *len, "No SPC-3 Reservation holder\n");
                spin_unlock(&dev->dev_reservation_lock);
                return *len;
@@ -1093,7 +1084,7 @@ static ssize_t target_core_dev_pr_show_spc3_res(
                                PR_REG_ISID_ID_LEN);
 
        *len += sprintf(page + *len, "SPC-3 Reservation: %s Initiator: %s%s\n",
-               TPG_TFO(se_nacl->se_tpg)->get_fabric_name(),
+               se_nacl->se_tpg->se_tpg_tfo->get_fabric_name(),
                se_nacl->initiatorname, (prf_isid) ? &i_buf[0] : "");
        spin_unlock(&dev->dev_reservation_lock);
 
@@ -1109,13 +1100,13 @@ static ssize_t target_core_dev_pr_show_spc2_res(
 
        spin_lock(&dev->dev_reservation_lock);
        se_nacl = dev->dev_reserved_node_acl;
-       if (!(se_nacl)) {
+       if (!se_nacl) {
                *len += sprintf(page + *len, "No SPC-2 Reservation holder\n");
                spin_unlock(&dev->dev_reservation_lock);
                return *len;
        }
        *len += sprintf(page + *len, "SPC-2 Reservation: %s Initiator: %s\n",
-               TPG_TFO(se_nacl->se_tpg)->get_fabric_name(),
+               se_nacl->se_tpg->se_tpg_tfo->get_fabric_name(),
                se_nacl->initiatorname);
        spin_unlock(&dev->dev_reservation_lock);
 
@@ -1128,10 +1119,10 @@ static ssize_t target_core_dev_pr_show_attr_res_holder(
 {
        ssize_t len = 0;
 
-       if (!(su_dev->se_dev_ptr))
+       if (!su_dev->se_dev_ptr)
                return -ENODEV;
 
-       switch (T10_RES(su_dev)->res_type) {
+       switch (su_dev->t10_pr.res_type) {
        case SPC3_PERSISTENT_RESERVATIONS:
                target_core_dev_pr_show_spc3_res(su_dev->se_dev_ptr,
                                page, &len);
@@ -1165,15 +1156,15 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_all_tgt_pts(
        ssize_t len = 0;
 
        dev = su_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return len;
 
        spin_lock(&dev->dev_reservation_lock);
        pr_reg = dev->dev_pr_res_holder;
-       if (!(pr_reg)) {
+       if (!pr_reg) {
                len = sprintf(page, "No SPC-3 Reservation holder\n");
                spin_unlock(&dev->dev_reservation_lock);
                return len;
@@ -1202,13 +1193,13 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_generation(
        struct se_subsystem_dev *su_dev,
        char *page)
 {
-       if (!(su_dev->se_dev_ptr))
+       if (!su_dev->se_dev_ptr)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return 0;
 
-       return sprintf(page, "0x%08x\n", T10_RES(su_dev)->pr_generation);
+       return sprintf(page, "0x%08x\n", su_dev->t10_pr.pr_generation);
 }
 
 SE_DEV_PR_ATTR_RO(res_pr_generation);
@@ -1229,15 +1220,15 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_holder_tg_port(
        ssize_t len = 0;
 
        dev = su_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return len;
 
        spin_lock(&dev->dev_reservation_lock);
        pr_reg = dev->dev_pr_res_holder;
-       if (!(pr_reg)) {
+       if (!pr_reg) {
                len = sprintf(page, "No SPC-3 Reservation holder\n");
                spin_unlock(&dev->dev_reservation_lock);
                return len;
@@ -1245,7 +1236,7 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_holder_tg_port(
        se_nacl = pr_reg->pr_reg_nacl;
        se_tpg = se_nacl->se_tpg;
        lun = pr_reg->pr_reg_tg_pt_lun;
-       tfo = TPG_TFO(se_tpg);
+       tfo = se_tpg->se_tpg_tfo;
 
        len += sprintf(page+len, "SPC-3 Reservation: %s"
                " Target Node Endpoint: %s\n", tfo->get_fabric_name(),
@@ -1276,16 +1267,16 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_registered_i_pts(
        ssize_t len = 0;
        int reg_count = 0, prf_isid;
 
-       if (!(su_dev->se_dev_ptr))
+       if (!su_dev->se_dev_ptr)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return len;
 
        len += sprintf(page+len, "SPC-3 PR Registrations:\n");
 
-       spin_lock(&T10_RES(su_dev)->registration_lock);
-       list_for_each_entry(pr_reg, &T10_RES(su_dev)->registration_list,
+       spin_lock(&su_dev->t10_pr.registration_lock);
+       list_for_each_entry(pr_reg, &su_dev->t10_pr.registration_list,
                        pr_reg_list) {
 
                memset(buf, 0, 384);
@@ -1299,15 +1290,15 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_registered_i_pts(
                        &i_buf[0] : "", pr_reg->pr_res_key,
                        pr_reg->pr_res_generation);
 
-               if ((len + strlen(buf) >= PAGE_SIZE))
+               if (len + strlen(buf) >= PAGE_SIZE)
                        break;
 
                len += sprintf(page+len, "%s", buf);
                reg_count++;
        }
-       spin_unlock(&T10_RES(su_dev)->registration_lock);
+       spin_unlock(&su_dev->t10_pr.registration_lock);
 
-       if (!(reg_count))
+       if (!reg_count)
                len += sprintf(page+len, "None\n");
 
        return len;
@@ -1327,15 +1318,15 @@ static ssize_t target_core_dev_pr_show_attr_res_pr_type(
        ssize_t len = 0;
 
        dev = su_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return len;
 
        spin_lock(&dev->dev_reservation_lock);
        pr_reg = dev->dev_pr_res_holder;
-       if (!(pr_reg)) {
+       if (!pr_reg) {
                len = sprintf(page, "No SPC-3 Reservation holder\n");
                spin_unlock(&dev->dev_reservation_lock);
                return len;
@@ -1358,10 +1349,10 @@ static ssize_t target_core_dev_pr_show_attr_res_type(
 {
        ssize_t len = 0;
 
-       if (!(su_dev->se_dev_ptr))
+       if (!su_dev->se_dev_ptr)
                return -ENODEV;
 
-       switch (T10_RES(su_dev)->res_type) {
+       switch (su_dev->t10_pr.res_type) {
        case SPC3_PERSISTENT_RESERVATIONS:
                len = sprintf(page, "SPC3_PERSISTENT_RESERVATIONS\n");
                break;
@@ -1389,14 +1380,14 @@ static ssize_t target_core_dev_pr_show_attr_res_aptpl_active(
        struct se_subsystem_dev *su_dev,
        char *page)
 {
-       if (!(su_dev->se_dev_ptr))
+       if (!su_dev->se_dev_ptr)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return 0;
 
        return sprintf(page, "APTPL Bit Status: %s\n",
-               (T10_RES(su_dev)->pr_aptpl_active) ? "Activated" : "Disabled");
+               (su_dev->t10_pr.pr_aptpl_active) ? "Activated" : "Disabled");
 }
 
 SE_DEV_PR_ATTR_RO(res_aptpl_active);
@@ -1408,10 +1399,10 @@ static ssize_t target_core_dev_pr_show_attr_res_aptpl_metadata(
        struct se_subsystem_dev *su_dev,
        char *page)
 {
-       if (!(su_dev->se_dev_ptr))
+       if (!su_dev->se_dev_ptr)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return 0;
 
        return sprintf(page, "Ready to process PR APTPL metadata..\n");
@@ -1460,14 +1451,14 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
        u8 type = 0, scope;
 
        dev = su_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return 0;
 
        if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-               printk(KERN_INFO "Unable to process APTPL metadata while"
+               pr_debug("Unable to process APTPL metadata while"
                        " active fabric exports exist\n");
                return -EINVAL;
        }
@@ -1497,7 +1488,7 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
                                goto out;
                        }
                        if (strlen(i_port) >= PR_APTPL_MAX_IPORT_LEN) {
-                               printk(KERN_ERR "APTPL metadata initiator_node="
+                               pr_err("APTPL metadata initiator_node="
                                        " exceeds PR_APTPL_MAX_IPORT_LEN: %d\n",
                                        PR_APTPL_MAX_IPORT_LEN);
                                ret = -EINVAL;
@@ -1511,7 +1502,7 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
                                goto out;
                        }
                        if (strlen(isid) >= PR_REG_ISID_LEN) {
-                               printk(KERN_ERR "APTPL metadata initiator_isid"
+                               pr_err("APTPL metadata initiator_isid"
                                        "= exceeds PR_REG_ISID_LEN: %d\n",
                                        PR_REG_ISID_LEN);
                                ret = -EINVAL;
@@ -1526,7 +1517,7 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
                        }
                        ret = strict_strtoull(arg_p, 0, &tmp_ll);
                        if (ret < 0) {
-                               printk(KERN_ERR "strict_strtoull() failed for"
+                               pr_err("strict_strtoull() failed for"
                                        " sa_res_key=\n");
                                goto out;
                        }
@@ -1572,7 +1563,7 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
                                goto out;
                        }
                        if (strlen(t_port) >= PR_APTPL_MAX_TPORT_LEN) {
-                               printk(KERN_ERR "APTPL metadata target_node="
+                               pr_err("APTPL metadata target_node="
                                        " exceeds PR_APTPL_MAX_TPORT_LEN: %d\n",
                                        PR_APTPL_MAX_TPORT_LEN);
                                ret = -EINVAL;
@@ -1596,20 +1587,20 @@ static ssize_t target_core_dev_pr_store_attr_res_aptpl_metadata(
                }
        }
 
-       if (!(i_port) || !(t_port) || !(sa_res_key)) {
-               printk(KERN_ERR "Illegal parameters for APTPL registration\n");
+       if (!i_port || !t_port || !sa_res_key) {
+               pr_err("Illegal parameters for APTPL registration\n");
                ret = -EINVAL;
                goto out;
        }
 
        if (res_holder && !(type)) {
-               printk(KERN_ERR "Illegal PR type: 0x%02x for reservation"
+               pr_err("Illegal PR type: 0x%02x for reservation"
                                " holder\n", type);
                ret = -EINVAL;
                goto out;
        }
 
-       ret = core_scsi3_alloc_aptpl_registration(T10_RES(su_dev), sa_res_key,
+       ret = core_scsi3_alloc_aptpl_registration(&su_dev->t10_pr, sa_res_key,
                        i_port, isid, mapped_lun, t_port, tpgt, target_lun,
                        res_holder, all_tg_pt, type);
 out:
@@ -1662,7 +1653,7 @@ static ssize_t target_core_show_dev_info(void *p, char *page)
        int bl = 0;
        ssize_t read_bytes = 0;
 
-       if (!(se_dev->se_dev_ptr))
+       if (!se_dev->se_dev_ptr)
                return -ENODEV;
 
        transport_dump_dev_state(se_dev->se_dev_ptr, page, &bl);
@@ -1688,8 +1679,8 @@ static ssize_t target_core_store_dev_control(
        struct se_hba *hba = se_dev->se_dev_hba;
        struct se_subsystem_api *t = hba->transport;
 
-       if (!(se_dev->se_dev_su_ptr)) {
-               printk(KERN_ERR "Unable to locate struct se_subsystem_dev>se"
+       if (!se_dev->se_dev_su_ptr) {
+               pr_err("Unable to locate struct se_subsystem_dev>se"
                                "_dev_su_ptr\n");
                return -EINVAL;
        }
@@ -1725,7 +1716,7 @@ static ssize_t target_core_store_dev_alias(
        ssize_t read_bytes;
 
        if (count > (SE_DEV_ALIAS_LEN-1)) {
-               printk(KERN_ERR "alias count: %d exceeds"
+               pr_err("alias count: %d exceeds"
                        " SE_DEV_ALIAS_LEN-1: %u\n", (int)count,
                        SE_DEV_ALIAS_LEN-1);
                return -EINVAL;
@@ -1735,7 +1726,7 @@ static ssize_t target_core_store_dev_alias(
        read_bytes = snprintf(&se_dev->se_dev_alias[0], SE_DEV_ALIAS_LEN,
                        "%s", page);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: %s/%s set alias: %s\n",
+       pr_debug("Target_Core_ConfigFS: %s/%s set alias: %s\n",
                config_item_name(&hba->hba_group.cg_item),
                config_item_name(&se_dev->se_dev_group.cg_item),
                se_dev->se_dev_alias);
@@ -1771,7 +1762,7 @@ static ssize_t target_core_store_dev_udev_path(
        ssize_t read_bytes;
 
        if (count > (SE_UDEV_PATH_LEN-1)) {
-               printk(KERN_ERR "udev_path count: %d exceeds"
+               pr_err("udev_path count: %d exceeds"
                        " SE_UDEV_PATH_LEN-1: %u\n", (int)count,
                        SE_UDEV_PATH_LEN-1);
                return -EINVAL;
@@ -1781,7 +1772,7 @@ static ssize_t target_core_store_dev_udev_path(
        read_bytes = snprintf(&se_dev->se_dev_udev_path[0], SE_UDEV_PATH_LEN,
                        "%s", page);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: %s/%s set udev_path: %s\n",
+       pr_debug("Target_Core_ConfigFS: %s/%s set udev_path: %s\n",
                config_item_name(&hba->hba_group.cg_item),
                config_item_name(&se_dev->se_dev_group.cg_item),
                se_dev->se_dev_udev_path);
@@ -1809,13 +1800,13 @@ static ssize_t target_core_store_dev_enable(
        char *ptr;
 
        ptr = strstr(page, "1");
-       if (!(ptr)) {
-               printk(KERN_ERR "For dev_enable ops, only valid value"
+       if (!ptr) {
+               pr_err("For dev_enable ops, only valid value"
                                " is \"1\"\n");
                return -EINVAL;
        }
-       if ((se_dev->se_dev_ptr)) {
-               printk(KERN_ERR "se_dev->se_dev_ptr already set for storage"
+       if (se_dev->se_dev_ptr) {
+               pr_err("se_dev->se_dev_ptr already set for storage"
                                " object\n");
                return -EEXIST;
        }
@@ -1830,7 +1821,7 @@ static ssize_t target_core_store_dev_enable(
                return -EINVAL;
 
        se_dev->se_dev_ptr = dev;
-       printk(KERN_INFO "Target_Core_ConfigFS: Registered se_dev->se_dev_ptr:"
+       pr_debug("Target_Core_ConfigFS: Registered se_dev->se_dev_ptr:"
                " %p\n", se_dev->se_dev_ptr);
 
        return count;
@@ -1854,22 +1845,22 @@ static ssize_t target_core_show_alua_lu_gp(void *p, char *page)
        ssize_t len = 0;
 
        dev = su_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
-       if (T10_ALUA(su_dev)->alua_type != SPC3_ALUA_EMULATED)
+       if (su_dev->t10_alua.alua_type != SPC3_ALUA_EMULATED)
                return len;
 
        lu_gp_mem = dev->dev_alua_lu_gp_mem;
-       if (!(lu_gp_mem)) {
-               printk(KERN_ERR "NULL struct se_device->dev_alua_lu_gp_mem"
+       if (!lu_gp_mem) {
+               pr_err("NULL struct se_device->dev_alua_lu_gp_mem"
                                " pointer\n");
                return -EINVAL;
        }
 
        spin_lock(&lu_gp_mem->lu_gp_mem_lock);
        lu_gp = lu_gp_mem->lu_gp;
-       if ((lu_gp)) {
+       if (lu_gp) {
                lu_ci = &lu_gp->lu_gp_group.cg_item;
                len += sprintf(page, "LU Group Alias: %s\nLU Group ID: %hu\n",
                        config_item_name(lu_ci), lu_gp->lu_gp_id);
@@ -1893,17 +1884,17 @@ static ssize_t target_core_store_alua_lu_gp(
        int move = 0;
 
        dev = su_dev->se_dev_ptr;
-       if (!(dev))
+       if (!dev)
                return -ENODEV;
 
-       if (T10_ALUA(su_dev)->alua_type != SPC3_ALUA_EMULATED) {
-               printk(KERN_WARNING "SPC3_ALUA_EMULATED not enabled for %s/%s\n",
+       if (su_dev->t10_alua.alua_type != SPC3_ALUA_EMULATED) {
+               pr_warn("SPC3_ALUA_EMULATED not enabled for %s/%s\n",
                        config_item_name(&hba->hba_group.cg_item),
                        config_item_name(&su_dev->se_dev_group.cg_item));
                return -EINVAL;
        }
        if (count > LU_GROUP_NAME_BUF) {
-               printk(KERN_ERR "ALUA LU Group Alias too large!\n");
+               pr_err("ALUA LU Group Alias too large!\n");
                return -EINVAL;
        }
        memset(buf, 0, LU_GROUP_NAME_BUF);
@@ -1919,27 +1910,27 @@ static ssize_t target_core_store_alua_lu_gp(
                 * core_alua_get_lu_gp_by_name below().
                 */
                lu_gp_new = core_alua_get_lu_gp_by_name(strstrip(buf));
-               if (!(lu_gp_new))
+               if (!lu_gp_new)
                        return -ENODEV;
        }
        lu_gp_mem = dev->dev_alua_lu_gp_mem;
-       if (!(lu_gp_mem)) {
+       if (!lu_gp_mem) {
                if (lu_gp_new)
                        core_alua_put_lu_gp_from_name(lu_gp_new);
-               printk(KERN_ERR "NULL struct se_device->dev_alua_lu_gp_mem"
+               pr_err("NULL struct se_device->dev_alua_lu_gp_mem"
                                " pointer\n");
                return -EINVAL;
        }
 
        spin_lock(&lu_gp_mem->lu_gp_mem_lock);
        lu_gp = lu_gp_mem->lu_gp;
-       if ((lu_gp)) {
+       if (lu_gp) {
                /*
                 * Clearing an existing lu_gp association, and replacing
                 * with NULL
                 */
-               if (!(lu_gp_new)) {
-                       printk(KERN_INFO "Target_Core_ConfigFS: Releasing %s/%s"
+               if (!lu_gp_new) {
+                       pr_debug("Target_Core_ConfigFS: Releasing %s/%s"
                                " from ALUA LU Group: core/alua/lu_gps/%s, ID:"
                                " %hu\n",
                                config_item_name(&hba->hba_group.cg_item),
@@ -1964,7 +1955,7 @@ static ssize_t target_core_store_alua_lu_gp(
        __core_alua_attach_lu_gp_mem(lu_gp_mem, lu_gp_new);
        spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: %s %s/%s to ALUA LU Group:"
+       pr_debug("Target_Core_ConfigFS: %s %s/%s to ALUA LU Group:"
                " core/alua/lu_gps/%s, ID: %hu\n",
                (move) ? "Moving" : "Adding",
                config_item_name(&hba->hba_group.cg_item),
@@ -2008,7 +1999,7 @@ static void target_core_dev_release(struct config_item *item)
         *`echo 1 > $CONFIGFS/core/$HBA/$DEV/dev_enable`
         */
        if (se_dev->se_dev_ptr) {
-               printk(KERN_INFO "Target_Core_ConfigFS: Calling se_free_"
+               pr_debug("Target_Core_ConfigFS: Calling se_free_"
                        "virtual_device() for se_dev_ptr: %p\n",
                        se_dev->se_dev_ptr);
 
@@ -2017,14 +2008,14 @@ static void target_core_dev_release(struct config_item *item)
                /*
                 * Release struct se_subsystem_dev->se_dev_su_ptr..
                 */
-               printk(KERN_INFO "Target_Core_ConfigFS: Calling t->free_"
+               pr_debug("Target_Core_ConfigFS: Calling t->free_"
                        "device() for se_dev_su_ptr: %p\n",
                        se_dev->se_dev_su_ptr);
 
                t->free_device(se_dev->se_dev_su_ptr);
        }
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Deallocating se_subsystem"
+       pr_debug("Target_Core_ConfigFS: Deallocating se_subsystem"
                        "_dev_t: %p\n", se_dev);
        kfree(se_dev);
 }
@@ -2039,10 +2030,10 @@ static ssize_t target_core_dev_show(struct config_item *item,
        struct target_core_configfs_attribute *tc_attr = container_of(
                        attr, struct target_core_configfs_attribute, attr);
 
-       if (!(tc_attr->show))
+       if (!tc_attr->show)
                return -EINVAL;
 
-       return tc_attr->show((void *)se_dev, page);
+       return tc_attr->show(se_dev, page);
 }
 
 static ssize_t target_core_dev_store(struct config_item *item,
@@ -2055,10 +2046,10 @@ static ssize_t target_core_dev_store(struct config_item *item,
        struct target_core_configfs_attribute *tc_attr = container_of(
                        attr, struct target_core_configfs_attribute, attr);
 
-       if (!(tc_attr->store))
+       if (!tc_attr->store)
                return -EINVAL;
 
-       return tc_attr->store((void *)se_dev, page, count);
+       return tc_attr->store(se_dev, page, count);
 }
 
 static struct configfs_item_operations target_core_dev_item_ops = {
@@ -2098,7 +2089,7 @@ static ssize_t target_core_alua_lu_gp_show_attr_lu_gp_id(
        struct t10_alua_lu_gp *lu_gp,
        char *page)
 {
-       if (!(lu_gp->lu_gp_valid_id))
+       if (!lu_gp->lu_gp_valid_id)
                return 0;
 
        return sprintf(page, "%hu\n", lu_gp->lu_gp_id);
@@ -2115,12 +2106,12 @@ static ssize_t target_core_alua_lu_gp_store_attr_lu_gp_id(
 
        ret = strict_strtoul(page, 0, &lu_gp_id);
        if (ret < 0) {
-               printk(KERN_ERR "strict_strtoul() returned %d for"
+               pr_err("strict_strtoul() returned %d for"
                        " lu_gp_id\n", ret);
                return -EINVAL;
        }
        if (lu_gp_id > 0x0000ffff) {
-               printk(KERN_ERR "ALUA lu_gp_id: %lu exceeds maximum:"
+               pr_err("ALUA lu_gp_id: %lu exceeds maximum:"
                        " 0x0000ffff\n", lu_gp_id);
                return -EINVAL;
        }
@@ -2129,7 +2120,7 @@ static ssize_t target_core_alua_lu_gp_store_attr_lu_gp_id(
        if (ret < 0)
                return -EINVAL;
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Set ALUA Logical Unit"
+       pr_debug("Target_Core_ConfigFS: Set ALUA Logical Unit"
                " Group: core/alua/lu_gps/%s to ID: %hu\n",
                config_item_name(&alua_lu_gp_cg->cg_item),
                lu_gp->lu_gp_id);
@@ -2167,7 +2158,7 @@ static ssize_t target_core_alua_lu_gp_show_attr_members(
                cur_len++; /* Extra byte for NULL terminator */
 
                if ((cur_len + len) > PAGE_SIZE) {
-                       printk(KERN_WARNING "Ran out of lu_gp_show_attr"
+                       pr_warn("Ran out of lu_gp_show_attr"
                                "_members buffer\n");
                        break;
                }
@@ -2231,7 +2222,7 @@ static struct config_group *target_core_alua_create_lu_gp(
        config_group_init_type_name(alua_lu_gp_cg, name,
                        &target_core_alua_lu_gp_cit);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Allocated ALUA Logical Unit"
+       pr_debug("Target_Core_ConfigFS: Allocated ALUA Logical Unit"
                " Group: core/alua/lu_gps/%s\n",
                config_item_name(alua_lu_gp_ci));
 
@@ -2246,7 +2237,7 @@ static void target_core_alua_drop_lu_gp(
        struct t10_alua_lu_gp *lu_gp = container_of(to_config_group(item),
                        struct t10_alua_lu_gp, lu_gp_group);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Releasing ALUA Logical Unit"
+       pr_debug("Target_Core_ConfigFS: Releasing ALUA Logical Unit"
                " Group: core/alua/lu_gps/%s, ID: %hu\n",
                config_item_name(item), lu_gp->lu_gp_id);
        /*
@@ -2305,22 +2296,22 @@ static ssize_t target_core_alua_tg_pt_gp_store_attr_alua_access_state(
        unsigned long tmp;
        int new_state, ret;
 
-       if (!(tg_pt_gp->tg_pt_gp_valid_id)) {
-               printk(KERN_ERR "Unable to do implict ALUA on non valid"
+       if (!tg_pt_gp->tg_pt_gp_valid_id) {
+               pr_err("Unable to do implict ALUA on non valid"
                        " tg_pt_gp ID: %hu\n", tg_pt_gp->tg_pt_gp_valid_id);
                return -EINVAL;
        }
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk("Unable to extract new ALUA access state from"
+               pr_err("Unable to extract new ALUA access state from"
                                " %s\n", page);
                return -EINVAL;
        }
        new_state = (int)tmp;
 
        if (!(tg_pt_gp->tg_pt_gp_alua_access_type & TPGS_IMPLICT_ALUA)) {
-               printk(KERN_ERR "Unable to process implict configfs ALUA"
+               pr_err("Unable to process implict configfs ALUA"
                        " transition while TPGS_IMPLICT_ALUA is diabled\n");
                return -EINVAL;
        }
@@ -2351,8 +2342,8 @@ static ssize_t target_core_alua_tg_pt_gp_store_attr_alua_access_status(
        unsigned long tmp;
        int new_status, ret;
 
-       if (!(tg_pt_gp->tg_pt_gp_valid_id)) {
-               printk(KERN_ERR "Unable to do set ALUA access status on non"
+       if (!tg_pt_gp->tg_pt_gp_valid_id) {
+               pr_err("Unable to do set ALUA access status on non"
                        " valid tg_pt_gp ID: %hu\n",
                        tg_pt_gp->tg_pt_gp_valid_id);
                return -EINVAL;
@@ -2360,7 +2351,7 @@ static ssize_t target_core_alua_tg_pt_gp_store_attr_alua_access_status(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract new ALUA access status"
+               pr_err("Unable to extract new ALUA access status"
                                " from %s\n", page);
                return -EINVAL;
        }
@@ -2369,7 +2360,7 @@ static ssize_t target_core_alua_tg_pt_gp_store_attr_alua_access_status(
        if ((new_status != ALUA_STATUS_NONE) &&
            (new_status != ALUA_STATUS_ALTERED_BY_EXPLICT_STPG) &&
            (new_status != ALUA_STATUS_ALTERED_BY_IMPLICT_ALUA)) {
-               printk(KERN_ERR "Illegal ALUA access status: 0x%02x\n",
+               pr_err("Illegal ALUA access status: 0x%02x\n",
                                new_status);
                return -EINVAL;
        }
@@ -2420,12 +2411,12 @@ static ssize_t target_core_alua_tg_pt_gp_store_attr_alua_write_metadata(
 
        ret = strict_strtoul(page, 0, &tmp);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract alua_write_metadata\n");
+               pr_err("Unable to extract alua_write_metadata\n");
                return -EINVAL;
        }
 
        if ((tmp != 0) && (tmp != 1)) {
-               printk(KERN_ERR "Illegal value for alua_write_metadata:"
+               pr_err("Illegal value for alua_write_metadata:"
                        " %lu\n", tmp);
                return -EINVAL;
        }
@@ -2507,7 +2498,7 @@ static ssize_t target_core_alua_tg_pt_gp_show_attr_tg_pt_gp_id(
        struct t10_alua_tg_pt_gp *tg_pt_gp,
        char *page)
 {
-       if (!(tg_pt_gp->tg_pt_gp_valid_id))
+       if (!tg_pt_gp->tg_pt_gp_valid_id)
                return 0;
 
        return sprintf(page, "%hu\n", tg_pt_gp->tg_pt_gp_id);
@@ -2524,12 +2515,12 @@ static ssize_t target_core_alua_tg_pt_gp_store_attr_tg_pt_gp_id(
 
        ret = strict_strtoul(page, 0, &tg_pt_gp_id);
        if (ret < 0) {
-               printk(KERN_ERR "strict_strtoul() returned %d for"
+               pr_err("strict_strtoul() returned %d for"
                        " tg_pt_gp_id\n", ret);
                return -EINVAL;
        }
        if (tg_pt_gp_id > 0x0000ffff) {
-               printk(KERN_ERR "ALUA tg_pt_gp_id: %lu exceeds maximum:"
+               pr_err("ALUA tg_pt_gp_id: %lu exceeds maximum:"
                        " 0x0000ffff\n", tg_pt_gp_id);
                return -EINVAL;
        }
@@ -2538,7 +2529,7 @@ static ssize_t target_core_alua_tg_pt_gp_store_attr_tg_pt_gp_id(
        if (ret < 0)
                return -EINVAL;
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Set ALUA Target Port Group: "
+       pr_debug("Target_Core_ConfigFS: Set ALUA Target Port Group: "
                "core/alua/tg_pt_gps/%s to ID: %hu\n",
                config_item_name(&alua_tg_pt_gp_cg->cg_item),
                tg_pt_gp->tg_pt_gp_id);
@@ -2572,14 +2563,14 @@ static ssize_t target_core_alua_tg_pt_gp_show_attr_members(
                lun = port->sep_lun;
 
                cur_len = snprintf(buf, TG_PT_GROUP_NAME_BUF, "%s/%s/tpgt_%hu"
-                       "/%s\n", TPG_TFO(tpg)->get_fabric_name(),
-                       TPG_TFO(tpg)->tpg_get_wwn(tpg),
-                       TPG_TFO(tpg)->tpg_get_tag(tpg),
+                       "/%s\n", tpg->se_tpg_tfo->get_fabric_name(),
+                       tpg->se_tpg_tfo->tpg_get_wwn(tpg),
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg),
                        config_item_name(&lun->lun_group.cg_item));
                cur_len++; /* Extra byte for NULL terminator */
 
                if ((cur_len + len) > PAGE_SIZE) {
-                       printk(KERN_WARNING "Ran out of lu_gp_show_attr"
+                       pr_warn("Ran out of lu_gp_show_attr"
                                "_members buffer\n");
                        break;
                }
@@ -2645,7 +2636,7 @@ static struct config_group *target_core_alua_create_tg_pt_gp(
        struct config_item *alua_tg_pt_gp_ci = NULL;
 
        tg_pt_gp = core_alua_allocate_tg_pt_gp(su_dev, name, 0);
-       if (!(tg_pt_gp))
+       if (!tg_pt_gp)
                return NULL;
 
        alua_tg_pt_gp_cg = &tg_pt_gp->tg_pt_gp_group;
@@ -2654,7 +2645,7 @@ static struct config_group *target_core_alua_create_tg_pt_gp(
        config_group_init_type_name(alua_tg_pt_gp_cg, name,
                        &target_core_alua_tg_pt_gp_cit);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Allocated ALUA Target Port"
+       pr_debug("Target_Core_ConfigFS: Allocated ALUA Target Port"
                " Group: alua/tg_pt_gps/%s\n",
                config_item_name(alua_tg_pt_gp_ci));
 
@@ -2668,7 +2659,7 @@ static void target_core_alua_drop_tg_pt_gp(
        struct t10_alua_tg_pt_gp *tg_pt_gp = container_of(to_config_group(item),
                        struct t10_alua_tg_pt_gp, tg_pt_gp_group);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Releasing ALUA Target Port"
+       pr_debug("Target_Core_ConfigFS: Releasing ALUA Target Port"
                " Group: alua/tg_pt_gps/%s, ID: %hu\n",
                config_item_name(item), tg_pt_gp->tg_pt_gp_id);
        /*
@@ -2759,21 +2750,21 @@ static struct config_group *target_core_make_subdev(
 
        se_dev = kzalloc(sizeof(struct se_subsystem_dev), GFP_KERNEL);
        if (!se_dev) {
-               printk(KERN_ERR "Unable to allocate memory for"
+               pr_err("Unable to allocate memory for"
                                " struct se_subsystem_dev\n");
                goto unlock;
        }
-       INIT_LIST_HEAD(&se_dev->g_se_dev_list);
+       INIT_LIST_HEAD(&se_dev->se_dev_node);
        INIT_LIST_HEAD(&se_dev->t10_wwn.t10_vpd_list);
        spin_lock_init(&se_dev->t10_wwn.t10_vpd_lock);
-       INIT_LIST_HEAD(&se_dev->t10_reservation.registration_list);
-       INIT_LIST_HEAD(&se_dev->t10_reservation.aptpl_reg_list);
-       spin_lock_init(&se_dev->t10_reservation.registration_lock);
-       spin_lock_init(&se_dev->t10_reservation.aptpl_reg_lock);
+       INIT_LIST_HEAD(&se_dev->t10_pr.registration_list);
+       INIT_LIST_HEAD(&se_dev->t10_pr.aptpl_reg_list);
+       spin_lock_init(&se_dev->t10_pr.registration_lock);
+       spin_lock_init(&se_dev->t10_pr.aptpl_reg_lock);
        INIT_LIST_HEAD(&se_dev->t10_alua.tg_pt_gps_list);
        spin_lock_init(&se_dev->t10_alua.tg_pt_gps_lock);
        spin_lock_init(&se_dev->se_dev_lock);
-       se_dev->t10_reservation.pr_aptpl_buf_len = PR_APTPL_BUF_LEN;
+       se_dev->t10_pr.pr_aptpl_buf_len = PR_APTPL_BUF_LEN;
        se_dev->t10_wwn.t10_sub_dev = se_dev;
        se_dev->t10_alua.t10_sub_dev = se_dev;
        se_dev->se_dev_attrib.da_sub_dev = se_dev;
@@ -2783,7 +2774,7 @@ static struct config_group *target_core_make_subdev(
 
        dev_cg->default_groups = kzalloc(sizeof(struct config_group) * 7,
                        GFP_KERNEL);
-       if (!(dev_cg->default_groups))
+       if (!dev_cg->default_groups)
                goto out;
        /*
         * Set se_dev_su_ptr from struct se_subsystem_api returned void ptr
@@ -2794,14 +2785,14 @@ static struct config_group *target_core_make_subdev(
         * configfs tree for device object's struct config_group.
         */
        se_dev->se_dev_su_ptr = t->allocate_virtdevice(hba, name);
-       if (!(se_dev->se_dev_su_ptr)) {
-               printk(KERN_ERR "Unable to locate subsystem dependent pointer"
+       if (!se_dev->se_dev_su_ptr) {
+               pr_err("Unable to locate subsystem dependent pointer"
                        " from allocate_virtdevice()\n");
                goto out;
        }
-       spin_lock(&se_global->g_device_lock);
-       list_add_tail(&se_dev->g_se_dev_list, &se_global->g_se_dev_list);
-       spin_unlock(&se_global->g_device_lock);
+       spin_lock(&se_device_lock);
+       list_add_tail(&se_dev->se_dev_node, &se_dev_list);
+       spin_unlock(&se_device_lock);
 
        config_group_init_type_name(&se_dev->se_dev_group, name,
                        &target_core_dev_cit);
@@ -2826,14 +2817,14 @@ static struct config_group *target_core_make_subdev(
         * Add core/$HBA/$DEV/alua/default_tg_pt_gp
         */
        tg_pt_gp = core_alua_allocate_tg_pt_gp(se_dev, "default_tg_pt_gp", 1);
-       if (!(tg_pt_gp))
+       if (!tg_pt_gp)
                goto out;
 
-       tg_pt_gp_cg = &T10_ALUA(se_dev)->alua_tg_pt_gps_group;
+       tg_pt_gp_cg = &se_dev->t10_alua.alua_tg_pt_gps_group;
        tg_pt_gp_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
                                GFP_KERNEL);
-       if (!(tg_pt_gp_cg->default_groups)) {
-               printk(KERN_ERR "Unable to allocate tg_pt_gp_cg->"
+       if (!tg_pt_gp_cg->default_groups) {
+               pr_err("Unable to allocate tg_pt_gp_cg->"
                                "default_groups\n");
                goto out;
        }
@@ -2842,28 +2833,28 @@ static struct config_group *target_core_make_subdev(
                        "default_tg_pt_gp", &target_core_alua_tg_pt_gp_cit);
        tg_pt_gp_cg->default_groups[0] = &tg_pt_gp->tg_pt_gp_group;
        tg_pt_gp_cg->default_groups[1] = NULL;
-       T10_ALUA(se_dev)->default_tg_pt_gp = tg_pt_gp;
+       se_dev->t10_alua.default_tg_pt_gp = tg_pt_gp;
        /*
         * Add core/$HBA/$DEV/statistics/ default groups
         */
-       dev_stat_grp = &DEV_STAT_GRP(se_dev)->stat_group;
+       dev_stat_grp = &se_dev->dev_stat_grps.stat_group;
        dev_stat_grp->default_groups = kzalloc(sizeof(struct config_group) * 4,
                                GFP_KERNEL);
        if (!dev_stat_grp->default_groups) {
-               printk(KERN_ERR "Unable to allocate dev_stat_grp->default_groups\n");
+               pr_err("Unable to allocate dev_stat_grp->default_groups\n");
                goto out;
        }
        target_stat_setup_dev_default_groups(se_dev);
 
-       printk(KERN_INFO "Target_Core_ConfigFS: Allocated struct se_subsystem_dev:"
+       pr_debug("Target_Core_ConfigFS: Allocated struct se_subsystem_dev:"
                " %p se_dev_su_ptr: %p\n", se_dev, se_dev->se_dev_su_ptr);
 
        mutex_unlock(&hba->hba_access_mutex);
        return &se_dev->se_dev_group;
 out:
-       if (T10_ALUA(se_dev)->default_tg_pt_gp) {
-               core_alua_free_tg_pt_gp(T10_ALUA(se_dev)->default_tg_pt_gp);
-               T10_ALUA(se_dev)->default_tg_pt_gp = NULL;
+       if (se_dev->t10_alua.default_tg_pt_gp) {
+               core_alua_free_tg_pt_gp(se_dev->t10_alua.default_tg_pt_gp);
+               se_dev->t10_alua.default_tg_pt_gp = NULL;
        }
        if (dev_stat_grp)
                kfree(dev_stat_grp->default_groups);
@@ -2896,11 +2887,11 @@ static void target_core_drop_subdev(
        mutex_lock(&hba->hba_access_mutex);
        t = hba->transport;
 
-       spin_lock(&se_global->g_device_lock);
-       list_del(&se_dev->g_se_dev_list);
-       spin_unlock(&se_global->g_device_lock);
+       spin_lock(&se_device_lock);
+       list_del(&se_dev->se_dev_node);
+       spin_unlock(&se_device_lock);
 
-       dev_stat_grp = &DEV_STAT_GRP(se_dev)->stat_group;
+       dev_stat_grp = &se_dev->dev_stat_grps.stat_group;
        for (i = 0; dev_stat_grp->default_groups[i]; i++) {
                df_item = &dev_stat_grp->default_groups[i]->cg_item;
                dev_stat_grp->default_groups[i] = NULL;
@@ -2908,7 +2899,7 @@ static void target_core_drop_subdev(
        }
        kfree(dev_stat_grp->default_groups);
 
-       tg_pt_gp_cg = &T10_ALUA(se_dev)->alua_tg_pt_gps_group;
+       tg_pt_gp_cg = &se_dev->t10_alua.alua_tg_pt_gps_group;
        for (i = 0; tg_pt_gp_cg->default_groups[i]; i++) {
                df_item = &tg_pt_gp_cg->default_groups[i]->cg_item;
                tg_pt_gp_cg->default_groups[i] = NULL;
@@ -2919,7 +2910,7 @@ static void target_core_drop_subdev(
         * core_alua_free_tg_pt_gp() is called from ->default_tg_pt_gp
         * directly from target_core_alua_tg_pt_gp_release().
         */
-       T10_ALUA(se_dev)->default_tg_pt_gp = NULL;
+       se_dev->t10_alua.default_tg_pt_gp = NULL;
 
        dev_cg = &se_dev->se_dev_group;
        for (i = 0; dev_cg->default_groups[i]; i++) {
@@ -2988,13 +2979,13 @@ static ssize_t target_core_hba_store_attr_hba_mode(struct se_hba *hba,
 
        ret = strict_strtoul(page, 0, &mode_flag);
        if (ret < 0) {
-               printk(KERN_ERR "Unable to extract hba mode flag: %d\n", ret);
+               pr_err("Unable to extract hba mode flag: %d\n", ret);
                return -EINVAL;
        }
 
        spin_lock(&hba->device_lock);
-       if (!(list_empty(&hba->hba_dev_list))) {
-               printk(KERN_ERR "Unable to set hba_mode with active devices\n");
+       if (!list_empty(&hba->hba_dev_list)) {
+               pr_err("Unable to set hba_mode with active devices\n");
                spin_unlock(&hba->device_lock);
                return -EINVAL;
        }
@@ -3053,7 +3044,7 @@ static struct config_group *target_core_call_addhbatotarget(
 
        memset(buf, 0, TARGET_CORE_NAME_MAX_LEN);
        if (strlen(name) >= TARGET_CORE_NAME_MAX_LEN) {
-               printk(KERN_ERR "Passed *name strlen(): %d exceeds"
+               pr_err("Passed *name strlen(): %d exceeds"
                        " TARGET_CORE_NAME_MAX_LEN: %d\n", (int)strlen(name),
                        TARGET_CORE_NAME_MAX_LEN);
                return ERR_PTR(-ENAMETOOLONG);
@@ -3061,8 +3052,8 @@ static struct config_group *target_core_call_addhbatotarget(
        snprintf(buf, TARGET_CORE_NAME_MAX_LEN, "%s", name);
 
        str = strstr(buf, "_");
-       if (!(str)) {
-               printk(KERN_ERR "Unable to locate \"_\" for $SUBSYSTEM_PLUGIN_$HOST_ID\n");
+       if (!str) {
+               pr_err("Unable to locate \"_\" for $SUBSYSTEM_PLUGIN_$HOST_ID\n");
                return ERR_PTR(-EINVAL);
        }
        se_plugin_str = buf;
@@ -3071,7 +3062,7 @@ static struct config_group *target_core_call_addhbatotarget(
         * Namely rd_direct and rd_mcp..
         */
        str2 = strstr(str+1, "_");
-       if ((str2)) {
+       if (str2) {
                *str2 = '\0'; /* Terminate for *se_plugin_str */
                str2++; /* Skip to start of plugin dependent ID */
                str = str2;
@@ -3082,7 +3073,7 @@ static struct config_group *target_core_call_addhbatotarget(
 
        ret = strict_strtoul(str, 0, &plugin_dep_id);
        if (ret < 0) {
-               printk(KERN_ERR "strict_strtoul() returned %d for"
+               pr_err("strict_strtoul() returned %d for"
                                " plugin_dep_id\n", ret);
                return ERR_PTR(-EINVAL);
        }
@@ -3135,7 +3126,7 @@ static int __init target_core_init_configfs(void)
        struct t10_alua_lu_gp *lu_gp;
        int ret;
 
-       printk(KERN_INFO "TARGET_CORE[0]: Loading Generic Kernel Storage"
+       pr_debug("TARGET_CORE[0]: Loading Generic Kernel Storage"
                " Engine: %s on %s/%s on "UTS_RELEASE"\n",
                TARGET_CORE_VERSION, utsname()->sysname, utsname()->machine);
 
@@ -3145,10 +3136,9 @@ static int __init target_core_init_configfs(void)
 
        INIT_LIST_HEAD(&g_tf_list);
        mutex_init(&g_tf_lock);
-       init_scsi_index_table();
-       ret = init_se_global();
+       ret = init_se_kmem_caches();
        if (ret < 0)
-               return -1;
+               return ret;
        /*
         * Create $CONFIGFS/target/core default group for HBA <-> Storage Object
         * and ALUA Logical Unit Group and Target Port Group infrastructure.
@@ -3156,44 +3146,44 @@ static int __init target_core_init_configfs(void)
        target_cg = &subsys->su_group;
        target_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
                                GFP_KERNEL);
-       if (!(target_cg->default_groups)) {
-               printk(KERN_ERR "Unable to allocate target_cg->default_groups\n");
+       if (!target_cg->default_groups) {
+               pr_err("Unable to allocate target_cg->default_groups\n");
                goto out_global;
        }
 
-       config_group_init_type_name(&se_global->target_core_hbagroup,
+       config_group_init_type_name(&target_core_hbagroup,
                        "core", &target_core_cit);
-       target_cg->default_groups[0] = &se_global->target_core_hbagroup;
+       target_cg->default_groups[0] = &target_core_hbagroup;
        target_cg->default_groups[1] = NULL;
        /*
         * Create ALUA infrastructure under /sys/kernel/config/target/core/alua/
         */
-       hba_cg = &se_global->target_core_hbagroup;
+       hba_cg = &target_core_hbagroup;
        hba_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
                                GFP_KERNEL);
-       if (!(hba_cg->default_groups)) {
-               printk(KERN_ERR "Unable to allocate hba_cg->default_groups\n");
+       if (!hba_cg->default_groups) {
+               pr_err("Unable to allocate hba_cg->default_groups\n");
                goto out_global;
        }
-       config_group_init_type_name(&se_global->alua_group,
+       config_group_init_type_name(&alua_group,
                        "alua", &target_core_alua_cit);
-       hba_cg->default_groups[0] = &se_global->alua_group;
+       hba_cg->default_groups[0] = &alua_group;
        hba_cg->default_groups[1] = NULL;
        /*
         * Add ALUA Logical Unit Group and Target Port Group ConfigFS
         * groups under /sys/kernel/config/target/core/alua/
         */
-       alua_cg = &se_global->alua_group;
+       alua_cg = &alua_group;
        alua_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
                        GFP_KERNEL);
-       if (!(alua_cg->default_groups)) {
-               printk(KERN_ERR "Unable to allocate alua_cg->default_groups\n");
+       if (!alua_cg->default_groups) {
+               pr_err("Unable to allocate alua_cg->default_groups\n");
                goto out_global;
        }
 
-       config_group_init_type_name(&se_global->alua_lu_gps_group,
+       config_group_init_type_name(&alua_lu_gps_group,
                        "lu_gps", &target_core_alua_lu_gps_cit);
-       alua_cg->default_groups[0] = &se_global->alua_lu_gps_group;
+       alua_cg->default_groups[0] = &alua_lu_gps_group;
        alua_cg->default_groups[1] = NULL;
        /*
         * Add core/alua/lu_gps/default_lu_gp
@@ -3202,11 +3192,11 @@ static int __init target_core_init_configfs(void)
        if (IS_ERR(lu_gp))
                goto out_global;
 
-       lu_gp_cg = &se_global->alua_lu_gps_group;
+       lu_gp_cg = &alua_lu_gps_group;
        lu_gp_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
                        GFP_KERNEL);
-       if (!(lu_gp_cg->default_groups)) {
-               printk(KERN_ERR "Unable to allocate lu_gp_cg->default_groups\n");
+       if (!lu_gp_cg->default_groups) {
+               pr_err("Unable to allocate lu_gp_cg->default_groups\n");
                goto out_global;
        }
 
@@ -3214,17 +3204,17 @@ static int __init target_core_init_configfs(void)
                                &target_core_alua_lu_gp_cit);
        lu_gp_cg->default_groups[0] = &lu_gp->lu_gp_group;
        lu_gp_cg->default_groups[1] = NULL;
-       se_global->default_lu_gp = lu_gp;
+       default_lu_gp = lu_gp;
        /*
         * Register the target_core_mod subsystem with configfs.
         */
        ret = configfs_register_subsystem(subsys);
        if (ret < 0) {
-               printk(KERN_ERR "Error %d while registering subsystem %s\n",
+               pr_err("Error %d while registering subsystem %s\n",
                        ret, subsys->su_group.cg_item.ci_namebuf);
                goto out_global;
        }
-       printk(KERN_INFO "TARGET_CORE[0]: Initialized ConfigFS Fabric"
+       pr_debug("TARGET_CORE[0]: Initialized ConfigFS Fabric"
                " Infrastructure: "TARGET_CORE_CONFIGFS_VERSION" on %s/%s"
                " on "UTS_RELEASE"\n", utsname()->sysname, utsname()->machine);
        /*
@@ -3244,9 +3234,9 @@ out:
        core_dev_release_virtual_lun0();
        rd_module_exit();
 out_global:
-       if (se_global->default_lu_gp) {
-               core_alua_free_lu_gp(se_global->default_lu_gp);
-               se_global->default_lu_gp = NULL;
+       if (default_lu_gp) {
+               core_alua_free_lu_gp(default_lu_gp);
+               default_lu_gp = NULL;
        }
        if (lu_gp_cg)
                kfree(lu_gp_cg->default_groups);
@@ -3255,8 +3245,8 @@ out_global:
        if (hba_cg)
                kfree(hba_cg->default_groups);
        kfree(target_cg->default_groups);
-       release_se_global();
-       return -1;
+       release_se_kmem_caches();
+       return ret;
 }
 
 static void __exit target_core_exit_configfs(void)
@@ -3266,10 +3256,9 @@ static void __exit target_core_exit_configfs(void)
        struct config_item *item;
        int i;
 
-       se_global->in_shutdown = 1;
        subsys = target_core_subsystem[0];
 
-       lu_gp_cg = &se_global->alua_lu_gps_group;
+       lu_gp_cg = &alua_lu_gps_group;
        for (i = 0; lu_gp_cg->default_groups[i]; i++) {
                item = &lu_gp_cg->default_groups[i]->cg_item;
                lu_gp_cg->default_groups[i] = NULL;
@@ -3278,7 +3267,7 @@ static void __exit target_core_exit_configfs(void)
        kfree(lu_gp_cg->default_groups);
        lu_gp_cg->default_groups = NULL;
 
-       alua_cg = &se_global->alua_group;
+       alua_cg = &alua_group;
        for (i = 0; alua_cg->default_groups[i]; i++) {
                item = &alua_cg->default_groups[i]->cg_item;
                alua_cg->default_groups[i] = NULL;
@@ -3287,7 +3276,7 @@ static void __exit target_core_exit_configfs(void)
        kfree(alua_cg->default_groups);
        alua_cg->default_groups = NULL;
 
-       hba_cg = &se_global->target_core_hbagroup;
+       hba_cg = &target_core_hbagroup;
        for (i = 0; hba_cg->default_groups[i]; i++) {
                item = &hba_cg->default_groups[i]->cg_item;
                hba_cg->default_groups[i] = NULL;
@@ -3302,17 +3291,15 @@ static void __exit target_core_exit_configfs(void)
        configfs_unregister_subsystem(subsys);
        kfree(subsys->su_group.default_groups);
 
-       core_alua_free_lu_gp(se_global->default_lu_gp);
-       se_global->default_lu_gp = NULL;
+       core_alua_free_lu_gp(default_lu_gp);
+       default_lu_gp = NULL;
 
-       printk(KERN_INFO "TARGET_CORE[0]: Released ConfigFS Fabric"
+       pr_debug("TARGET_CORE[0]: Released ConfigFS Fabric"
                        " Infrastructure\n");
 
        core_dev_release_virtual_lun0();
        rd_module_exit();
-       release_se_global();
-
-       return;
+       release_se_kmem_caches();
 }
 
 MODULE_DESCRIPTION("Target_Core_Mod/ConfigFS");
index ba698ea62bb2a52c19d87a6d1df7de10fcba9ecd..b38b6c993e6555855be8cd830451d41a0ab6e1d5 100644 (file)
@@ -1,7 +1,7 @@
 /*******************************************************************************
  * Filename:  target_core_device.c (based on iscsi_target_device.c)
  *
- * This file contains the iSCSI Virtual Device and Disk Transport
+ * This file contains the TCM Virtual Device and Disk Transport
  * agnostic related functions.
  *
  * Copyright (c) 2003, 2004, 2005 PyX Technologies, Inc.
 static void se_dev_start(struct se_device *dev);
 static void se_dev_stop(struct se_device *dev);
 
-int transport_get_lun_for_cmd(
-       struct se_cmd *se_cmd,
-       unsigned char *cdb,
-       u32 unpacked_lun)
+static struct se_hba *lun0_hba;
+static struct se_subsystem_dev *lun0_su_dev;
+/* not static, needed by tpg.c */
+struct se_device *g_lun0_dev;
+
+int transport_lookup_cmd_lun(struct se_cmd *se_cmd, u32 unpacked_lun)
 {
-       struct se_dev_entry *deve;
        struct se_lun *se_lun = NULL;
-       struct se_session *se_sess = SE_SESS(se_cmd);
+       struct se_session *se_sess = se_cmd->se_sess;
+       struct se_device *dev;
        unsigned long flags;
-       int read_only = 0;
 
-       spin_lock_irq(&SE_NODE_ACL(se_sess)->device_list_lock);
-       deve = se_cmd->se_deve =
-                       &SE_NODE_ACL(se_sess)->device_list[unpacked_lun];
-       if (deve->lun_flags & TRANSPORT_LUNFLAGS_INITIATOR_ACCESS) {
-               if (se_cmd) {
-                       deve->total_cmds++;
-                       deve->total_bytes += se_cmd->data_length;
-
-                       if (se_cmd->data_direction == DMA_TO_DEVICE) {
-                               if (deve->lun_flags &
-                                               TRANSPORT_LUNFLAGS_READ_ONLY) {
-                                       read_only = 1;
-                                       goto out;
-                               }
-                               deve->write_bytes += se_cmd->data_length;
-                       } else if (se_cmd->data_direction ==
-                                  DMA_FROM_DEVICE) {
-                               deve->read_bytes += se_cmd->data_length;
-                       }
+       if (unpacked_lun >= TRANSPORT_MAX_LUNS_PER_TPG) {
+               se_cmd->scsi_sense_reason = TCM_NON_EXISTENT_LUN;
+               se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+               return -ENODEV;
+       }
+
+       spin_lock_irqsave(&se_sess->se_node_acl->device_list_lock, flags);
+       se_cmd->se_deve = &se_sess->se_node_acl->device_list[unpacked_lun];
+       if (se_cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_INITIATOR_ACCESS) {
+               struct se_dev_entry *deve = se_cmd->se_deve;
+
+               deve->total_cmds++;
+               deve->total_bytes += se_cmd->data_length;
+
+               if ((se_cmd->data_direction == DMA_TO_DEVICE) &&
+                   (deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)) {
+                       se_cmd->scsi_sense_reason = TCM_WRITE_PROTECTED;
+                       se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+                       pr_err("TARGET_CORE[%s]: Detected WRITE_PROTECTED LUN"
+                               " Access for 0x%08x\n",
+                               se_cmd->se_tfo->get_fabric_name(),
+                               unpacked_lun);
+                       spin_unlock_irqrestore(&se_sess->se_node_acl->device_list_lock, flags);
+                       return -EACCES;
                }
+
+               if (se_cmd->data_direction == DMA_TO_DEVICE)
+                       deve->write_bytes += se_cmd->data_length;
+               else if (se_cmd->data_direction == DMA_FROM_DEVICE)
+                       deve->read_bytes += se_cmd->data_length;
+
                deve->deve_cmds++;
 
-               se_lun = se_cmd->se_lun = deve->se_lun;
+               se_lun = deve->se_lun;
+               se_cmd->se_lun = deve->se_lun;
                se_cmd->pr_res_key = deve->pr_res_key;
                se_cmd->orig_fe_lun = unpacked_lun;
-               se_cmd->se_orig_obj_ptr = SE_LUN(se_cmd)->lun_se_dev;
+               se_cmd->se_orig_obj_ptr = se_cmd->se_lun->lun_se_dev;
                se_cmd->se_cmd_flags |= SCF_SE_LUN_CMD;
        }
-out:
-       spin_unlock_irq(&SE_NODE_ACL(se_sess)->device_list_lock);
+       spin_unlock_irqrestore(&se_sess->se_node_acl->device_list_lock, flags);
 
        if (!se_lun) {
-               if (read_only) {
-                       se_cmd->scsi_sense_reason = TCM_WRITE_PROTECTED;
+               /*
+                * Use the se_portal_group->tpg_virt_lun0 to allow for
+                * REPORT_LUNS, et al to be returned when no active
+                * MappedLUN=0 exists for this Initiator Port.
+                */
+               if (unpacked_lun != 0) {
+                       se_cmd->scsi_sense_reason = TCM_NON_EXISTENT_LUN;
                        se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-                       printk("TARGET_CORE[%s]: Detected WRITE_PROTECTED LUN"
+                       pr_err("TARGET_CORE[%s]: Detected NON_EXISTENT_LUN"
                                " Access for 0x%08x\n",
-                               CMD_TFO(se_cmd)->get_fabric_name(),
+                               se_cmd->se_tfo->get_fabric_name(),
                                unpacked_lun);
-                       return -1;
-               } else {
-                       /*
-                        * Use the se_portal_group->tpg_virt_lun0 to allow for
-                        * REPORT_LUNS, et al to be returned when no active
-                        * MappedLUN=0 exists for this Initiator Port.
-                        */
-                       if (unpacked_lun != 0) {
-                               se_cmd->scsi_sense_reason = TCM_NON_EXISTENT_LUN;
-                               se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-                               printk("TARGET_CORE[%s]: Detected NON_EXISTENT_LUN"
-                                       " Access for 0x%08x\n",
-                                       CMD_TFO(se_cmd)->get_fabric_name(),
-                                       unpacked_lun);
-                               return -1;
-                       }
-                       /*
-                        * Force WRITE PROTECT for virtual LUN 0
-                        */
-                       if ((se_cmd->data_direction != DMA_FROM_DEVICE) &&
-                           (se_cmd->data_direction != DMA_NONE)) {
-                               se_cmd->scsi_sense_reason = TCM_WRITE_PROTECTED;
-                               se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-                               return -1;
-                       }
-#if 0
-                       printk("TARGET_CORE[%s]: Using virtual LUN0! :-)\n",
-                               CMD_TFO(se_cmd)->get_fabric_name());
-#endif
-                       se_lun = se_cmd->se_lun = &se_sess->se_tpg->tpg_virt_lun0;
-                       se_cmd->orig_fe_lun = 0;
-                       se_cmd->se_orig_obj_ptr = SE_LUN(se_cmd)->lun_se_dev;
-                       se_cmd->se_cmd_flags |= SCF_SE_LUN_CMD;
+                       return -ENODEV;
                }
+               /*
+                * Force WRITE PROTECT for virtual LUN 0
+                */
+               if ((se_cmd->data_direction != DMA_FROM_DEVICE) &&
+                   (se_cmd->data_direction != DMA_NONE)) {
+                       se_cmd->scsi_sense_reason = TCM_WRITE_PROTECTED;
+                       se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+                       return -EACCES;
+               }
+
+               se_lun = &se_sess->se_tpg->tpg_virt_lun0;
+               se_cmd->se_lun = &se_sess->se_tpg->tpg_virt_lun0;
+               se_cmd->orig_fe_lun = 0;
+               se_cmd->se_orig_obj_ptr = se_cmd->se_lun->lun_se_dev;
+               se_cmd->se_cmd_flags |= SCF_SE_LUN_CMD;
        }
        /*
         * Determine if the struct se_lun is online.
+        * FIXME: Check for LUN_RESET + UNIT Attention
         */
-/* #warning FIXME: Check for LUN_RESET + UNIT Attention */
        if (se_dev_check_online(se_lun->lun_se_dev) != 0) {
                se_cmd->scsi_sense_reason = TCM_NON_EXISTENT_LUN;
                se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-               return -1;
+               return -ENODEV;
        }
 
-       {
-       struct se_device *dev = se_lun->lun_se_dev;
-       spin_lock_irq(&dev->stats_lock);
+       /* Directly associate cmd with se_dev */
+       se_cmd->se_dev = se_lun->lun_se_dev;
+
+       /* TODO: get rid of this and use atomics for stats */
+       dev = se_lun->lun_se_dev;
+       spin_lock_irqsave(&dev->stats_lock, flags);
        dev->num_cmds++;
        if (se_cmd->data_direction == DMA_TO_DEVICE)
                dev->write_bytes += se_cmd->data_length;
        else if (se_cmd->data_direction == DMA_FROM_DEVICE)
                dev->read_bytes += se_cmd->data_length;
-       spin_unlock_irq(&dev->stats_lock);
-       }
+       spin_unlock_irqrestore(&dev->stats_lock, flags);
 
        /*
         * Add the iscsi_cmd_t to the struct se_lun's cmd list.  This list is used
         * for tracking state of struct se_cmds during LUN shutdown events.
         */
        spin_lock_irqsave(&se_lun->lun_cmd_lock, flags);
-       list_add_tail(&se_cmd->se_lun_list, &se_lun->lun_cmd_list);
-       atomic_set(&T_TASK(se_cmd)->transport_lun_active, 1);
-#if 0
-       printk(KERN_INFO "Adding ITT: 0x%08x to LUN LIST[%d]\n",
-               CMD_TFO(se_cmd)->get_task_tag(se_cmd), se_lun->unpacked_lun);
-#endif
+       list_add_tail(&se_cmd->se_lun_node, &se_lun->lun_cmd_list);
+       atomic_set(&se_cmd->transport_lun_active, 1);
        spin_unlock_irqrestore(&se_lun->lun_cmd_lock, flags);
 
        return 0;
 }
-EXPORT_SYMBOL(transport_get_lun_for_cmd);
+EXPORT_SYMBOL(transport_lookup_cmd_lun);
 
-int transport_get_lun_for_tmr(
-       struct se_cmd *se_cmd,
-       u32 unpacked_lun)
+int transport_lookup_tmr_lun(struct se_cmd *se_cmd, u32 unpacked_lun)
 {
-       struct se_device *dev = NULL;
        struct se_dev_entry *deve;
        struct se_lun *se_lun = NULL;
-       struct se_session *se_sess = SE_SESS(se_cmd);
+       struct se_session *se_sess = se_cmd->se_sess;
        struct se_tmr_req *se_tmr = se_cmd->se_tmr_req;
+       unsigned long flags;
+
+       if (unpacked_lun >= TRANSPORT_MAX_LUNS_PER_TPG) {
+               se_cmd->scsi_sense_reason = TCM_NON_EXISTENT_LUN;
+               se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+               return -ENODEV;
+       }
+
+       spin_lock_irqsave(&se_sess->se_node_acl->device_list_lock, flags);
+       se_cmd->se_deve = &se_sess->se_node_acl->device_list[unpacked_lun];
+       deve = se_cmd->se_deve;
 
-       spin_lock_irq(&SE_NODE_ACL(se_sess)->device_list_lock);
-       deve = se_cmd->se_deve =
-                       &SE_NODE_ACL(se_sess)->device_list[unpacked_lun];
        if (deve->lun_flags & TRANSPORT_LUNFLAGS_INITIATOR_ACCESS) {
-               se_lun = se_cmd->se_lun = se_tmr->tmr_lun = deve->se_lun;
-               dev = se_lun->lun_se_dev;
+               se_tmr->tmr_lun = deve->se_lun;
+               se_cmd->se_lun = deve->se_lun;
+               se_lun = deve->se_lun;
                se_cmd->pr_res_key = deve->pr_res_key;
                se_cmd->orig_fe_lun = unpacked_lun;
-               se_cmd->se_orig_obj_ptr = SE_LUN(se_cmd)->lun_se_dev;
-/*             se_cmd->se_cmd_flags |= SCF_SE_LUN_CMD; */
+               se_cmd->se_orig_obj_ptr = se_cmd->se_dev;
        }
-       spin_unlock_irq(&SE_NODE_ACL(se_sess)->device_list_lock);
+       spin_unlock_irqrestore(&se_sess->se_node_acl->device_list_lock, flags);
 
        if (!se_lun) {
-               printk(KERN_INFO "TARGET_CORE[%s]: Detected NON_EXISTENT_LUN"
+               pr_debug("TARGET_CORE[%s]: Detected NON_EXISTENT_LUN"
                        " Access for 0x%08x\n",
-                       CMD_TFO(se_cmd)->get_fabric_name(),
+                       se_cmd->se_tfo->get_fabric_name(),
                        unpacked_lun);
                se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-               return -1;
+               return -ENODEV;
        }
        /*
         * Determine if the struct se_lun is online.
+        * FIXME: Check for LUN_RESET + UNIT Attention
         */
-/* #warning FIXME: Check for LUN_RESET + UNIT Attention */
        if (se_dev_check_online(se_lun->lun_se_dev) != 0) {
                se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-               return -1;
+               return -ENODEV;
        }
-       se_tmr->tmr_dev = dev;
 
-       spin_lock(&dev->se_tmr_lock);
-       list_add_tail(&se_tmr->tmr_list, &dev->dev_tmr_list);
-       spin_unlock(&dev->se_tmr_lock);
+       /* Directly associate cmd with se_dev */
+       se_cmd->se_dev = se_lun->lun_se_dev;
+       se_tmr->tmr_dev = se_lun->lun_se_dev;
+
+       spin_lock_irqsave(&se_tmr->tmr_dev->se_tmr_lock, flags);
+       list_add_tail(&se_tmr->tmr_list, &se_tmr->tmr_dev->dev_tmr_list);
+       spin_unlock_irqrestore(&se_tmr->tmr_dev->se_tmr_lock, flags);
 
        return 0;
 }
-EXPORT_SYMBOL(transport_get_lun_for_tmr);
+EXPORT_SYMBOL(transport_lookup_tmr_lun);
 
 /*
  * This function is called from core_scsi3_emulate_pro_register_and_move()
@@ -249,17 +255,17 @@ struct se_dev_entry *core_get_se_deve_from_rtpi(
                        continue;
 
                lun = deve->se_lun;
-               if (!(lun)) {
-                       printk(KERN_ERR "%s device entries device pointer is"
+               if (!lun) {
+                       pr_err("%s device entries device pointer is"
                                " NULL, but Initiator has access.\n",
-                               TPG_TFO(tpg)->get_fabric_name());
+                               tpg->se_tpg_tfo->get_fabric_name());
                        continue;
                }
                port = lun->lun_sep;
-               if (!(port)) {
-                       printk(KERN_ERR "%s device entries device pointer is"
+               if (!port) {
+                       pr_err("%s device entries device pointer is"
                                " NULL, but Initiator has access.\n",
-                               TPG_TFO(tpg)->get_fabric_name());
+                               tpg->se_tpg_tfo->get_fabric_name());
                        continue;
                }
                if (port->sep_rtpi != rtpi)
@@ -295,9 +301,9 @@ int core_free_device_list_for_node(
                        continue;
 
                if (!deve->se_lun) {
-                       printk(KERN_ERR "%s device entries device pointer is"
+                       pr_err("%s device entries device pointer is"
                                " NULL, but Initiator has access.\n",
-                               TPG_TFO(tpg)->get_fabric_name());
+                               tpg->se_tpg_tfo->get_fabric_name());
                        continue;
                }
                lun = deve->se_lun;
@@ -323,8 +329,6 @@ void core_dec_lacl_count(struct se_node_acl *se_nacl, struct se_cmd *se_cmd)
        deve = &se_nacl->device_list[se_cmd->orig_fe_lun];
        deve->deve_cmds--;
        spin_unlock_irq(&se_nacl->device_list_lock);
-
-       return;
 }
 
 void core_update_device_list_access(
@@ -344,8 +348,6 @@ void core_update_device_list_access(
                deve->lun_flags |= TRANSPORT_LUNFLAGS_READ_ONLY;
        }
        spin_unlock_irq(&nacl->device_list_lock);
-
-       return;
 }
 
 /*      core_update_device_list_for_node():
@@ -370,7 +372,7 @@ int core_update_device_list_for_node(
         * struct se_dev_entry pointers below as logic in
         * core_alua_do_transition_tg_pt() depends on these being present.
         */
-       if (!(enable)) {
+       if (!enable) {
                /*
                 * deve->se_lun_acl will be NULL for demo-mode created LUNs
                 * that have not been explicitly concerted to MappedLUNs ->
@@ -393,18 +395,18 @@ int core_update_device_list_for_node(
                 */
                if (deve->lun_flags & TRANSPORT_LUNFLAGS_INITIATOR_ACCESS) {
                        if (deve->se_lun_acl != NULL) {
-                               printk(KERN_ERR "struct se_dev_entry->se_lun_acl"
+                               pr_err("struct se_dev_entry->se_lun_acl"
                                        " already set for demo mode -> explict"
                                        " LUN ACL transition\n");
                                spin_unlock_irq(&nacl->device_list_lock);
-                               return -1;
+                               return -EINVAL;
                        }
                        if (deve->se_lun != lun) {
-                               printk(KERN_ERR "struct se_dev_entry->se_lun does"
+                               pr_err("struct se_dev_entry->se_lun does"
                                        " match passed struct se_lun for demo mode"
                                        " -> explict LUN ACL transition\n");
                                spin_unlock_irq(&nacl->device_list_lock);
-                               return -1;
+                               return -EINVAL;
                        }
                        deve->se_lun_acl = lun_acl;
                        trans = 1;
@@ -492,8 +494,6 @@ void core_clear_lun_from_tpg(struct se_lun *lun, struct se_portal_group *tpg)
                spin_lock_bh(&tpg->acl_node_lock);
        }
        spin_unlock_bh(&tpg->acl_node_lock);
-
-       return;
 }
 
 static struct se_port *core_alloc_port(struct se_device *dev)
@@ -501,9 +501,9 @@ static struct se_port *core_alloc_port(struct se_device *dev)
        struct se_port *port, *port_tmp;
 
        port = kzalloc(sizeof(struct se_port), GFP_KERNEL);
-       if (!(port)) {
-               printk(KERN_ERR "Unable to allocate struct se_port\n");
-               return NULL;
+       if (!port) {
+               pr_err("Unable to allocate struct se_port\n");
+               return ERR_PTR(-ENOMEM);
        }
        INIT_LIST_HEAD(&port->sep_alua_list);
        INIT_LIST_HEAD(&port->sep_list);
@@ -513,10 +513,10 @@ static struct se_port *core_alloc_port(struct se_device *dev)
 
        spin_lock(&dev->se_port_lock);
        if (dev->dev_port_count == 0x0000ffff) {
-               printk(KERN_WARNING "Reached dev->dev_port_count =="
+               pr_warn("Reached dev->dev_port_count =="
                                " 0x0000ffff\n");
                spin_unlock(&dev->se_port_lock);
-               return NULL;
+               return ERR_PTR(-ENOSPC);
        }
 again:
        /*
@@ -532,7 +532,7 @@ again:
         * 3h to FFFFh    Relative port 3 through 65 535
         */
        port->sep_rtpi = dev->dev_rpti_counter++;
-       if (!(port->sep_rtpi))
+       if (!port->sep_rtpi)
                goto again;
 
        list_for_each_entry(port_tmp, &dev->dev_sep_list, sep_list) {
@@ -554,7 +554,7 @@ static void core_export_port(
        struct se_port *port,
        struct se_lun *lun)
 {
-       struct se_subsystem_dev *su_dev = SU_DEV(dev);
+       struct se_subsystem_dev *su_dev = dev->se_sub_dev;
        struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem = NULL;
 
        spin_lock(&dev->se_port_lock);
@@ -567,20 +567,20 @@ static void core_export_port(
        list_add_tail(&port->sep_list, &dev->dev_sep_list);
        spin_unlock(&dev->se_port_lock);
 
-       if (T10_ALUA(su_dev)->alua_type == SPC3_ALUA_EMULATED) {
+       if (su_dev->t10_alua.alua_type == SPC3_ALUA_EMULATED) {
                tg_pt_gp_mem = core_alua_allocate_tg_pt_gp_mem(port);
                if (IS_ERR(tg_pt_gp_mem) || !tg_pt_gp_mem) {
-                       printk(KERN_ERR "Unable to allocate t10_alua_tg_pt"
+                       pr_err("Unable to allocate t10_alua_tg_pt"
                                        "_gp_member_t\n");
                        return;
                }
                spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
                __core_alua_attach_tg_pt_gp_mem(tg_pt_gp_mem,
-                       T10_ALUA(su_dev)->default_tg_pt_gp);
+                       su_dev->t10_alua.default_tg_pt_gp);
                spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
-               printk(KERN_INFO "%s/%s: Adding to default ALUA Target Port"
+               pr_debug("%s/%s: Adding to default ALUA Target Port"
                        " Group: alua/default_tg_pt_gp\n",
-                       TRANSPORT(dev)->name, TPG_TFO(tpg)->get_fabric_name());
+                       dev->transport->name, tpg->se_tpg_tfo->get_fabric_name());
        }
 
        dev->dev_port_count++;
@@ -607,8 +607,6 @@ static void core_release_port(struct se_device *dev, struct se_port *port)
        list_del(&port->sep_list);
        dev->dev_port_count--;
        kfree(port);
-
-       return;
 }
 
 int core_dev_export(
@@ -619,8 +617,8 @@ int core_dev_export(
        struct se_port *port;
 
        port = core_alloc_port(dev);
-       if (!(port))
-               return -1;
+       if (IS_ERR(port))
+               return PTR_ERR(port);
 
        lun->lun_se_dev = dev;
        se_dev_start(dev);
@@ -657,33 +655,35 @@ int transport_core_report_lun_response(struct se_cmd *se_cmd)
 {
        struct se_dev_entry *deve;
        struct se_lun *se_lun;
-       struct se_session *se_sess = SE_SESS(se_cmd);
+       struct se_session *se_sess = se_cmd->se_sess;
        struct se_task *se_task;
-       unsigned char *buf = (unsigned char *)T_TASK(se_cmd)->t_task_buf;
+       unsigned char *buf;
        u32 cdb_offset = 0, lun_count = 0, offset = 8, i;
 
-       list_for_each_entry(se_task, &T_TASK(se_cmd)->t_task_list, t_list)
+       list_for_each_entry(se_task, &se_cmd->t_task_list, t_list)
                break;
 
-       if (!(se_task)) {
-               printk(KERN_ERR "Unable to locate struct se_task for struct se_cmd\n");
+       if (!se_task) {
+               pr_err("Unable to locate struct se_task for struct se_cmd\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
 
+       buf = transport_kmap_first_data_page(se_cmd);
+
        /*
         * If no struct se_session pointer is present, this struct se_cmd is
         * coming via a target_core_mod PASSTHROUGH op, and not through
         * a $FABRIC_MOD.  In that case, report LUN=0 only.
         */
-       if (!(se_sess)) {
+       if (!se_sess) {
                int_to_scsilun(0, (struct scsi_lun *)&buf[offset]);
                lun_count = 1;
                goto done;
        }
 
-       spin_lock_irq(&SE_NODE_ACL(se_sess)->device_list_lock);
+       spin_lock_irq(&se_sess->se_node_acl->device_list_lock);
        for (i = 0; i < TRANSPORT_MAX_LUNS_PER_TPG; i++) {
-               deve = &SE_NODE_ACL(se_sess)->device_list[i];
+               deve = &se_sess->se_node_acl->device_list[i];
                if (!(deve->lun_flags & TRANSPORT_LUNFLAGS_INITIATOR_ACCESS))
                        continue;
                se_lun = deve->se_lun;
@@ -700,12 +700,13 @@ int transport_core_report_lun_response(struct se_cmd *se_cmd)
                offset += 8;
                cdb_offset += 8;
        }
-       spin_unlock_irq(&SE_NODE_ACL(se_sess)->device_list_lock);
+       spin_unlock_irq(&se_sess->se_node_acl->device_list_lock);
 
        /*
         * See SPC3 r07, page 159.
         */
 done:
+       transport_kunmap_first_data_page(se_cmd);
        lun_count *= 8;
        buf[0] = ((lun_count >> 24) & 0xff);
        buf[1] = ((lun_count >> 16) & 0xff);
@@ -744,26 +745,20 @@ void se_release_device_for_hba(struct se_device *dev)
        core_scsi3_free_all_registrations(dev);
        se_release_vpd_for_dev(dev);
 
-       kfree(dev->dev_status_queue_obj);
-       kfree(dev->dev_queue_obj);
        kfree(dev);
-
-       return;
 }
 
 void se_release_vpd_for_dev(struct se_device *dev)
 {
        struct t10_vpd *vpd, *vpd_tmp;
 
-       spin_lock(&DEV_T10_WWN(dev)->t10_vpd_lock);
+       spin_lock(&dev->se_sub_dev->t10_wwn.t10_vpd_lock);
        list_for_each_entry_safe(vpd, vpd_tmp,
-                       &DEV_T10_WWN(dev)->t10_vpd_list, vpd_list) {
+                       &dev->se_sub_dev->t10_wwn.t10_vpd_list, vpd_list) {
                list_del(&vpd->vpd_list);
                kfree(vpd);
        }
-       spin_unlock(&DEV_T10_WWN(dev)->t10_vpd_lock);
-
-       return;
+       spin_unlock(&dev->se_sub_dev->t10_wwn.t10_vpd_lock);
 }
 
 /*     se_free_virtual_device():
@@ -822,12 +817,13 @@ static void se_dev_stop(struct se_device *dev)
 
 int se_dev_check_online(struct se_device *dev)
 {
+       unsigned long flags;
        int ret;
 
-       spin_lock_irq(&dev->dev_status_lock);
+       spin_lock_irqsave(&dev->dev_status_lock, flags);
        ret = ((dev->dev_status & TRANSPORT_DEVICE_ACTIVATED) ||
               (dev->dev_status & TRANSPORT_DEVICE_DEACTIVATED)) ? 0 : 1;
-       spin_unlock_irq(&dev->dev_status_lock);
+       spin_unlock_irqrestore(&dev->dev_status_lock, flags);
 
        return ret;
 }
@@ -849,59 +845,61 @@ void se_dev_set_default_attribs(
 {
        struct queue_limits *limits = &dev_limits->limits;
 
-       DEV_ATTRIB(dev)->emulate_dpo = DA_EMULATE_DPO;
-       DEV_ATTRIB(dev)->emulate_fua_write = DA_EMULATE_FUA_WRITE;
-       DEV_ATTRIB(dev)->emulate_fua_read = DA_EMULATE_FUA_READ;
-       DEV_ATTRIB(dev)->emulate_write_cache = DA_EMULATE_WRITE_CACHE;
-       DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl = DA_EMULATE_UA_INTLLCK_CTRL;
-       DEV_ATTRIB(dev)->emulate_tas = DA_EMULATE_TAS;
-       DEV_ATTRIB(dev)->emulate_tpu = DA_EMULATE_TPU;
-       DEV_ATTRIB(dev)->emulate_tpws = DA_EMULATE_TPWS;
-       DEV_ATTRIB(dev)->emulate_reservations = DA_EMULATE_RESERVATIONS;
-       DEV_ATTRIB(dev)->emulate_alua = DA_EMULATE_ALUA;
-       DEV_ATTRIB(dev)->enforce_pr_isids = DA_ENFORCE_PR_ISIDS;
+       dev->se_sub_dev->se_dev_attrib.emulate_dpo = DA_EMULATE_DPO;
+       dev->se_sub_dev->se_dev_attrib.emulate_fua_write = DA_EMULATE_FUA_WRITE;
+       dev->se_sub_dev->se_dev_attrib.emulate_fua_read = DA_EMULATE_FUA_READ;
+       dev->se_sub_dev->se_dev_attrib.emulate_write_cache = DA_EMULATE_WRITE_CACHE;
+       dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl = DA_EMULATE_UA_INTLLCK_CTRL;
+       dev->se_sub_dev->se_dev_attrib.emulate_tas = DA_EMULATE_TAS;
+       dev->se_sub_dev->se_dev_attrib.emulate_tpu = DA_EMULATE_TPU;
+       dev->se_sub_dev->se_dev_attrib.emulate_tpws = DA_EMULATE_TPWS;
+       dev->se_sub_dev->se_dev_attrib.emulate_reservations = DA_EMULATE_RESERVATIONS;
+       dev->se_sub_dev->se_dev_attrib.emulate_alua = DA_EMULATE_ALUA;
+       dev->se_sub_dev->se_dev_attrib.enforce_pr_isids = DA_ENFORCE_PR_ISIDS;
+       dev->se_sub_dev->se_dev_attrib.is_nonrot = DA_IS_NONROT;
+       dev->se_sub_dev->se_dev_attrib.emulate_rest_reord = DA_EMULATE_REST_REORD;
        /*
         * The TPU=1 and TPWS=1 settings will be set in TCM/IBLOCK
         * iblock_create_virtdevice() from struct queue_limits values
         * if blk_queue_discard()==1
         */
-       DEV_ATTRIB(dev)->max_unmap_lba_count = DA_MAX_UNMAP_LBA_COUNT;
-       DEV_ATTRIB(dev)->max_unmap_block_desc_count =
-                               DA_MAX_UNMAP_BLOCK_DESC_COUNT;
-       DEV_ATTRIB(dev)->unmap_granularity = DA_UNMAP_GRANULARITY_DEFAULT;
-       DEV_ATTRIB(dev)->unmap_granularity_alignment =
+       dev->se_sub_dev->se_dev_attrib.max_unmap_lba_count = DA_MAX_UNMAP_LBA_COUNT;
+       dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count =
+               DA_MAX_UNMAP_BLOCK_DESC_COUNT;
+       dev->se_sub_dev->se_dev_attrib.unmap_granularity = DA_UNMAP_GRANULARITY_DEFAULT;
+       dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment =
                                DA_UNMAP_GRANULARITY_ALIGNMENT_DEFAULT;
        /*
         * block_size is based on subsystem plugin dependent requirements.
         */
-       DEV_ATTRIB(dev)->hw_block_size = limits->logical_block_size;
-       DEV_ATTRIB(dev)->block_size = limits->logical_block_size;
+       dev->se_sub_dev->se_dev_attrib.hw_block_size = limits->logical_block_size;
+       dev->se_sub_dev->se_dev_attrib.block_size = limits->logical_block_size;
        /*
         * max_sectors is based on subsystem plugin dependent requirements.
         */
-       DEV_ATTRIB(dev)->hw_max_sectors = limits->max_hw_sectors;
-       DEV_ATTRIB(dev)->max_sectors = limits->max_sectors;
+       dev->se_sub_dev->se_dev_attrib.hw_max_sectors = limits->max_hw_sectors;
+       dev->se_sub_dev->se_dev_attrib.max_sectors = limits->max_sectors;
        /*
         * Set optimal_sectors from max_sectors, which can be lowered via
         * configfs.
         */
-       DEV_ATTRIB(dev)->optimal_sectors = limits->max_sectors;
+       dev->se_sub_dev->se_dev_attrib.optimal_sectors = limits->max_sectors;
        /*
         * queue_depth is based on subsystem plugin dependent requirements.
         */
-       DEV_ATTRIB(dev)->hw_queue_depth = dev_limits->hw_queue_depth;
-       DEV_ATTRIB(dev)->queue_depth = dev_limits->queue_depth;
+       dev->se_sub_dev->se_dev_attrib.hw_queue_depth = dev_limits->hw_queue_depth;
+       dev->se_sub_dev->se_dev_attrib.queue_depth = dev_limits->queue_depth;
 }
 
 int se_dev_set_task_timeout(struct se_device *dev, u32 task_timeout)
 {
        if (task_timeout > DA_TASK_TIMEOUT_MAX) {
-               printk(KERN_ERR "dev[%p]: Passed task_timeout: %u larger then"
+               pr_err("dev[%p]: Passed task_timeout: %u larger then"
                        " DA_TASK_TIMEOUT_MAX\n", dev, task_timeout);
-               return -1;
+               return -EINVAL;
        } else {
-               DEV_ATTRIB(dev)->task_timeout = task_timeout;
-               printk(KERN_INFO "dev[%p]: Set SE Device task_timeout: %u\n",
+               dev->se_sub_dev->se_dev_attrib.task_timeout = task_timeout;
+               pr_debug("dev[%p]: Set SE Device task_timeout: %u\n",
                        dev, task_timeout);
        }
 
@@ -912,9 +910,9 @@ int se_dev_set_max_unmap_lba_count(
        struct se_device *dev,
        u32 max_unmap_lba_count)
 {
-       DEV_ATTRIB(dev)->max_unmap_lba_count = max_unmap_lba_count;
-       printk(KERN_INFO "dev[%p]: Set max_unmap_lba_count: %u\n",
-                       dev, DEV_ATTRIB(dev)->max_unmap_lba_count);
+       dev->se_sub_dev->se_dev_attrib.max_unmap_lba_count = max_unmap_lba_count;
+       pr_debug("dev[%p]: Set max_unmap_lba_count: %u\n",
+                       dev, dev->se_sub_dev->se_dev_attrib.max_unmap_lba_count);
        return 0;
 }
 
@@ -922,9 +920,10 @@ int se_dev_set_max_unmap_block_desc_count(
        struct se_device *dev,
        u32 max_unmap_block_desc_count)
 {
-       DEV_ATTRIB(dev)->max_unmap_block_desc_count = max_unmap_block_desc_count;
-       printk(KERN_INFO "dev[%p]: Set max_unmap_block_desc_count: %u\n",
-                       dev, DEV_ATTRIB(dev)->max_unmap_block_desc_count);
+       dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count =
+               max_unmap_block_desc_count;
+       pr_debug("dev[%p]: Set max_unmap_block_desc_count: %u\n",
+                       dev, dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count);
        return 0;
 }
 
@@ -932,9 +931,9 @@ int se_dev_set_unmap_granularity(
        struct se_device *dev,
        u32 unmap_granularity)
 {
-       DEV_ATTRIB(dev)->unmap_granularity = unmap_granularity;
-       printk(KERN_INFO "dev[%p]: Set unmap_granularity: %u\n",
-                       dev, DEV_ATTRIB(dev)->unmap_granularity);
+       dev->se_sub_dev->se_dev_attrib.unmap_granularity = unmap_granularity;
+       pr_debug("dev[%p]: Set unmap_granularity: %u\n",
+                       dev, dev->se_sub_dev->se_dev_attrib.unmap_granularity);
        return 0;
 }
 
@@ -942,109 +941,109 @@ int se_dev_set_unmap_granularity_alignment(
        struct se_device *dev,
        u32 unmap_granularity_alignment)
 {
-       DEV_ATTRIB(dev)->unmap_granularity_alignment = unmap_granularity_alignment;
-       printk(KERN_INFO "dev[%p]: Set unmap_granularity_alignment: %u\n",
-                       dev, DEV_ATTRIB(dev)->unmap_granularity_alignment);
+       dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment = unmap_granularity_alignment;
+       pr_debug("dev[%p]: Set unmap_granularity_alignment: %u\n",
+                       dev, dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment);
        return 0;
 }
 
 int se_dev_set_emulate_dpo(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->dpo_emulated == NULL) {
-               printk(KERN_ERR "TRANSPORT(dev)->dpo_emulated is NULL\n");
-               return -1;
+       if (dev->transport->dpo_emulated == NULL) {
+               pr_err("dev->transport->dpo_emulated is NULL\n");
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->dpo_emulated(dev) == 0) {
-               printk(KERN_ERR "TRANSPORT(dev)->dpo_emulated not supported\n");
-               return -1;
+       if (dev->transport->dpo_emulated(dev) == 0) {
+               pr_err("dev->transport->dpo_emulated not supported\n");
+               return -EINVAL;
        }
-       DEV_ATTRIB(dev)->emulate_dpo = flag;
-       printk(KERN_INFO "dev[%p]: SE Device Page Out (DPO) Emulation"
-                       " bit: %d\n", dev, DEV_ATTRIB(dev)->emulate_dpo);
+       dev->se_sub_dev->se_dev_attrib.emulate_dpo = flag;
+       pr_debug("dev[%p]: SE Device Page Out (DPO) Emulation"
+                       " bit: %d\n", dev, dev->se_sub_dev->se_dev_attrib.emulate_dpo);
        return 0;
 }
 
 int se_dev_set_emulate_fua_write(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->fua_write_emulated == NULL) {
-               printk(KERN_ERR "TRANSPORT(dev)->fua_write_emulated is NULL\n");
-               return -1;
+       if (dev->transport->fua_write_emulated == NULL) {
+               pr_err("dev->transport->fua_write_emulated is NULL\n");
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->fua_write_emulated(dev) == 0) {
-               printk(KERN_ERR "TRANSPORT(dev)->fua_write_emulated not supported\n");
-               return -1;
+       if (dev->transport->fua_write_emulated(dev) == 0) {
+               pr_err("dev->transport->fua_write_emulated not supported\n");
+               return -EINVAL;
        }
-       DEV_ATTRIB(dev)->emulate_fua_write = flag;
-       printk(KERN_INFO "dev[%p]: SE Device Forced Unit Access WRITEs: %d\n",
-                       dev, DEV_ATTRIB(dev)->emulate_fua_write);
+       dev->se_sub_dev->se_dev_attrib.emulate_fua_write = flag;
+       pr_debug("dev[%p]: SE Device Forced Unit Access WRITEs: %d\n",
+                       dev, dev->se_sub_dev->se_dev_attrib.emulate_fua_write);
        return 0;
 }
 
 int se_dev_set_emulate_fua_read(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->fua_read_emulated == NULL) {
-               printk(KERN_ERR "TRANSPORT(dev)->fua_read_emulated is NULL\n");
-               return -1;
+       if (dev->transport->fua_read_emulated == NULL) {
+               pr_err("dev->transport->fua_read_emulated is NULL\n");
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->fua_read_emulated(dev) == 0) {
-               printk(KERN_ERR "TRANSPORT(dev)->fua_read_emulated not supported\n");
-               return -1;
+       if (dev->transport->fua_read_emulated(dev) == 0) {
+               pr_err("dev->transport->fua_read_emulated not supported\n");
+               return -EINVAL;
        }
-       DEV_ATTRIB(dev)->emulate_fua_read = flag;
-       printk(KERN_INFO "dev[%p]: SE Device Forced Unit Access READs: %d\n",
-                       dev, DEV_ATTRIB(dev)->emulate_fua_read);
+       dev->se_sub_dev->se_dev_attrib.emulate_fua_read = flag;
+       pr_debug("dev[%p]: SE Device Forced Unit Access READs: %d\n",
+                       dev, dev->se_sub_dev->se_dev_attrib.emulate_fua_read);
        return 0;
 }
 
 int se_dev_set_emulate_write_cache(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->write_cache_emulated == NULL) {
-               printk(KERN_ERR "TRANSPORT(dev)->write_cache_emulated is NULL\n");
-               return -1;
+       if (dev->transport->write_cache_emulated == NULL) {
+               pr_err("dev->transport->write_cache_emulated is NULL\n");
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->write_cache_emulated(dev) == 0) {
-               printk(KERN_ERR "TRANSPORT(dev)->write_cache_emulated not supported\n");
-               return -1;
+       if (dev->transport->write_cache_emulated(dev) == 0) {
+               pr_err("dev->transport->write_cache_emulated not supported\n");
+               return -EINVAL;
        }
-       DEV_ATTRIB(dev)->emulate_write_cache = flag;
-       printk(KERN_INFO "dev[%p]: SE Device WRITE_CACHE_EMULATION flag: %d\n",
-                       dev, DEV_ATTRIB(dev)->emulate_write_cache);
+       dev->se_sub_dev->se_dev_attrib.emulate_write_cache = flag;
+       pr_debug("dev[%p]: SE Device WRITE_CACHE_EMULATION flag: %d\n",
+                       dev, dev->se_sub_dev->se_dev_attrib.emulate_write_cache);
        return 0;
 }
 
 int se_dev_set_emulate_ua_intlck_ctrl(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1) && (flag != 2)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
 
        if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-               printk(KERN_ERR "dev[%p]: Unable to change SE Device"
+               pr_err("dev[%p]: Unable to change SE Device"
                        " UA_INTRLCK_CTRL while dev_export_obj: %d count"
                        " exists\n", dev,
                        atomic_read(&dev->dev_export_obj.obj_access_count));
-               return -1;
+               return -EINVAL;
        }
-       DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl = flag;
-       printk(KERN_INFO "dev[%p]: SE Device UA_INTRLCK_CTRL flag: %d\n",
-               dev, DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl);
+       dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl = flag;
+       pr_debug("dev[%p]: SE Device UA_INTRLCK_CTRL flag: %d\n",
+               dev, dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl);
 
        return 0;
 }
@@ -1052,19 +1051,19 @@ int se_dev_set_emulate_ua_intlck_ctrl(struct se_device *dev, int flag)
 int se_dev_set_emulate_tas(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
 
        if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-               printk(KERN_ERR "dev[%p]: Unable to change SE Device TAS while"
+               pr_err("dev[%p]: Unable to change SE Device TAS while"
                        " dev_export_obj: %d count exists\n", dev,
                        atomic_read(&dev->dev_export_obj.obj_access_count));
-               return -1;
+               return -EINVAL;
        }
-       DEV_ATTRIB(dev)->emulate_tas = flag;
-       printk(KERN_INFO "dev[%p]: SE Device TASK_ABORTED status bit: %s\n",
-               dev, (DEV_ATTRIB(dev)->emulate_tas) ? "Enabled" : "Disabled");
+       dev->se_sub_dev->se_dev_attrib.emulate_tas = flag;
+       pr_debug("dev[%p]: SE Device TASK_ABORTED status bit: %s\n",
+               dev, (dev->se_sub_dev->se_dev_attrib.emulate_tas) ? "Enabled" : "Disabled");
 
        return 0;
 }
@@ -1072,20 +1071,20 @@ int se_dev_set_emulate_tas(struct se_device *dev, int flag)
 int se_dev_set_emulate_tpu(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
        /*
         * We expect this value to be non-zero when generic Block Layer
         * Discard supported is detected iblock_create_virtdevice().
         */
-       if (!(DEV_ATTRIB(dev)->max_unmap_block_desc_count)) {
-               printk(KERN_ERR "Generic Block Discard not supported\n");
+       if (!dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count) {
+               pr_err("Generic Block Discard not supported\n");
                return -ENOSYS;
        }
 
-       DEV_ATTRIB(dev)->emulate_tpu = flag;
-       printk(KERN_INFO "dev[%p]: SE Device Thin Provisioning UNMAP bit: %d\n",
+       dev->se_sub_dev->se_dev_attrib.emulate_tpu = flag;
+       pr_debug("dev[%p]: SE Device Thin Provisioning UNMAP bit: %d\n",
                                dev, flag);
        return 0;
 }
@@ -1093,33 +1092,57 @@ int se_dev_set_emulate_tpu(struct se_device *dev, int flag)
 int se_dev_set_emulate_tpws(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
        }
        /*
         * We expect this value to be non-zero when generic Block Layer
         * Discard supported is detected iblock_create_virtdevice().
         */
-       if (!(DEV_ATTRIB(dev)->max_unmap_block_desc_count)) {
-               printk(KERN_ERR "Generic Block Discard not supported\n");
+       if (!dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count) {
+               pr_err("Generic Block Discard not supported\n");
                return -ENOSYS;
        }
 
-       DEV_ATTRIB(dev)->emulate_tpws = flag;
-       printk(KERN_INFO "dev[%p]: SE Device Thin Provisioning WRITE_SAME: %d\n",
+       dev->se_sub_dev->se_dev_attrib.emulate_tpws = flag;
+       pr_debug("dev[%p]: SE Device Thin Provisioning WRITE_SAME: %d\n",
                                dev, flag);
        return 0;
 }
 
 int se_dev_set_enforce_pr_isids(struct se_device *dev, int flag)
+{
+       if ((flag != 0) && (flag != 1)) {
+               pr_err("Illegal value %d\n", flag);
+               return -EINVAL;
+       }
+       dev->se_sub_dev->se_dev_attrib.enforce_pr_isids = flag;
+       pr_debug("dev[%p]: SE Device enforce_pr_isids bit: %s\n", dev,
+               (dev->se_sub_dev->se_dev_attrib.enforce_pr_isids) ? "Enabled" : "Disabled");
+       return 0;
+}
+
+int se_dev_set_is_nonrot(struct se_device *dev, int flag)
 {
        if ((flag != 0) && (flag != 1)) {
                printk(KERN_ERR "Illegal value %d\n", flag);
-               return -1;
+               return -EINVAL;
+       }
+       dev->se_sub_dev->se_dev_attrib.is_nonrot = flag;
+       pr_debug("dev[%p]: SE Device is_nonrot bit: %d\n",
+              dev, flag);
+       return 0;
+}
+
+int se_dev_set_emulate_rest_reord(struct se_device *dev, int flag)
+{
+       if (flag != 0) {
+               printk(KERN_ERR "dev[%p]: SE Device emulatation of restricted"
+                       " reordering not implemented\n", dev);
+               return -ENOSYS;
        }
-       DEV_ATTRIB(dev)->enforce_pr_isids = flag;
-       printk(KERN_INFO "dev[%p]: SE Device enforce_pr_isids bit: %s\n", dev,
-               (DEV_ATTRIB(dev)->enforce_pr_isids) ? "Enabled" : "Disabled");
+       dev->se_sub_dev->se_dev_attrib.emulate_rest_reord = flag;
+       pr_debug("dev[%p]: SE Device emulate_rest_reord: %d\n", dev, flag);
        return 0;
 }
 
@@ -1131,44 +1154,44 @@ int se_dev_set_queue_depth(struct se_device *dev, u32 queue_depth)
        u32 orig_queue_depth = dev->queue_depth;
 
        if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-               printk(KERN_ERR "dev[%p]: Unable to change SE Device TCQ while"
+               pr_err("dev[%p]: Unable to change SE Device TCQ while"
                        " dev_export_obj: %d count exists\n", dev,
                        atomic_read(&dev->dev_export_obj.obj_access_count));
-               return -1;
+               return -EINVAL;
        }
-       if (!(queue_depth)) {
-               printk(KERN_ERR "dev[%p]: Illegal ZERO value for queue"
+       if (!queue_depth) {
+               pr_err("dev[%p]: Illegal ZERO value for queue"
                        "_depth\n", dev);
-               return -1;
+               return -EINVAL;
        }
 
-       if (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
-               if (queue_depth > DEV_ATTRIB(dev)->hw_queue_depth) {
-                       printk(KERN_ERR "dev[%p]: Passed queue_depth: %u"
+       if (dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
+               if (queue_depth > dev->se_sub_dev->se_dev_attrib.hw_queue_depth) {
+                       pr_err("dev[%p]: Passed queue_depth: %u"
                                " exceeds TCM/SE_Device TCQ: %u\n",
                                dev, queue_depth,
-                               DEV_ATTRIB(dev)->hw_queue_depth);
-                       return -1;
+                               dev->se_sub_dev->se_dev_attrib.hw_queue_depth);
+                       return -EINVAL;
                }
        } else {
-               if (queue_depth > DEV_ATTRIB(dev)->queue_depth) {
-                       if (queue_depth > DEV_ATTRIB(dev)->hw_queue_depth) {
-                               printk(KERN_ERR "dev[%p]: Passed queue_depth:"
+               if (queue_depth > dev->se_sub_dev->se_dev_attrib.queue_depth) {
+                       if (queue_depth > dev->se_sub_dev->se_dev_attrib.hw_queue_depth) {
+                               pr_err("dev[%p]: Passed queue_depth:"
                                        " %u exceeds TCM/SE_Device MAX"
                                        " TCQ: %u\n", dev, queue_depth,
-                                       DEV_ATTRIB(dev)->hw_queue_depth);
-                               return -1;
+                                       dev->se_sub_dev->se_dev_attrib.hw_queue_depth);
+                               return -EINVAL;
                        }
                }
        }
 
-       DEV_ATTRIB(dev)->queue_depth = dev->queue_depth = queue_depth;
+       dev->se_sub_dev->se_dev_attrib.queue_depth = dev->queue_depth = queue_depth;
        if (queue_depth > orig_queue_depth)
                atomic_add(queue_depth - orig_queue_depth, &dev->depth_left);
        else if (queue_depth < orig_queue_depth)
                atomic_sub(orig_queue_depth - queue_depth, &dev->depth_left);
 
-       printk(KERN_INFO "dev[%p]: SE Device TCQ Depth changed to: %u\n",
+       pr_debug("dev[%p]: SE Device TCQ Depth changed to: %u\n",
                        dev, queue_depth);
        return 0;
 }
@@ -1178,50 +1201,50 @@ int se_dev_set_max_sectors(struct se_device *dev, u32 max_sectors)
        int force = 0; /* Force setting for VDEVS */
 
        if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-               printk(KERN_ERR "dev[%p]: Unable to change SE Device"
+               pr_err("dev[%p]: Unable to change SE Device"
                        " max_sectors while dev_export_obj: %d count exists\n",
                        dev, atomic_read(&dev->dev_export_obj.obj_access_count));
-               return -1;
+               return -EINVAL;
        }
-       if (!(max_sectors)) {
-               printk(KERN_ERR "dev[%p]: Illegal ZERO value for"
+       if (!max_sectors) {
+               pr_err("dev[%p]: Illegal ZERO value for"
                        " max_sectors\n", dev);
-               return -1;
+               return -EINVAL;
        }
        if (max_sectors < DA_STATUS_MAX_SECTORS_MIN) {
-               printk(KERN_ERR "dev[%p]: Passed max_sectors: %u less than"
+               pr_err("dev[%p]: Passed max_sectors: %u less than"
                        " DA_STATUS_MAX_SECTORS_MIN: %u\n", dev, max_sectors,
                                DA_STATUS_MAX_SECTORS_MIN);
-               return -1;
+               return -EINVAL;
        }
-       if (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
-               if (max_sectors > DEV_ATTRIB(dev)->hw_max_sectors) {
-                       printk(KERN_ERR "dev[%p]: Passed max_sectors: %u"
+       if (dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
+               if (max_sectors > dev->se_sub_dev->se_dev_attrib.hw_max_sectors) {
+                       pr_err("dev[%p]: Passed max_sectors: %u"
                                " greater than TCM/SE_Device max_sectors:"
                                " %u\n", dev, max_sectors,
-                               DEV_ATTRIB(dev)->hw_max_sectors);
-                        return -1;
+                               dev->se_sub_dev->se_dev_attrib.hw_max_sectors);
+                        return -EINVAL;
                }
        } else {
-               if (!(force) && (max_sectors >
-                                DEV_ATTRIB(dev)->hw_max_sectors)) {
-                       printk(KERN_ERR "dev[%p]: Passed max_sectors: %u"
+               if (!force && (max_sectors >
+                                dev->se_sub_dev->se_dev_attrib.hw_max_sectors)) {
+                       pr_err("dev[%p]: Passed max_sectors: %u"
                                " greater than TCM/SE_Device max_sectors"
                                ": %u, use force=1 to override.\n", dev,
-                               max_sectors, DEV_ATTRIB(dev)->hw_max_sectors);
-                       return -1;
+                               max_sectors, dev->se_sub_dev->se_dev_attrib.hw_max_sectors);
+                       return -EINVAL;
                }
                if (max_sectors > DA_STATUS_MAX_SECTORS_MAX) {
-                       printk(KERN_ERR "dev[%p]: Passed max_sectors: %u"
+                       pr_err("dev[%p]: Passed max_sectors: %u"
                                " greater than DA_STATUS_MAX_SECTORS_MAX:"
                                " %u\n", dev, max_sectors,
                                DA_STATUS_MAX_SECTORS_MAX);
-                       return -1;
+                       return -EINVAL;
                }
        }
 
-       DEV_ATTRIB(dev)->max_sectors = max_sectors;
-       printk("dev[%p]: SE Device max_sectors changed to %u\n",
+       dev->se_sub_dev->se_dev_attrib.max_sectors = max_sectors;
+       pr_debug("dev[%p]: SE Device max_sectors changed to %u\n",
                        dev, max_sectors);
        return 0;
 }
@@ -1229,25 +1252,25 @@ int se_dev_set_max_sectors(struct se_device *dev, u32 max_sectors)
 int se_dev_set_optimal_sectors(struct se_device *dev, u32 optimal_sectors)
 {
        if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-               printk(KERN_ERR "dev[%p]: Unable to change SE Device"
+               pr_err("dev[%p]: Unable to change SE Device"
                        " optimal_sectors while dev_export_obj: %d count exists\n",
                        dev, atomic_read(&dev->dev_export_obj.obj_access_count));
                return -EINVAL;
        }
-       if (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
-               printk(KERN_ERR "dev[%p]: Passed optimal_sectors cannot be"
+       if (dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
+               pr_err("dev[%p]: Passed optimal_sectors cannot be"
                                " changed for TCM/pSCSI\n", dev);
                return -EINVAL;
        }
-       if (optimal_sectors > DEV_ATTRIB(dev)->max_sectors) {
-               printk(KERN_ERR "dev[%p]: Passed optimal_sectors %u cannot be"
+       if (optimal_sectors > dev->se_sub_dev->se_dev_attrib.max_sectors) {
+               pr_err("dev[%p]: Passed optimal_sectors %u cannot be"
                        " greater than max_sectors: %u\n", dev,
-                       optimal_sectors, DEV_ATTRIB(dev)->max_sectors);
+                       optimal_sectors, dev->se_sub_dev->se_dev_attrib.max_sectors);
                return -EINVAL;
        }
 
-       DEV_ATTRIB(dev)->optimal_sectors = optimal_sectors;
-       printk(KERN_INFO "dev[%p]: SE Device optimal_sectors changed to %u\n",
+       dev->se_sub_dev->se_dev_attrib.optimal_sectors = optimal_sectors;
+       pr_debug("dev[%p]: SE Device optimal_sectors changed to %u\n",
                        dev, optimal_sectors);
        return 0;
 }
@@ -1255,31 +1278,31 @@ int se_dev_set_optimal_sectors(struct se_device *dev, u32 optimal_sectors)
 int se_dev_set_block_size(struct se_device *dev, u32 block_size)
 {
        if (atomic_read(&dev->dev_export_obj.obj_access_count)) {
-               printk(KERN_ERR "dev[%p]: Unable to change SE Device block_size"
+               pr_err("dev[%p]: Unable to change SE Device block_size"
                        " while dev_export_obj: %d count exists\n", dev,
                        atomic_read(&dev->dev_export_obj.obj_access_count));
-               return -1;
+               return -EINVAL;
        }
 
        if ((block_size != 512) &&
            (block_size != 1024) &&
            (block_size != 2048) &&
            (block_size != 4096)) {
-               printk(KERN_ERR "dev[%p]: Illegal value for block_device: %u"
+               pr_err("dev[%p]: Illegal value for block_device: %u"
                        " for SE device, must be 512, 1024, 2048 or 4096\n",
                        dev, block_size);
-               return -1;
+               return -EINVAL;
        }
 
-       if (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
-               printk(KERN_ERR "dev[%p]: Not allowed to change block_size for"
+       if (dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
+               pr_err("dev[%p]: Not allowed to change block_size for"
                        " Physical Device, use for Linux/SCSI to change"
                        " block_size for underlying hardware\n", dev);
-               return -1;
+               return -EINVAL;
        }
 
-       DEV_ATTRIB(dev)->block_size = block_size;
-       printk(KERN_INFO "dev[%p]: SE Device block_size changed to %u\n",
+       dev->se_sub_dev->se_dev_attrib.block_size = block_size;
+       pr_debug("dev[%p]: SE Device block_size changed to %u\n",
                        dev, block_size);
        return 0;
 }
@@ -1294,13 +1317,13 @@ struct se_lun *core_dev_add_lun(
        u32 lun_access = 0;
 
        if (atomic_read(&dev->dev_access_obj.obj_access_count) != 0) {
-               printk(KERN_ERR "Unable to export struct se_device while dev_access_obj: %d\n",
+               pr_err("Unable to export struct se_device while dev_access_obj: %d\n",
                        atomic_read(&dev->dev_access_obj.obj_access_count));
                return NULL;
        }
 
        lun_p = core_tpg_pre_addlun(tpg, lun);
-       if ((IS_ERR(lun_p)) || !(lun_p))
+       if ((IS_ERR(lun_p)) || !lun_p)
                return NULL;
 
        if (dev->dev_flags & DF_READ_ONLY)
@@ -1311,15 +1334,15 @@ struct se_lun *core_dev_add_lun(
        if (core_tpg_post_addlun(tpg, lun_p, lun_access, dev) < 0)
                return NULL;
 
-       printk(KERN_INFO "%s_TPG[%u]_LUN[%u] - Activated %s Logical Unit from"
-               " CORE HBA: %u\n", TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg), lun_p->unpacked_lun,
-               TPG_TFO(tpg)->get_fabric_name(), hba->hba_id);
+       pr_debug("%s_TPG[%u]_LUN[%u] - Activated %s Logical Unit from"
+               " CORE HBA: %u\n", tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg), lun_p->unpacked_lun,
+               tpg->se_tpg_tfo->get_fabric_name(), hba->hba_id);
        /*
         * Update LUN maps for dynamically added initiators when
         * generate_node_acl is enabled.
         */
-       if (TPG_TFO(tpg)->tpg_check_demo_mode(tpg)) {
+       if (tpg->se_tpg_tfo->tpg_check_demo_mode(tpg)) {
                struct se_node_acl *acl;
                spin_lock_bh(&tpg->acl_node_lock);
                list_for_each_entry(acl, &tpg->acl_node_list, acl_list) {
@@ -1347,15 +1370,15 @@ int core_dev_del_lun(
        int ret = 0;
 
        lun = core_tpg_pre_dellun(tpg, unpacked_lun, &ret);
-       if (!(lun))
+       if (!lun)
                return ret;
 
        core_tpg_post_dellun(tpg, lun);
 
-       printk(KERN_INFO "%s_TPG[%u]_LUN[%u] - Deactivated %s Logical Unit from"
-               " device object\n", TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg), unpacked_lun,
-               TPG_TFO(tpg)->get_fabric_name());
+       pr_debug("%s_TPG[%u]_LUN[%u] - Deactivated %s Logical Unit from"
+               " device object\n", tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg), unpacked_lun,
+               tpg->se_tpg_tfo->get_fabric_name());
 
        return 0;
 }
@@ -1366,21 +1389,21 @@ struct se_lun *core_get_lun_from_tpg(struct se_portal_group *tpg, u32 unpacked_l
 
        spin_lock(&tpg->tpg_lun_lock);
        if (unpacked_lun > (TRANSPORT_MAX_LUNS_PER_TPG-1)) {
-               printk(KERN_ERR "%s LUN: %u exceeds TRANSPORT_MAX_LUNS"
+               pr_err("%s LUN: %u exceeds TRANSPORT_MAX_LUNS"
                        "_PER_TPG-1: %u for Target Portal Group: %hu\n",
-                       TPG_TFO(tpg)->get_fabric_name(), unpacked_lun,
+                       tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
                        TRANSPORT_MAX_LUNS_PER_TPG-1,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock(&tpg->tpg_lun_lock);
                return NULL;
        }
        lun = &tpg->tpg_lun_list[unpacked_lun];
 
        if (lun->lun_status != TRANSPORT_LUN_STATUS_FREE) {
-               printk(KERN_ERR "%s Logical Unit Number: %u is not free on"
+               pr_err("%s Logical Unit Number: %u is not free on"
                        " Target Portal Group: %hu, ignoring request.\n",
-                       TPG_TFO(tpg)->get_fabric_name(), unpacked_lun,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock(&tpg->tpg_lun_lock);
                return NULL;
        }
@@ -1399,21 +1422,21 @@ static struct se_lun *core_dev_get_lun(struct se_portal_group *tpg, u32 unpacked
 
        spin_lock(&tpg->tpg_lun_lock);
        if (unpacked_lun > (TRANSPORT_MAX_LUNS_PER_TPG-1)) {
-               printk(KERN_ERR "%s LUN: %u exceeds TRANSPORT_MAX_LUNS_PER"
+               pr_err("%s LUN: %u exceeds TRANSPORT_MAX_LUNS_PER"
                        "_TPG-1: %u for Target Portal Group: %hu\n",
-                       TPG_TFO(tpg)->get_fabric_name(), unpacked_lun,
+                       tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
                        TRANSPORT_MAX_LUNS_PER_TPG-1,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock(&tpg->tpg_lun_lock);
                return NULL;
        }
        lun = &tpg->tpg_lun_list[unpacked_lun];
 
        if (lun->lun_status != TRANSPORT_LUN_STATUS_ACTIVE) {
-               printk(KERN_ERR "%s Logical Unit Number: %u is not active on"
+               pr_err("%s Logical Unit Number: %u is not active on"
                        " Target Portal Group: %hu, ignoring request.\n",
-                       TPG_TFO(tpg)->get_fabric_name(), unpacked_lun,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock(&tpg->tpg_lun_lock);
                return NULL;
        }
@@ -1432,19 +1455,19 @@ struct se_lun_acl *core_dev_init_initiator_node_lun_acl(
        struct se_node_acl *nacl;
 
        if (strlen(initiatorname) >= TRANSPORT_IQN_LEN) {
-               printk(KERN_ERR "%s InitiatorName exceeds maximum size.\n",
-                       TPG_TFO(tpg)->get_fabric_name());
+               pr_err("%s InitiatorName exceeds maximum size.\n",
+                       tpg->se_tpg_tfo->get_fabric_name());
                *ret = -EOVERFLOW;
                return NULL;
        }
        nacl = core_tpg_get_initiator_node_acl(tpg, initiatorname);
-       if (!(nacl)) {
+       if (!nacl) {
                *ret = -EINVAL;
                return NULL;
        }
        lacl = kzalloc(sizeof(struct se_lun_acl), GFP_KERNEL);
-       if (!(lacl)) {
-               printk(KERN_ERR "Unable to allocate memory for struct se_lun_acl.\n");
+       if (!lacl) {
+               pr_err("Unable to allocate memory for struct se_lun_acl.\n");
                *ret = -ENOMEM;
                return NULL;
        }
@@ -1467,16 +1490,16 @@ int core_dev_add_initiator_node_lun_acl(
        struct se_node_acl *nacl;
 
        lun = core_dev_get_lun(tpg, unpacked_lun);
-       if (!(lun)) {
-               printk(KERN_ERR "%s Logical Unit Number: %u is not active on"
+       if (!lun) {
+               pr_err("%s Logical Unit Number: %u is not active on"
                        " Target Portal Group: %hu, ignoring request.\n",
-                       TPG_TFO(tpg)->get_fabric_name(), unpacked_lun,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                return -EINVAL;
        }
 
        nacl = lacl->se_lun_nacl;
-       if (!(nacl))
+       if (!nacl)
                return -EINVAL;
 
        if ((lun->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) &&
@@ -1495,9 +1518,9 @@ int core_dev_add_initiator_node_lun_acl(
        smp_mb__after_atomic_inc();
        spin_unlock(&lun->lun_acl_lock);
 
-       printk(KERN_INFO "%s_TPG[%hu]_LUN[%u->%u] - Added %s ACL for "
-               " InitiatorNode: %s\n", TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg), unpacked_lun, lacl->mapped_lun,
+       pr_debug("%s_TPG[%hu]_LUN[%u->%u] - Added %s ACL for "
+               " InitiatorNode: %s\n", tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg), unpacked_lun, lacl->mapped_lun,
                (lun_access & TRANSPORT_LUNFLAGS_READ_WRITE) ? "RW" : "RO",
                lacl->initiatorname);
        /*
@@ -1520,7 +1543,7 @@ int core_dev_del_initiator_node_lun_acl(
        struct se_node_acl *nacl;
 
        nacl = lacl->se_lun_nacl;
-       if (!(nacl))
+       if (!nacl)
                return -EINVAL;
 
        spin_lock(&lun->lun_acl_lock);
@@ -1534,10 +1557,10 @@ int core_dev_del_initiator_node_lun_acl(
 
        lacl->se_lun = NULL;
 
-       printk(KERN_INFO "%s_TPG[%hu]_LUN[%u] - Removed ACL for"
+       pr_debug("%s_TPG[%hu]_LUN[%u] - Removed ACL for"
                " InitiatorNode: %s Mapped LUN: %u\n",
-               TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg), lun->unpacked_lun,
+               tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg), lun->unpacked_lun,
                lacl->initiatorname, lacl->mapped_lun);
 
        return 0;
@@ -1547,10 +1570,10 @@ void core_dev_free_initiator_node_lun_acl(
        struct se_portal_group *tpg,
        struct se_lun_acl *lacl)
 {
-       printk("%s_TPG[%hu] - Freeing ACL for %s InitiatorNode: %s"
-               " Mapped LUN: %u\n", TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg),
-               TPG_TFO(tpg)->get_fabric_name(),
+       pr_debug("%s_TPG[%hu] - Freeing ACL for %s InitiatorNode: %s"
+               " Mapped LUN: %u\n", tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg),
+               tpg->se_tpg_tfo->get_fabric_name(),
                lacl->initiatorname, lacl->mapped_lun);
 
        kfree(lacl);
@@ -1565,64 +1588,64 @@ int core_dev_setup_virtual_lun0(void)
        char buf[16];
        int ret;
 
-       hba = core_alloc_hba("rd_dr", 0, HBA_FLAGS_INTERNAL_USE);
+       hba = core_alloc_hba("rd_mcp", 0, HBA_FLAGS_INTERNAL_USE);
        if (IS_ERR(hba))
                return PTR_ERR(hba);
 
-       se_global->g_lun0_hba = hba;
+       lun0_hba = hba;
        t = hba->transport;
 
        se_dev = kzalloc(sizeof(struct se_subsystem_dev), GFP_KERNEL);
-       if (!(se_dev)) {
-               printk(KERN_ERR "Unable to allocate memory for"
+       if (!se_dev) {
+               pr_err("Unable to allocate memory for"
                                " struct se_subsystem_dev\n");
                ret = -ENOMEM;
                goto out;
        }
-       INIT_LIST_HEAD(&se_dev->g_se_dev_list);
+       INIT_LIST_HEAD(&se_dev->se_dev_node);
        INIT_LIST_HEAD(&se_dev->t10_wwn.t10_vpd_list);
        spin_lock_init(&se_dev->t10_wwn.t10_vpd_lock);
-       INIT_LIST_HEAD(&se_dev->t10_reservation.registration_list);
-       INIT_LIST_HEAD(&se_dev->t10_reservation.aptpl_reg_list);
-       spin_lock_init(&se_dev->t10_reservation.registration_lock);
-       spin_lock_init(&se_dev->t10_reservation.aptpl_reg_lock);
+       INIT_LIST_HEAD(&se_dev->t10_pr.registration_list);
+       INIT_LIST_HEAD(&se_dev->t10_pr.aptpl_reg_list);
+       spin_lock_init(&se_dev->t10_pr.registration_lock);
+       spin_lock_init(&se_dev->t10_pr.aptpl_reg_lock);
        INIT_LIST_HEAD(&se_dev->t10_alua.tg_pt_gps_list);
        spin_lock_init(&se_dev->t10_alua.tg_pt_gps_lock);
        spin_lock_init(&se_dev->se_dev_lock);
-       se_dev->t10_reservation.pr_aptpl_buf_len = PR_APTPL_BUF_LEN;
+       se_dev->t10_pr.pr_aptpl_buf_len = PR_APTPL_BUF_LEN;
        se_dev->t10_wwn.t10_sub_dev = se_dev;
        se_dev->t10_alua.t10_sub_dev = se_dev;
        se_dev->se_dev_attrib.da_sub_dev = se_dev;
        se_dev->se_dev_hba = hba;
 
        se_dev->se_dev_su_ptr = t->allocate_virtdevice(hba, "virt_lun0");
-       if (!(se_dev->se_dev_su_ptr)) {
-               printk(KERN_ERR "Unable to locate subsystem dependent pointer"
+       if (!se_dev->se_dev_su_ptr) {
+               pr_err("Unable to locate subsystem dependent pointer"
                        " from allocate_virtdevice()\n");
                ret = -ENOMEM;
                goto out;
        }
-       se_global->g_lun0_su_dev = se_dev;
+       lun0_su_dev = se_dev;
 
        memset(buf, 0, 16);
        sprintf(buf, "rd_pages=8");
        t->set_configfs_dev_params(hba, se_dev, buf, sizeof(buf));
 
        dev = t->create_virtdevice(hba, se_dev, se_dev->se_dev_su_ptr);
-       if (!(dev) || IS_ERR(dev)) {
-               ret = -ENOMEM;
+       if (IS_ERR(dev)) {
+               ret = PTR_ERR(dev);
                goto out;
        }
        se_dev->se_dev_ptr = dev;
-       se_global->g_lun0_dev = dev;
+       g_lun0_dev = dev;
 
        return 0;
 out:
-       se_global->g_lun0_su_dev = NULL;
+       lun0_su_dev = NULL;
        kfree(se_dev);
-       if (se_global->g_lun0_hba) {
-               core_delete_hba(se_global->g_lun0_hba);
-               se_global->g_lun0_hba = NULL;
+       if (lun0_hba) {
+               core_delete_hba(lun0_hba);
+               lun0_hba = NULL;
        }
        return ret;
 }
@@ -1630,14 +1653,14 @@ out:
 
 void core_dev_release_virtual_lun0(void)
 {
-       struct se_hba *hba = se_global->g_lun0_hba;
-       struct se_subsystem_dev *su_dev = se_global->g_lun0_su_dev;
+       struct se_hba *hba = lun0_hba;
+       struct se_subsystem_dev *su_dev = lun0_su_dev;
 
-       if (!(hba))
+       if (!hba)
                return;
 
-       if (se_global->g_lun0_dev)
-               se_free_virtual_device(se_global->g_lun0_dev, hba);
+       if (g_lun0_dev)
+               se_free_virtual_device(g_lun0_dev, hba);
 
        kfree(su_dev);
        core_delete_hba(hba);
index 07ab5a3bb8e8d56a97a76c3c21af3fa8a8e96792..f1654694f4ea4e3b33c1ef3e5793b6e372b70156 100644 (file)
@@ -60,7 +60,7 @@ static void target_fabric_setup_##_name##_cit(struct target_fabric_configfs *tf)
        cit->ct_group_ops = _group_ops;                                 \
        cit->ct_attrs = _attrs;                                         \
        cit->ct_owner = tf->tf_module;                                  \
-       printk("Setup generic %s\n", __stringify(_name));               \
+       pr_debug("Setup generic %s\n", __stringify(_name));             \
 }
 
 /* Start of tfc_tpg_mappedlun_cit */
@@ -80,8 +80,8 @@ static int target_fabric_mappedlun_link(
        /*
         * Ensure that the source port exists
         */
-       if (!(lun->lun_sep) || !(lun->lun_sep->sep_tpg)) {
-               printk(KERN_ERR "Source se_lun->lun_sep or lun->lun_sep->sep"
+       if (!lun->lun_sep || !lun->lun_sep->sep_tpg) {
+               pr_err("Source se_lun->lun_sep or lun->lun_sep->sep"
                                "_tpg does not exist\n");
                return -EINVAL;
        }
@@ -96,12 +96,12 @@ static int target_fabric_mappedlun_link(
         * Make sure the SymLink is going to the same $FABRIC/$WWN/tpgt_$TPGT
         */
        if (strcmp(config_item_name(wwn_ci), config_item_name(wwn_ci_s))) {
-               printk(KERN_ERR "Illegal Initiator ACL SymLink outside of %s\n",
+               pr_err("Illegal Initiator ACL SymLink outside of %s\n",
                        config_item_name(wwn_ci));
                return -EINVAL;
        }
        if (strcmp(config_item_name(tpg_ci), config_item_name(tpg_ci_s))) {
-               printk(KERN_ERR "Illegal Initiator ACL Symlink outside of %s"
+               pr_err("Illegal Initiator ACL Symlink outside of %s"
                        " TPGT: %s\n", config_item_name(wwn_ci),
                        config_item_name(tpg_ci));
                return -EINVAL;
@@ -118,7 +118,7 @@ static int target_fabric_mappedlun_link(
                lun_access = deve->lun_flags;
        else
                lun_access =
-                       (TPG_TFO(se_tpg)->tpg_check_prod_mode_write_protect(
+                       (se_tpg->se_tpg_tfo->tpg_check_prod_mode_write_protect(
                                se_tpg)) ? TRANSPORT_LUNFLAGS_READ_ONLY :
                                           TRANSPORT_LUNFLAGS_READ_WRITE;
        spin_unlock_irq(&lacl->se_lun_nacl->device_list_lock);
@@ -147,7 +147,7 @@ static int target_fabric_mappedlun_unlink(
        /*
         * Determine if the underlying MappedLUN has already been released..
         */
-       if (!(deve->se_lun))
+       if (!deve->se_lun)
                return 0;
 
        lun = container_of(to_config_group(lun_ci), struct se_lun, lun_group);
@@ -202,9 +202,9 @@ static ssize_t target_fabric_mappedlun_store_write_protect(
                        TRANSPORT_LUNFLAGS_READ_WRITE,
                        lacl->se_lun_nacl);
 
-       printk(KERN_INFO "%s_ConfigFS: Changed Initiator ACL: %s"
+       pr_debug("%s_ConfigFS: Changed Initiator ACL: %s"
                " Mapped LUN: %u Write Protect bit to %s\n",
-               TPG_TFO(se_tpg)->get_fabric_name(),
+               se_tpg->se_tpg_tfo->get_fabric_name(),
                lacl->initiatorname, lacl->mapped_lun, (op) ? "ON" : "OFF");
 
        return count;
@@ -327,14 +327,14 @@ static struct config_group *target_fabric_make_mappedlun(
        int ret = 0;
 
        acl_ci = &group->cg_item;
-       if (!(acl_ci)) {
-               printk(KERN_ERR "Unable to locatel acl_ci\n");
+       if (!acl_ci) {
+               pr_err("Unable to locatel acl_ci\n");
                return NULL;
        }
 
        buf = kzalloc(strlen(name) + 1, GFP_KERNEL);
-       if (!(buf)) {
-               printk(KERN_ERR "Unable to allocate memory for name buf\n");
+       if (!buf) {
+               pr_err("Unable to allocate memory for name buf\n");
                return ERR_PTR(-ENOMEM);
        }
        snprintf(buf, strlen(name) + 1, "%s", name);
@@ -342,7 +342,7 @@ static struct config_group *target_fabric_make_mappedlun(
         * Make sure user is creating iscsi/$IQN/$TPGT/acls/$INITIATOR/lun_$ID.
         */
        if (strstr(buf, "lun_") != buf) {
-               printk(KERN_ERR "Unable to locate \"lun_\" from buf: %s"
+               pr_err("Unable to locate \"lun_\" from buf: %s"
                        " name: %s\n", buf, name);
                ret = -EINVAL;
                goto out;
@@ -358,7 +358,7 @@ static struct config_group *target_fabric_make_mappedlun(
 
        lacl = core_dev_init_initiator_node_lun_acl(se_tpg, mapped_lun,
                        config_item_name(acl_ci), &ret);
-       if (!(lacl)) {
+       if (!lacl) {
                ret = -EINVAL;
                goto out;
        }
@@ -367,7 +367,7 @@ static struct config_group *target_fabric_make_mappedlun(
        lacl_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
                                GFP_KERNEL);
        if (!lacl_cg->default_groups) {
-               printk(KERN_ERR "Unable to allocate lacl_cg->default_groups\n");
+               pr_err("Unable to allocate lacl_cg->default_groups\n");
                ret = -ENOMEM;
                goto out;
        }
@@ -379,11 +379,11 @@ static struct config_group *target_fabric_make_mappedlun(
        lacl_cg->default_groups[0] = &lacl->ml_stat_grps.stat_group;
        lacl_cg->default_groups[1] = NULL;
 
-       ml_stat_grp = &ML_STAT_GRPS(lacl)->stat_group;
+       ml_stat_grp = &lacl->ml_stat_grps.stat_group;
        ml_stat_grp->default_groups = kzalloc(sizeof(struct config_group) * 3,
                                GFP_KERNEL);
        if (!ml_stat_grp->default_groups) {
-               printk(KERN_ERR "Unable to allocate ml_stat_grp->default_groups\n");
+               pr_err("Unable to allocate ml_stat_grp->default_groups\n");
                ret = -ENOMEM;
                goto out;
        }
@@ -408,7 +408,7 @@ static void target_fabric_drop_mappedlun(
        struct config_group *lacl_cg = NULL, *ml_stat_grp = NULL;
        int i;
 
-       ml_stat_grp = &ML_STAT_GRPS(lacl)->stat_group;
+       ml_stat_grp = &lacl->ml_stat_grps.stat_group;
        for (i = 0; ml_stat_grp->default_groups[i]; i++) {
                df_item = &ml_stat_grp->default_groups[i]->cg_item;
                ml_stat_grp->default_groups[i] = NULL;
@@ -474,8 +474,8 @@ static struct config_group *target_fabric_make_nodeacl(
        struct se_node_acl *se_nacl;
        struct config_group *nacl_cg;
 
-       if (!(tf->tf_ops.fabric_make_nodeacl)) {
-               printk(KERN_ERR "tf->tf_ops.fabric_make_nodeacl is NULL\n");
+       if (!tf->tf_ops.fabric_make_nodeacl) {
+               pr_err("tf->tf_ops.fabric_make_nodeacl is NULL\n");
                return ERR_PTR(-ENOSYS);
        }
 
@@ -572,13 +572,13 @@ static struct config_group *target_fabric_make_np(
        struct target_fabric_configfs *tf = se_tpg->se_tpg_wwn->wwn_tf;
        struct se_tpg_np *se_tpg_np;
 
-       if (!(tf->tf_ops.fabric_make_np)) {
-               printk(KERN_ERR "tf->tf_ops.fabric_make_np is NULL\n");
+       if (!tf->tf_ops.fabric_make_np) {
+               pr_err("tf->tf_ops.fabric_make_np is NULL\n");
                return ERR_PTR(-ENOSYS);
        }
 
        se_tpg_np = tf->tf_ops.fabric_make_np(se_tpg, group, name);
-       if (!(se_tpg_np) || IS_ERR(se_tpg_np))
+       if (!se_tpg_np || IS_ERR(se_tpg_np))
                return ERR_PTR(-EINVAL);
 
        se_tpg_np->tpg_np_parent = se_tpg;
@@ -627,10 +627,7 @@ static ssize_t target_fabric_port_show_attr_alua_tg_pt_gp(
        struct se_lun *lun,
        char *page)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_show_tg_pt_gp_info(lun->lun_sep, page);
@@ -641,10 +638,7 @@ static ssize_t target_fabric_port_store_attr_alua_tg_pt_gp(
        const char *page,
        size_t count)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_store_tg_pt_gp_info(lun->lun_sep, page, count);
@@ -659,10 +653,7 @@ static ssize_t target_fabric_port_show_attr_alua_tg_pt_offline(
        struct se_lun *lun,
        char *page)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_show_offline_bit(lun, page);
@@ -673,10 +664,7 @@ static ssize_t target_fabric_port_store_attr_alua_tg_pt_offline(
        const char *page,
        size_t count)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_store_offline_bit(lun, page, count);
@@ -691,10 +679,7 @@ static ssize_t target_fabric_port_show_attr_alua_tg_pt_status(
        struct se_lun *lun,
        char *page)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_show_secondary_status(lun, page);
@@ -705,10 +690,7 @@ static ssize_t target_fabric_port_store_attr_alua_tg_pt_status(
        const char *page,
        size_t count)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_store_secondary_status(lun, page, count);
@@ -723,10 +705,7 @@ static ssize_t target_fabric_port_show_attr_alua_tg_pt_write_md(
        struct se_lun *lun,
        char *page)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_show_secondary_write_metadata(lun, page);
@@ -737,10 +716,7 @@ static ssize_t target_fabric_port_store_attr_alua_tg_pt_write_md(
        const char *page,
        size_t count)
 {
-       if (!(lun))
-               return -ENODEV;
-
-       if (!(lun->lun_sep))
+       if (!lun || !lun->lun_sep)
                return -ENODEV;
 
        return core_alua_store_secondary_write_metadata(lun, page, count);
@@ -781,13 +757,13 @@ static int target_fabric_port_link(
        tf = se_tpg->se_tpg_wwn->wwn_tf;
 
        if (lun->lun_se_dev !=  NULL) {
-               printk(KERN_ERR "Port Symlink already exists\n");
+               pr_err("Port Symlink already exists\n");
                return -EEXIST;
        }
 
        dev = se_dev->se_dev_ptr;
-       if (!(dev)) {
-               printk(KERN_ERR "Unable to locate struct se_device pointer from"
+       if (!dev) {
+               pr_err("Unable to locate struct se_device pointer from"
                        " %s\n", config_item_name(se_dev_ci));
                ret = -ENODEV;
                goto out;
@@ -795,8 +771,8 @@ static int target_fabric_port_link(
 
        lun_p = core_dev_add_lun(se_tpg, dev->se_hba, dev,
                                lun->unpacked_lun);
-       if ((IS_ERR(lun_p)) || !(lun_p)) {
-               printk(KERN_ERR "core_dev_add_lun() failed\n");
+       if (IS_ERR(lun_p) || !lun_p) {
+               pr_err("core_dev_add_lun() failed\n");
                ret = -EINVAL;
                goto out;
        }
@@ -888,7 +864,7 @@ static struct config_group *target_fabric_make_lun(
        int errno;
 
        if (strstr(name, "lun_") != name) {
-               printk(KERN_ERR "Unable to locate \'_\" in"
+               pr_err("Unable to locate \'_\" in"
                                " \"lun_$LUN_NUMBER\"\n");
                return ERR_PTR(-EINVAL);
        }
@@ -896,14 +872,14 @@ static struct config_group *target_fabric_make_lun(
                return ERR_PTR(-EINVAL);
 
        lun = core_get_lun_from_tpg(se_tpg, unpacked_lun);
-       if (!(lun))
+       if (!lun)
                return ERR_PTR(-EINVAL);
 
        lun_cg = &lun->lun_group;
        lun_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
                                GFP_KERNEL);
        if (!lun_cg->default_groups) {
-               printk(KERN_ERR "Unable to allocate lun_cg->default_groups\n");
+               pr_err("Unable to allocate lun_cg->default_groups\n");
                return ERR_PTR(-ENOMEM);
        }
 
@@ -914,11 +890,11 @@ static struct config_group *target_fabric_make_lun(
        lun_cg->default_groups[0] = &lun->port_stat_grps.stat_group;
        lun_cg->default_groups[1] = NULL;
 
-       port_stat_grp = &PORT_STAT_GRP(lun)->stat_group;
+       port_stat_grp = &lun->port_stat_grps.stat_group;
        port_stat_grp->default_groups =  kzalloc(sizeof(struct config_group) * 3,
                                GFP_KERNEL);
        if (!port_stat_grp->default_groups) {
-               printk(KERN_ERR "Unable to allocate port_stat_grp->default_groups\n");
+               pr_err("Unable to allocate port_stat_grp->default_groups\n");
                errno = -ENOMEM;
                goto out;
        }
@@ -941,7 +917,7 @@ static void target_fabric_drop_lun(
        struct config_group *lun_cg, *port_stat_grp;
        int i;
 
-       port_stat_grp = &PORT_STAT_GRP(lun)->stat_group;
+       port_stat_grp = &lun->port_stat_grps.stat_group;
        for (i = 0; port_stat_grp->default_groups[i]; i++) {
                df_item = &port_stat_grp->default_groups[i]->cg_item;
                port_stat_grp->default_groups[i] = NULL;
@@ -1031,13 +1007,13 @@ static struct config_group *target_fabric_make_tpg(
        struct target_fabric_configfs *tf = wwn->wwn_tf;
        struct se_portal_group *se_tpg;
 
-       if (!(tf->tf_ops.fabric_make_tpg)) {
-               printk(KERN_ERR "tf->tf_ops.fabric_make_tpg is NULL\n");
+       if (!tf->tf_ops.fabric_make_tpg) {
+               pr_err("tf->tf_ops.fabric_make_tpg is NULL\n");
                return ERR_PTR(-ENOSYS);
        }
 
        se_tpg = tf->tf_ops.fabric_make_tpg(wwn, group, name);
-       if (!(se_tpg) || IS_ERR(se_tpg))
+       if (!se_tpg || IS_ERR(se_tpg))
                return ERR_PTR(-EINVAL);
        /*
         * Setup default groups from pre-allocated se_tpg->tpg_default_groups
@@ -1130,13 +1106,13 @@ static struct config_group *target_fabric_make_wwn(
                                struct target_fabric_configfs, tf_group);
        struct se_wwn *wwn;
 
-       if (!(tf->tf_ops.fabric_make_wwn)) {
-               printk(KERN_ERR "tf->tf_ops.fabric_make_wwn is NULL\n");
+       if (!tf->tf_ops.fabric_make_wwn) {
+               pr_err("tf->tf_ops.fabric_make_wwn is NULL\n");
                return ERR_PTR(-ENOSYS);
        }
 
        wwn = tf->tf_ops.fabric_make_wwn(tf, group, name);
-       if (!(wwn) || IS_ERR(wwn))
+       if (!wwn || IS_ERR(wwn))
                return ERR_PTR(-EINVAL);
 
        wwn->wwn_tf = tf;
index 1e193f32489578c736ae463d9cae9aed28c21aea..c4ea3a9a555b4c35fc9d20495d8d3d492104e5e8 100644 (file)
@@ -25,6 +25,7 @@
  *
  ******************************************************************************/
 
+#include <linux/kernel.h>
 #include <linux/string.h>
 #include <linux/ctype.h>
 #include <linux/spinlock.h>
@@ -61,9 +62,8 @@ u32 sas_get_pr_transport_id(
        int *format_code,
        unsigned char *buf)
 {
-       unsigned char binary, *ptr;
-       int i;
-       u32 off = 4;
+       unsigned char *ptr;
+
        /*
         * Set PROTOCOL IDENTIFIER to 6h for SAS
         */
@@ -74,10 +74,8 @@ u32 sas_get_pr_transport_id(
         */
        ptr = &se_nacl->initiatorname[4]; /* Skip over 'naa. prefix */
 
-       for (i = 0; i < 16; i += 2) {
-               binary = transport_asciihex_to_binaryhex(&ptr[i]);
-               buf[off++] = binary;
-       }
+       hex2bin(&buf[4], ptr, 8);
+
        /*
         * The SAS Transport ID is a hardcoded 24-byte length
         */
@@ -157,7 +155,7 @@ u32 fc_get_pr_transport_id(
        int *format_code,
        unsigned char *buf)
 {
-       unsigned char binary, *ptr;
+       unsigned char *ptr;
        int i;
        u32 off = 8;
        /*
@@ -172,12 +170,11 @@ u32 fc_get_pr_transport_id(
        ptr = &se_nacl->initiatorname[0];
 
        for (i = 0; i < 24; ) {
-               if (!(strncmp(&ptr[i], ":", 1))) {
+               if (!strncmp(&ptr[i], ":", 1)) {
                        i++;
                        continue;
                }
-               binary = transport_asciihex_to_binaryhex(&ptr[i]);
-               buf[off++] = binary;
+               hex2bin(&buf[off++], &ptr[i], 1);
                i += 2;
        }
        /*
@@ -386,7 +383,7 @@ char *iscsi_parse_pr_out_transport_id(
         *            Reserved
         */
        if ((format_code != 0x00) && (format_code != 0x40)) {
-               printk(KERN_ERR "Illegal format code: 0x%02x for iSCSI"
+               pr_err("Illegal format code: 0x%02x for iSCSI"
                        " Initiator Transport ID\n", format_code);
                return NULL;
        }
@@ -406,7 +403,7 @@ char *iscsi_parse_pr_out_transport_id(
                        tid_len += padding;
 
                if ((add_len + 4) != tid_len) {
-                       printk(KERN_INFO "LIO-Target Extracted add_len: %hu "
+                       pr_debug("LIO-Target Extracted add_len: %hu "
                                "does not match calculated tid_len: %u,"
                                " using tid_len instead\n", add_len+4, tid_len);
                        *out_tid_len = tid_len;
@@ -420,8 +417,8 @@ char *iscsi_parse_pr_out_transport_id(
         */
        if (format_code == 0x40) {
                p = strstr((char *)&buf[4], ",i,0x");
-               if (!(p)) {
-                       printk(KERN_ERR "Unable to locate \",i,0x\" seperator"
+               if (!p) {
+                       pr_err("Unable to locate \",i,0x\" seperator"
                                " for Initiator port identifier: %s\n",
                                (char *)&buf[4]);
                        return NULL;
index 150c4305f385ea9c003341d8d853cac0fe4ab6bd..bc1b33639b8d8d0e16c2772ab859da2b63a1c1aa 100644 (file)
 
 #include "target_core_file.h"
 
-#if 1
-#define DEBUG_FD_CACHE(x...) printk(x)
-#else
-#define DEBUG_FD_CACHE(x...)
-#endif
-
-#if 1
-#define DEBUG_FD_FUA(x...) printk(x)
-#else
-#define DEBUG_FD_FUA(x...)
-#endif
-
 static struct se_subsystem_api fileio_template;
 
 /*     fd_attach_hba(): (Part of se_subsystem_api_t template)
@@ -65,24 +53,21 @@ static int fd_attach_hba(struct se_hba *hba, u32 host_id)
        struct fd_host *fd_host;
 
        fd_host = kzalloc(sizeof(struct fd_host), GFP_KERNEL);
-       if (!(fd_host)) {
-               printk(KERN_ERR "Unable to allocate memory for struct fd_host\n");
-               return -1;
+       if (!fd_host) {
+               pr_err("Unable to allocate memory for struct fd_host\n");
+               return -ENOMEM;
        }
 
        fd_host->fd_host_id = host_id;
 
-       atomic_set(&hba->left_queue_depth, FD_HBA_QUEUE_DEPTH);
-       atomic_set(&hba->max_queue_depth, FD_HBA_QUEUE_DEPTH);
-       hba->hba_ptr = (void *) fd_host;
+       hba->hba_ptr = fd_host;
 
-       printk(KERN_INFO "CORE_HBA[%d] - TCM FILEIO HBA Driver %s on Generic"
+       pr_debug("CORE_HBA[%d] - TCM FILEIO HBA Driver %s on Generic"
                " Target Core Stack %s\n", hba->hba_id, FD_VERSION,
                TARGET_CORE_MOD_VERSION);
-       printk(KERN_INFO "CORE_HBA[%d] - Attached FILEIO HBA: %u to Generic"
-               " Target Core with TCQ Depth: %d MaxSectors: %u\n",
-               hba->hba_id, fd_host->fd_host_id,
-               atomic_read(&hba->max_queue_depth), FD_MAX_SECTORS);
+       pr_debug("CORE_HBA[%d] - Attached FILEIO HBA: %u to Generic"
+               " MaxSectors: %u\n",
+               hba->hba_id, fd_host->fd_host_id, FD_MAX_SECTORS);
 
        return 0;
 }
@@ -91,7 +76,7 @@ static void fd_detach_hba(struct se_hba *hba)
 {
        struct fd_host *fd_host = hba->hba_ptr;
 
-       printk(KERN_INFO "CORE_HBA[%d] - Detached FILEIO HBA: %u from Generic"
+       pr_debug("CORE_HBA[%d] - Detached FILEIO HBA: %u from Generic"
                " Target Core\n", hba->hba_id, fd_host->fd_host_id);
 
        kfree(fd_host);
@@ -104,14 +89,14 @@ static void *fd_allocate_virtdevice(struct se_hba *hba, const char *name)
        struct fd_host *fd_host = (struct fd_host *) hba->hba_ptr;
 
        fd_dev = kzalloc(sizeof(struct fd_dev), GFP_KERNEL);
-       if (!(fd_dev)) {
-               printk(KERN_ERR "Unable to allocate memory for struct fd_dev\n");
+       if (!fd_dev) {
+               pr_err("Unable to allocate memory for struct fd_dev\n");
                return NULL;
        }
 
        fd_dev->fd_host = fd_host;
 
-       printk(KERN_INFO "FILEIO: Allocated fd_dev for %p\n", name);
+       pr_debug("FILEIO: Allocated fd_dev for %p\n", name);
 
        return fd_dev;
 }
@@ -144,7 +129,7 @@ static struct se_device *fd_create_virtdevice(
        set_fs(old_fs);
 
        if (IS_ERR(dev_p)) {
-               printk(KERN_ERR "getname(%s) failed: %lu\n",
+               pr_err("getname(%s) failed: %lu\n",
                        fd_dev->fd_dev_name, IS_ERR(dev_p));
                ret = PTR_ERR(dev_p);
                goto fail;
@@ -167,12 +152,12 @@ static struct se_device *fd_create_virtdevice(
 
        file = filp_open(dev_p, flags, 0600);
        if (IS_ERR(file)) {
-               printk(KERN_ERR "filp_open(%s) failed\n", dev_p);
+               pr_err("filp_open(%s) failed\n", dev_p);
                ret = PTR_ERR(file);
                goto fail;
        }
        if (!file || !file->f_dentry) {
-               printk(KERN_ERR "filp_open(%s) failed\n", dev_p);
+               pr_err("filp_open(%s) failed\n", dev_p);
                goto fail;
        }
        fd_dev->fd_file = file;
@@ -202,14 +187,14 @@ static struct se_device *fd_create_virtdevice(
                fd_dev->fd_dev_size = (i_size_read(file->f_mapping->host) -
                                       fd_dev->fd_block_size);
 
-               printk(KERN_INFO "FILEIO: Using size: %llu bytes from struct"
+               pr_debug("FILEIO: Using size: %llu bytes from struct"
                        " block_device blocks: %llu logical_block_size: %d\n",
                        fd_dev->fd_dev_size,
                        div_u64(fd_dev->fd_dev_size, fd_dev->fd_block_size),
                        fd_dev->fd_block_size);
        } else {
                if (!(fd_dev->fbd_flags & FBDF_HAS_SIZE)) {
-                       printk(KERN_ERR "FILEIO: Missing fd_dev_size="
+                       pr_err("FILEIO: Missing fd_dev_size="
                                " parameter, and no backing struct"
                                " block_device\n");
                        goto fail;
@@ -226,15 +211,15 @@ static struct se_device *fd_create_virtdevice(
        dev_limits.queue_depth = FD_DEVICE_QUEUE_DEPTH;
 
        dev = transport_add_device_to_core_hba(hba, &fileio_template,
-                               se_dev, dev_flags, (void *)fd_dev,
+                               se_dev, dev_flags, fd_dev,
                                &dev_limits, "FILEIO", FD_VERSION);
-       if (!(dev))
+       if (!dev)
                goto fail;
 
        fd_dev->fd_dev_id = fd_host->fd_host_dev_id_count++;
        fd_dev->fd_queue_depth = dev->queue_depth;
 
-       printk(KERN_INFO "CORE_FILE[%u] - Added TCM FILEIO Device ID: %u at %s,"
+       pr_debug("CORE_FILE[%u] - Added TCM FILEIO Device ID: %u at %s,"
                " %llu total bytes\n", fd_host->fd_host_id, fd_dev->fd_dev_id,
                        fd_dev->fd_dev_name, fd_dev->fd_dev_size);
 
@@ -272,45 +257,45 @@ static inline struct fd_request *FILE_REQ(struct se_task *task)
 
 
 static struct se_task *
-fd_alloc_task(struct se_cmd *cmd)
+fd_alloc_task(unsigned char *cdb)
 {
        struct fd_request *fd_req;
 
        fd_req = kzalloc(sizeof(struct fd_request), GFP_KERNEL);
-       if (!(fd_req)) {
-               printk(KERN_ERR "Unable to allocate struct fd_request\n");
+       if (!fd_req) {
+               pr_err("Unable to allocate struct fd_request\n");
                return NULL;
        }
 
-       fd_req->fd_dev = SE_DEV(cmd)->dev_ptr;
-
        return &fd_req->fd_task;
 }
 
 static int fd_do_readv(struct se_task *task)
 {
        struct fd_request *req = FILE_REQ(task);
-       struct file *fd = req->fd_dev->fd_file;
+       struct fd_dev *dev = req->fd_task.se_dev->dev_ptr;
+       struct file *fd = dev->fd_file;
        struct scatterlist *sg = task->task_sg;
        struct iovec *iov;
        mm_segment_t old_fs;
-       loff_t pos = (task->task_lba * DEV_ATTRIB(task->se_dev)->block_size);
+       loff_t pos = (task->task_lba *
+                     task->se_dev->se_sub_dev->se_dev_attrib.block_size);
        int ret = 0, i;
 
-       iov = kzalloc(sizeof(struct iovec) * task->task_sg_num, GFP_KERNEL);
-       if (!(iov)) {
-               printk(KERN_ERR "Unable to allocate fd_do_readv iov[]\n");
-               return -1;
+       iov = kzalloc(sizeof(struct iovec) * task->task_sg_nents, GFP_KERNEL);
+       if (!iov) {
+               pr_err("Unable to allocate fd_do_readv iov[]\n");
+               return -ENOMEM;
        }
 
-       for (i = 0; i < task->task_sg_num; i++) {
+       for (i = 0; i < task->task_sg_nents; i++) {
                iov[i].iov_len = sg[i].length;
                iov[i].iov_base = sg_virt(&sg[i]);
        }
 
        old_fs = get_fs();
        set_fs(get_ds());
-       ret = vfs_readv(fd, &iov[0], task->task_sg_num, &pos);
+       ret = vfs_readv(fd, &iov[0], task->task_sg_nents, &pos);
        set_fs(old_fs);
 
        kfree(iov);
@@ -321,16 +306,16 @@ static int fd_do_readv(struct se_task *task)
         */
        if (S_ISBLK(fd->f_dentry->d_inode->i_mode)) {
                if (ret < 0 || ret != task->task_size) {
-                       printk(KERN_ERR "vfs_readv() returned %d,"
+                       pr_err("vfs_readv() returned %d,"
                                " expecting %d for S_ISBLK\n", ret,
                                (int)task->task_size);
-                       return -1;
+                       return (ret < 0 ? ret : -EINVAL);
                }
        } else {
                if (ret < 0) {
-                       printk(KERN_ERR "vfs_readv() returned %d for non"
+                       pr_err("vfs_readv() returned %d for non"
                                " S_ISBLK\n", ret);
-                       return -1;
+                       return ret;
                }
        }
 
@@ -340,34 +325,36 @@ static int fd_do_readv(struct se_task *task)
 static int fd_do_writev(struct se_task *task)
 {
        struct fd_request *req = FILE_REQ(task);
-       struct file *fd = req->fd_dev->fd_file;
+       struct fd_dev *dev = req->fd_task.se_dev->dev_ptr;
+       struct file *fd = dev->fd_file;
        struct scatterlist *sg = task->task_sg;
        struct iovec *iov;
        mm_segment_t old_fs;
-       loff_t pos = (task->task_lba * DEV_ATTRIB(task->se_dev)->block_size);
+       loff_t pos = (task->task_lba *
+                     task->se_dev->se_sub_dev->se_dev_attrib.block_size);
        int ret, i = 0;
 
-       iov = kzalloc(sizeof(struct iovec) * task->task_sg_num, GFP_KERNEL);
-       if (!(iov)) {
-               printk(KERN_ERR "Unable to allocate fd_do_writev iov[]\n");
-               return -1;
+       iov = kzalloc(sizeof(struct iovec) * task->task_sg_nents, GFP_KERNEL);
+       if (!iov) {
+               pr_err("Unable to allocate fd_do_writev iov[]\n");
+               return -ENOMEM;
        }
 
-       for (i = 0; i < task->task_sg_num; i++) {
+       for (i = 0; i < task->task_sg_nents; i++) {
                iov[i].iov_len = sg[i].length;
                iov[i].iov_base = sg_virt(&sg[i]);
        }
 
        old_fs = get_fs();
        set_fs(get_ds());
-       ret = vfs_writev(fd, &iov[0], task->task_sg_num, &pos);
+       ret = vfs_writev(fd, &iov[0], task->task_sg_nents, &pos);
        set_fs(old_fs);
 
        kfree(iov);
 
        if (ret < 0 || ret != task->task_size) {
-               printk(KERN_ERR "vfs_writev() returned %d\n", ret);
-               return -1;
+               pr_err("vfs_writev() returned %d\n", ret);
+               return (ret < 0 ? ret : -EINVAL);
        }
 
        return 1;
@@ -375,10 +362,10 @@ static int fd_do_writev(struct se_task *task)
 
 static void fd_emulate_sync_cache(struct se_task *task)
 {
-       struct se_cmd *cmd = TASK_CMD(task);
+       struct se_cmd *cmd = task->task_se_cmd;
        struct se_device *dev = cmd->se_dev;
        struct fd_dev *fd_dev = dev->dev_ptr;
-       int immed = (cmd->t_task->t_task_cdb[1] & 0x2);
+       int immed = (cmd->t_task_cdb[1] & 0x2);
        loff_t start, end;
        int ret;
 
@@ -392,11 +379,11 @@ static void fd_emulate_sync_cache(struct se_task *task)
        /*
         * Determine if we will be flushing the entire device.
         */
-       if (cmd->t_task->t_task_lba == 0 && cmd->data_length == 0) {
+       if (cmd->t_task_lba == 0 && cmd->data_length == 0) {
                start = 0;
                end = LLONG_MAX;
        } else {
-               start = cmd->t_task->t_task_lba * DEV_ATTRIB(dev)->block_size;
+               start = cmd->t_task_lba * dev->se_sub_dev->se_dev_attrib.block_size;
                if (cmd->data_length)
                        end = start + cmd->data_length;
                else
@@ -405,7 +392,7 @@ static void fd_emulate_sync_cache(struct se_task *task)
 
        ret = vfs_fsync_range(fd_dev->fd_file, start, end, 1);
        if (ret != 0)
-               printk(KERN_ERR "FILEIO: vfs_fsync_range() failed: %d\n", ret);
+               pr_err("FILEIO: vfs_fsync_range() failed: %d\n", ret);
 
        if (!immed)
                transport_complete_sync_cache(cmd, ret == 0);
@@ -446,16 +433,16 @@ static void fd_emulate_write_fua(struct se_cmd *cmd, struct se_task *task)
 {
        struct se_device *dev = cmd->se_dev;
        struct fd_dev *fd_dev = dev->dev_ptr;
-       loff_t start = task->task_lba * DEV_ATTRIB(dev)->block_size;
+       loff_t start = task->task_lba * dev->se_sub_dev->se_dev_attrib.block_size;
        loff_t end = start + task->task_size;
        int ret;
 
-       DEBUG_FD_CACHE("FILEIO: FUA WRITE LBA: %llu, bytes: %u\n",
+       pr_debug("FILEIO: FUA WRITE LBA: %llu, bytes: %u\n",
                        task->task_lba, task->task_size);
 
        ret = vfs_fsync_range(fd_dev->fd_file, start, end, 1);
        if (ret != 0)
-               printk(KERN_ERR "FILEIO: vfs_fsync_range() failed: %d\n", ret);
+               pr_err("FILEIO: vfs_fsync_range() failed: %d\n", ret);
 }
 
 static int fd_do_task(struct se_task *task)
@@ -474,9 +461,9 @@ static int fd_do_task(struct se_task *task)
                ret = fd_do_writev(task);
 
                if (ret > 0 &&
-                   DEV_ATTRIB(dev)->emulate_write_cache > 0 &&
-                   DEV_ATTRIB(dev)->emulate_fua_write > 0 &&
-                   T_TASK(cmd)->t_tasks_fua) {
+                   dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0 &&
+                   dev->se_sub_dev->se_dev_attrib.emulate_fua_write > 0 &&
+                   cmd->t_tasks_fua) {
                        /*
                         * We might need to be a bit smarter here
                         * and return some sense data to let the initiator
@@ -549,7 +536,7 @@ static ssize_t fd_set_configfs_dev_params(
                        snprintf(fd_dev->fd_dev_name, FD_MAX_DEV_NAME,
                                        "%s", arg_p);
                        kfree(arg_p);
-                       printk(KERN_INFO "FILEIO: Referencing Path: %s\n",
+                       pr_debug("FILEIO: Referencing Path: %s\n",
                                        fd_dev->fd_dev_name);
                        fd_dev->fbd_flags |= FBDF_HAS_PATH;
                        break;
@@ -562,23 +549,23 @@ static ssize_t fd_set_configfs_dev_params(
                        ret = strict_strtoull(arg_p, 0, &fd_dev->fd_dev_size);
                        kfree(arg_p);
                        if (ret < 0) {
-                               printk(KERN_ERR "strict_strtoull() failed for"
+                               pr_err("strict_strtoull() failed for"
                                                " fd_dev_size=\n");
                                goto out;
                        }
-                       printk(KERN_INFO "FILEIO: Referencing Size: %llu"
+                       pr_debug("FILEIO: Referencing Size: %llu"
                                        " bytes\n", fd_dev->fd_dev_size);
                        fd_dev->fbd_flags |= FBDF_HAS_SIZE;
                        break;
                case Opt_fd_buffered_io:
                        match_int(args, &arg);
                        if (arg != 1) {
-                               printk(KERN_ERR "bogus fd_buffered_io=%d value\n", arg);
+                               pr_err("bogus fd_buffered_io=%d value\n", arg);
                                ret = -EINVAL;
                                goto out;
                        }
 
-                       printk(KERN_INFO "FILEIO: Using buffered I/O"
+                       pr_debug("FILEIO: Using buffered I/O"
                                " operations for struct fd_dev\n");
 
                        fd_dev->fbd_flags |= FDBD_USE_BUFFERED_IO;
@@ -598,8 +585,8 @@ static ssize_t fd_check_configfs_dev_params(struct se_hba *hba, struct se_subsys
        struct fd_dev *fd_dev = (struct fd_dev *) se_dev->se_dev_su_ptr;
 
        if (!(fd_dev->fbd_flags & FBDF_HAS_PATH)) {
-               printk(KERN_ERR "Missing fd_dev_name=\n");
-               return -1;
+               pr_err("Missing fd_dev_name=\n");
+               return -EINVAL;
        }
 
        return 0;
@@ -654,7 +641,7 @@ static sector_t fd_get_blocks(struct se_device *dev)
 {
        struct fd_dev *fd_dev = dev->dev_ptr;
        unsigned long long blocks_long = div_u64(fd_dev->fd_dev_size,
-                       DEV_ATTRIB(dev)->block_size);
+                       dev->se_sub_dev->se_dev_attrib.block_size);
 
        return blocks_long;
 }
index ef4de2b4bd4666792ea7099e9b8e458273c48765..daebd710b8939bcfc744f35d3a6c25ad872bbd35 100644 (file)
@@ -4,8 +4,6 @@
 #define FD_VERSION             "4.0"
 
 #define FD_MAX_DEV_NAME                256
-/* Maximum queuedepth for the FILEIO HBA */
-#define FD_HBA_QUEUE_DEPTH     256
 #define FD_DEVICE_QUEUE_DEPTH  32
 #define FD_MAX_DEVICE_QUEUE_DEPTH 128
 #define FD_BLOCKSIZE           512
@@ -18,8 +16,6 @@ struct fd_request {
        struct se_task  fd_task;
        /* SCSI CDB from iSCSI Command PDU */
        unsigned char   fd_scsi_cdb[TCM_MAX_COMMAND_SIZE];
-       /* FILEIO device */
-       struct fd_dev   *fd_dev;
 } ____cacheline_aligned;
 
 #define FBDF_HAS_PATH          0x01
index 0b8f8da890199aa4807958a1f101ab14ede03a3d..0639b975d6f55c79944583c4260e3c1722bf7909 100644 (file)
@@ -1,7 +1,7 @@
 /*******************************************************************************
  * Filename:  target_core_hba.c
  *
- * This file copntains the iSCSI HBA Transport related functions.
+ * This file contains the TCM HBA Transport related functions.
  *
  * Copyright (c) 2003, 2004, 2005 PyX Technologies, Inc.
  * Copyright (c) 2005, 2006, 2007 SBE, Inc.
 static LIST_HEAD(subsystem_list);
 static DEFINE_MUTEX(subsystem_mutex);
 
+static u32 hba_id_counter;
+
+static DEFINE_SPINLOCK(hba_lock);
+static LIST_HEAD(hba_list);
+
 int transport_subsystem_register(struct se_subsystem_api *sub_api)
 {
        struct se_subsystem_api *s;
@@ -53,8 +58,8 @@ int transport_subsystem_register(struct se_subsystem_api *sub_api)
 
        mutex_lock(&subsystem_mutex);
        list_for_each_entry(s, &subsystem_list, sub_api_list) {
-               if (!(strcmp(s->name, sub_api->name))) {
-                       printk(KERN_ERR "%p is already registered with"
+               if (!strcmp(s->name, sub_api->name)) {
+                       pr_err("%p is already registered with"
                                " duplicate name %s, unable to process"
                                " request\n", s, s->name);
                        mutex_unlock(&subsystem_mutex);
@@ -64,7 +69,7 @@ int transport_subsystem_register(struct se_subsystem_api *sub_api)
        list_add_tail(&sub_api->sub_api_list, &subsystem_list);
        mutex_unlock(&subsystem_mutex);
 
-       printk(KERN_INFO "TCM: Registered subsystem plugin: %s struct module:"
+       pr_debug("TCM: Registered subsystem plugin: %s struct module:"
                        " %p\n", sub_api->name, sub_api->owner);
        return 0;
 }
@@ -104,21 +109,17 @@ core_alloc_hba(const char *plugin_name, u32 plugin_dep_id, u32 hba_flags)
 
        hba = kzalloc(sizeof(*hba), GFP_KERNEL);
        if (!hba) {
-               printk(KERN_ERR "Unable to allocate struct se_hba\n");
+               pr_err("Unable to allocate struct se_hba\n");
                return ERR_PTR(-ENOMEM);
        }
 
        INIT_LIST_HEAD(&hba->hba_dev_list);
        spin_lock_init(&hba->device_lock);
-       spin_lock_init(&hba->hba_queue_lock);
        mutex_init(&hba->hba_access_mutex);
 
        hba->hba_index = scsi_get_new_index(SCSI_INST_INDEX);
        hba->hba_flags |= hba_flags;
 
-       atomic_set(&hba->max_queue_depth, 0);
-       atomic_set(&hba->left_queue_depth, 0);
-
        hba->transport = core_get_backend(plugin_name);
        if (!hba->transport) {
                ret = -EINVAL;
@@ -129,12 +130,12 @@ core_alloc_hba(const char *plugin_name, u32 plugin_dep_id, u32 hba_flags)
        if (ret < 0)
                goto out_module_put;
 
-       spin_lock(&se_global->hba_lock);
-       hba->hba_id = se_global->g_hba_id_counter++;
-       list_add_tail(&hba->hba_list, &se_global->g_hba_list);
-       spin_unlock(&se_global->hba_lock);
+       spin_lock(&hba_lock);
+       hba->hba_id = hba_id_counter++;
+       list_add_tail(&hba->hba_node, &hba_list);
+       spin_unlock(&hba_lock);
 
-       printk(KERN_INFO "CORE_HBA[%d] - Attached HBA to Generic Target"
+       pr_debug("CORE_HBA[%d] - Attached HBA to Generic Target"
                        " Core\n", hba->hba_id);
 
        return hba;
@@ -156,11 +157,11 @@ core_delete_hba(struct se_hba *hba)
 
        hba->transport->detach_hba(hba);
 
-       spin_lock(&se_global->hba_lock);
-       list_del(&hba->hba_list);
-       spin_unlock(&se_global->hba_lock);
+       spin_lock(&hba_lock);
+       list_del(&hba->hba_node);
+       spin_unlock(&hba_lock);
 
-       printk(KERN_INFO "CORE_HBA[%d] - Detached HBA from Generic Target"
+       pr_debug("CORE_HBA[%d] - Detached HBA from Generic Target"
                        " Core\n", hba->hba_id);
 
        if (hba->transport->owner)
index 86639004af9eed7a7acf899d66157a05dd98ffc6..7e12341054426404ddd010d793450ca5ac0d7265 100644 (file)
 
 #include "target_core_iblock.h"
 
-#if 0
-#define DEBUG_IBLOCK(x...) printk(x)
-#else
-#define DEBUG_IBLOCK(x...)
-#endif
-
 static struct se_subsystem_api iblock_template;
 
 static void iblock_bio_done(struct bio *, int);
@@ -66,25 +60,22 @@ static int iblock_attach_hba(struct se_hba *hba, u32 host_id)
        struct iblock_hba *ib_host;
 
        ib_host = kzalloc(sizeof(struct iblock_hba), GFP_KERNEL);
-       if (!(ib_host)) {
-               printk(KERN_ERR "Unable to allocate memory for"
+       if (!ib_host) {
+               pr_err("Unable to allocate memory for"
                                " struct iblock_hba\n");
                return -ENOMEM;
        }
 
        ib_host->iblock_host_id = host_id;
 
-       atomic_set(&hba->left_queue_depth, IBLOCK_HBA_QUEUE_DEPTH);
-       atomic_set(&hba->max_queue_depth, IBLOCK_HBA_QUEUE_DEPTH);
-       hba->hba_ptr = (void *) ib_host;
+       hba->hba_ptr = ib_host;
 
-       printk(KERN_INFO "CORE_HBA[%d] - TCM iBlock HBA Driver %s on"
+       pr_debug("CORE_HBA[%d] - TCM iBlock HBA Driver %s on"
                " Generic Target Core Stack %s\n", hba->hba_id,
                IBLOCK_VERSION, TARGET_CORE_MOD_VERSION);
 
-       printk(KERN_INFO "CORE_HBA[%d] - Attached iBlock HBA: %u to Generic"
-               " Target Core TCQ Depth: %d\n", hba->hba_id,
-               ib_host->iblock_host_id, atomic_read(&hba->max_queue_depth));
+       pr_debug("CORE_HBA[%d] - Attached iBlock HBA: %u to Generic\n",
+               hba->hba_id, ib_host->iblock_host_id);
 
        return 0;
 }
@@ -93,7 +84,7 @@ static void iblock_detach_hba(struct se_hba *hba)
 {
        struct iblock_hba *ib_host = hba->hba_ptr;
 
-       printk(KERN_INFO "CORE_HBA[%d] - Detached iBlock HBA: %u from Generic"
+       pr_debug("CORE_HBA[%d] - Detached iBlock HBA: %u from Generic"
                " Target Core\n", hba->hba_id, ib_host->iblock_host_id);
 
        kfree(ib_host);
@@ -106,13 +97,13 @@ static void *iblock_allocate_virtdevice(struct se_hba *hba, const char *name)
        struct iblock_hba *ib_host = hba->hba_ptr;
 
        ib_dev = kzalloc(sizeof(struct iblock_dev), GFP_KERNEL);
-       if (!(ib_dev)) {
-               printk(KERN_ERR "Unable to allocate struct iblock_dev\n");
+       if (!ib_dev) {
+               pr_err("Unable to allocate struct iblock_dev\n");
                return NULL;
        }
        ib_dev->ibd_host = ib_host;
 
-       printk(KERN_INFO  "IBLOCK: Allocated ib_dev for %s\n", name);
+       pr_debug( "IBLOCK: Allocated ib_dev for %s\n", name);
 
        return ib_dev;
 }
@@ -131,8 +122,8 @@ static struct se_device *iblock_create_virtdevice(
        u32 dev_flags = 0;
        int ret = -EINVAL;
 
-       if (!(ib_dev)) {
-               printk(KERN_ERR "Unable to locate struct iblock_dev parameter\n");
+       if (!ib_dev) {
+               pr_err("Unable to locate struct iblock_dev parameter\n");
                return ERR_PTR(ret);
        }
        memset(&dev_limits, 0, sizeof(struct se_dev_limits));
@@ -140,16 +131,16 @@ static struct se_device *iblock_create_virtdevice(
         * These settings need to be made tunable..
         */
        ib_dev->ibd_bio_set = bioset_create(32, 64);
-       if (!(ib_dev->ibd_bio_set)) {
-               printk(KERN_ERR "IBLOCK: Unable to create bioset()\n");
+       if (!ib_dev->ibd_bio_set) {
+               pr_err("IBLOCK: Unable to create bioset()\n");
                return ERR_PTR(-ENOMEM);
        }
-       printk(KERN_INFO "IBLOCK: Created bio_set()\n");
+       pr_debug("IBLOCK: Created bio_set()\n");
        /*
         * iblock_check_configfs_dev_params() ensures that ib_dev->ibd_udev_path
         * must already have been set in order for echo 1 > $HBA/$DEV/enable to run.
         */
-       printk(KERN_INFO  "IBLOCK: Claiming struct block_device: %s\n",
+       pr_debug( "IBLOCK: Claiming struct block_device: %s\n",
                        ib_dev->ibd_udev_path);
 
        bd = blkdev_get_by_path(ib_dev->ibd_udev_path,
@@ -167,42 +158,41 @@ static struct se_device *iblock_create_virtdevice(
        limits->logical_block_size = bdev_logical_block_size(bd);
        limits->max_hw_sectors = queue_max_hw_sectors(q);
        limits->max_sectors = queue_max_sectors(q);
-       dev_limits.hw_queue_depth = IBLOCK_MAX_DEVICE_QUEUE_DEPTH;
-       dev_limits.queue_depth = IBLOCK_DEVICE_QUEUE_DEPTH;
+       dev_limits.hw_queue_depth = q->nr_requests;
+       dev_limits.queue_depth = q->nr_requests;
 
-       ib_dev->ibd_major = MAJOR(bd->bd_dev);
-       ib_dev->ibd_minor = MINOR(bd->bd_dev);
        ib_dev->ibd_bd = bd;
 
        dev = transport_add_device_to_core_hba(hba,
-                       &iblock_template, se_dev, dev_flags, (void *)ib_dev,
+                       &iblock_template, se_dev, dev_flags, ib_dev,
                        &dev_limits, "IBLOCK", IBLOCK_VERSION);
-       if (!(dev))
+       if (!dev)
                goto failed;
 
-       ib_dev->ibd_depth = dev->queue_depth;
-
        /*
         * Check if the underlying struct block_device request_queue supports
         * the QUEUE_FLAG_DISCARD bit for UNMAP/WRITE_SAME in SCSI + TRIM
         * in ATA and we need to set TPE=1
         */
        if (blk_queue_discard(q)) {
-               DEV_ATTRIB(dev)->max_unmap_lba_count =
+               dev->se_sub_dev->se_dev_attrib.max_unmap_lba_count =
                                q->limits.max_discard_sectors;
                /*
                 * Currently hardcoded to 1 in Linux/SCSI code..
                 */
-               DEV_ATTRIB(dev)->max_unmap_block_desc_count = 1;
-               DEV_ATTRIB(dev)->unmap_granularity =
+               dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count = 1;
+               dev->se_sub_dev->se_dev_attrib.unmap_granularity =
                                q->limits.discard_granularity;
-               DEV_ATTRIB(dev)->unmap_granularity_alignment =
+               dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment =
                                q->limits.discard_alignment;
 
-               printk(KERN_INFO "IBLOCK: BLOCK Discard support available,"
+               pr_debug("IBLOCK: BLOCK Discard support available,"
                                " disabled by default\n");
        }
 
+       if (blk_queue_nonrot(q))
+               dev->se_sub_dev->se_dev_attrib.is_nonrot = 1;
+
        return dev;
 
 failed:
@@ -211,8 +201,6 @@ failed:
                ib_dev->ibd_bio_set = NULL;
        }
        ib_dev->ibd_bd = NULL;
-       ib_dev->ibd_major = 0;
-       ib_dev->ibd_minor = 0;
        return ERR_PTR(ret);
 }
 
@@ -233,17 +221,16 @@ static inline struct iblock_req *IBLOCK_REQ(struct se_task *task)
 }
 
 static struct se_task *
-iblock_alloc_task(struct se_cmd *cmd)
+iblock_alloc_task(unsigned char *cdb)
 {
        struct iblock_req *ib_req;
 
        ib_req = kzalloc(sizeof(struct iblock_req), GFP_KERNEL);
-       if (!(ib_req)) {
-               printk(KERN_ERR "Unable to allocate memory for struct iblock_req\n");
+       if (!ib_req) {
+               pr_err("Unable to allocate memory for struct iblock_req\n");
                return NULL;
        }
 
-       ib_req->ib_dev = SE_DEV(cmd)->dev_ptr;
        atomic_set(&ib_req->ib_bio_cnt, 0);
        return &ib_req->ib_task;
 }
@@ -257,12 +244,12 @@ static unsigned long long iblock_emulate_read_cap_with_block_size(
                                        bdev_logical_block_size(bd)) - 1);
        u32 block_size = bdev_logical_block_size(bd);
 
-       if (block_size == DEV_ATTRIB(dev)->block_size)
+       if (block_size == dev->se_sub_dev->se_dev_attrib.block_size)
                return blocks_long;
 
        switch (block_size) {
        case 4096:
-               switch (DEV_ATTRIB(dev)->block_size) {
+               switch (dev->se_sub_dev->se_dev_attrib.block_size) {
                case 2048:
                        blocks_long <<= 1;
                        break;
@@ -276,7 +263,7 @@ static unsigned long long iblock_emulate_read_cap_with_block_size(
                }
                break;
        case 2048:
-               switch (DEV_ATTRIB(dev)->block_size) {
+               switch (dev->se_sub_dev->se_dev_attrib.block_size) {
                case 4096:
                        blocks_long >>= 1;
                        break;
@@ -291,7 +278,7 @@ static unsigned long long iblock_emulate_read_cap_with_block_size(
                }
                break;
        case 1024:
-               switch (DEV_ATTRIB(dev)->block_size) {
+               switch (dev->se_sub_dev->se_dev_attrib.block_size) {
                case 4096:
                        blocks_long >>= 2;
                        break;
@@ -306,7 +293,7 @@ static unsigned long long iblock_emulate_read_cap_with_block_size(
                }
                break;
        case 512:
-               switch (DEV_ATTRIB(dev)->block_size) {
+               switch (dev->se_sub_dev->se_dev_attrib.block_size) {
                case 4096:
                        blocks_long >>= 3;
                        break;
@@ -332,9 +319,9 @@ static unsigned long long iblock_emulate_read_cap_with_block_size(
  */
 static void iblock_emulate_sync_cache(struct se_task *task)
 {
-       struct se_cmd *cmd = TASK_CMD(task);
+       struct se_cmd *cmd = task->task_se_cmd;
        struct iblock_dev *ib_dev = cmd->se_dev->dev_ptr;
-       int immed = (T_TASK(cmd)->t_task_cdb[1] & 0x2);
+       int immed = (cmd->t_task_cdb[1] & 0x2);
        sector_t error_sector;
        int ret;
 
@@ -351,7 +338,7 @@ static void iblock_emulate_sync_cache(struct se_task *task)
         */
        ret = blkdev_issue_flush(ib_dev->ibd_bd, GFP_KERNEL, &error_sector);
        if (ret != 0) {
-               printk(KERN_ERR "IBLOCK: block_issue_flush() failed: %d "
+               pr_err("IBLOCK: block_issue_flush() failed: %d "
                        " error_sector: %llu\n", ret,
                        (unsigned long long)error_sector);
        }
@@ -401,9 +388,9 @@ static int iblock_do_task(struct se_task *task)
                 * Force data to disk if we pretend to not have a volatile
                 * write cache, or the initiator set the Force Unit Access bit.
                 */
-               if (DEV_ATTRIB(dev)->emulate_write_cache == 0 ||
-                   (DEV_ATTRIB(dev)->emulate_fua_write > 0 &&
-                    T_TASK(task->task_se_cmd)->t_tasks_fua))
+               if (dev->se_sub_dev->se_dev_attrib.emulate_write_cache == 0 ||
+                   (dev->se_sub_dev->se_dev_attrib.emulate_fua_write > 0 &&
+                    task->task_se_cmd->t_tasks_fua))
                        rw = WRITE_FUA;
                else
                        rw = WRITE;
@@ -415,8 +402,9 @@ static int iblock_do_task(struct se_task *task)
        while (bio) {
                nbio = bio->bi_next;
                bio->bi_next = NULL;
-               DEBUG_IBLOCK("Calling submit_bio() task: %p bio: %p"
-                       " bio->bi_sector: %llu\n", task, bio, bio->bi_sector);
+               pr_debug("Calling submit_bio() task: %p bio: %p"
+                       " bio->bi_sector: %llu\n", task, bio,
+                        (unsigned long long)bio->bi_sector);
 
                submit_bio(rw, bio);
                bio = nbio;
@@ -470,7 +458,7 @@ static ssize_t iblock_set_configfs_dev_params(struct se_hba *hba,
        struct iblock_dev *ib_dev = se_dev->se_dev_su_ptr;
        char *orig, *ptr, *arg_p, *opts;
        substring_t args[MAX_OPT_ARGS];
-       int ret = 0, arg, token;
+       int ret = 0, token;
 
        opts = kstrdup(page, GFP_KERNEL);
        if (!opts)
@@ -486,7 +474,7 @@ static ssize_t iblock_set_configfs_dev_params(struct se_hba *hba,
                switch (token) {
                case Opt_udev_path:
                        if (ib_dev->ibd_bd) {
-                               printk(KERN_ERR "Unable to set udev_path= while"
+                               pr_err("Unable to set udev_path= while"
                                        " ib_dev->ibd_bd exists\n");
                                ret = -EEXIST;
                                goto out;
@@ -499,15 +487,11 @@ static ssize_t iblock_set_configfs_dev_params(struct se_hba *hba,
                        snprintf(ib_dev->ibd_udev_path, SE_UDEV_PATH_LEN,
                                        "%s", arg_p);
                        kfree(arg_p);
-                       printk(KERN_INFO "IBLOCK: Referencing UDEV path: %s\n",
+                       pr_debug("IBLOCK: Referencing UDEV path: %s\n",
                                        ib_dev->ibd_udev_path);
                        ib_dev->ibd_flags |= IBDF_HAS_UDEV_PATH;
                        break;
                case Opt_force:
-                       match_int(args, &arg);
-                       ib_dev->ibd_force = arg;
-                       printk(KERN_INFO "IBLOCK: Set force=%d\n",
-                               ib_dev->ibd_force);
                        break;
                default:
                        break;
@@ -526,8 +510,8 @@ static ssize_t iblock_check_configfs_dev_params(
        struct iblock_dev *ibd = se_dev->se_dev_su_ptr;
 
        if (!(ibd->ibd_flags & IBDF_HAS_UDEV_PATH)) {
-               printk(KERN_ERR "Missing udev_path= parameters for IBLOCK\n");
-               return -1;
+               pr_err("Missing udev_path= parameters for IBLOCK\n");
+               return -EINVAL;
        }
 
        return 0;
@@ -555,12 +539,11 @@ static ssize_t iblock_show_configfs_dev_params(
        bl += sprintf(b + bl, "        ");
        if (bd) {
                bl += sprintf(b + bl, "Major: %d Minor: %d  %s\n",
-                       ibd->ibd_major, ibd->ibd_minor, (!bd->bd_contains) ?
+                       MAJOR(bd->bd_dev), MINOR(bd->bd_dev), (!bd->bd_contains) ?
                        "" : (bd->bd_holder == (struct iblock_dev *)ibd) ?
                        "CLAIMED: IBLOCK" : "CLAIMED: OS");
        } else {
-               bl += sprintf(b + bl, "Major: %d Minor: %d\n",
-                       ibd->ibd_major, ibd->ibd_minor);
+               bl += sprintf(b + bl, "Major: 0 Minor: 0\n");
        }
 
        return bl;
@@ -585,103 +568,103 @@ static struct bio *iblock_get_bio(
        struct bio *bio;
 
        bio = bio_alloc_bioset(GFP_NOIO, sg_num, ib_dev->ibd_bio_set);
-       if (!(bio)) {
-               printk(KERN_ERR "Unable to allocate memory for bio\n");
+       if (!bio) {
+               pr_err("Unable to allocate memory for bio\n");
                *ret = PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
                return NULL;
        }
 
-       DEBUG_IBLOCK("Allocated bio: %p task_sg_num: %u using ibd_bio_set:"
-               " %p\n", bio, task->task_sg_num, ib_dev->ibd_bio_set);
-       DEBUG_IBLOCK("Allocated bio: %p task_size: %u\n", bio, task->task_size);
+       pr_debug("Allocated bio: %p task_sg_nents: %u using ibd_bio_set:"
+               " %p\n", bio, task->task_sg_nents, ib_dev->ibd_bio_set);
+       pr_debug("Allocated bio: %p task_size: %u\n", bio, task->task_size);
 
        bio->bi_bdev = ib_dev->ibd_bd;
-       bio->bi_private = (void *) task;
+       bio->bi_private = task;
        bio->bi_destructor = iblock_bio_destructor;
        bio->bi_end_io = &iblock_bio_done;
        bio->bi_sector = lba;
        atomic_inc(&ib_req->ib_bio_cnt);
 
-       DEBUG_IBLOCK("Set bio->bi_sector: %llu\n", bio->bi_sector);
-       DEBUG_IBLOCK("Set ib_req->ib_bio_cnt: %d\n",
+       pr_debug("Set bio->bi_sector: %llu\n", (unsigned long long)bio->bi_sector);
+       pr_debug("Set ib_req->ib_bio_cnt: %d\n",
                        atomic_read(&ib_req->ib_bio_cnt));
        return bio;
 }
 
-static int iblock_map_task_SG(struct se_task *task)
+static int iblock_map_data_SG(struct se_task *task)
 {
        struct se_cmd *cmd = task->task_se_cmd;
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct iblock_dev *ib_dev = task->se_dev->dev_ptr;
        struct iblock_req *ib_req = IBLOCK_REQ(task);
        struct bio *bio = NULL, *hbio = NULL, *tbio = NULL;
        struct scatterlist *sg;
        int ret = 0;
-       u32 i, sg_num = task->task_sg_num;
+       u32 i, sg_num = task->task_sg_nents;
        sector_t block_lba;
        /*
         * Do starting conversion up from non 512-byte blocksize with
         * struct se_task SCSI blocksize into Linux/Block 512 units for BIO.
         */
-       if (DEV_ATTRIB(dev)->block_size == 4096)
+       if (dev->se_sub_dev->se_dev_attrib.block_size == 4096)
                block_lba = (task->task_lba << 3);
-       else if (DEV_ATTRIB(dev)->block_size == 2048)
+       else if (dev->se_sub_dev->se_dev_attrib.block_size == 2048)
                block_lba = (task->task_lba << 2);
-       else if (DEV_ATTRIB(dev)->block_size == 1024)
+       else if (dev->se_sub_dev->se_dev_attrib.block_size == 1024)
                block_lba = (task->task_lba << 1);
-       else if (DEV_ATTRIB(dev)->block_size == 512)
+       else if (dev->se_sub_dev->se_dev_attrib.block_size == 512)
                block_lba = task->task_lba;
        else {
-               printk(KERN_ERR "Unsupported SCSI -> BLOCK LBA conversion:"
-                               " %u\n", DEV_ATTRIB(dev)->block_size);
+               pr_err("Unsupported SCSI -> BLOCK LBA conversion:"
+                               " %u\n", dev->se_sub_dev->se_dev_attrib.block_size);
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
 
        bio = iblock_get_bio(task, ib_req, ib_dev, &ret, block_lba, sg_num);
-       if (!(bio))
+       if (!bio)
                return ret;
 
        ib_req->ib_bio = bio;
        hbio = tbio = bio;
        /*
         * Use fs/bio.c:bio_add_pages() to setup the bio_vec maplist
-        * from TCM struct se_mem -> task->task_sg -> struct scatterlist memory.
+        * from task->task_sg -> struct scatterlist memory.
         */
-       for_each_sg(task->task_sg, sg, task->task_sg_num, i) {
-               DEBUG_IBLOCK("task: %p bio: %p Calling bio_add_page(): page:"
+       for_each_sg(task->task_sg, sg, task->task_sg_nents, i) {
+               pr_debug("task: %p bio: %p Calling bio_add_page(): page:"
                        " %p len: %u offset: %u\n", task, bio, sg_page(sg),
                                sg->length, sg->offset);
 again:
                ret = bio_add_page(bio, sg_page(sg), sg->length, sg->offset);
                if (ret != sg->length) {
 
-                       DEBUG_IBLOCK("*** Set bio->bi_sector: %llu\n",
-                                       bio->bi_sector);
-                       DEBUG_IBLOCK("** task->task_size: %u\n",
+                       pr_debug("*** Set bio->bi_sector: %llu\n",
+                                (unsigned long long)bio->bi_sector);
+                       pr_debug("** task->task_size: %u\n",
                                        task->task_size);
-                       DEBUG_IBLOCK("*** bio->bi_max_vecs: %u\n",
+                       pr_debug("*** bio->bi_max_vecs: %u\n",
                                        bio->bi_max_vecs);
-                       DEBUG_IBLOCK("*** bio->bi_vcnt: %u\n",
+                       pr_debug("*** bio->bi_vcnt: %u\n",
                                        bio->bi_vcnt);
 
                        bio = iblock_get_bio(task, ib_req, ib_dev, &ret,
                                                block_lba, sg_num);
-                       if (!(bio))
+                       if (!bio)
                                goto fail;
 
                        tbio = tbio->bi_next = bio;
-                       DEBUG_IBLOCK("-----------------> Added +1 bio: %p to"
+                       pr_debug("-----------------> Added +1 bio: %p to"
                                " list, Going to again\n", bio);
                        goto again;
                }
                /* Always in 512 byte units for Linux/Block */
                block_lba += sg->length >> IBLOCK_LBA_SHIFT;
                sg_num--;
-               DEBUG_IBLOCK("task: %p bio-add_page() passed!, decremented"
+               pr_debug("task: %p bio-add_page() passed!, decremented"
                        " sg_num to %u\n", task, sg_num);
-               DEBUG_IBLOCK("task: %p bio_add_page() passed!, increased lba"
-                               " to %llu\n", task, block_lba);
-               DEBUG_IBLOCK("task: %p bio_add_page() passed!, bio->bi_vcnt:"
+               pr_debug("task: %p bio_add_page() passed!, increased lba"
+                        " to %llu\n", task, (unsigned long long)block_lba);
+               pr_debug("task: %p bio_add_page() passed!, bio->bi_vcnt:"
                                " %u\n", task, bio->bi_vcnt);
        }
 
@@ -727,11 +710,11 @@ static void iblock_bio_done(struct bio *bio, int err)
        /*
         * Set -EIO if !BIO_UPTODATE and the passed is still err=0
         */
-       if (!(test_bit(BIO_UPTODATE, &bio->bi_flags)) && !(err))
+       if (!test_bit(BIO_UPTODATE, &bio->bi_flags) && !err)
                err = -EIO;
 
        if (err != 0) {
-               printk(KERN_ERR "test_bit(BIO_UPTODATE) failed for bio: %p,"
+               pr_err("test_bit(BIO_UPTODATE) failed for bio: %p,"
                        " err: %d\n", bio, err);
                /*
                 * Bump the ib_bio_err_cnt and release bio.
@@ -742,15 +725,15 @@ static void iblock_bio_done(struct bio *bio, int err)
                /*
                 * Wait to complete the task until the last bio as completed.
                 */
-               if (!(atomic_dec_and_test(&ibr->ib_bio_cnt)))
+               if (!atomic_dec_and_test(&ibr->ib_bio_cnt))
                        return;
 
                ibr->ib_bio = NULL;
                transport_complete_task(task, 0);
                return;
        }
-       DEBUG_IBLOCK("done[%p] bio: %p task_lba: %llu bio_lba: %llu err=%d\n",
-               task, bio, task->task_lba, bio->bi_sector, err);
+       pr_debug("done[%p] bio: %p task_lba: %llu bio_lba: %llu err=%d\n",
+                task, bio, task->task_lba, (unsigned long long)bio->bi_sector, err);
        /*
         * bio_put() will call iblock_bio_destructor() to release the bio back
         * to ibr->ib_bio_set.
@@ -759,7 +742,7 @@ static void iblock_bio_done(struct bio *bio, int err)
        /*
         * Wait to complete the task until the last bio as completed.
         */
-       if (!(atomic_dec_and_test(&ibr->ib_bio_cnt)))
+       if (!atomic_dec_and_test(&ibr->ib_bio_cnt))
                return;
        /*
         * Return GOOD status for task if zero ib_bio_err_cnt exists.
@@ -772,7 +755,7 @@ static struct se_subsystem_api iblock_template = {
        .name                   = "iblock",
        .owner                  = THIS_MODULE,
        .transport_type         = TRANSPORT_PLUGIN_VHBA_PDEV,
-       .map_task_SG            = iblock_map_task_SG,
+       .map_data_SG            = iblock_map_data_SG,
        .attach_hba             = iblock_attach_hba,
        .detach_hba             = iblock_detach_hba,
        .allocate_virtdevice    = iblock_allocate_virtdevice,
index 64c1f4d69f76fd8936bd87fddecabc859b7ca588..a121cd1b657517de83c7931a3e9884b271922786 100644 (file)
@@ -3,9 +3,6 @@
 
 #define IBLOCK_VERSION         "4.0"
 
-#define IBLOCK_HBA_QUEUE_DEPTH 512
-#define IBLOCK_DEVICE_QUEUE_DEPTH      32
-#define IBLOCK_MAX_DEVICE_QUEUE_DEPTH  128
 #define IBLOCK_MAX_CDBS                16
 #define IBLOCK_LBA_SHIFT       9
 
@@ -15,18 +12,12 @@ struct iblock_req {
        atomic_t ib_bio_cnt;
        atomic_t ib_bio_err_cnt;
        struct bio *ib_bio;
-       struct iblock_dev *ib_dev;
 } ____cacheline_aligned;
 
 #define IBDF_HAS_UDEV_PATH             0x01
-#define IBDF_HAS_FORCE                 0x02
 
 struct iblock_dev {
        unsigned char ibd_udev_path[SE_UDEV_PATH_LEN];
-       int     ibd_force;
-       int     ibd_major;
-       int     ibd_minor;
-       u32     ibd_depth;
        u32     ibd_flags;
        struct bio_set  *ibd_bio_set;
        struct block_device *ibd_bd;
index b662db3a320bb96f272cbf9a831c1c8813e89952..1c1b849cd4fb9c24799ce95d399b07012c328a53 100644 (file)
@@ -62,7 +62,7 @@ int core_pr_dump_initiator_port(
        char *buf,
        u32 size)
 {
-       if (!(pr_reg->isid_present_at_reg))
+       if (!pr_reg->isid_present_at_reg)
                return 0;
 
        snprintf(buf, size, ",i,0x%s", &pr_reg->pr_reg_isid[0]);
@@ -95,7 +95,7 @@ static int core_scsi2_reservation_check(struct se_cmd *cmd, u32 *pr_reg_type)
        struct se_session *sess = cmd->se_sess;
        int ret;
 
-       if (!(sess))
+       if (!sess)
                return 0;
 
        spin_lock(&dev->dev_reservation_lock);
@@ -105,13 +105,13 @@ static int core_scsi2_reservation_check(struct se_cmd *cmd, u32 *pr_reg_type)
        }
        if (dev->dev_reserved_node_acl != sess->se_node_acl) {
                spin_unlock(&dev->dev_reservation_lock);
-               return -1;
+               return -EINVAL;
        }
        if (!(dev->dev_flags & DF_SPC2_RESERVATIONS_WITH_ISID)) {
                spin_unlock(&dev->dev_reservation_lock);
                return 0;
        }
-       ret = (dev->dev_res_bin_isid == sess->sess_bin_isid) ? 0 : -1;
+       ret = (dev->dev_res_bin_isid == sess->sess_bin_isid) ? 0 : -EINVAL;
        spin_unlock(&dev->dev_reservation_lock);
 
        return ret;
@@ -123,7 +123,7 @@ static int core_scsi2_reservation_release(struct se_cmd *cmd)
        struct se_session *sess = cmd->se_sess;
        struct se_portal_group *tpg = sess->se_tpg;
 
-       if (!(sess) || !(tpg))
+       if (!sess || !tpg)
                return 0;
 
        spin_lock(&dev->dev_reservation_lock);
@@ -142,9 +142,9 @@ static int core_scsi2_reservation_release(struct se_cmd *cmd)
                dev->dev_res_bin_isid = 0;
                dev->dev_flags &= ~DF_SPC2_RESERVATIONS_WITH_ISID;
        }
-       printk(KERN_INFO "SCSI-2 Released reservation for %s LUN: %u ->"
-               " MAPPED LUN: %u for %s\n", TPG_TFO(tpg)->get_fabric_name(),
-               SE_LUN(cmd)->unpacked_lun, cmd->se_deve->mapped_lun,
+       pr_debug("SCSI-2 Released reservation for %s LUN: %u ->"
+               " MAPPED LUN: %u for %s\n", tpg->se_tpg_tfo->get_fabric_name(),
+               cmd->se_lun->unpacked_lun, cmd->se_deve->mapped_lun,
                sess->se_node_acl->initiatorname);
        spin_unlock(&dev->dev_reservation_lock);
 
@@ -157,9 +157,9 @@ static int core_scsi2_reservation_reserve(struct se_cmd *cmd)
        struct se_session *sess = cmd->se_sess;
        struct se_portal_group *tpg = sess->se_tpg;
 
-       if ((T_TASK(cmd)->t_task_cdb[1] & 0x01) &&
-           (T_TASK(cmd)->t_task_cdb[1] & 0x02)) {
-               printk(KERN_ERR "LongIO and Obselete Bits set, returning"
+       if ((cmd->t_task_cdb[1] & 0x01) &&
+           (cmd->t_task_cdb[1] & 0x02)) {
+               pr_err("LongIO and Obselete Bits set, returning"
                                " ILLEGAL_REQUEST\n");
                return PYX_TRANSPORT_ILLEGAL_REQUEST;
        }
@@ -167,19 +167,19 @@ static int core_scsi2_reservation_reserve(struct se_cmd *cmd)
         * This is currently the case for target_core_mod passthrough struct se_cmd
         * ops
         */
-       if (!(sess) || !(tpg))
+       if (!sess || !tpg)
                return 0;
 
        spin_lock(&dev->dev_reservation_lock);
        if (dev->dev_reserved_node_acl &&
           (dev->dev_reserved_node_acl != sess->se_node_acl)) {
-               printk(KERN_ERR "SCSI-2 RESERVATION CONFLIFT for %s fabric\n",
-                       TPG_TFO(tpg)->get_fabric_name());
-               printk(KERN_ERR "Original reserver LUN: %u %s\n",
-                       SE_LUN(cmd)->unpacked_lun,
+               pr_err("SCSI-2 RESERVATION CONFLIFT for %s fabric\n",
+                       tpg->se_tpg_tfo->get_fabric_name());
+               pr_err("Original reserver LUN: %u %s\n",
+                       cmd->se_lun->unpacked_lun,
                        dev->dev_reserved_node_acl->initiatorname);
-               printk(KERN_ERR "Current attempt - LUN: %u -> MAPPED LUN: %u"
-                       " from %s \n", SE_LUN(cmd)->unpacked_lun,
+               pr_err("Current attempt - LUN: %u -> MAPPED LUN: %u"
+                       " from %s \n", cmd->se_lun->unpacked_lun,
                        cmd->se_deve->mapped_lun,
                        sess->se_node_acl->initiatorname);
                spin_unlock(&dev->dev_reservation_lock);
@@ -192,9 +192,9 @@ static int core_scsi2_reservation_reserve(struct se_cmd *cmd)
                dev->dev_res_bin_isid = sess->sess_bin_isid;
                dev->dev_flags |= DF_SPC2_RESERVATIONS_WITH_ISID;
        }
-       printk(KERN_INFO "SCSI-2 Reserved %s LUN: %u -> MAPPED LUN: %u"
-               " for %s\n", TPG_TFO(tpg)->get_fabric_name(),
-               SE_LUN(cmd)->unpacked_lun, cmd->se_deve->mapped_lun,
+       pr_debug("SCSI-2 Reserved %s LUN: %u -> MAPPED LUN: %u"
+               " for %s\n", tpg->se_tpg_tfo->get_fabric_name(),
+               cmd->se_lun->unpacked_lun, cmd->se_deve->mapped_lun,
                sess->se_node_acl->initiatorname);
        spin_unlock(&dev->dev_reservation_lock);
 
@@ -215,15 +215,15 @@ int core_scsi2_emulate_crh(struct se_cmd *cmd)
        struct se_session *se_sess = cmd->se_sess;
        struct se_subsystem_dev *su_dev = cmd->se_dev->se_sub_dev;
        struct t10_pr_registration *pr_reg;
-       struct t10_reservation_template *pr_tmpl = &su_dev->t10_reservation;
-       unsigned char *cdb = &T_TASK(cmd)->t_task_cdb[0];
-       int crh = (T10_RES(su_dev)->res_type == SPC3_PERSISTENT_RESERVATIONS);
+       struct t10_reservation *pr_tmpl = &su_dev->t10_pr;
+       unsigned char *cdb = &cmd->t_task_cdb[0];
+       int crh = (su_dev->t10_pr.res_type == SPC3_PERSISTENT_RESERVATIONS);
        int conflict = 0;
 
-       if (!(se_sess))
+       if (!se_sess)
                return 0;
 
-       if (!(crh))
+       if (!crh)
                goto after_crh;
 
        pr_reg = core_scsi3_locate_pr_reg(cmd->se_dev, se_sess->se_node_acl,
@@ -280,7 +280,7 @@ int core_scsi2_emulate_crh(struct se_cmd *cmd)
        }
 
        if (conflict) {
-               printk(KERN_ERR "Received legacy SPC-2 RESERVE/RELEASE"
+               pr_err("Received legacy SPC-2 RESERVE/RELEASE"
                        " while active SPC-3 registrations exist,"
                        " returning RESERVATION_CONFLICT\n");
                return PYX_TRANSPORT_RESERVATION_CONFLICT;
@@ -307,7 +307,7 @@ static int core_scsi3_pr_seq_non_holder(
        u32 pr_reg_type)
 {
        struct se_dev_entry *se_deve;
-       struct se_session *se_sess = SE_SESS(cmd);
+       struct se_session *se_sess = cmd->se_sess;
        int other_cdb = 0, ignore_reg;
        int registered_nexus = 0, ret = 1; /* Conflict by default */
        int all_reg = 0, reg_only = 0; /* ALL_REG, REG_ONLY */
@@ -362,7 +362,7 @@ static int core_scsi3_pr_seq_non_holder(
                        registered_nexus = 1;
                break;
        default:
-               return -1;
+               return -EINVAL;
        }
        /*
         * Referenced from spc4r17 table 45 for *NON* PR holder access
@@ -412,9 +412,9 @@ static int core_scsi3_pr_seq_non_holder(
                        ret = (registered_nexus) ? 0 : 1;
                        break;
                default:
-                       printk(KERN_ERR "Unknown PERSISTENT_RESERVE_OUT service"
+                       pr_err("Unknown PERSISTENT_RESERVE_OUT service"
                                " action: 0x%02x\n", cdb[1] & 0x1f);
-                       return -1;
+                       return -EINVAL;
                }
                break;
        case RELEASE:
@@ -459,9 +459,9 @@ static int core_scsi3_pr_seq_non_holder(
                        ret = 0; /* Allowed */
                        break;
                default:
-                       printk(KERN_ERR "Unknown MI Service Action: 0x%02x\n",
+                       pr_err("Unknown MI Service Action: 0x%02x\n",
                                (cdb[1] & 0x1f));
-                       return -1;
+                       return -EINVAL;
                }
                break;
        case ACCESS_CONTROL_IN:
@@ -481,9 +481,9 @@ static int core_scsi3_pr_seq_non_holder(
         * Case where the CDB is explicitly allowed in the above switch
         * statement.
         */
-       if (!(ret) && !(other_cdb)) {
+       if (!ret && !other_cdb) {
 #if 0
-               printk(KERN_INFO "Allowing explict CDB: 0x%02x for %s"
+               pr_debug("Allowing explict CDB: 0x%02x for %s"
                        " reservation holder\n", cdb[0],
                        core_scsi3_pr_dump_type(pr_reg_type));
 #endif
@@ -498,7 +498,7 @@ static int core_scsi3_pr_seq_non_holder(
                        /*
                         * Conflict for write exclusive
                         */
-                       printk(KERN_INFO "%s Conflict for unregistered nexus"
+                       pr_debug("%s Conflict for unregistered nexus"
                                " %s CDB: 0x%02x to %s reservation\n",
                                transport_dump_cmd_direction(cmd),
                                se_sess->se_node_acl->initiatorname, cdb[0],
@@ -515,8 +515,8 @@ static int core_scsi3_pr_seq_non_holder(
                         * nexuses to issue CDBs.
                         */
 #if 0
-                       if (!(registered_nexus)) {
-                               printk(KERN_INFO "Allowing implict CDB: 0x%02x"
+                       if (!registered_nexus) {
+                               pr_debug("Allowing implict CDB: 0x%02x"
                                        " for %s reservation on unregistered"
                                        " nexus\n", cdb[0],
                                        core_scsi3_pr_dump_type(pr_reg_type));
@@ -531,14 +531,14 @@ static int core_scsi3_pr_seq_non_holder(
                         * allow commands from registered nexuses.
                         */
 #if 0
-                       printk(KERN_INFO "Allowing implict CDB: 0x%02x for %s"
+                       pr_debug("Allowing implict CDB: 0x%02x for %s"
                                " reservation\n", cdb[0],
                                core_scsi3_pr_dump_type(pr_reg_type));
 #endif
                        return 0;
                }
        }
-       printk(KERN_INFO "%s Conflict for %sregistered nexus %s CDB: 0x%2x"
+       pr_debug("%s Conflict for %sregistered nexus %s CDB: 0x%2x"
                " for %s reservation\n", transport_dump_cmd_direction(cmd),
                (registered_nexus) ? "" : "un",
                se_sess->se_node_acl->initiatorname, cdb[0],
@@ -549,7 +549,7 @@ static int core_scsi3_pr_seq_non_holder(
 
 static u32 core_scsi3_pr_generation(struct se_device *dev)
 {
-       struct se_subsystem_dev *su_dev = SU_DEV(dev);
+       struct se_subsystem_dev *su_dev = dev->se_sub_dev;
        u32 prg;
        /*
         * PRGeneration field shall contain the value of a 32-bit wrapping
@@ -561,7 +561,7 @@ static u32 core_scsi3_pr_generation(struct se_device *dev)
         * See spc4r17 section 6.3.12 READ_KEYS service action
         */
        spin_lock(&dev->dev_reservation_lock);
-       prg = T10_RES(su_dev)->pr_generation++;
+       prg = su_dev->t10_pr.pr_generation++;
        spin_unlock(&dev->dev_reservation_lock);
 
        return prg;
@@ -575,7 +575,7 @@ static int core_scsi3_pr_reservation_check(
        struct se_session *sess = cmd->se_sess;
        int ret;
 
-       if (!(sess))
+       if (!sess)
                return 0;
        /*
         * A legacy SPC-2 reservation is being held.
@@ -584,7 +584,7 @@ static int core_scsi3_pr_reservation_check(
                return core_scsi2_reservation_check(cmd, pr_reg_type);
 
        spin_lock(&dev->dev_reservation_lock);
-       if (!(dev->dev_pr_res_holder)) {
+       if (!dev->dev_pr_res_holder) {
                spin_unlock(&dev->dev_reservation_lock);
                return 0;
        }
@@ -592,14 +592,14 @@ static int core_scsi3_pr_reservation_check(
        cmd->pr_res_key = dev->dev_pr_res_holder->pr_res_key;
        if (dev->dev_pr_res_holder->pr_reg_nacl != sess->se_node_acl) {
                spin_unlock(&dev->dev_reservation_lock);
-               return -1;
+               return -EINVAL;
        }
-       if (!(dev->dev_pr_res_holder->isid_present_at_reg)) {
+       if (!dev->dev_pr_res_holder->isid_present_at_reg) {
                spin_unlock(&dev->dev_reservation_lock);
                return 0;
        }
        ret = (dev->dev_pr_res_holder->pr_reg_bin_isid ==
-              sess->sess_bin_isid) ? 0 : -1;
+              sess->sess_bin_isid) ? 0 : -EINVAL;
        /*
         * Use bit in *pr_reg_type to notify ISID mismatch in
         * core_scsi3_pr_seq_non_holder().
@@ -620,19 +620,19 @@ static struct t10_pr_registration *__core_scsi3_do_alloc_registration(
        int all_tg_pt,
        int aptpl)
 {
-       struct se_subsystem_dev *su_dev = SU_DEV(dev);
+       struct se_subsystem_dev *su_dev = dev->se_sub_dev;
        struct t10_pr_registration *pr_reg;
 
        pr_reg = kmem_cache_zalloc(t10_pr_reg_cache, GFP_ATOMIC);
-       if (!(pr_reg)) {
-               printk(KERN_ERR "Unable to allocate struct t10_pr_registration\n");
+       if (!pr_reg) {
+               pr_err("Unable to allocate struct t10_pr_registration\n");
                return NULL;
        }
 
-       pr_reg->pr_aptpl_buf = kzalloc(T10_RES(su_dev)->pr_aptpl_buf_len,
+       pr_reg->pr_aptpl_buf = kzalloc(su_dev->t10_pr.pr_aptpl_buf_len,
                                        GFP_ATOMIC);
-       if (!(pr_reg->pr_aptpl_buf)) {
-               printk(KERN_ERR "Unable to allocate pr_reg->pr_aptpl_buf\n");
+       if (!pr_reg->pr_aptpl_buf) {
+               pr_err("Unable to allocate pr_reg->pr_aptpl_buf\n");
                kmem_cache_free(t10_pr_reg_cache, pr_reg);
                return NULL;
        }
@@ -692,12 +692,12 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
         */
        pr_reg = __core_scsi3_do_alloc_registration(dev, nacl, deve, isid,
                        sa_res_key, all_tg_pt, aptpl);
-       if (!(pr_reg))
+       if (!pr_reg)
                return NULL;
        /*
         * Return pointer to pr_reg for ALL_TG_PT=0
         */
-       if (!(all_tg_pt))
+       if (!all_tg_pt)
                return pr_reg;
        /*
         * Create list of matching SCSI Initiator Port registrations
@@ -717,7 +717,7 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
                         * that have not been make explict via a ConfigFS
                         * MappedLUN group for the SCSI Initiator Node ACL.
                         */
-                       if (!(deve_tmp->se_lun_acl))
+                       if (!deve_tmp->se_lun_acl)
                                continue;
 
                        nacl_tmp = deve_tmp->se_lun_acl->se_lun_nacl;
@@ -751,7 +751,7 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
                         */
                        ret = core_scsi3_lunacl_depend_item(deve_tmp);
                        if (ret < 0) {
-                               printk(KERN_ERR "core_scsi3_lunacl_depend"
+                               pr_err("core_scsi3_lunacl_depend"
                                                "_item() failed\n");
                                atomic_dec(&port->sep_tg_pt_ref_cnt);
                                smp_mb__after_atomic_dec();
@@ -769,7 +769,7 @@ static struct t10_pr_registration *__core_scsi3_alloc_registration(
                        pr_reg_atp = __core_scsi3_do_alloc_registration(dev,
                                                nacl_tmp, deve_tmp, NULL,
                                                sa_res_key, all_tg_pt, aptpl);
-                       if (!(pr_reg_atp)) {
+                       if (!pr_reg_atp) {
                                atomic_dec(&port->sep_tg_pt_ref_cnt);
                                smp_mb__after_atomic_dec();
                                atomic_dec(&deve_tmp->pr_ref_count);
@@ -803,7 +803,7 @@ out:
 }
 
 int core_scsi3_alloc_aptpl_registration(
-       struct t10_reservation_template *pr_tmpl,
+       struct t10_reservation *pr_tmpl,
        u64 sa_res_key,
        unsigned char *i_port,
        unsigned char *isid,
@@ -817,15 +817,15 @@ int core_scsi3_alloc_aptpl_registration(
 {
        struct t10_pr_registration *pr_reg;
 
-       if (!(i_port) || !(t_port) || !(sa_res_key)) {
-               printk(KERN_ERR "Illegal parameters for APTPL registration\n");
-               return -1;
+       if (!i_port || !t_port || !sa_res_key) {
+               pr_err("Illegal parameters for APTPL registration\n");
+               return -EINVAL;
        }
 
        pr_reg = kmem_cache_zalloc(t10_pr_reg_cache, GFP_KERNEL);
-       if (!(pr_reg)) {
-               printk(KERN_ERR "Unable to allocate struct t10_pr_registration\n");
-               return -1;
+       if (!pr_reg) {
+               pr_err("Unable to allocate struct t10_pr_registration\n");
+               return -ENOMEM;
        }
        pr_reg->pr_aptpl_buf = kzalloc(pr_tmpl->pr_aptpl_buf_len, GFP_KERNEL);
 
@@ -869,7 +869,7 @@ int core_scsi3_alloc_aptpl_registration(
        pr_reg->pr_res_holder = res_holder;
 
        list_add_tail(&pr_reg->pr_reg_aptpl_list, &pr_tmpl->aptpl_reg_list);
-       printk(KERN_INFO "SPC-3 PR APTPL Successfully added registration%s from"
+       pr_debug("SPC-3 PR APTPL Successfully added registration%s from"
                        " metadata\n", (res_holder) ? "+reservation" : "");
        return 0;
 }
@@ -891,13 +891,13 @@ static void core_scsi3_aptpl_reserve(
        dev->dev_pr_res_holder = pr_reg;
        spin_unlock(&dev->dev_reservation_lock);
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: APTPL RESERVE created"
+       pr_debug("SPC-3 PR [%s] Service Action: APTPL RESERVE created"
                " new reservation holder TYPE: %s ALL_TG_PT: %d\n",
-               TPG_TFO(tpg)->get_fabric_name(),
+               tpg->se_tpg_tfo->get_fabric_name(),
                core_scsi3_pr_dump_type(pr_reg->pr_res_type),
                (pr_reg->pr_reg_all_tg_pt) ? 1 : 0);
-       printk(KERN_INFO "SPC-3 PR [%s] RESERVE Node: %s%s\n",
-               TPG_TFO(tpg)->get_fabric_name(), node_acl->initiatorname,
+       pr_debug("SPC-3 PR [%s] RESERVE Node: %s%s\n",
+               tpg->se_tpg_tfo->get_fabric_name(), node_acl->initiatorname,
                (prf_isid) ? &i_buf[0] : "");
 }
 
@@ -913,7 +913,7 @@ static int __core_scsi3_check_aptpl_registration(
        struct se_dev_entry *deve)
 {
        struct t10_pr_registration *pr_reg, *pr_reg_tmp;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        unsigned char i_port[PR_APTPL_MAX_IPORT_LEN];
        unsigned char t_port[PR_APTPL_MAX_TPORT_LEN];
        u16 tpgt;
@@ -925,8 +925,8 @@ static int __core_scsi3_check_aptpl_registration(
         */
        snprintf(i_port, PR_APTPL_MAX_IPORT_LEN, "%s", nacl->initiatorname);
        snprintf(t_port, PR_APTPL_MAX_TPORT_LEN, "%s",
-                       TPG_TFO(tpg)->tpg_get_wwn(tpg));
-       tpgt = TPG_TFO(tpg)->tpg_get_tag(tpg);
+                       tpg->se_tpg_tfo->tpg_get_wwn(tpg));
+       tpgt = tpg->se_tpg_tfo->tpg_get_tag(tpg);
        /*
         * Look for the matching registrations+reservation from those
         * created from APTPL metadata.  Note that multiple registrations
@@ -936,7 +936,7 @@ static int __core_scsi3_check_aptpl_registration(
        spin_lock(&pr_tmpl->aptpl_reg_lock);
        list_for_each_entry_safe(pr_reg, pr_reg_tmp, &pr_tmpl->aptpl_reg_list,
                                pr_reg_aptpl_list) {
-               if (!(strcmp(pr_reg->pr_iport, i_port)) &&
+               if (!strcmp(pr_reg->pr_iport, i_port) &&
                     (pr_reg->pr_res_mapped_lun == deve->mapped_lun) &&
                    !(strcmp(pr_reg->pr_tport, t_port)) &&
                     (pr_reg->pr_reg_tpgt == tpgt) &&
@@ -980,11 +980,11 @@ int core_scsi3_check_aptpl_registration(
        struct se_lun *lun,
        struct se_lun_acl *lun_acl)
 {
-       struct se_subsystem_dev *su_dev = SU_DEV(dev);
+       struct se_subsystem_dev *su_dev = dev->se_sub_dev;
        struct se_node_acl *nacl = lun_acl->se_lun_nacl;
        struct se_dev_entry *deve = &nacl->device_list[lun_acl->mapped_lun];
 
-       if (T10_RES(su_dev)->res_type != SPC3_PERSISTENT_RESERVATIONS)
+       if (su_dev->t10_pr.res_type != SPC3_PERSISTENT_RESERVATIONS)
                return 0;
 
        return __core_scsi3_check_aptpl_registration(dev, tpg, lun,
@@ -1006,19 +1006,19 @@ static void __core_scsi3_dump_registration(
        prf_isid = core_pr_dump_initiator_port(pr_reg, &i_buf[0],
                                PR_REG_ISID_ID_LEN);
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: REGISTER%s Initiator"
+       pr_debug("SPC-3 PR [%s] Service Action: REGISTER%s Initiator"
                " Node: %s%s\n", tfo->get_fabric_name(), (register_type == 2) ?
                "_AND_MOVE" : (register_type == 1) ?
                "_AND_IGNORE_EXISTING_KEY" : "", nacl->initiatorname,
                (prf_isid) ? i_buf : "");
-       printk(KERN_INFO "SPC-3 PR [%s] registration on Target Port: %s,0x%04x\n",
+       pr_debug("SPC-3 PR [%s] registration on Target Port: %s,0x%04x\n",
                 tfo->get_fabric_name(), tfo->tpg_get_wwn(se_tpg),
                tfo->tpg_get_tag(se_tpg));
-       printk(KERN_INFO "SPC-3 PR [%s] for %s TCM Subsystem %s Object Target"
+       pr_debug("SPC-3 PR [%s] for %s TCM Subsystem %s Object Target"
                " Port(s)\n",  tfo->get_fabric_name(),
                (pr_reg->pr_reg_all_tg_pt) ? "ALL" : "SINGLE",
-               TRANSPORT(dev)->name);
-       printk(KERN_INFO "SPC-3 PR [%s] SA Res Key: 0x%016Lx PRgeneration:"
+               dev->transport->name);
+       pr_debug("SPC-3 PR [%s] SA Res Key: 0x%016Lx PRgeneration:"
                " 0x%08x  APTPL: %d\n", tfo->get_fabric_name(),
                pr_reg->pr_res_key, pr_reg->pr_res_generation,
                pr_reg->pr_reg_aptpl);
@@ -1035,10 +1035,10 @@ static void __core_scsi3_add_registration(
        int register_type,
        int register_move)
 {
-       struct se_subsystem_dev *su_dev = SU_DEV(dev);
+       struct se_subsystem_dev *su_dev = dev->se_sub_dev;
        struct target_core_fabric_ops *tfo = nacl->se_tpg->se_tpg_tfo;
        struct t10_pr_registration *pr_reg_tmp, *pr_reg_tmp_safe;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
 
        /*
         * Increment PRgeneration counter for struct se_device upon a successful
@@ -1050,7 +1050,7 @@ static void __core_scsi3_add_registration(
         * for the REGISTER.
         */
        pr_reg->pr_res_generation = (register_move) ?
-                       T10_RES(su_dev)->pr_generation++ :
+                       su_dev->t10_pr.pr_generation++ :
                        core_scsi3_pr_generation(dev);
 
        spin_lock(&pr_tmpl->registration_lock);
@@ -1062,7 +1062,7 @@ static void __core_scsi3_add_registration(
        /*
         * Skip extra processing for ALL_TG_PT=0 or REGISTER_AND_MOVE.
         */
-       if (!(pr_reg->pr_reg_all_tg_pt) || (register_move))
+       if (!pr_reg->pr_reg_all_tg_pt || register_move)
                return;
        /*
         * Walk pr_reg->pr_reg_atp_list and add registrations for ALL_TG_PT=1
@@ -1106,8 +1106,8 @@ static int core_scsi3_alloc_registration(
 
        pr_reg = __core_scsi3_alloc_registration(dev, nacl, deve, isid,
                        sa_res_key, all_tg_pt, aptpl);
-       if (!(pr_reg))
-               return -1;
+       if (!pr_reg)
+               return -EPERM;
 
        __core_scsi3_add_registration(dev, nacl, pr_reg,
                        register_type, register_move);
@@ -1119,7 +1119,7 @@ static struct t10_pr_registration *__core_scsi3_locate_pr_reg(
        struct se_node_acl *nacl,
        unsigned char *isid)
 {
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        struct t10_pr_registration *pr_reg, *pr_reg_tmp;
        struct se_portal_group *tpg;
 
@@ -1137,14 +1137,14 @@ static struct t10_pr_registration *__core_scsi3_locate_pr_reg(
                 * If this registration does NOT contain a fabric provided
                 * ISID, then we have found a match.
                 */
-               if (!(pr_reg->isid_present_at_reg)) {
+               if (!pr_reg->isid_present_at_reg) {
                        /*
                         * Determine if this SCSI device server requires that
                         * SCSI Intiatior TransportID w/ ISIDs is enforced
                         * for fabric modules (iSCSI) requiring them.
                         */
-                       if (TPG_TFO(tpg)->sess_get_initiator_sid != NULL) {
-                               if (DEV_ATTRIB(dev)->enforce_pr_isids)
+                       if (tpg->se_tpg_tfo->sess_get_initiator_sid != NULL) {
+                               if (dev->se_sub_dev->se_dev_attrib.enforce_pr_isids)
                                        continue;
                        }
                        atomic_inc(&pr_reg->pr_res_holders);
@@ -1157,7 +1157,7 @@ static struct t10_pr_registration *__core_scsi3_locate_pr_reg(
                 * SCSI Initiator Port TransportIDs, then we expect a valid
                 * matching ISID to be provided by the local SCSI Initiator Port.
                 */
-               if (!(isid))
+               if (!isid)
                        continue;
                if (strcmp(isid, pr_reg->pr_reg_isid))
                        continue;
@@ -1180,9 +1180,9 @@ static struct t10_pr_registration *core_scsi3_locate_pr_reg(
        struct se_portal_group *tpg = nacl->se_tpg;
        unsigned char buf[PR_REG_ISID_LEN], *isid_ptr = NULL;
 
-       if (TPG_TFO(tpg)->sess_get_initiator_sid != NULL) {
+       if (tpg->se_tpg_tfo->sess_get_initiator_sid != NULL) {
                memset(&buf[0], 0, PR_REG_ISID_LEN);
-               TPG_TFO(tpg)->sess_get_initiator_sid(sess, &buf[0],
+               tpg->se_tpg_tfo->sess_get_initiator_sid(sess, &buf[0],
                                        PR_REG_ISID_LEN);
                isid_ptr = &buf[0];
        }
@@ -1206,7 +1206,7 @@ static int core_scsi3_check_implict_release(
 
        spin_lock(&dev->dev_reservation_lock);
        pr_res_holder = dev->dev_pr_res_holder;
-       if (!(pr_res_holder)) {
+       if (!pr_res_holder) {
                spin_unlock(&dev->dev_reservation_lock);
                return ret;
        }
@@ -1236,11 +1236,11 @@ static int core_scsi3_check_implict_release(
                  (!strcmp(pr_res_holder->pr_reg_nacl->initiatorname,
                          pr_reg->pr_reg_nacl->initiatorname)) &&
                  (pr_res_holder->pr_res_key == pr_reg->pr_res_key)) {
-               printk(KERN_ERR "SPC-3 PR: Unable to perform ALL_TG_PT=1"
+               pr_err("SPC-3 PR: Unable to perform ALL_TG_PT=1"
                        " UNREGISTER while existing reservation with matching"
                        " key 0x%016Lx is present from another SCSI Initiator"
                        " Port\n", pr_reg->pr_res_key);
-               ret = -1;
+               ret = -EPERM;
        }
        spin_unlock(&dev->dev_reservation_lock);
 
@@ -1248,7 +1248,7 @@ static int core_scsi3_check_implict_release(
 }
 
 /*
- * Called with struct t10_reservation_template->registration_lock held.
+ * Called with struct t10_reservation->registration_lock held.
  */
 static void __core_scsi3_free_registration(
        struct se_device *dev,
@@ -1258,7 +1258,7 @@ static void __core_scsi3_free_registration(
 {
        struct target_core_fabric_ops *tfo =
                        pr_reg->pr_reg_nacl->se_tpg->se_tpg_tfo;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        char i_buf[PR_REG_ISID_ID_LEN];
        int prf_isid;
 
@@ -1283,25 +1283,25 @@ static void __core_scsi3_free_registration(
         */
        while (atomic_read(&pr_reg->pr_res_holders) != 0) {
                spin_unlock(&pr_tmpl->registration_lock);
-               printk("SPC-3 PR [%s] waiting for pr_res_holders\n",
+               pr_debug("SPC-3 PR [%s] waiting for pr_res_holders\n",
                                tfo->get_fabric_name());
                cpu_relax();
                spin_lock(&pr_tmpl->registration_lock);
        }
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: UNREGISTER Initiator"
+       pr_debug("SPC-3 PR [%s] Service Action: UNREGISTER Initiator"
                " Node: %s%s\n", tfo->get_fabric_name(),
                pr_reg->pr_reg_nacl->initiatorname,
                (prf_isid) ? &i_buf[0] : "");
-       printk(KERN_INFO "SPC-3 PR [%s] for %s TCM Subsystem %s Object Target"
+       pr_debug("SPC-3 PR [%s] for %s TCM Subsystem %s Object Target"
                " Port(s)\n", tfo->get_fabric_name(),
                (pr_reg->pr_reg_all_tg_pt) ? "ALL" : "SINGLE",
-               TRANSPORT(dev)->name);
-       printk(KERN_INFO "SPC-3 PR [%s] SA Res Key: 0x%016Lx PRgeneration:"
+               dev->transport->name);
+       pr_debug("SPC-3 PR [%s] SA Res Key: 0x%016Lx PRgeneration:"
                " 0x%08x\n", tfo->get_fabric_name(), pr_reg->pr_res_key,
                pr_reg->pr_res_generation);
 
-       if (!(preempt_and_abort_list)) {
+       if (!preempt_and_abort_list) {
                pr_reg->pr_reg_deve = NULL;
                pr_reg->pr_reg_nacl = NULL;
                kfree(pr_reg->pr_aptpl_buf);
@@ -1319,7 +1319,7 @@ void core_scsi3_free_pr_reg_from_nacl(
        struct se_device *dev,
        struct se_node_acl *nacl)
 {
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        struct t10_pr_registration *pr_reg, *pr_reg_tmp, *pr_res_holder;
        /*
         * If the passed se_node_acl matches the reservation holder,
@@ -1349,7 +1349,7 @@ void core_scsi3_free_pr_reg_from_nacl(
 void core_scsi3_free_all_registrations(
        struct se_device *dev)
 {
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        struct t10_pr_registration *pr_reg, *pr_reg_tmp, *pr_res_holder;
 
        spin_lock(&dev->dev_reservation_lock);
@@ -1381,13 +1381,13 @@ void core_scsi3_free_all_registrations(
 
 static int core_scsi3_tpg_depend_item(struct se_portal_group *tpg)
 {
-       return configfs_depend_item(TPG_TFO(tpg)->tf_subsys,
+       return configfs_depend_item(tpg->se_tpg_tfo->tf_subsys,
                        &tpg->tpg_group.cg_item);
 }
 
 static void core_scsi3_tpg_undepend_item(struct se_portal_group *tpg)
 {
-       configfs_undepend_item(TPG_TFO(tpg)->tf_subsys,
+       configfs_undepend_item(tpg->se_tpg_tfo->tf_subsys,
                        &tpg->tpg_group.cg_item);
 
        atomic_dec(&tpg->tpg_pr_ref_count);
@@ -1401,7 +1401,7 @@ static int core_scsi3_nodeacl_depend_item(struct se_node_acl *nacl)
        if (nacl->dynamic_node_acl)
                return 0;
 
-       return configfs_depend_item(TPG_TFO(tpg)->tf_subsys,
+       return configfs_depend_item(tpg->se_tpg_tfo->tf_subsys,
                        &nacl->acl_group.cg_item);
 }
 
@@ -1415,7 +1415,7 @@ static void core_scsi3_nodeacl_undepend_item(struct se_node_acl *nacl)
                return;
        }
 
-       configfs_undepend_item(TPG_TFO(tpg)->tf_subsys,
+       configfs_undepend_item(tpg->se_tpg_tfo->tf_subsys,
                        &nacl->acl_group.cg_item);
 
        atomic_dec(&nacl->acl_pr_ref_count);
@@ -1430,13 +1430,13 @@ static int core_scsi3_lunacl_depend_item(struct se_dev_entry *se_deve)
        /*
         * For nacl->dynamic_node_acl=1
         */
-       if (!(lun_acl))
+       if (!lun_acl)
                return 0;
 
        nacl = lun_acl->se_lun_nacl;
        tpg = nacl->se_tpg;
 
-       return configfs_depend_item(TPG_TFO(tpg)->tf_subsys,
+       return configfs_depend_item(tpg->se_tpg_tfo->tf_subsys,
                        &lun_acl->se_lun_group.cg_item);
 }
 
@@ -1448,7 +1448,7 @@ static void core_scsi3_lunacl_undepend_item(struct se_dev_entry *se_deve)
        /*
         * For nacl->dynamic_node_acl=1
         */
-       if (!(lun_acl)) {
+       if (!lun_acl) {
                atomic_dec(&se_deve->pr_ref_count);
                smp_mb__after_atomic_dec();
                return;
@@ -1456,7 +1456,7 @@ static void core_scsi3_lunacl_undepend_item(struct se_dev_entry *se_deve)
        nacl = lun_acl->se_lun_nacl;
        tpg = nacl->se_tpg;
 
-       configfs_undepend_item(TPG_TFO(tpg)->tf_subsys,
+       configfs_undepend_item(tpg->se_tpg_tfo->tf_subsys,
                        &lun_acl->se_lun_group.cg_item);
 
        atomic_dec(&se_deve->pr_ref_count);
@@ -1471,10 +1471,10 @@ static int core_scsi3_decode_spec_i_port(
        int all_tg_pt,
        int aptpl)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct se_port *tmp_port;
        struct se_portal_group *dest_tpg = NULL, *tmp_tpg;
-       struct se_session *se_sess = SE_SESS(cmd);
+       struct se_session *se_sess = cmd->se_sess;
        struct se_node_acl *dest_node_acl = NULL;
        struct se_dev_entry *dest_se_deve = NULL, *local_se_deve;
        struct t10_pr_registration *dest_pr_reg, *local_pr_reg, *pr_reg_e;
@@ -1482,7 +1482,7 @@ static int core_scsi3_decode_spec_i_port(
        struct list_head tid_dest_list;
        struct pr_transport_id_holder *tidh_new, *tidh, *tidh_tmp;
        struct target_core_fabric_ops *tmp_tf_ops;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       unsigned char *buf;
        unsigned char *ptr, *i_str = NULL, proto_ident, tmp_proto_ident;
        char *iport_ptr = NULL, dest_iport[64], i_buf[PR_REG_ISID_ID_LEN];
        u32 tpdl, tid_len = 0;
@@ -1500,8 +1500,8 @@ static int core_scsi3_decode_spec_i_port(
         * processing in the loop of tid_dest_list below.
         */
        tidh_new = kzalloc(sizeof(struct pr_transport_id_holder), GFP_KERNEL);
-       if (!(tidh_new)) {
-               printk(KERN_ERR "Unable to allocate tidh_new\n");
+       if (!tidh_new) {
+               pr_err("Unable to allocate tidh_new\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
        INIT_LIST_HEAD(&tidh_new->dest_list);
@@ -1509,10 +1509,10 @@ static int core_scsi3_decode_spec_i_port(
        tidh_new->dest_node_acl = se_sess->se_node_acl;
        tidh_new->dest_se_deve = local_se_deve;
 
-       local_pr_reg = __core_scsi3_alloc_registration(SE_DEV(cmd),
+       local_pr_reg = __core_scsi3_alloc_registration(cmd->se_dev,
                                se_sess->se_node_acl, local_se_deve, l_isid,
                                sa_res_key, all_tg_pt, aptpl);
-       if (!(local_pr_reg)) {
+       if (!local_pr_reg) {
                kfree(tidh_new);
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
@@ -1524,6 +1524,8 @@ static int core_scsi3_decode_spec_i_port(
         */
        tidh_new->dest_local_nexus = 1;
        list_add_tail(&tidh_new->dest_list, &tid_dest_list);
+
+       buf = transport_kmap_first_data_page(cmd);
        /*
         * For a PERSISTENT RESERVE OUT specify initiator ports payload,
         * first extract TransportID Parameter Data Length, and make sure
@@ -1535,7 +1537,7 @@ static int core_scsi3_decode_spec_i_port(
        tpdl |= buf[27] & 0xff;
 
        if ((tpdl + 28) != cmd->data_length) {
-               printk(KERN_ERR "SPC-3 PR: Illegal tpdl: %u + 28 byte header"
+               pr_err("SPC-3 PR: Illegal tpdl: %u + 28 byte header"
                        " does not equal CDB data_length: %u\n", tpdl,
                        cmd->data_length);
                ret = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
@@ -1555,13 +1557,13 @@ static int core_scsi3_decode_spec_i_port(
                spin_lock(&dev->se_port_lock);
                list_for_each_entry(tmp_port, &dev->dev_sep_list, sep_list) {
                        tmp_tpg = tmp_port->sep_tpg;
-                       if (!(tmp_tpg))
+                       if (!tmp_tpg)
                                continue;
-                       tmp_tf_ops = TPG_TFO(tmp_tpg);
-                       if (!(tmp_tf_ops))
+                       tmp_tf_ops = tmp_tpg->se_tpg_tfo;
+                       if (!tmp_tf_ops)
                                continue;
-                       if (!(tmp_tf_ops->get_fabric_proto_ident) ||
-                           !(tmp_tf_ops->tpg_parse_pr_out_transport_id))
+                       if (!tmp_tf_ops->get_fabric_proto_ident ||
+                           !tmp_tf_ops->tpg_parse_pr_out_transport_id)
                                continue;
                        /*
                         * Look for the matching proto_ident provided by
@@ -1575,7 +1577,7 @@ static int core_scsi3_decode_spec_i_port(
                        i_str = tmp_tf_ops->tpg_parse_pr_out_transport_id(
                                        tmp_tpg, (const char *)ptr, &tid_len,
                                        &iport_ptr);
-                       if (!(i_str))
+                       if (!i_str)
                                continue;
 
                        atomic_inc(&tmp_tpg->tpg_pr_ref_count);
@@ -1584,7 +1586,7 @@ static int core_scsi3_decode_spec_i_port(
 
                        ret = core_scsi3_tpg_depend_item(tmp_tpg);
                        if (ret != 0) {
-                               printk(KERN_ERR " core_scsi3_tpg_depend_item()"
+                               pr_err(" core_scsi3_tpg_depend_item()"
                                        " for tmp_tpg\n");
                                atomic_dec(&tmp_tpg->tpg_pr_ref_count);
                                smp_mb__after_atomic_dec();
@@ -1605,7 +1607,7 @@ static int core_scsi3_decode_spec_i_port(
                        }
                        spin_unlock_bh(&tmp_tpg->acl_node_lock);
 
-                       if (!(dest_node_acl)) {
+                       if (!dest_node_acl) {
                                core_scsi3_tpg_undepend_item(tmp_tpg);
                                spin_lock(&dev->se_port_lock);
                                continue;
@@ -1613,7 +1615,7 @@ static int core_scsi3_decode_spec_i_port(
 
                        ret = core_scsi3_nodeacl_depend_item(dest_node_acl);
                        if (ret != 0) {
-                               printk(KERN_ERR "configfs_depend_item() failed"
+                               pr_err("configfs_depend_item() failed"
                                        " for dest_node_acl->acl_group\n");
                                atomic_dec(&dest_node_acl->acl_pr_ref_count);
                                smp_mb__after_atomic_dec();
@@ -1623,9 +1625,9 @@ static int core_scsi3_decode_spec_i_port(
                        }
 
                        dest_tpg = tmp_tpg;
-                       printk(KERN_INFO "SPC-3 PR SPEC_I_PT: Located %s Node:"
+                       pr_debug("SPC-3 PR SPEC_I_PT: Located %s Node:"
                                " %s Port RTPI: %hu\n",
-                               TPG_TFO(dest_tpg)->get_fabric_name(),
+                               dest_tpg->se_tpg_tfo->get_fabric_name(),
                                dest_node_acl->initiatorname, dest_rtpi);
 
                        spin_lock(&dev->se_port_lock);
@@ -1633,20 +1635,20 @@ static int core_scsi3_decode_spec_i_port(
                }
                spin_unlock(&dev->se_port_lock);
 
-               if (!(dest_tpg)) {
-                       printk(KERN_ERR "SPC-3 PR SPEC_I_PT: Unable to locate"
+               if (!dest_tpg) {
+                       pr_err("SPC-3 PR SPEC_I_PT: Unable to locate"
                                        " dest_tpg\n");
                        ret = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
                        goto out;
                }
 #if 0
-               printk("SPC-3 PR SPEC_I_PT: Got %s data_length: %u tpdl: %u"
+               pr_debug("SPC-3 PR SPEC_I_PT: Got %s data_length: %u tpdl: %u"
                        " tid_len: %d for %s + %s\n",
-                       TPG_TFO(dest_tpg)->get_fabric_name(), cmd->data_length,
+                       dest_tpg->se_tpg_tfo->get_fabric_name(), cmd->data_length,
                        tpdl, tid_len, i_str, iport_ptr);
 #endif
                if (tid_len > tpdl) {
-                       printk(KERN_ERR "SPC-3 PR SPEC_I_PT: Illegal tid_len:"
+                       pr_err("SPC-3 PR SPEC_I_PT: Illegal tid_len:"
                                " %u for Transport ID: %s\n", tid_len, ptr);
                        core_scsi3_nodeacl_undepend_item(dest_node_acl);
                        core_scsi3_tpg_undepend_item(dest_tpg);
@@ -1660,10 +1662,10 @@ static int core_scsi3_decode_spec_i_port(
                 */
                dest_se_deve = core_get_se_deve_from_rtpi(dest_node_acl,
                                        dest_rtpi);
-               if (!(dest_se_deve)) {
-                       printk(KERN_ERR "Unable to locate %s dest_se_deve"
+               if (!dest_se_deve) {
+                       pr_err("Unable to locate %s dest_se_deve"
                                " from destination RTPI: %hu\n",
-                               TPG_TFO(dest_tpg)->get_fabric_name(),
+                               dest_tpg->se_tpg_tfo->get_fabric_name(),
                                dest_rtpi);
 
                        core_scsi3_nodeacl_undepend_item(dest_node_acl);
@@ -1674,7 +1676,7 @@ static int core_scsi3_decode_spec_i_port(
 
                ret = core_scsi3_lunacl_depend_item(dest_se_deve);
                if (ret < 0) {
-                       printk(KERN_ERR "core_scsi3_lunacl_depend_item()"
+                       pr_err("core_scsi3_lunacl_depend_item()"
                                        " failed\n");
                        atomic_dec(&dest_se_deve->pr_ref_count);
                        smp_mb__after_atomic_dec();
@@ -1684,9 +1686,9 @@ static int core_scsi3_decode_spec_i_port(
                        goto out;
                }
 #if 0
-               printk(KERN_INFO "SPC-3 PR SPEC_I_PT: Located %s Node: %s"
+               pr_debug("SPC-3 PR SPEC_I_PT: Located %s Node: %s"
                        " dest_se_deve mapped_lun: %u\n",
-                       TPG_TFO(dest_tpg)->get_fabric_name(),
+                       dest_tpg->se_tpg_tfo->get_fabric_name(),
                        dest_node_acl->initiatorname, dest_se_deve->mapped_lun);
 #endif
                /*
@@ -1712,8 +1714,8 @@ static int core_scsi3_decode_spec_i_port(
                 */
                tidh_new = kzalloc(sizeof(struct pr_transport_id_holder),
                                GFP_KERNEL);
-               if (!(tidh_new)) {
-                       printk(KERN_ERR "Unable to allocate tidh_new\n");
+               if (!tidh_new) {
+                       pr_err("Unable to allocate tidh_new\n");
                        core_scsi3_lunacl_undepend_item(dest_se_deve);
                        core_scsi3_nodeacl_undepend_item(dest_node_acl);
                        core_scsi3_tpg_undepend_item(dest_tpg);
@@ -1741,10 +1743,10 @@ static int core_scsi3_decode_spec_i_port(
                 * and then call __core_scsi3_add_registration() in the
                 * 2nd loop which will never fail.
                 */
-               dest_pr_reg = __core_scsi3_alloc_registration(SE_DEV(cmd),
+               dest_pr_reg = __core_scsi3_alloc_registration(cmd->se_dev,
                                dest_node_acl, dest_se_deve, iport_ptr,
                                sa_res_key, all_tg_pt, aptpl);
-               if (!(dest_pr_reg)) {
+               if (!dest_pr_reg) {
                        core_scsi3_lunacl_undepend_item(dest_se_deve);
                        core_scsi3_nodeacl_undepend_item(dest_node_acl);
                        core_scsi3_tpg_undepend_item(dest_tpg);
@@ -1760,6 +1762,9 @@ static int core_scsi3_decode_spec_i_port(
                tid_len = 0;
 
        }
+
+       transport_kunmap_first_data_page(cmd);
+
        /*
         * Go ahead and create a registrations from tid_dest_list for the
         * SPEC_I_PT provided TransportID for the *tidh referenced dest_node_acl
@@ -1787,12 +1792,12 @@ static int core_scsi3_decode_spec_i_port(
                prf_isid = core_pr_dump_initiator_port(dest_pr_reg, &i_buf[0],
                                                PR_REG_ISID_ID_LEN);
 
-               __core_scsi3_add_registration(SE_DEV(cmd), dest_node_acl,
+               __core_scsi3_add_registration(cmd->se_dev, dest_node_acl,
                                        dest_pr_reg, 0, 0);
 
-               printk(KERN_INFO "SPC-3 PR [%s] SPEC_I_PT: Successfully"
+               pr_debug("SPC-3 PR [%s] SPEC_I_PT: Successfully"
                        " registered Transport ID for Node: %s%s Mapped LUN:"
-                       " %u\n", TPG_TFO(dest_tpg)->get_fabric_name(),
+                       " %u\n", dest_tpg->se_tpg_tfo->get_fabric_name(),
                        dest_node_acl->initiatorname, (prf_isid) ?
                        &i_buf[0] : "", dest_se_deve->mapped_lun);
 
@@ -1806,6 +1811,7 @@ static int core_scsi3_decode_spec_i_port(
 
        return 0;
 out:
+       transport_kunmap_first_data_page(cmd);
        /*
         * For the failure case, release everything from tid_dest_list
         * including *dest_pr_reg and the configfs dependances..
@@ -1855,7 +1861,7 @@ static int __core_scsi3_update_aptpl_buf(
 {
        struct se_lun *lun;
        struct se_portal_group *tpg;
-       struct se_subsystem_dev *su_dev = SU_DEV(dev);
+       struct se_subsystem_dev *su_dev = dev->se_sub_dev;
        struct t10_pr_registration *pr_reg;
        unsigned char tmp[512], isid_buf[32];
        ssize_t len = 0;
@@ -1873,8 +1879,8 @@ static int __core_scsi3_update_aptpl_buf(
        /*
         * Walk the registration list..
         */
-       spin_lock(&T10_RES(su_dev)->registration_lock);
-       list_for_each_entry(pr_reg, &T10_RES(su_dev)->registration_list,
+       spin_lock(&su_dev->t10_pr.registration_lock);
+       list_for_each_entry(pr_reg, &su_dev->t10_pr.registration_list,
                        pr_reg_list) {
 
                tmp[0] = '\0';
@@ -1900,7 +1906,7 @@ static int __core_scsi3_update_aptpl_buf(
                                "res_holder=1\nres_type=%02x\n"
                                "res_scope=%02x\nres_all_tg_pt=%d\n"
                                "mapped_lun=%u\n", reg_count,
-                               TPG_TFO(tpg)->get_fabric_name(),
+                               tpg->se_tpg_tfo->get_fabric_name(),
                                pr_reg->pr_reg_nacl->initiatorname, isid_buf,
                                pr_reg->pr_res_key, pr_reg->pr_res_type,
                                pr_reg->pr_res_scope, pr_reg->pr_reg_all_tg_pt,
@@ -1910,17 +1916,17 @@ static int __core_scsi3_update_aptpl_buf(
                                "initiator_fabric=%s\ninitiator_node=%s\n%s"
                                "sa_res_key=%llu\nres_holder=0\n"
                                "res_all_tg_pt=%d\nmapped_lun=%u\n",
-                               reg_count, TPG_TFO(tpg)->get_fabric_name(),
+                               reg_count, tpg->se_tpg_tfo->get_fabric_name(),
                                pr_reg->pr_reg_nacl->initiatorname, isid_buf,
                                pr_reg->pr_res_key, pr_reg->pr_reg_all_tg_pt,
                                pr_reg->pr_res_mapped_lun);
                }
 
                if ((len + strlen(tmp) >= pr_aptpl_buf_len)) {
-                       printk(KERN_ERR "Unable to update renaming"
+                       pr_err("Unable to update renaming"
                                " APTPL metadata\n");
-                       spin_unlock(&T10_RES(su_dev)->registration_lock);
-                       return -1;
+                       spin_unlock(&su_dev->t10_pr.registration_lock);
+                       return -EMSGSIZE;
                }
                len += sprintf(buf+len, "%s", tmp);
 
@@ -1929,23 +1935,23 @@ static int __core_scsi3_update_aptpl_buf(
                 */
                snprintf(tmp, 512, "target_fabric=%s\ntarget_node=%s\n"
                        "tpgt=%hu\nport_rtpi=%hu\ntarget_lun=%u\nPR_REG_END:"
-                       " %d\n", TPG_TFO(tpg)->get_fabric_name(),
-                       TPG_TFO(tpg)->tpg_get_wwn(tpg),
-                       TPG_TFO(tpg)->tpg_get_tag(tpg),
+                       " %d\n", tpg->se_tpg_tfo->get_fabric_name(),
+                       tpg->se_tpg_tfo->tpg_get_wwn(tpg),
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg),
                        lun->lun_sep->sep_rtpi, lun->unpacked_lun, reg_count);
 
                if ((len + strlen(tmp) >= pr_aptpl_buf_len)) {
-                       printk(KERN_ERR "Unable to update renaming"
+                       pr_err("Unable to update renaming"
                                " APTPL metadata\n");
-                       spin_unlock(&T10_RES(su_dev)->registration_lock);
-                       return -1;
+                       spin_unlock(&su_dev->t10_pr.registration_lock);
+                       return -EMSGSIZE;
                }
                len += sprintf(buf+len, "%s", tmp);
                reg_count++;
        }
-       spin_unlock(&T10_RES(su_dev)->registration_lock);
+       spin_unlock(&su_dev->t10_pr.registration_lock);
 
-       if (!(reg_count))
+       if (!reg_count)
                len += sprintf(buf+len, "No Registrations or Reservations");
 
        return 0;
@@ -1975,7 +1981,7 @@ static int __core_scsi3_write_aptpl_to_file(
        unsigned char *buf,
        u32 pr_aptpl_buf_len)
 {
-       struct t10_wwn *wwn = &SU_DEV(dev)->t10_wwn;
+       struct t10_wwn *wwn = &dev->se_sub_dev->t10_wwn;
        struct file *file;
        struct iovec iov[1];
        mm_segment_t old_fs;
@@ -1987,21 +1993,21 @@ static int __core_scsi3_write_aptpl_to_file(
        memset(path, 0, 512);
 
        if (strlen(&wwn->unit_serial[0]) >= 512) {
-               printk(KERN_ERR "WWN value for struct se_device does not fit"
+               pr_err("WWN value for struct se_device does not fit"
                        " into path buffer\n");
-               return -1;
+               return -EMSGSIZE;
        }
 
        snprintf(path, 512, "/var/target/pr/aptpl_%s", &wwn->unit_serial[0]);
        file = filp_open(path, flags, 0600);
        if (IS_ERR(file) || !file || !file->f_dentry) {
-               printk(KERN_ERR "filp_open(%s) for APTPL metadata"
+               pr_err("filp_open(%s) for APTPL metadata"
                        " failed\n", path);
-               return -1;
+               return (PTR_ERR(file) < 0 ? PTR_ERR(file) : -ENOENT);
        }
 
        iov[0].iov_base = &buf[0];
-       if (!(pr_aptpl_buf_len))
+       if (!pr_aptpl_buf_len)
                iov[0].iov_len = (strlen(&buf[0]) + 1); /* Add extra for NULL */
        else
                iov[0].iov_len = pr_aptpl_buf_len;
@@ -2012,9 +2018,9 @@ static int __core_scsi3_write_aptpl_to_file(
        set_fs(old_fs);
 
        if (ret < 0) {
-               printk("Error writing APTPL metadata file: %s\n", path);
+               pr_debug("Error writing APTPL metadata file: %s\n", path);
                filp_close(file, NULL);
-               return -1;
+               return -EIO;
        }
        filp_close(file, NULL);
 
@@ -2032,7 +2038,7 @@ static int core_scsi3_update_and_write_aptpl(
        /*
         * Can be called with a NULL pointer from PROUT service action CLEAR
         */
-       if (!(in_buf)) {
+       if (!in_buf) {
                memset(null_buf, 0, 64);
                buf = &null_buf[0];
                /*
@@ -2049,14 +2055,14 @@ static int core_scsi3_update_and_write_aptpl(
        ret = core_scsi3_update_aptpl_buf(dev, buf, pr_aptpl_buf_len,
                                clear_aptpl_metadata);
        if (ret != 0)
-               return -1;
+               return ret;
        /*
         * __core_scsi3_write_aptpl_to_file() will call strlen()
         * on the passed buf to determine pr_aptpl_buf_len.
         */
        ret = __core_scsi3_write_aptpl_to_file(dev, buf, 0);
        if (ret != 0)
-               return -1;
+               return ret;
 
        return ret;
 }
@@ -2070,28 +2076,28 @@ static int core_scsi3_emulate_pro_register(
        int spec_i_pt,
        int ignore_key)
 {
-       struct se_session *se_sess = SE_SESS(cmd);
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_session *se_sess = cmd->se_sess;
+       struct se_device *dev = cmd->se_dev;
        struct se_dev_entry *se_deve;
-       struct se_lun *se_lun = SE_LUN(cmd);
+       struct se_lun *se_lun = cmd->se_lun;
        struct se_portal_group *se_tpg;
        struct t10_pr_registration *pr_reg, *pr_reg_p, *pr_reg_tmp, *pr_reg_e;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        /* Used for APTPL metadata w/ UNREGISTER */
        unsigned char *pr_aptpl_buf = NULL;
        unsigned char isid_buf[PR_REG_ISID_LEN], *isid_ptr = NULL;
        int pr_holder = 0, ret = 0, type;
 
-       if (!(se_sess) || !(se_lun)) {
-               printk(KERN_ERR "SPC-3 PR: se_sess || struct se_lun is NULL!\n");
+       if (!se_sess || !se_lun) {
+               pr_err("SPC-3 PR: se_sess || struct se_lun is NULL!\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
        se_tpg = se_sess->se_tpg;
        se_deve = &se_sess->se_node_acl->device_list[cmd->orig_fe_lun];
 
-       if (TPG_TFO(se_tpg)->sess_get_initiator_sid != NULL) {
+       if (se_tpg->se_tpg_tfo->sess_get_initiator_sid) {
                memset(&isid_buf[0], 0, PR_REG_ISID_LEN);
-               TPG_TFO(se_tpg)->sess_get_initiator_sid(se_sess, &isid_buf[0],
+               se_tpg->se_tpg_tfo->sess_get_initiator_sid(se_sess, &isid_buf[0],
                                PR_REG_ISID_LEN);
                isid_ptr = &isid_buf[0];
        }
@@ -2099,30 +2105,30 @@ static int core_scsi3_emulate_pro_register(
         * Follow logic from spc4r17 Section 5.7.7, Register Behaviors Table 47
         */
        pr_reg_e = core_scsi3_locate_pr_reg(dev, se_sess->se_node_acl, se_sess);
-       if (!(pr_reg_e)) {
+       if (!pr_reg_e) {
                if (res_key) {
-                       printk(KERN_WARNING "SPC-3 PR: Reservation Key non-zero"
+                       pr_warn("SPC-3 PR: Reservation Key non-zero"
                                " for SA REGISTER, returning CONFLICT\n");
                        return PYX_TRANSPORT_RESERVATION_CONFLICT;
                }
                /*
                 * Do nothing but return GOOD status.
                 */
-               if (!(sa_res_key))
+               if (!sa_res_key)
                        return PYX_TRANSPORT_SENT_TO_TRANSPORT;
 
-               if (!(spec_i_pt)) {
+               if (!spec_i_pt) {
                        /*
                         * Perform the Service Action REGISTER on the Initiator
                         * Port Endpoint that the PRO was received from on the
                         * Logical Unit of the SCSI device server.
                         */
-                       ret = core_scsi3_alloc_registration(SE_DEV(cmd),
+                       ret = core_scsi3_alloc_registration(cmd->se_dev,
                                        se_sess->se_node_acl, se_deve, isid_ptr,
                                        sa_res_key, all_tg_pt, aptpl,
                                        ignore_key, 0);
                        if (ret != 0) {
-                               printk(KERN_ERR "Unable to allocate"
+                               pr_err("Unable to allocate"
                                        " struct t10_pr_registration\n");
                                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
                        }
@@ -2143,10 +2149,10 @@ static int core_scsi3_emulate_pro_register(
                /*
                 * Nothing left to do for the APTPL=0 case.
                 */
-               if (!(aptpl)) {
+               if (!aptpl) {
                        pr_tmpl->pr_aptpl_active = 0;
-                       core_scsi3_update_and_write_aptpl(SE_DEV(cmd), NULL, 0);
-                       printk("SPC-3 PR: Set APTPL Bit Deactivated for"
+                       core_scsi3_update_and_write_aptpl(cmd->se_dev, NULL, 0);
+                       pr_debug("SPC-3 PR: Set APTPL Bit Deactivated for"
                                        " REGISTER\n");
                        return 0;
                }
@@ -2155,15 +2161,15 @@ static int core_scsi3_emulate_pro_register(
                 * update the APTPL metadata information using its
                 * preallocated *pr_reg->pr_aptpl_buf.
                 */
-               pr_reg = core_scsi3_locate_pr_reg(SE_DEV(cmd),
+               pr_reg = core_scsi3_locate_pr_reg(cmd->se_dev,
                                se_sess->se_node_acl, se_sess);
 
-               ret = core_scsi3_update_and_write_aptpl(SE_DEV(cmd),
+               ret = core_scsi3_update_and_write_aptpl(cmd->se_dev,
                                &pr_reg->pr_aptpl_buf[0],
                                pr_tmpl->pr_aptpl_buf_len);
-               if (!(ret)) {
+               if (!ret) {
                        pr_tmpl->pr_aptpl_active = 1;
-                       printk("SPC-3 PR: Set APTPL Bit Activated for REGISTER\n");
+                       pr_debug("SPC-3 PR: Set APTPL Bit Activated for REGISTER\n");
                }
 
                core_scsi3_put_pr_reg(pr_reg);
@@ -2175,9 +2181,9 @@ static int core_scsi3_emulate_pro_register(
                pr_reg = pr_reg_e;
                type = pr_reg->pr_res_type;
 
-               if (!(ignore_key)) {
+               if (!ignore_key) {
                        if (res_key != pr_reg->pr_res_key) {
-                               printk(KERN_ERR "SPC-3 PR REGISTER: Received"
+                               pr_err("SPC-3 PR REGISTER: Received"
                                        " res_key: 0x%016Lx does not match"
                                        " existing SA REGISTER res_key:"
                                        " 0x%016Lx\n", res_key,
@@ -2187,7 +2193,7 @@ static int core_scsi3_emulate_pro_register(
                        }
                }
                if (spec_i_pt) {
-                       printk(KERN_ERR "SPC-3 PR UNREGISTER: SPEC_I_PT"
+                       pr_err("SPC-3 PR UNREGISTER: SPEC_I_PT"
                                " set while sa_res_key=0\n");
                        core_scsi3_put_pr_reg(pr_reg);
                        return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
@@ -2197,7 +2203,7 @@ static int core_scsi3_emulate_pro_register(
                 * must also set ALL_TG_PT=1 in the incoming PROUT.
                 */
                if (pr_reg->pr_reg_all_tg_pt && !(all_tg_pt)) {
-                       printk(KERN_ERR "SPC-3 PR UNREGISTER: ALL_TG_PT=1"
+                       pr_err("SPC-3 PR UNREGISTER: ALL_TG_PT=1"
                                " registration exists, but ALL_TG_PT=1 bit not"
                                " present in received PROUT\n");
                        core_scsi3_put_pr_reg(pr_reg);
@@ -2209,8 +2215,8 @@ static int core_scsi3_emulate_pro_register(
                if (aptpl) {
                        pr_aptpl_buf = kzalloc(pr_tmpl->pr_aptpl_buf_len,
                                                GFP_KERNEL);
-                       if (!(pr_aptpl_buf)) {
-                               printk(KERN_ERR "Unable to allocate"
+                       if (!pr_aptpl_buf) {
+                               pr_err("Unable to allocate"
                                        " pr_aptpl_buf\n");
                                core_scsi3_put_pr_reg(pr_reg);
                                return PYX_TRANSPORT_LU_COMM_FAILURE;
@@ -2221,9 +2227,9 @@ static int core_scsi3_emulate_pro_register(
                 * Nexus sa_res_key=1 Change Reservation Key for registered I_T
                 * Nexus.
                 */
-               if (!(sa_res_key)) {
+               if (!sa_res_key) {
                        pr_holder = core_scsi3_check_implict_release(
-                                       SE_DEV(cmd), pr_reg);
+                                       cmd->se_dev, pr_reg);
                        if (pr_holder < 0) {
                                kfree(pr_aptpl_buf);
                                core_scsi3_put_pr_reg(pr_reg);
@@ -2240,7 +2246,7 @@ static int core_scsi3_emulate_pro_register(
                                                &pr_tmpl->registration_list,
                                                pr_reg_list) {
 
-                                       if (!(pr_reg_p->pr_reg_all_tg_pt))
+                                       if (!pr_reg_p->pr_reg_all_tg_pt)
                                                continue;
 
                                        if (pr_reg_p->pr_res_key != res_key)
@@ -2260,7 +2266,7 @@ static int core_scsi3_emulate_pro_register(
                        /*
                         * Release the calling I_T Nexus registration now..
                         */
-                       __core_scsi3_free_registration(SE_DEV(cmd), pr_reg,
+                       __core_scsi3_free_registration(cmd->se_dev, pr_reg,
                                                        NULL, 1);
                        /*
                         * From spc4r17, section 5.7.11.3 Unregistering
@@ -2289,10 +2295,10 @@ static int core_scsi3_emulate_pro_register(
                        }
                        spin_unlock(&pr_tmpl->registration_lock);
 
-                       if (!(aptpl)) {
+                       if (!aptpl) {
                                pr_tmpl->pr_aptpl_active = 0;
                                core_scsi3_update_and_write_aptpl(dev, NULL, 0);
-                               printk("SPC-3 PR: Set APTPL Bit Deactivated"
+                               pr_debug("SPC-3 PR: Set APTPL Bit Deactivated"
                                                " for UNREGISTER\n");
                                return 0;
                        }
@@ -2300,9 +2306,9 @@ static int core_scsi3_emulate_pro_register(
                        ret = core_scsi3_update_and_write_aptpl(dev,
                                        &pr_aptpl_buf[0],
                                        pr_tmpl->pr_aptpl_buf_len);
-                       if (!(ret)) {
+                       if (!ret) {
                                pr_tmpl->pr_aptpl_active = 1;
-                               printk("SPC-3 PR: Set APTPL Bit Activated"
+                               pr_debug("SPC-3 PR: Set APTPL Bit Activated"
                                                " for UNREGISTER\n");
                        }
 
@@ -2315,20 +2321,20 @@ static int core_scsi3_emulate_pro_register(
                         * READ_KEYS service action.
                         */
                        pr_reg->pr_res_generation = core_scsi3_pr_generation(
-                                                       SE_DEV(cmd));
+                                                       cmd->se_dev);
                        pr_reg->pr_res_key = sa_res_key;
-                       printk("SPC-3 PR [%s] REGISTER%s: Changed Reservation"
+                       pr_debug("SPC-3 PR [%s] REGISTER%s: Changed Reservation"
                                " Key for %s to: 0x%016Lx PRgeneration:"
-                               " 0x%08x\n", CMD_TFO(cmd)->get_fabric_name(),
+                               " 0x%08x\n", cmd->se_tfo->get_fabric_name(),
                                (ignore_key) ? "_AND_IGNORE_EXISTING_KEY" : "",
                                pr_reg->pr_reg_nacl->initiatorname,
                                pr_reg->pr_res_key, pr_reg->pr_res_generation);
 
-                       if (!(aptpl)) {
+                       if (!aptpl) {
                                pr_tmpl->pr_aptpl_active = 0;
                                core_scsi3_update_and_write_aptpl(dev, NULL, 0);
                                core_scsi3_put_pr_reg(pr_reg);
-                               printk("SPC-3 PR: Set APTPL Bit Deactivated"
+                               pr_debug("SPC-3 PR: Set APTPL Bit Deactivated"
                                                " for REGISTER\n");
                                return 0;
                        }
@@ -2336,9 +2342,9 @@ static int core_scsi3_emulate_pro_register(
                        ret = core_scsi3_update_and_write_aptpl(dev,
                                        &pr_aptpl_buf[0],
                                        pr_tmpl->pr_aptpl_buf_len);
-                       if (!(ret)) {
+                       if (!ret) {
                                pr_tmpl->pr_aptpl_active = 1;
-                               printk("SPC-3 PR: Set APTPL Bit Activated"
+                               pr_debug("SPC-3 PR: Set APTPL Bit Activated"
                                                " for REGISTER\n");
                        }
 
@@ -2378,19 +2384,19 @@ static int core_scsi3_pro_reserve(
        int scope,
        u64 res_key)
 {
-       struct se_session *se_sess = SE_SESS(cmd);
+       struct se_session *se_sess = cmd->se_sess;
        struct se_dev_entry *se_deve;
-       struct se_lun *se_lun = SE_LUN(cmd);
+       struct se_lun *se_lun = cmd->se_lun;
        struct se_portal_group *se_tpg;
        struct t10_pr_registration *pr_reg, *pr_res_holder;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        char i_buf[PR_REG_ISID_ID_LEN];
        int ret, prf_isid;
 
        memset(i_buf, 0, PR_REG_ISID_ID_LEN);
 
-       if (!(se_sess) || !(se_lun)) {
-               printk(KERN_ERR "SPC-3 PR: se_sess || struct se_lun is NULL!\n");
+       if (!se_sess || !se_lun) {
+               pr_err("SPC-3 PR: se_sess || struct se_lun is NULL!\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
        se_tpg = se_sess->se_tpg;
@@ -2398,10 +2404,10 @@ static int core_scsi3_pro_reserve(
        /*
         * Locate the existing *pr_reg via struct se_node_acl pointers
         */
-       pr_reg = core_scsi3_locate_pr_reg(SE_DEV(cmd), se_sess->se_node_acl,
+       pr_reg = core_scsi3_locate_pr_reg(cmd->se_dev, se_sess->se_node_acl,
                                se_sess);
-       if (!(pr_reg)) {
-               printk(KERN_ERR "SPC-3 PR: Unable to locate"
+       if (!pr_reg) {
+               pr_err("SPC-3 PR: Unable to locate"
                        " PR_REGISTERED *pr_reg for RESERVE\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
@@ -2415,7 +2421,7 @@ static int core_scsi3_pro_reserve(
         *       registered with the logical unit for the I_T nexus; and
         */
        if (res_key != pr_reg->pr_res_key) {
-               printk(KERN_ERR "SPC-3 PR RESERVE: Received res_key: 0x%016Lx"
+               pr_err("SPC-3 PR RESERVE: Received res_key: 0x%016Lx"
                        " does not match existing SA REGISTER res_key:"
                        " 0x%016Lx\n", res_key, pr_reg->pr_res_key);
                core_scsi3_put_pr_reg(pr_reg);
@@ -2432,7 +2438,7 @@ static int core_scsi3_pro_reserve(
         * and that persistent reservation has a scope of LU_SCOPE.
         */
        if (scope != PR_SCOPE_LU_SCOPE) {
-               printk(KERN_ERR "SPC-3 PR: Illegal SCOPE: 0x%02x\n", scope);
+               pr_err("SPC-3 PR: Illegal SCOPE: 0x%02x\n", scope);
                core_scsi3_put_pr_reg(pr_reg);
                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
        }
@@ -2456,12 +2462,12 @@ static int core_scsi3_pro_reserve(
                 */
                if (pr_res_holder != pr_reg) {
                        struct se_node_acl *pr_res_nacl = pr_res_holder->pr_reg_nacl;
-                       printk(KERN_ERR "SPC-3 PR: Attempted RESERVE from"
+                       pr_err("SPC-3 PR: Attempted RESERVE from"
                                " [%s]: %s while reservation already held by"
                                " [%s]: %s, returning RESERVATION_CONFLICT\n",
-                               CMD_TFO(cmd)->get_fabric_name(),
+                               cmd->se_tfo->get_fabric_name(),
                                se_sess->se_node_acl->initiatorname,
-                               TPG_TFO(pr_res_nacl->se_tpg)->get_fabric_name(),
+                               pr_res_nacl->se_tpg->se_tpg_tfo->get_fabric_name(),
                                pr_res_holder->pr_reg_nacl->initiatorname);
 
                        spin_unlock(&dev->dev_reservation_lock);
@@ -2478,13 +2484,13 @@ static int core_scsi3_pro_reserve(
                if ((pr_res_holder->pr_res_type != type) ||
                    (pr_res_holder->pr_res_scope != scope)) {
                        struct se_node_acl *pr_res_nacl = pr_res_holder->pr_reg_nacl;
-                       printk(KERN_ERR "SPC-3 PR: Attempted RESERVE from"
+                       pr_err("SPC-3 PR: Attempted RESERVE from"
                                " [%s]: %s trying to change TYPE and/or SCOPE,"
                                " while reservation already held by [%s]: %s,"
                                " returning RESERVATION_CONFLICT\n",
-                               CMD_TFO(cmd)->get_fabric_name(),
+                               cmd->se_tfo->get_fabric_name(),
                                se_sess->se_node_acl->initiatorname,
-                               TPG_TFO(pr_res_nacl->se_tpg)->get_fabric_name(),
+                               pr_res_nacl->se_tpg->se_tpg_tfo->get_fabric_name(),
                                pr_res_holder->pr_reg_nacl->initiatorname);
 
                        spin_unlock(&dev->dev_reservation_lock);
@@ -2516,22 +2522,22 @@ static int core_scsi3_pro_reserve(
        prf_isid = core_pr_dump_initiator_port(pr_reg, &i_buf[0],
                                PR_REG_ISID_ID_LEN);
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: RESERVE created new"
+       pr_debug("SPC-3 PR [%s] Service Action: RESERVE created new"
                " reservation holder TYPE: %s ALL_TG_PT: %d\n",
-               CMD_TFO(cmd)->get_fabric_name(), core_scsi3_pr_dump_type(type),
+               cmd->se_tfo->get_fabric_name(), core_scsi3_pr_dump_type(type),
                (pr_reg->pr_reg_all_tg_pt) ? 1 : 0);
-       printk(KERN_INFO "SPC-3 PR [%s] RESERVE Node: %s%s\n",
-                       CMD_TFO(cmd)->get_fabric_name(),
+       pr_debug("SPC-3 PR [%s] RESERVE Node: %s%s\n",
+                       cmd->se_tfo->get_fabric_name(),
                        se_sess->se_node_acl->initiatorname,
                        (prf_isid) ? &i_buf[0] : "");
        spin_unlock(&dev->dev_reservation_lock);
 
        if (pr_tmpl->pr_aptpl_active) {
-               ret = core_scsi3_update_and_write_aptpl(SE_DEV(cmd),
+               ret = core_scsi3_update_and_write_aptpl(cmd->se_dev,
                                &pr_reg->pr_aptpl_buf[0],
                                pr_tmpl->pr_aptpl_buf_len);
-               if (!(ret))
-                       printk(KERN_INFO "SPC-3 PR: Updated APTPL metadata"
+               if (!ret)
+                       pr_debug("SPC-3 PR: Updated APTPL metadata"
                                        " for RESERVE\n");
        }
 
@@ -2558,7 +2564,7 @@ static int core_scsi3_emulate_pro_reserve(
                ret = core_scsi3_pro_reserve(cmd, dev, type, scope, res_key);
                break;
        default:
-               printk(KERN_ERR "SPC-3 PR: Unknown Service Action RESERVE Type:"
+               pr_err("SPC-3 PR: Unknown Service Action RESERVE Type:"
                        " 0x%02x\n", type);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
@@ -2587,12 +2593,12 @@ static void __core_scsi3_complete_pro_release(
         */
        dev->dev_pr_res_holder = NULL;
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: %s RELEASE cleared"
+       pr_debug("SPC-3 PR [%s] Service Action: %s RELEASE cleared"
                " reservation holder TYPE: %s ALL_TG_PT: %d\n",
                tfo->get_fabric_name(), (explict) ? "explict" : "implict",
                core_scsi3_pr_dump_type(pr_reg->pr_res_type),
                (pr_reg->pr_reg_all_tg_pt) ? 1 : 0);
-       printk(KERN_INFO "SPC-3 PR [%s] RELEASE Node: %s%s\n",
+       pr_debug("SPC-3 PR [%s] RELEASE Node: %s%s\n",
                tfo->get_fabric_name(), se_nacl->initiatorname,
                (prf_isid) ? &i_buf[0] : "");
        /*
@@ -2608,22 +2614,22 @@ static int core_scsi3_emulate_pro_release(
        u64 res_key)
 {
        struct se_device *dev = cmd->se_dev;
-       struct se_session *se_sess = SE_SESS(cmd);
-       struct se_lun *se_lun = SE_LUN(cmd);
+       struct se_session *se_sess = cmd->se_sess;
+       struct se_lun *se_lun = cmd->se_lun;
        struct t10_pr_registration *pr_reg, *pr_reg_p, *pr_res_holder;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        int ret, all_reg = 0;
 
-       if (!(se_sess) || !(se_lun)) {
-               printk(KERN_ERR "SPC-3 PR: se_sess || struct se_lun is NULL!\n");
+       if (!se_sess || !se_lun) {
+               pr_err("SPC-3 PR: se_sess || struct se_lun is NULL!\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
        /*
         * Locate the existing *pr_reg via struct se_node_acl pointers
         */
        pr_reg = core_scsi3_locate_pr_reg(dev, se_sess->se_node_acl, se_sess);
-       if (!(pr_reg)) {
-               printk(KERN_ERR "SPC-3 PR: Unable to locate"
+       if (!pr_reg) {
+               pr_err("SPC-3 PR: Unable to locate"
                        " PR_REGISTERED *pr_reg for RELEASE\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
@@ -2641,7 +2647,7 @@ static int core_scsi3_emulate_pro_release(
         */
        spin_lock(&dev->dev_reservation_lock);
        pr_res_holder = dev->dev_pr_res_holder;
-       if (!(pr_res_holder)) {
+       if (!pr_res_holder) {
                /*
                 * No persistent reservation, return GOOD status.
                 */
@@ -2678,7 +2684,7 @@ static int core_scsi3_emulate_pro_release(
         *        that is registered with the logical unit for the I_T nexus;
         */
        if (res_key != pr_reg->pr_res_key) {
-               printk(KERN_ERR "SPC-3 PR RELEASE: Received res_key: 0x%016Lx"
+               pr_err("SPC-3 PR RELEASE: Received res_key: 0x%016Lx"
                        " does not match existing SA REGISTER res_key:"
                        " 0x%016Lx\n", res_key, pr_reg->pr_res_key);
                spin_unlock(&dev->dev_reservation_lock);
@@ -2694,13 +2700,13 @@ static int core_scsi3_emulate_pro_release(
        if ((pr_res_holder->pr_res_type != type) ||
            (pr_res_holder->pr_res_scope != scope)) {
                struct se_node_acl *pr_res_nacl = pr_res_holder->pr_reg_nacl;
-               printk(KERN_ERR "SPC-3 PR RELEASE: Attempted to release"
+               pr_err("SPC-3 PR RELEASE: Attempted to release"
                        " reservation from [%s]: %s with different TYPE "
                        "and/or SCOPE  while reservation already held by"
                        " [%s]: %s, returning RESERVATION_CONFLICT\n",
-                       CMD_TFO(cmd)->get_fabric_name(),
+                       cmd->se_tfo->get_fabric_name(),
                        se_sess->se_node_acl->initiatorname,
-                       TPG_TFO(pr_res_nacl->se_tpg)->get_fabric_name(),
+                       pr_res_nacl->se_tpg->se_tpg_tfo->get_fabric_name(),
                        pr_res_holder->pr_reg_nacl->initiatorname);
 
                spin_unlock(&dev->dev_reservation_lock);
@@ -2758,11 +2764,11 @@ static int core_scsi3_emulate_pro_release(
 
 write_aptpl:
        if (pr_tmpl->pr_aptpl_active) {
-               ret = core_scsi3_update_and_write_aptpl(SE_DEV(cmd),
+               ret = core_scsi3_update_and_write_aptpl(cmd->se_dev,
                                &pr_reg->pr_aptpl_buf[0],
                                pr_tmpl->pr_aptpl_buf_len);
-               if (!(ret))
-                       printk("SPC-3 PR: Updated APTPL metadata for RELEASE\n");
+               if (!ret)
+                       pr_debug("SPC-3 PR: Updated APTPL metadata for RELEASE\n");
        }
 
        core_scsi3_put_pr_reg(pr_reg);
@@ -2775,18 +2781,18 @@ static int core_scsi3_emulate_pro_clear(
 {
        struct se_device *dev = cmd->se_dev;
        struct se_node_acl *pr_reg_nacl;
-       struct se_session *se_sess = SE_SESS(cmd);
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct se_session *se_sess = cmd->se_sess;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        struct t10_pr_registration *pr_reg, *pr_reg_tmp, *pr_reg_n, *pr_res_holder;
        u32 pr_res_mapped_lun = 0;
        int calling_it_nexus = 0;
        /*
         * Locate the existing *pr_reg via struct se_node_acl pointers
         */
-       pr_reg_n = core_scsi3_locate_pr_reg(SE_DEV(cmd),
+       pr_reg_n = core_scsi3_locate_pr_reg(cmd->se_dev,
                        se_sess->se_node_acl, se_sess);
-       if (!(pr_reg_n)) {
-               printk(KERN_ERR "SPC-3 PR: Unable to locate"
+       if (!pr_reg_n) {
+               pr_err("SPC-3 PR: Unable to locate"
                        " PR_REGISTERED *pr_reg for CLEAR\n");
                        return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
@@ -2802,7 +2808,7 @@ static int core_scsi3_emulate_pro_clear(
         *         that is registered with the logical unit for the I_T nexus.
         */
        if (res_key != pr_reg_n->pr_res_key) {
-               printk(KERN_ERR "SPC-3 PR REGISTER: Received"
+               pr_err("SPC-3 PR REGISTER: Received"
                        " res_key: 0x%016Lx does not match"
                        " existing SA REGISTER res_key:"
                        " 0x%016Lx\n", res_key, pr_reg_n->pr_res_key);
@@ -2839,18 +2845,18 @@ static int core_scsi3_emulate_pro_clear(
                 *    command with CLEAR service action was received, with the
                 *    additional sense code set to RESERVATIONS PREEMPTED.
                 */
-               if (!(calling_it_nexus))
+               if (!calling_it_nexus)
                        core_scsi3_ua_allocate(pr_reg_nacl, pr_res_mapped_lun,
                                0x2A, ASCQ_2AH_RESERVATIONS_PREEMPTED);
        }
        spin_unlock(&pr_tmpl->registration_lock);
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: CLEAR complete\n",
-               CMD_TFO(cmd)->get_fabric_name());
+       pr_debug("SPC-3 PR [%s] Service Action: CLEAR complete\n",
+               cmd->se_tfo->get_fabric_name());
 
        if (pr_tmpl->pr_aptpl_active) {
-               core_scsi3_update_and_write_aptpl(SE_DEV(cmd), NULL, 0);
-               printk(KERN_INFO "SPC-3 PR: Updated APTPL metadata"
+               core_scsi3_update_and_write_aptpl(cmd->se_dev, NULL, 0);
+               pr_debug("SPC-3 PR: Updated APTPL metadata"
                                " for CLEAR\n");
        }
 
@@ -2889,12 +2895,12 @@ static void __core_scsi3_complete_pro_preempt(
        pr_reg->pr_res_type = type;
        pr_reg->pr_res_scope = scope;
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: PREEMPT%s created new"
+       pr_debug("SPC-3 PR [%s] Service Action: PREEMPT%s created new"
                " reservation holder TYPE: %s ALL_TG_PT: %d\n",
                tfo->get_fabric_name(), (abort) ? "_AND_ABORT" : "",
                core_scsi3_pr_dump_type(type),
                (pr_reg->pr_reg_all_tg_pt) ? 1 : 0);
-       printk(KERN_INFO "SPC-3 PR [%s] PREEMPT%s from Node: %s%s\n",
+       pr_debug("SPC-3 PR [%s] PREEMPT%s from Node: %s%s\n",
                tfo->get_fabric_name(), (abort) ? "_AND_ABORT" : "",
                nacl->initiatorname, (prf_isid) ? &i_buf[0] : "");
        /*
@@ -2920,7 +2926,7 @@ static void core_scsi3_release_preempt_and_abort(
                if (pr_reg_holder == pr_reg)
                        continue;
                if (pr_reg->pr_res_holder) {
-                       printk(KERN_WARNING "pr_reg->pr_res_holder still set\n");
+                       pr_warn("pr_reg->pr_res_holder still set\n");
                        continue;
                }
 
@@ -2954,25 +2960,25 @@ static int core_scsi3_pro_preempt(
        u64 sa_res_key,
        int abort)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct se_dev_entry *se_deve;
        struct se_node_acl *pr_reg_nacl;
-       struct se_session *se_sess = SE_SESS(cmd);
+       struct se_session *se_sess = cmd->se_sess;
        struct list_head preempt_and_abort_list;
        struct t10_pr_registration *pr_reg, *pr_reg_tmp, *pr_reg_n, *pr_res_holder;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
        u32 pr_res_mapped_lun = 0;
        int all_reg = 0, calling_it_nexus = 0, released_regs = 0;
        int prh_type = 0, prh_scope = 0, ret;
 
-       if (!(se_sess))
+       if (!se_sess)
                return PYX_TRANSPORT_LU_COMM_FAILURE;
 
        se_deve = &se_sess->se_node_acl->device_list[cmd->orig_fe_lun];
-       pr_reg_n = core_scsi3_locate_pr_reg(SE_DEV(cmd), se_sess->se_node_acl,
+       pr_reg_n = core_scsi3_locate_pr_reg(cmd->se_dev, se_sess->se_node_acl,
                                se_sess);
-       if (!(pr_reg_n)) {
-               printk(KERN_ERR "SPC-3 PR: Unable to locate"
+       if (!pr_reg_n) {
+               pr_err("SPC-3 PR: Unable to locate"
                        " PR_REGISTERED *pr_reg for PREEMPT%s\n",
                        (abort) ? "_AND_ABORT" : "");
                return PYX_TRANSPORT_RESERVATION_CONFLICT;
@@ -2982,7 +2988,7 @@ static int core_scsi3_pro_preempt(
                return PYX_TRANSPORT_RESERVATION_CONFLICT;
        }
        if (scope != PR_SCOPE_LU_SCOPE) {
-               printk(KERN_ERR "SPC-3 PR: Illegal SCOPE: 0x%02x\n", scope);
+               pr_err("SPC-3 PR: Illegal SCOPE: 0x%02x\n", scope);
                core_scsi3_put_pr_reg(pr_reg_n);
                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
        }
@@ -2995,7 +3001,7 @@ static int core_scsi3_pro_preempt(
            (pr_res_holder->pr_res_type == PR_TYPE_EXCLUSIVE_ACCESS_ALLREG)))
                all_reg = 1;
 
-       if (!(all_reg) && !(sa_res_key)) {
+       if (!all_reg && !sa_res_key) {
                spin_unlock(&dev->dev_reservation_lock);
                core_scsi3_put_pr_reg(pr_reg_n);
                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
@@ -3009,7 +3015,7 @@ static int core_scsi3_pro_preempt(
         * server shall perform a preempt by doing the following in an
         * uninterrupted series of actions. (See below..)
         */
-       if (!(pr_res_holder) || (pr_res_holder->pr_res_key != sa_res_key)) {
+       if (!pr_res_holder || (pr_res_holder->pr_res_key != sa_res_key)) {
                /*
                 * No existing or SA Reservation Key matching reservations..
                 *
@@ -3036,7 +3042,7 @@ static int core_scsi3_pro_preempt(
                         *    was received, with the additional sense code set
                         *    to REGISTRATIONS PREEMPTED.
                         */
-                       if (!(all_reg)) {
+                       if (!all_reg) {
                                if (pr_reg->pr_res_key != sa_res_key)
                                        continue;
 
@@ -3076,7 +3082,7 @@ static int core_scsi3_pro_preempt(
                                                NULL, 0);
                                released_regs++;
                        }
-                       if (!(calling_it_nexus))
+                       if (!calling_it_nexus)
                                core_scsi3_ua_allocate(pr_reg_nacl,
                                        pr_res_mapped_lun, 0x2A,
                                        ASCQ_2AH_RESERVATIONS_PREEMPTED);
@@ -3089,7 +3095,7 @@ static int core_scsi3_pro_preempt(
                 * registered reservation key, then the device server shall
                 * complete the command with RESERVATION CONFLICT status.
                 */
-               if (!(released_regs)) {
+               if (!released_regs) {
                        spin_unlock(&dev->dev_reservation_lock);
                        core_scsi3_put_pr_reg(pr_reg_n);
                        return PYX_TRANSPORT_RESERVATION_CONFLICT;
@@ -3111,17 +3117,17 @@ static int core_scsi3_pro_preempt(
                spin_unlock(&dev->dev_reservation_lock);
 
                if (pr_tmpl->pr_aptpl_active) {
-                       ret = core_scsi3_update_and_write_aptpl(SE_DEV(cmd),
+                       ret = core_scsi3_update_and_write_aptpl(cmd->se_dev,
                                        &pr_reg_n->pr_aptpl_buf[0],
                                        pr_tmpl->pr_aptpl_buf_len);
-                       if (!(ret))
-                               printk(KERN_INFO "SPC-3 PR: Updated APTPL"
+                       if (!ret)
+                               pr_debug("SPC-3 PR: Updated APTPL"
                                        " metadata for  PREEMPT%s\n", (abort) ?
                                        "_AND_ABORT" : "");
                }
 
                core_scsi3_put_pr_reg(pr_reg_n);
-               core_scsi3_pr_generation(SE_DEV(cmd));
+               core_scsi3_pr_generation(cmd->se_dev);
                return 0;
        }
        /*
@@ -3247,16 +3253,16 @@ static int core_scsi3_pro_preempt(
        }
 
        if (pr_tmpl->pr_aptpl_active) {
-               ret = core_scsi3_update_and_write_aptpl(SE_DEV(cmd),
+               ret = core_scsi3_update_and_write_aptpl(cmd->se_dev,
                                &pr_reg_n->pr_aptpl_buf[0],
                                pr_tmpl->pr_aptpl_buf_len);
-               if (!(ret))
-                       printk("SPC-3 PR: Updated APTPL metadata for PREEMPT"
+               if (!ret)
+                       pr_debug("SPC-3 PR: Updated APTPL metadata for PREEMPT"
                                "%s\n", (abort) ? "_AND_ABORT" : "");
        }
 
        core_scsi3_put_pr_reg(pr_reg_n);
-       core_scsi3_pr_generation(SE_DEV(cmd));
+       core_scsi3_pr_generation(cmd->se_dev);
        return 0;
 }
 
@@ -3281,7 +3287,7 @@ static int core_scsi3_emulate_pro_preempt(
                                res_key, sa_res_key, abort);
                break;
        default:
-               printk(KERN_ERR "SPC-3 PR: Unknown Service Action PREEMPT%s"
+               pr_err("SPC-3 PR: Unknown Service Action PREEMPT%s"
                        " Type: 0x%02x\n", (abort) ? "_AND_ABORT" : "", type);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
@@ -3297,17 +3303,17 @@ static int core_scsi3_emulate_pro_register_and_move(
        int aptpl,
        int unreg)
 {
-       struct se_session *se_sess = SE_SESS(cmd);
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_session *se_sess = cmd->se_sess;
+       struct se_device *dev = cmd->se_dev;
        struct se_dev_entry *se_deve, *dest_se_deve = NULL;
-       struct se_lun *se_lun = SE_LUN(cmd);
+       struct se_lun *se_lun = cmd->se_lun;
        struct se_node_acl *pr_res_nacl, *pr_reg_nacl, *dest_node_acl = NULL;
        struct se_port *se_port;
        struct se_portal_group *se_tpg, *dest_se_tpg = NULL;
        struct target_core_fabric_ops *dest_tf_ops = NULL, *tf_ops;
        struct t10_pr_registration *pr_reg, *pr_res_holder, *dest_pr_reg;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
+       unsigned char *buf;
        unsigned char *initiator_str;
        char *iport_ptr = NULL, dest_iport[64], i_buf[PR_REG_ISID_ID_LEN];
        u32 tid_len, tmp_tid_len;
@@ -3315,14 +3321,14 @@ static int core_scsi3_emulate_pro_register_and_move(
        unsigned short rtpi;
        unsigned char proto_ident;
 
-       if (!(se_sess) || !(se_lun)) {
-               printk(KERN_ERR "SPC-3 PR: se_sess || struct se_lun is NULL!\n");
+       if (!se_sess || !se_lun) {
+               pr_err("SPC-3 PR: se_sess || struct se_lun is NULL!\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
        memset(dest_iport, 0, 64);
        memset(i_buf, 0, PR_REG_ISID_ID_LEN);
        se_tpg = se_sess->se_tpg;
-       tf_ops = TPG_TFO(se_tpg);
+       tf_ops = se_tpg->se_tpg_tfo;
        se_deve = &se_sess->se_node_acl->device_list[cmd->orig_fe_lun];
        /*
         * Follow logic from spc4r17 Section 5.7.8, Table 50 --
@@ -3330,10 +3336,10 @@ static int core_scsi3_emulate_pro_register_and_move(
         *
         * Locate the existing *pr_reg via struct se_node_acl pointers
         */
-       pr_reg = core_scsi3_locate_pr_reg(SE_DEV(cmd), se_sess->se_node_acl,
+       pr_reg = core_scsi3_locate_pr_reg(cmd->se_dev, se_sess->se_node_acl,
                                se_sess);
-       if (!(pr_reg)) {
-               printk(KERN_ERR "SPC-3 PR: Unable to locate PR_REGISTERED"
+       if (!pr_reg) {
+               pr_err("SPC-3 PR: Unable to locate PR_REGISTERED"
                        " *pr_reg for REGISTER_AND_MOVE\n");
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
@@ -3342,7 +3348,7 @@ static int core_scsi3_emulate_pro_register_and_move(
         * provided during this initiator's I_T nexus registration.
         */
        if (res_key != pr_reg->pr_res_key) {
-               printk(KERN_WARNING "SPC-3 PR REGISTER_AND_MOVE: Received"
+               pr_warn("SPC-3 PR REGISTER_AND_MOVE: Received"
                        " res_key: 0x%016Lx does not match existing SA REGISTER"
                        " res_key: 0x%016Lx\n", res_key, pr_reg->pr_res_key);
                core_scsi3_put_pr_reg(pr_reg);
@@ -3351,26 +3357,30 @@ static int core_scsi3_emulate_pro_register_and_move(
        /*
         * The service active reservation key needs to be non zero
         */
-       if (!(sa_res_key)) {
-               printk(KERN_WARNING "SPC-3 PR REGISTER_AND_MOVE: Received zero"
+       if (!sa_res_key) {
+               pr_warn("SPC-3 PR REGISTER_AND_MOVE: Received zero"
                        " sa_res_key\n");
                core_scsi3_put_pr_reg(pr_reg);
                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
        }
+
        /*
         * Determine the Relative Target Port Identifier where the reservation
         * will be moved to for the TransportID containing SCSI initiator WWN
         * information.
         */
+       buf = transport_kmap_first_data_page(cmd);
        rtpi = (buf[18] & 0xff) << 8;
        rtpi |= buf[19] & 0xff;
        tid_len = (buf[20] & 0xff) << 24;
        tid_len |= (buf[21] & 0xff) << 16;
        tid_len |= (buf[22] & 0xff) << 8;
        tid_len |= buf[23] & 0xff;
+       transport_kunmap_first_data_page(cmd);
+       buf = NULL;
 
        if ((tid_len + 24) != cmd->data_length) {
-               printk(KERN_ERR "SPC-3 PR: Illegal tid_len: %u + 24 byte header"
+               pr_err("SPC-3 PR: Illegal tid_len: %u + 24 byte header"
                        " does not equal CDB data_length: %u\n", tid_len,
                        cmd->data_length);
                core_scsi3_put_pr_reg(pr_reg);
@@ -3382,10 +3392,10 @@ static int core_scsi3_emulate_pro_register_and_move(
                if (se_port->sep_rtpi != rtpi)
                        continue;
                dest_se_tpg = se_port->sep_tpg;
-               if (!(dest_se_tpg))
+               if (!dest_se_tpg)
                        continue;
-               dest_tf_ops = TPG_TFO(dest_se_tpg);
-               if (!(dest_tf_ops))
+               dest_tf_ops = dest_se_tpg->se_tpg_tfo;
+               if (!dest_tf_ops)
                        continue;
 
                atomic_inc(&dest_se_tpg->tpg_pr_ref_count);
@@ -3394,7 +3404,7 @@ static int core_scsi3_emulate_pro_register_and_move(
 
                ret = core_scsi3_tpg_depend_item(dest_se_tpg);
                if (ret != 0) {
-                       printk(KERN_ERR "core_scsi3_tpg_depend_item() failed"
+                       pr_err("core_scsi3_tpg_depend_item() failed"
                                " for dest_se_tpg\n");
                        atomic_dec(&dest_se_tpg->tpg_pr_ref_count);
                        smp_mb__after_atomic_dec();
@@ -3407,20 +3417,22 @@ static int core_scsi3_emulate_pro_register_and_move(
        }
        spin_unlock(&dev->se_port_lock);
 
-       if (!(dest_se_tpg) || (!dest_tf_ops)) {
-               printk(KERN_ERR "SPC-3 PR REGISTER_AND_MOVE: Unable to locate"
+       if (!dest_se_tpg || !dest_tf_ops) {
+               pr_err("SPC-3 PR REGISTER_AND_MOVE: Unable to locate"
                        " fabric ops from Relative Target Port Identifier:"
                        " %hu\n", rtpi);
                core_scsi3_put_pr_reg(pr_reg);
                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
        }
+
+       buf = transport_kmap_first_data_page(cmd);
        proto_ident = (buf[24] & 0x0f);
 #if 0
-       printk("SPC-3 PR REGISTER_AND_MOVE: Extracted Protocol Identifier:"
+       pr_debug("SPC-3 PR REGISTER_AND_MOVE: Extracted Protocol Identifier:"
                        " 0x%02x\n", proto_ident);
 #endif
        if (proto_ident != dest_tf_ops->get_fabric_proto_ident(dest_se_tpg)) {
-               printk(KERN_ERR "SPC-3 PR REGISTER_AND_MOVE: Received"
+               pr_err("SPC-3 PR REGISTER_AND_MOVE: Received"
                        " proto_ident: 0x%02x does not match ident: 0x%02x"
                        " from fabric: %s\n", proto_ident,
                        dest_tf_ops->get_fabric_proto_ident(dest_se_tpg),
@@ -3429,7 +3441,7 @@ static int core_scsi3_emulate_pro_register_and_move(
                goto out;
        }
        if (dest_tf_ops->tpg_parse_pr_out_transport_id == NULL) {
-               printk(KERN_ERR "SPC-3 PR REGISTER_AND_MOVE: Fabric does not"
+               pr_err("SPC-3 PR REGISTER_AND_MOVE: Fabric does not"
                        " containg a valid tpg_parse_pr_out_transport_id"
                        " function pointer\n");
                ret = PYX_TRANSPORT_LU_COMM_FAILURE;
@@ -3437,14 +3449,17 @@ static int core_scsi3_emulate_pro_register_and_move(
        }
        initiator_str = dest_tf_ops->tpg_parse_pr_out_transport_id(dest_se_tpg,
                        (const char *)&buf[24], &tmp_tid_len, &iport_ptr);
-       if (!(initiator_str)) {
-               printk(KERN_ERR "SPC-3 PR REGISTER_AND_MOVE: Unable to locate"
+       if (!initiator_str) {
+               pr_err("SPC-3 PR REGISTER_AND_MOVE: Unable to locate"
                        " initiator_str from Transport ID\n");
                ret = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
                goto out;
        }
 
-       printk(KERN_INFO "SPC-3 PR [%s] Extracted initiator %s identifier: %s"
+       transport_kunmap_first_data_page(cmd);
+       buf = NULL;
+
+       pr_debug("SPC-3 PR [%s] Extracted initiator %s identifier: %s"
                " %s\n", dest_tf_ops->get_fabric_name(), (iport_ptr != NULL) ?
                "port" : "device", initiator_str, (iport_ptr != NULL) ?
                iport_ptr : "");
@@ -3459,18 +3474,18 @@ static int core_scsi3_emulate_pro_register_and_move(
        pr_reg_nacl = pr_reg->pr_reg_nacl;
        matching_iname = (!strcmp(initiator_str,
                                  pr_reg_nacl->initiatorname)) ? 1 : 0;
-       if (!(matching_iname))
+       if (!matching_iname)
                goto after_iport_check;
 
-       if (!(iport_ptr) || !(pr_reg->isid_present_at_reg)) {
-               printk(KERN_ERR "SPC-3 PR REGISTER_AND_MOVE: TransportID: %s"
+       if (!iport_ptr || !pr_reg->isid_present_at_reg) {
+               pr_err("SPC-3 PR REGISTER_AND_MOVE: TransportID: %s"
                        " matches: %s on received I_T Nexus\n", initiator_str,
                        pr_reg_nacl->initiatorname);
                ret = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
                goto out;
        }
-       if (!(strcmp(iport_ptr, pr_reg->pr_reg_isid))) {
-               printk(KERN_ERR "SPC-3 PR REGISTER_AND_MOVE: TransportID: %s %s"
+       if (!strcmp(iport_ptr, pr_reg->pr_reg_isid)) {
+               pr_err("SPC-3 PR REGISTER_AND_MOVE: TransportID: %s %s"
                        " matches: %s %s on received I_T Nexus\n",
                        initiator_str, iport_ptr, pr_reg_nacl->initiatorname,
                        pr_reg->pr_reg_isid);
@@ -3490,8 +3505,8 @@ after_iport_check:
        }
        spin_unlock_bh(&dest_se_tpg->acl_node_lock);
 
-       if (!(dest_node_acl)) {
-               printk(KERN_ERR "Unable to locate %s dest_node_acl for"
+       if (!dest_node_acl) {
+               pr_err("Unable to locate %s dest_node_acl for"
                        " TransportID%s\n", dest_tf_ops->get_fabric_name(),
                        initiator_str);
                ret = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
@@ -3499,7 +3514,7 @@ after_iport_check:
        }
        ret = core_scsi3_nodeacl_depend_item(dest_node_acl);
        if (ret != 0) {
-               printk(KERN_ERR "core_scsi3_nodeacl_depend_item() for"
+               pr_err("core_scsi3_nodeacl_depend_item() for"
                        " dest_node_acl\n");
                atomic_dec(&dest_node_acl->acl_pr_ref_count);
                smp_mb__after_atomic_dec();
@@ -3508,7 +3523,7 @@ after_iport_check:
                goto out;
        }
 #if 0
-       printk(KERN_INFO "SPC-3 PR REGISTER_AND_MOVE: Found %s dest_node_acl:"
+       pr_debug("SPC-3 PR REGISTER_AND_MOVE: Found %s dest_node_acl:"
                " %s from TransportID\n", dest_tf_ops->get_fabric_name(),
                dest_node_acl->initiatorname);
 #endif
@@ -3517,8 +3532,8 @@ after_iport_check:
         * PORT IDENTIFIER.
         */
        dest_se_deve = core_get_se_deve_from_rtpi(dest_node_acl, rtpi);
-       if (!(dest_se_deve)) {
-               printk(KERN_ERR "Unable to locate %s dest_se_deve from RTPI:"
+       if (!dest_se_deve) {
+               pr_err("Unable to locate %s dest_se_deve from RTPI:"
                        " %hu\n",  dest_tf_ops->get_fabric_name(), rtpi);
                ret = PYX_TRANSPORT_INVALID_PARAMETER_LIST;
                goto out;
@@ -3526,7 +3541,7 @@ after_iport_check:
 
        ret = core_scsi3_lunacl_depend_item(dest_se_deve);
        if (ret < 0) {
-               printk(KERN_ERR "core_scsi3_lunacl_depend_item() failed\n");
+               pr_err("core_scsi3_lunacl_depend_item() failed\n");
                atomic_dec(&dest_se_deve->pr_ref_count);
                smp_mb__after_atomic_dec();
                dest_se_deve = NULL;
@@ -3534,7 +3549,7 @@ after_iport_check:
                goto out;
        }
 #if 0
-       printk(KERN_INFO "SPC-3 PR REGISTER_AND_MOVE: Located %s node %s LUN"
+       pr_debug("SPC-3 PR REGISTER_AND_MOVE: Located %s node %s LUN"
                " ACL for dest_se_deve->mapped_lun: %u\n",
                dest_tf_ops->get_fabric_name(), dest_node_acl->initiatorname,
                dest_se_deve->mapped_lun);
@@ -3545,8 +3560,8 @@ after_iport_check:
         */
        spin_lock(&dev->dev_reservation_lock);
        pr_res_holder = dev->dev_pr_res_holder;
-       if (!(pr_res_holder)) {
-               printk(KERN_WARNING "SPC-3 PR REGISTER_AND_MOVE: No reservation"
+       if (!pr_res_holder) {
+               pr_warn("SPC-3 PR REGISTER_AND_MOVE: No reservation"
                        " currently held\n");
                spin_unlock(&dev->dev_reservation_lock);
                ret = PYX_TRANSPORT_INVALID_CDB_FIELD;
@@ -3559,7 +3574,7 @@ after_iport_check:
         *      Register behaviors for a REGISTER AND MOVE service action
         */
        if (pr_res_holder != pr_reg) {
-               printk(KERN_WARNING "SPC-3 PR REGISTER_AND_MOVE: Calling I_T"
+               pr_warn("SPC-3 PR REGISTER_AND_MOVE: Calling I_T"
                        " Nexus is not reservation holder\n");
                spin_unlock(&dev->dev_reservation_lock);
                ret = PYX_TRANSPORT_RESERVATION_CONFLICT;
@@ -3576,7 +3591,7 @@ after_iport_check:
         */
        if ((pr_res_holder->pr_res_type == PR_TYPE_WRITE_EXCLUSIVE_ALLREG) ||
            (pr_res_holder->pr_res_type == PR_TYPE_EXCLUSIVE_ACCESS_ALLREG)) {
-               printk(KERN_WARNING "SPC-3 PR REGISTER_AND_MOVE: Unable to move"
+               pr_warn("SPC-3 PR REGISTER_AND_MOVE: Unable to move"
                        " reservation for type: %s\n",
                        core_scsi3_pr_dump_type(pr_res_holder->pr_res_type));
                spin_unlock(&dev->dev_reservation_lock);
@@ -3611,8 +3626,8 @@ after_iport_check:
         */
        dest_pr_reg = __core_scsi3_locate_pr_reg(dev, dest_node_acl,
                                        iport_ptr);
-       if (!(dest_pr_reg)) {
-               ret = core_scsi3_alloc_registration(SE_DEV(cmd),
+       if (!dest_pr_reg) {
+               ret = core_scsi3_alloc_registration(cmd->se_dev,
                                dest_node_acl, dest_se_deve, iport_ptr,
                                sa_res_key, 0, aptpl, 2, 1);
                if (ret != 0) {
@@ -3644,16 +3659,16 @@ after_iport_check:
        /*
         * Increment PRGeneration for existing registrations..
         */
-       if (!(new_reg))
+       if (!new_reg)
                dest_pr_reg->pr_res_generation = pr_tmpl->pr_generation++;
        spin_unlock(&dev->dev_reservation_lock);
 
-       printk(KERN_INFO "SPC-3 PR [%s] Service Action: REGISTER_AND_MOVE"
+       pr_debug("SPC-3 PR [%s] Service Action: REGISTER_AND_MOVE"
                " created new reservation holder TYPE: %s on object RTPI:"
                " %hu  PRGeneration: 0x%08x\n", dest_tf_ops->get_fabric_name(),
                core_scsi3_pr_dump_type(type), rtpi,
                dest_pr_reg->pr_res_generation);
-       printk(KERN_INFO "SPC-3 PR Successfully moved reservation from"
+       pr_debug("SPC-3 PR Successfully moved reservation from"
                " %s Fabric Node: %s%s -> %s Fabric Node: %s %s\n",
                tf_ops->get_fabric_name(), pr_reg_nacl->initiatorname,
                (prf_isid) ? &i_buf[0] : "", dest_tf_ops->get_fabric_name(),
@@ -3681,24 +3696,28 @@ after_iport_check:
         * Clear the APTPL metadata if APTPL has been disabled, otherwise
         * write out the updated metadata to struct file for this SCSI device.
         */
-       if (!(aptpl)) {
+       if (!aptpl) {
                pr_tmpl->pr_aptpl_active = 0;
-               core_scsi3_update_and_write_aptpl(SE_DEV(cmd), NULL, 0);
-               printk("SPC-3 PR: Set APTPL Bit Deactivated for"
+               core_scsi3_update_and_write_aptpl(cmd->se_dev, NULL, 0);
+               pr_debug("SPC-3 PR: Set APTPL Bit Deactivated for"
                                " REGISTER_AND_MOVE\n");
        } else {
                pr_tmpl->pr_aptpl_active = 1;
-               ret = core_scsi3_update_and_write_aptpl(SE_DEV(cmd),
+               ret = core_scsi3_update_and_write_aptpl(cmd->se_dev,
                                &dest_pr_reg->pr_aptpl_buf[0],
                                pr_tmpl->pr_aptpl_buf_len);
-               if (!(ret))
-                       printk("SPC-3 PR: Set APTPL Bit Activated for"
+               if (!ret)
+                       pr_debug("SPC-3 PR: Set APTPL Bit Activated for"
                                        " REGISTER_AND_MOVE\n");
        }
 
+       transport_kunmap_first_data_page(cmd);
+
        core_scsi3_put_pr_reg(dest_pr_reg);
        return 0;
 out:
+       if (buf)
+               transport_kunmap_first_data_page(cmd);
        if (dest_se_deve)
                core_scsi3_lunacl_undepend_item(dest_se_deve);
        if (dest_node_acl)
@@ -3723,7 +3742,7 @@ static unsigned long long core_scsi3_extract_reservation_key(unsigned char *cdb)
  */
 static int core_scsi3_emulate_pr_out(struct se_cmd *cmd, unsigned char *cdb)
 {
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       unsigned char *buf;
        u64 res_key, sa_res_key;
        int sa, scope, type, aptpl;
        int spec_i_pt = 0, all_tg_pt = 0, unreg = 0;
@@ -3731,11 +3750,11 @@ static int core_scsi3_emulate_pr_out(struct se_cmd *cmd, unsigned char *cdb)
         * FIXME: A NULL struct se_session pointer means an this is not coming from
         * a $FABRIC_MOD's nexus, but from internal passthrough ops.
         */
-       if (!(SE_SESS(cmd)))
+       if (!cmd->se_sess)
                return PYX_TRANSPORT_LU_COMM_FAILURE;
 
        if (cmd->data_length < 24) {
-               printk(KERN_WARNING "SPC-PR: Received PR OUT parameter list"
+               pr_warn("SPC-PR: Received PR OUT parameter list"
                        " length too small: %u\n", cmd->data_length);
                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
        }
@@ -3745,6 +3764,8 @@ static int core_scsi3_emulate_pr_out(struct se_cmd *cmd, unsigned char *cdb)
        sa = (cdb[1] & 0x1f);
        scope = (cdb[2] & 0xf0);
        type = (cdb[2] & 0x0f);
+
+       buf = transport_kmap_first_data_page(cmd);
        /*
         * From PERSISTENT_RESERVE_OUT parameter list (payload)
         */
@@ -3762,6 +3783,9 @@ static int core_scsi3_emulate_pr_out(struct se_cmd *cmd, unsigned char *cdb)
                aptpl = (buf[17] & 0x01);
                unreg = (buf[17] & 0x02);
        }
+       transport_kunmap_first_data_page(cmd);
+       buf = NULL;
+
        /*
         * SPEC_I_PT=1 is only valid for Service action: REGISTER
         */
@@ -3776,9 +3800,9 @@ static int core_scsi3_emulate_pr_out(struct se_cmd *cmd, unsigned char *cdb)
         * the sense key set to ILLEGAL REQUEST, and the additional sense
         * code set to PARAMETER LIST LENGTH ERROR.
         */
-       if (!(spec_i_pt) && ((cdb[1] & 0x1f) != PRO_REGISTER_AND_MOVE) &&
+       if (!spec_i_pt && ((cdb[1] & 0x1f) != PRO_REGISTER_AND_MOVE) &&
            (cmd->data_length != 24)) {
-               printk(KERN_WARNING "SPC-PR: Received PR OUT illegal parameter"
+               pr_warn("SPC-PR: Received PR OUT illegal parameter"
                        " list length: %u\n", cmd->data_length);
                return PYX_TRANSPORT_INVALID_PARAMETER_LIST;
        }
@@ -3812,7 +3836,7 @@ static int core_scsi3_emulate_pr_out(struct se_cmd *cmd, unsigned char *cdb)
                return core_scsi3_emulate_pro_register_and_move(cmd, res_key,
                                sa_res_key, aptpl, unreg);
        default:
-               printk(KERN_ERR "Unknown PERSISTENT_RESERVE_OUT service"
+               pr_err("Unknown PERSISTENT_RESERVE_OUT service"
                        " action: 0x%02x\n", cdb[1] & 0x1f);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
@@ -3827,25 +3851,26 @@ static int core_scsi3_emulate_pr_out(struct se_cmd *cmd, unsigned char *cdb)
  */
 static int core_scsi3_pri_read_keys(struct se_cmd *cmd)
 {
-       struct se_device *se_dev = SE_DEV(cmd);
-       struct se_subsystem_dev *su_dev = SU_DEV(se_dev);
+       struct se_device *se_dev = cmd->se_dev;
+       struct se_subsystem_dev *su_dev = se_dev->se_sub_dev;
        struct t10_pr_registration *pr_reg;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       unsigned char *buf;
        u32 add_len = 0, off = 8;
 
        if (cmd->data_length < 8) {
-               printk(KERN_ERR "PRIN SA READ_KEYS SCSI Data Length: %u"
+               pr_err("PRIN SA READ_KEYS SCSI Data Length: %u"
                        " too small\n", cmd->data_length);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
 
-       buf[0] = ((T10_RES(su_dev)->pr_generation >> 24) & 0xff);
-       buf[1] = ((T10_RES(su_dev)->pr_generation >> 16) & 0xff);
-       buf[2] = ((T10_RES(su_dev)->pr_generation >> 8) & 0xff);
-       buf[3] = (T10_RES(su_dev)->pr_generation & 0xff);
+       buf = transport_kmap_first_data_page(cmd);
+       buf[0] = ((su_dev->t10_pr.pr_generation >> 24) & 0xff);
+       buf[1] = ((su_dev->t10_pr.pr_generation >> 16) & 0xff);
+       buf[2] = ((su_dev->t10_pr.pr_generation >> 8) & 0xff);
+       buf[3] = (su_dev->t10_pr.pr_generation & 0xff);
 
-       spin_lock(&T10_RES(su_dev)->registration_lock);
-       list_for_each_entry(pr_reg, &T10_RES(su_dev)->registration_list,
+       spin_lock(&su_dev->t10_pr.registration_lock);
+       list_for_each_entry(pr_reg, &su_dev->t10_pr.registration_list,
                        pr_reg_list) {
                /*
                 * Check for overflow of 8byte PRI READ_KEYS payload and
@@ -3865,13 +3890,15 @@ static int core_scsi3_pri_read_keys(struct se_cmd *cmd)
 
                add_len += 8;
        }
-       spin_unlock(&T10_RES(su_dev)->registration_lock);
+       spin_unlock(&su_dev->t10_pr.registration_lock);
 
        buf[4] = ((add_len >> 24) & 0xff);
        buf[5] = ((add_len >> 16) & 0xff);
        buf[6] = ((add_len >> 8) & 0xff);
        buf[7] = (add_len & 0xff);
 
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
@@ -3882,23 +3909,24 @@ static int core_scsi3_pri_read_keys(struct se_cmd *cmd)
  */
 static int core_scsi3_pri_read_reservation(struct se_cmd *cmd)
 {
-       struct se_device *se_dev = SE_DEV(cmd);
-       struct se_subsystem_dev *su_dev = SU_DEV(se_dev);
+       struct se_device *se_dev = cmd->se_dev;
+       struct se_subsystem_dev *su_dev = se_dev->se_sub_dev;
        struct t10_pr_registration *pr_reg;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       unsigned char *buf;
        u64 pr_res_key;
        u32 add_len = 16; /* Hardcoded to 16 when a reservation is held. */
 
        if (cmd->data_length < 8) {
-               printk(KERN_ERR "PRIN SA READ_RESERVATIONS SCSI Data Length: %u"
+               pr_err("PRIN SA READ_RESERVATIONS SCSI Data Length: %u"
                        " too small\n", cmd->data_length);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
 
-       buf[0] = ((T10_RES(su_dev)->pr_generation >> 24) & 0xff);
-       buf[1] = ((T10_RES(su_dev)->pr_generation >> 16) & 0xff);
-       buf[2] = ((T10_RES(su_dev)->pr_generation >> 8) & 0xff);
-       buf[3] = (T10_RES(su_dev)->pr_generation & 0xff);
+       buf = transport_kmap_first_data_page(cmd);
+       buf[0] = ((su_dev->t10_pr.pr_generation >> 24) & 0xff);
+       buf[1] = ((su_dev->t10_pr.pr_generation >> 16) & 0xff);
+       buf[2] = ((su_dev->t10_pr.pr_generation >> 8) & 0xff);
+       buf[3] = (su_dev->t10_pr.pr_generation & 0xff);
 
        spin_lock(&se_dev->dev_reservation_lock);
        pr_reg = se_dev->dev_pr_res_holder;
@@ -3911,10 +3939,9 @@ static int core_scsi3_pri_read_reservation(struct se_cmd *cmd)
                buf[6] = ((add_len >> 8) & 0xff);
                buf[7] = (add_len & 0xff);
 
-               if (cmd->data_length < 22) {
-                       spin_unlock(&se_dev->dev_reservation_lock);
-                       return 0;
-               }
+               if (cmd->data_length < 22)
+                       goto err;
+
                /*
                 * Set the Reservation key.
                 *
@@ -3951,7 +3978,10 @@ static int core_scsi3_pri_read_reservation(struct se_cmd *cmd)
                buf[21] = (pr_reg->pr_res_scope & 0xf0) |
                          (pr_reg->pr_res_type & 0x0f);
        }
+
+err:
        spin_unlock(&se_dev->dev_reservation_lock);
+       transport_kunmap_first_data_page(cmd);
 
        return 0;
 }
@@ -3963,17 +3993,19 @@ static int core_scsi3_pri_read_reservation(struct se_cmd *cmd)
  */
 static int core_scsi3_pri_report_capabilities(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(dev)->t10_reservation;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       struct se_device *dev = cmd->se_dev;
+       struct t10_reservation *pr_tmpl = &dev->se_sub_dev->t10_pr;
+       unsigned char *buf;
        u16 add_len = 8; /* Hardcoded to 8. */
 
        if (cmd->data_length < 6) {
-               printk(KERN_ERR "PRIN SA REPORT_CAPABILITIES SCSI Data Length:"
+               pr_err("PRIN SA REPORT_CAPABILITIES SCSI Data Length:"
                        " %u too small\n", cmd->data_length);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
 
+       buf = transport_kmap_first_data_page(cmd);
+
        buf[0] = ((add_len << 8) & 0xff);
        buf[1] = (add_len & 0xff);
        buf[2] |= 0x10; /* CRH: Compatible Reservation Hanlding bit. */
@@ -4004,6 +4036,8 @@ static int core_scsi3_pri_report_capabilities(struct se_cmd *cmd)
        buf[4] |= 0x02; /* PR_TYPE_WRITE_EXCLUSIVE */
        buf[5] |= 0x01; /* PR_TYPE_EXCLUSIVE_ACCESS_ALLREG */
 
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
@@ -4014,27 +4048,29 @@ static int core_scsi3_pri_report_capabilities(struct se_cmd *cmd)
  */
 static int core_scsi3_pri_read_full_status(struct se_cmd *cmd)
 {
-       struct se_device *se_dev = SE_DEV(cmd);
+       struct se_device *se_dev = cmd->se_dev;
        struct se_node_acl *se_nacl;
-       struct se_subsystem_dev *su_dev = SU_DEV(se_dev);
+       struct se_subsystem_dev *su_dev = se_dev->se_sub_dev;
        struct se_portal_group *se_tpg;
        struct t10_pr_registration *pr_reg, *pr_reg_tmp;
-       struct t10_reservation_template *pr_tmpl = &SU_DEV(se_dev)->t10_reservation;
-       unsigned char *buf = (unsigned char *)T_TASK(cmd)->t_task_buf;
+       struct t10_reservation *pr_tmpl = &se_dev->se_sub_dev->t10_pr;
+       unsigned char *buf;
        u32 add_desc_len = 0, add_len = 0, desc_len, exp_desc_len;
        u32 off = 8; /* off into first Full Status descriptor */
        int format_code = 0;
 
        if (cmd->data_length < 8) {
-               printk(KERN_ERR "PRIN SA READ_FULL_STATUS SCSI Data Length: %u"
+               pr_err("PRIN SA READ_FULL_STATUS SCSI Data Length: %u"
                        " too small\n", cmd->data_length);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
 
-       buf[0] = ((T10_RES(su_dev)->pr_generation >> 24) & 0xff);
-       buf[1] = ((T10_RES(su_dev)->pr_generation >> 16) & 0xff);
-       buf[2] = ((T10_RES(su_dev)->pr_generation >> 8) & 0xff);
-       buf[3] = (T10_RES(su_dev)->pr_generation & 0xff);
+       buf = transport_kmap_first_data_page(cmd);
+
+       buf[0] = ((su_dev->t10_pr.pr_generation >> 24) & 0xff);
+       buf[1] = ((su_dev->t10_pr.pr_generation >> 16) & 0xff);
+       buf[2] = ((su_dev->t10_pr.pr_generation >> 8) & 0xff);
+       buf[3] = (su_dev->t10_pr.pr_generation & 0xff);
 
        spin_lock(&pr_tmpl->registration_lock);
        list_for_each_entry_safe(pr_reg, pr_reg_tmp,
@@ -4051,11 +4087,11 @@ static int core_scsi3_pri_read_full_status(struct se_cmd *cmd)
                 * Determine expected length of $FABRIC_MOD specific
                 * TransportID full status descriptor..
                 */
-               exp_desc_len = TPG_TFO(se_tpg)->tpg_get_pr_transport_id_len(
+               exp_desc_len = se_tpg->se_tpg_tfo->tpg_get_pr_transport_id_len(
                                se_tpg, se_nacl, pr_reg, &format_code);
 
                if ((exp_desc_len + add_len) > cmd->data_length) {
-                       printk(KERN_WARNING "SPC-3 PRIN READ_FULL_STATUS ran"
+                       pr_warn("SPC-3 PRIN READ_FULL_STATUS ran"
                                " out of buffer: %d\n", cmd->data_length);
                        spin_lock(&pr_tmpl->registration_lock);
                        atomic_dec(&pr_reg->pr_res_holders);
@@ -4105,7 +4141,7 @@ static int core_scsi3_pri_read_full_status(struct se_cmd *cmd)
                 * bit is set to one, the contents of the RELATIVE TARGET PORT
                 * IDENTIFIER field are not defined by this standard.
                 */
-               if (!(pr_reg->pr_reg_all_tg_pt)) {
+               if (!pr_reg->pr_reg_all_tg_pt) {
                        struct se_port *port = pr_reg->pr_reg_tg_pt_lun->lun_sep;
 
                        buf[off++] = ((port->sep_rtpi >> 8) & 0xff);
@@ -4116,7 +4152,7 @@ static int core_scsi3_pri_read_full_status(struct se_cmd *cmd)
                /*
                 * Now, have the $FABRIC_MOD fill in the protocol identifier
                 */
-               desc_len = TPG_TFO(se_tpg)->tpg_get_pr_transport_id(se_tpg,
+               desc_len = se_tpg->se_tpg_tfo->tpg_get_pr_transport_id(se_tpg,
                                se_nacl, pr_reg, &format_code, &buf[off+4]);
 
                spin_lock(&pr_tmpl->registration_lock);
@@ -4150,6 +4186,8 @@ static int core_scsi3_pri_read_full_status(struct se_cmd *cmd)
        buf[6] = ((add_len >> 8) & 0xff);
        buf[7] = (add_len & 0xff);
 
+       transport_kunmap_first_data_page(cmd);
+
        return 0;
 }
 
@@ -4165,7 +4203,7 @@ static int core_scsi3_emulate_pr_in(struct se_cmd *cmd, unsigned char *cdb)
        case PRI_READ_FULL_STATUS:
                return core_scsi3_pri_read_full_status(cmd);
        default:
-               printk(KERN_ERR "Unknown PERSISTENT_RESERVE_IN service"
+               pr_err("Unknown PERSISTENT_RESERVE_IN service"
                        " action: 0x%02x\n", cdb[1] & 0x1f);
                return PYX_TRANSPORT_INVALID_CDB_FIELD;
        }
@@ -4174,7 +4212,7 @@ static int core_scsi3_emulate_pr_in(struct se_cmd *cmd, unsigned char *cdb)
 
 int core_scsi3_emulate_pr(struct se_cmd *cmd)
 {
-       unsigned char *cdb = &T_TASK(cmd)->t_task_cdb[0];
+       unsigned char *cdb = &cmd->t_task_cdb[0];
        struct se_device *dev = cmd->se_dev;
        /*
         * Following spc2r20 5.5.1 Reservations overview:
@@ -4186,7 +4224,7 @@ int core_scsi3_emulate_pr(struct se_cmd *cmd)
         * CONFLICT status.
         */
        if (dev->dev_flags & DF_SPC2_RESERVATIONS) {
-               printk(KERN_ERR "Received PERSISTENT_RESERVE CDB while legacy"
+               pr_err("Received PERSISTENT_RESERVE CDB while legacy"
                        " SPC-2 reservation is held, returning"
                        " RESERVATION_CONFLICT\n");
                return PYX_TRANSPORT_RESERVATION_CONFLICT;
@@ -4213,39 +4251,39 @@ static int core_pt_seq_non_holder(
 int core_setup_reservations(struct se_device *dev, int force_pt)
 {
        struct se_subsystem_dev *su_dev = dev->se_sub_dev;
-       struct t10_reservation_template *rest = &su_dev->t10_reservation;
+       struct t10_reservation *rest = &su_dev->t10_pr;
        /*
         * If this device is from Target_Core_Mod/pSCSI, use the reservations
         * of the Underlying SCSI hardware.  In Linux/SCSI terms, this can
         * cause a problem because libata and some SATA RAID HBAs appear
         * under Linux/SCSI, but to emulate reservations themselves.
         */
-       if (((TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) &&
-           !(DEV_ATTRIB(dev)->emulate_reservations)) || force_pt) {
+       if (((dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) &&
+           !(dev->se_sub_dev->se_dev_attrib.emulate_reservations)) || force_pt) {
                rest->res_type = SPC_PASSTHROUGH;
                rest->pr_ops.t10_reservation_check = &core_pt_reservation_check;
                rest->pr_ops.t10_seq_non_holder = &core_pt_seq_non_holder;
-               printk(KERN_INFO "%s: Using SPC_PASSTHROUGH, no reservation"
-                       " emulation\n", TRANSPORT(dev)->name);
+               pr_debug("%s: Using SPC_PASSTHROUGH, no reservation"
+                       " emulation\n", dev->transport->name);
                return 0;
        }
        /*
         * If SPC-3 or above is reported by real or emulated struct se_device,
         * use emulated Persistent Reservations.
         */
-       if (TRANSPORT(dev)->get_device_rev(dev) >= SCSI_3) {
+       if (dev->transport->get_device_rev(dev) >= SCSI_3) {
                rest->res_type = SPC3_PERSISTENT_RESERVATIONS;
                rest->pr_ops.t10_reservation_check = &core_scsi3_pr_reservation_check;
                rest->pr_ops.t10_seq_non_holder = &core_scsi3_pr_seq_non_holder;
-               printk(KERN_INFO "%s: Using SPC3_PERSISTENT_RESERVATIONS"
-                       " emulation\n", TRANSPORT(dev)->name);
+               pr_debug("%s: Using SPC3_PERSISTENT_RESERVATIONS"
+                       " emulation\n", dev->transport->name);
        } else {
                rest->res_type = SPC2_RESERVATIONS;
                rest->pr_ops.t10_reservation_check = &core_scsi2_reservation_check;
                rest->pr_ops.t10_seq_non_holder =
                                &core_scsi2_reservation_seq_non_holder;
-               printk(KERN_INFO "%s: Using SPC2_RESERVATIONS emulation\n",
-                       TRANSPORT(dev)->name);
+               pr_debug("%s: Using SPC2_RESERVATIONS emulation\n",
+                       dev->transport->name);
        }
 
        return 0;
index 5603bcfd86d3c033dd1642aa41464f9d84e59970..c8f47d06458494a3ecb9239a9259d2939ab279fa 100644 (file)
@@ -49,7 +49,7 @@ extern int core_pr_dump_initiator_port(struct t10_pr_registration *,
                        char *, u32);
 extern int core_scsi2_emulate_crh(struct se_cmd *);
 extern int core_scsi3_alloc_aptpl_registration(
-                       struct t10_reservation_template *, u64,
+                       struct t10_reservation *, u64,
                        unsigned char *, unsigned char *, u32,
                        unsigned char *, u16, u32, int, int, u8);
 extern int core_scsi3_check_aptpl_registration(struct se_device *,
index 331d423fd0e085bbc65949201661ddbfd06fca91..2b7b0da9146d3aa9ccf804975d6047afa70166e9 100644 (file)
@@ -55,24 +55,6 @@ static struct se_subsystem_api pscsi_template;
 
 static void pscsi_req_done(struct request *, int);
 
-/*     pscsi_get_sh():
- *
- *
- */
-static struct Scsi_Host *pscsi_get_sh(u32 host_no)
-{
-       struct Scsi_Host *sh = NULL;
-
-       sh = scsi_host_lookup(host_no);
-       if (IS_ERR(sh)) {
-               printk(KERN_ERR "Unable to locate SCSI HBA with Host ID:"
-                               " %u\n", host_no);
-               return NULL;
-       }
-
-       return sh;
-}
-
 /*     pscsi_attach_hba():
  *
  *     pscsi_get_sh() used scsi_host_lookup() to locate struct Scsi_Host.
@@ -80,28 +62,23 @@ static struct Scsi_Host *pscsi_get_sh(u32 host_no)
  */
 static int pscsi_attach_hba(struct se_hba *hba, u32 host_id)
 {
-       int hba_depth;
        struct pscsi_hba_virt *phv;
 
        phv = kzalloc(sizeof(struct pscsi_hba_virt), GFP_KERNEL);
-       if (!(phv)) {
-               printk(KERN_ERR "Unable to allocate struct pscsi_hba_virt\n");
-               return -1;
+       if (!phv) {
+               pr_err("Unable to allocate struct pscsi_hba_virt\n");
+               return -ENOMEM;
        }
        phv->phv_host_id = host_id;
        phv->phv_mode = PHV_VIRUTAL_HOST_ID;
-       hba_depth = PSCSI_VIRTUAL_HBA_DEPTH;
-       atomic_set(&hba->left_queue_depth, hba_depth);
-       atomic_set(&hba->max_queue_depth, hba_depth);
 
-       hba->hba_ptr = (void *)phv;
+       hba->hba_ptr = phv;
 
-       printk(KERN_INFO "CORE_HBA[%d] - TCM SCSI HBA Driver %s on"
+       pr_debug("CORE_HBA[%d] - TCM SCSI HBA Driver %s on"
                " Generic Target Core Stack %s\n", hba->hba_id,
                PSCSI_VERSION, TARGET_CORE_MOD_VERSION);
-       printk(KERN_INFO "CORE_HBA[%d] - Attached SCSI HBA to Generic"
-               " Target Core with TCQ Depth: %d\n", hba->hba_id,
-               atomic_read(&hba->max_queue_depth));
+       pr_debug("CORE_HBA[%d] - Attached SCSI HBA to Generic\n",
+              hba->hba_id);
 
        return 0;
 }
@@ -114,12 +91,12 @@ static void pscsi_detach_hba(struct se_hba *hba)
        if (scsi_host) {
                scsi_host_put(scsi_host);
 
-               printk(KERN_INFO "CORE_HBA[%d] - Detached SCSI HBA: %s from"
+               pr_debug("CORE_HBA[%d] - Detached SCSI HBA: %s from"
                        " Generic Target Core\n", hba->hba_id,
                        (scsi_host->hostt->name) ? (scsi_host->hostt->name) :
                        "Unknown");
        } else
-               printk(KERN_INFO "CORE_HBA[%d] - Detached Virtual SCSI HBA"
+               pr_debug("CORE_HBA[%d] - Detached Virtual SCSI HBA"
                        " from Generic Target Core\n", hba->hba_id);
 
        kfree(phv);
@@ -130,20 +107,17 @@ static int pscsi_pmode_enable_hba(struct se_hba *hba, unsigned long mode_flag)
 {
        struct pscsi_hba_virt *phv = (struct pscsi_hba_virt *)hba->hba_ptr;
        struct Scsi_Host *sh = phv->phv_lld_host;
-       int hba_depth = PSCSI_VIRTUAL_HBA_DEPTH;
        /*
         * Release the struct Scsi_Host
         */
-       if (!(mode_flag)) {
-               if (!(sh))
+       if (!mode_flag) {
+               if (!sh)
                        return 0;
 
                phv->phv_lld_host = NULL;
                phv->phv_mode = PHV_VIRUTAL_HOST_ID;
-               atomic_set(&hba->left_queue_depth, hba_depth);
-               atomic_set(&hba->max_queue_depth, hba_depth);
 
-               printk(KERN_INFO "CORE_HBA[%d] - Disabled pSCSI HBA Passthrough"
+               pr_debug("CORE_HBA[%d] - Disabled pSCSI HBA Passthrough"
                        " %s\n", hba->hba_id, (sh->hostt->name) ?
                        (sh->hostt->name) : "Unknown");
 
@@ -154,27 +128,17 @@ static int pscsi_pmode_enable_hba(struct se_hba *hba, unsigned long mode_flag)
         * Otherwise, locate struct Scsi_Host from the original passed
         * pSCSI Host ID and enable for phba mode
         */
-       sh = pscsi_get_sh(phv->phv_host_id);
-       if (!(sh)) {
-               printk(KERN_ERR "pSCSI: Unable to locate SCSI Host for"
+       sh = scsi_host_lookup(phv->phv_host_id);
+       if (IS_ERR(sh)) {
+               pr_err("pSCSI: Unable to locate SCSI Host for"
                        " phv_host_id: %d\n", phv->phv_host_id);
-               return -1;
+               return PTR_ERR(sh);
        }
-       /*
-        * Usually the SCSI LLD will use the hostt->can_queue value to define
-        * its HBA TCQ depth.  Some other drivers (like 2.6 megaraid) don't set
-        * this at all and set sh->can_queue at runtime.
-        */
-       hba_depth = (sh->hostt->can_queue > sh->can_queue) ?
-               sh->hostt->can_queue : sh->can_queue;
-
-       atomic_set(&hba->left_queue_depth, hba_depth);
-       atomic_set(&hba->max_queue_depth, hba_depth);
 
        phv->phv_lld_host = sh;
        phv->phv_mode = PHV_LLD_SCSI_HOST_NO;
 
-       printk(KERN_INFO "CORE_HBA[%d] - Enabled pSCSI HBA Passthrough %s\n",
+       pr_debug("CORE_HBA[%d] - Enabled pSCSI HBA Passthrough %s\n",
                hba->hba_id, (sh->hostt->name) ? (sh->hostt->name) : "Unknown");
 
        return 1;
@@ -236,7 +200,7 @@ pscsi_get_inquiry_vpd_serial(struct scsi_device *sdev, struct t10_wwn *wwn)
 
        buf = kzalloc(INQUIRY_VPD_SERIAL_LEN, GFP_KERNEL);
        if (!buf)
-               return -1;
+               return -ENOMEM;
 
        memset(cdb, 0, MAX_COMMAND_SIZE);
        cdb[0] = INQUIRY;
@@ -259,7 +223,7 @@ pscsi_get_inquiry_vpd_serial(struct scsi_device *sdev, struct t10_wwn *wwn)
 
 out_free:
        kfree(buf);
-       return -1;
+       return -EPERM;
 }
 
 static void
@@ -293,15 +257,15 @@ pscsi_get_inquiry_vpd_device_ident(struct scsi_device *sdev,
                page_83 = &buf[off];
                ident_len = page_83[3];
                if (!ident_len) {
-                       printk(KERN_ERR "page_83[3]: identifier"
+                       pr_err("page_83[3]: identifier"
                                        " length zero!\n");
                        break;
                }
-               printk(KERN_INFO "T10 VPD Identifer Length: %d\n", ident_len);
+               pr_debug("T10 VPD Identifer Length: %d\n", ident_len);
 
                vpd = kzalloc(sizeof(struct t10_vpd), GFP_KERNEL);
                if (!vpd) {
-                       printk(KERN_ERR "Unable to allocate memory for"
+                       pr_err("Unable to allocate memory for"
                                        " struct t10_vpd\n");
                        goto out;
                }
@@ -353,7 +317,7 @@ static struct se_device *pscsi_add_device_to_list(
        if (!sd->queue_depth) {
                sd->queue_depth = PSCSI_DEFAULT_QUEUEDEPTH;
 
-               printk(KERN_ERR "Set broken SCSI Device %d:%d:%d"
+               pr_err("Set broken SCSI Device %d:%d:%d"
                        " queue_depth to %d\n", sd->channel, sd->id,
                                sd->lun, sd->queue_depth);
        }
@@ -364,10 +328,8 @@ static struct se_device *pscsi_add_device_to_list(
        q = sd->request_queue;
        limits = &dev_limits.limits;
        limits->logical_block_size = sd->sector_size;
-       limits->max_hw_sectors = (sd->host->max_sectors > queue_max_hw_sectors(q)) ?
-                                 queue_max_hw_sectors(q) : sd->host->max_sectors;
-       limits->max_sectors = (sd->host->max_sectors > queue_max_sectors(q)) ?
-                                 queue_max_sectors(q) : sd->host->max_sectors;
+       limits->max_hw_sectors = min_t(int, sd->host->max_sectors, queue_max_hw_sectors(q));
+       limits->max_sectors = min_t(int, sd->host->max_sectors, queue_max_sectors(q));
        dev_limits.hw_queue_depth = sd->queue_depth;
        dev_limits.queue_depth = sd->queue_depth;
        /*
@@ -391,9 +353,9 @@ static struct se_device *pscsi_add_device_to_list(
        pdv->pdv_sd = sd;
 
        dev = transport_add_device_to_core_hba(hba, &pscsi_template,
-                               se_dev, dev_flags, (void *)pdv,
+                               se_dev, dev_flags, pdv,
                                &dev_limits, NULL, NULL);
-       if (!(dev)) {
+       if (!dev) {
                pdv->pdv_sd = NULL;
                return NULL;
        }
@@ -423,14 +385,14 @@ static void *pscsi_allocate_virtdevice(struct se_hba *hba, const char *name)
        struct pscsi_dev_virt *pdv;
 
        pdv = kzalloc(sizeof(struct pscsi_dev_virt), GFP_KERNEL);
-       if (!(pdv)) {
-               printk(KERN_ERR "Unable to allocate memory for struct pscsi_dev_virt\n");
+       if (!pdv) {
+               pr_err("Unable to allocate memory for struct pscsi_dev_virt\n");
                return NULL;
        }
        pdv->pdv_se_hba = hba;
 
-       printk(KERN_INFO "PSCSI: Allocated pdv: %p for %s\n", pdv, name);
-       return (void *)pdv;
+       pr_debug("PSCSI: Allocated pdv: %p for %s\n", pdv, name);
+       return pdv;
 }
 
 /*
@@ -450,7 +412,7 @@ static struct se_device *pscsi_create_type_disk(
        u32 dev_flags = 0;
 
        if (scsi_device_get(sd)) {
-               printk(KERN_ERR "scsi_device_get() failed for %d:%d:%d:%d\n",
+               pr_err("scsi_device_get() failed for %d:%d:%d:%d\n",
                        sh->host_no, sd->channel, sd->id, sd->lun);
                spin_unlock_irq(sh->host_lock);
                return NULL;
@@ -463,19 +425,19 @@ static struct se_device *pscsi_create_type_disk(
        bd = blkdev_get_by_path(se_dev->se_dev_udev_path,
                                FMODE_WRITE|FMODE_READ|FMODE_EXCL, pdv);
        if (IS_ERR(bd)) {
-               printk(KERN_ERR "pSCSI: blkdev_get_by_path() failed\n");
+               pr_err("pSCSI: blkdev_get_by_path() failed\n");
                scsi_device_put(sd);
                return NULL;
        }
        pdv->pdv_bd = bd;
 
        dev = pscsi_add_device_to_list(hba, se_dev, pdv, sd, dev_flags);
-       if (!(dev)) {
+       if (!dev) {
                blkdev_put(pdv->pdv_bd, FMODE_WRITE|FMODE_READ|FMODE_EXCL);
                scsi_device_put(sd);
                return NULL;
        }
-       printk(KERN_INFO "CORE_PSCSI[%d] - Added TYPE_DISK for %d:%d:%d:%d\n",
+       pr_debug("CORE_PSCSI[%d] - Added TYPE_DISK for %d:%d:%d:%d\n",
                phv->phv_host_id, sh->host_no, sd->channel, sd->id, sd->lun);
 
        return dev;
@@ -497,7 +459,7 @@ static struct se_device *pscsi_create_type_rom(
        u32 dev_flags = 0;
 
        if (scsi_device_get(sd)) {
-               printk(KERN_ERR "scsi_device_get() failed for %d:%d:%d:%d\n",
+               pr_err("scsi_device_get() failed for %d:%d:%d:%d\n",
                        sh->host_no, sd->channel, sd->id, sd->lun);
                spin_unlock_irq(sh->host_lock);
                return NULL;
@@ -505,11 +467,11 @@ static struct se_device *pscsi_create_type_rom(
        spin_unlock_irq(sh->host_lock);
 
        dev = pscsi_add_device_to_list(hba, se_dev, pdv, sd, dev_flags);
-       if (!(dev)) {
+       if (!dev) {
                scsi_device_put(sd);
                return NULL;
        }
-       printk(KERN_INFO "CORE_PSCSI[%d] - Added Type: %s for %d:%d:%d:%d\n",
+       pr_debug("CORE_PSCSI[%d] - Added Type: %s for %d:%d:%d:%d\n",
                phv->phv_host_id, scsi_device_type(sd->type), sh->host_no,
                sd->channel, sd->id, sd->lun);
 
@@ -533,10 +495,10 @@ static struct se_device *pscsi_create_type_other(
 
        spin_unlock_irq(sh->host_lock);
        dev = pscsi_add_device_to_list(hba, se_dev, pdv, sd, dev_flags);
-       if (!(dev))
+       if (!dev)
                return NULL;
 
-       printk(KERN_INFO "CORE_PSCSI[%d] - Added Type: %s for %d:%d:%d:%d\n",
+       pr_debug("CORE_PSCSI[%d] - Added Type: %s for %d:%d:%d:%d\n",
                phv->phv_host_id, scsi_device_type(sd->type), sh->host_no,
                sd->channel, sd->id, sd->lun);
 
@@ -555,8 +517,8 @@ static struct se_device *pscsi_create_virtdevice(
        struct Scsi_Host *sh = phv->phv_lld_host;
        int legacy_mode_enable = 0;
 
-       if (!(pdv)) {
-               printk(KERN_ERR "Unable to locate struct pscsi_dev_virt"
+       if (!pdv) {
+               pr_err("Unable to locate struct pscsi_dev_virt"
                                " parameter\n");
                return ERR_PTR(-EINVAL);
        }
@@ -564,9 +526,9 @@ static struct se_device *pscsi_create_virtdevice(
         * If not running in PHV_LLD_SCSI_HOST_NO mode, locate the
         * struct Scsi_Host we will need to bring the TCM/pSCSI object online
         */
-       if (!(sh)) {
+       if (!sh) {
                if (phv->phv_mode == PHV_LLD_SCSI_HOST_NO) {
-                       printk(KERN_ERR "pSCSI: Unable to locate struct"
+                       pr_err("pSCSI: Unable to locate struct"
                                " Scsi_Host for PHV_LLD_SCSI_HOST_NO\n");
                        return ERR_PTR(-ENODEV);
                }
@@ -575,7 +537,7 @@ static struct se_device *pscsi_create_virtdevice(
                 * reference, we enforce that udev_path has been set
                 */
                if (!(se_dev->su_dev_flags & SDF_USING_UDEV_PATH)) {
-                       printk(KERN_ERR "pSCSI: udev_path attribute has not"
+                       pr_err("pSCSI: udev_path attribute has not"
                                " been set before ENABLE=1\n");
                        return ERR_PTR(-EINVAL);
                }
@@ -586,8 +548,8 @@ static struct se_device *pscsi_create_virtdevice(
                 */
                if (!(pdv->pdv_flags & PDF_HAS_VIRT_HOST_ID)) {
                        spin_lock(&hba->device_lock);
-                       if (!(list_empty(&hba->hba_dev_list))) {
-                               printk(KERN_ERR "pSCSI: Unable to set hba_mode"
+                       if (!list_empty(&hba->hba_dev_list)) {
+                               pr_err("pSCSI: Unable to set hba_mode"
                                        " with active devices\n");
                                spin_unlock(&hba->device_lock);
                                return ERR_PTR(-EEXIST);
@@ -601,16 +563,16 @@ static struct se_device *pscsi_create_virtdevice(
                        hba->hba_flags |= HBA_FLAGS_PSCSI_MODE;
                        sh = phv->phv_lld_host;
                } else {
-                       sh = pscsi_get_sh(pdv->pdv_host_id);
-                       if (!(sh)) {
-                               printk(KERN_ERR "pSCSI: Unable to locate"
+                       sh = scsi_host_lookup(pdv->pdv_host_id);
+                       if (IS_ERR(sh)) {
+                               pr_err("pSCSI: Unable to locate"
                                        " pdv_host_id: %d\n", pdv->pdv_host_id);
-                               return ERR_PTR(-ENODEV);
+                               return (struct se_device *) sh;
                        }
                }
        } else {
                if (phv->phv_mode == PHV_VIRUTAL_HOST_ID) {
-                       printk(KERN_ERR "pSCSI: PHV_VIRUTAL_HOST_ID set while"
+                       pr_err("pSCSI: PHV_VIRUTAL_HOST_ID set while"
                                " struct Scsi_Host exists\n");
                        return ERR_PTR(-EEXIST);
                }
@@ -639,7 +601,7 @@ static struct se_device *pscsi_create_virtdevice(
                        break;
                }
 
-               if (!(dev)) {
+               if (!dev) {
                        if (phv->phv_mode == PHV_VIRUTAL_HOST_ID)
                                scsi_host_put(sh);
                        else if (legacy_mode_enable) {
@@ -653,7 +615,7 @@ static struct se_device *pscsi_create_virtdevice(
        }
        spin_unlock_irq(sh->host_lock);
 
-       printk(KERN_ERR "pSCSI: Unable to locate %d:%d:%d:%d\n", sh->host_no,
+       pr_err("pSCSI: Unable to locate %d:%d:%d:%d\n", sh->host_no,
                pdv->pdv_channel_id,  pdv->pdv_target_id, pdv->pdv_lun_id);
 
        if (phv->phv_mode == PHV_VIRUTAL_HOST_ID)
@@ -728,13 +690,12 @@ static int pscsi_transport_complete(struct se_task *task)
         */
        if (((cdb[0] == MODE_SENSE) || (cdb[0] == MODE_SENSE_10)) &&
             (status_byte(result) << 1) == SAM_STAT_GOOD) {
-               if (!TASK_CMD(task)->se_deve)
+               if (!task->task_se_cmd->se_deve)
                        goto after_mode_sense;
 
-               if (TASK_CMD(task)->se_deve->lun_flags &
+               if (task->task_se_cmd->se_deve->lun_flags &
                                TRANSPORT_LUNFLAGS_READ_ONLY) {
-                       unsigned char *buf = (unsigned char *)
-                               T_TASK(task->task_se_cmd)->t_task_buf;
+                       unsigned char *buf = transport_kmap_first_data_page(task->task_se_cmd);
 
                        if (cdb[0] == MODE_SENSE_10) {
                                if (!(buf[3] & 0x80))
@@ -743,6 +704,8 @@ static int pscsi_transport_complete(struct se_task *task)
                                if (!(buf[2] & 0x80))
                                        buf[2] |= 0x80;
                        }
+
+                       transport_kunmap_first_data_page(task->task_se_cmd);
                }
        }
 after_mode_sense:
@@ -766,8 +729,8 @@ after_mode_sense:
                u32 blocksize;
 
                buf = sg_virt(&sg[0]);
-               if (!(buf)) {
-                       printk(KERN_ERR "Unable to get buf for scatterlist\n");
+               if (!buf) {
+                       pr_err("Unable to get buf for scatterlist\n");
                        goto after_mode_select;
                }
 
@@ -797,34 +760,20 @@ after_mode_select:
 }
 
 static struct se_task *
-pscsi_alloc_task(struct se_cmd *cmd)
+pscsi_alloc_task(unsigned char *cdb)
 {
        struct pscsi_plugin_task *pt;
-       unsigned char *cdb = T_TASK(cmd)->t_task_cdb;
 
-       pt = kzalloc(sizeof(struct pscsi_plugin_task), GFP_KERNEL);
+       /*
+        * Dynamically alloc cdb space, since it may be larger than
+        * TCM_MAX_COMMAND_SIZE
+        */
+       pt = kzalloc(sizeof(*pt) + scsi_command_size(cdb), GFP_KERNEL);
        if (!pt) {
-               printk(KERN_ERR "Unable to allocate struct pscsi_plugin_task\n");
+               pr_err("Unable to allocate struct pscsi_plugin_task\n");
                return NULL;
        }
 
-       /*
-        * If TCM Core is signaling a > TCM_MAX_COMMAND_SIZE allocation,
-        * allocate the extended CDB buffer for per struct se_task context
-        * pt->pscsi_cdb now.
-        */
-       if (T_TASK(cmd)->t_task_cdb != T_TASK(cmd)->__t_task_cdb) {
-
-               pt->pscsi_cdb = kzalloc(scsi_command_size(cdb), GFP_KERNEL);
-               if (!(pt->pscsi_cdb)) {
-                       printk(KERN_ERR "pSCSI: Unable to allocate extended"
-                                       " pt->pscsi_cdb\n");
-                       kfree(pt);
-                       return NULL;
-               }
-       } else
-               pt->pscsi_cdb = &pt->__pscsi_cdb[0];
-
        return &pt->pscsi_task;
 }
 
@@ -849,7 +798,7 @@ static inline void pscsi_blk_init_request(
         * also set the end_io_data pointer.to struct se_task.
         */
        req->end_io = pscsi_req_done;
-       req->end_io_data = (void *)task;
+       req->end_io_data = task;
        /*
         * Load the referenced struct se_task's SCSI CDB into
         * include/linux/blkdev.h:struct request->cmd
@@ -859,7 +808,7 @@ static inline void pscsi_blk_init_request(
        /*
         * Setup pointer for outgoing sense data.
         */
-       req->sense = (void *)&pt->pscsi_sense[0];
+       req->sense = &pt->pscsi_sense[0];
        req->sense_len = 0;
 }
 
@@ -874,8 +823,8 @@ static int pscsi_blk_get_request(struct se_task *task)
        pt->pscsi_req = blk_get_request(pdv->pdv_sd->request_queue,
                        (task->task_data_direction == DMA_TO_DEVICE),
                        GFP_KERNEL);
-       if (!(pt->pscsi_req) || IS_ERR(pt->pscsi_req)) {
-               printk(KERN_ERR "PSCSI: blk_get_request() failed: %ld\n",
+       if (!pt->pscsi_req || IS_ERR(pt->pscsi_req)) {
+               pr_err("PSCSI: blk_get_request() failed: %ld\n",
                                IS_ERR(pt->pscsi_req));
                return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
@@ -920,14 +869,7 @@ static int pscsi_do_task(struct se_task *task)
 static void pscsi_free_task(struct se_task *task)
 {
        struct pscsi_plugin_task *pt = PSCSI_TASK(task);
-       struct se_cmd *cmd = task->task_se_cmd;
 
-       /*
-        * Release the extended CDB allocation from pscsi_alloc_task()
-        * if one exists.
-        */
-       if (T_TASK(cmd)->t_task_cdb != T_TASK(cmd)->__t_task_cdb)
-               kfree(pt->pscsi_cdb);
        /*
         * We do not release the bio(s) here associated with this task, as
         * this is handled by bio_put() and pscsi_bi_endio().
@@ -973,7 +915,7 @@ static ssize_t pscsi_set_configfs_dev_params(struct se_hba *hba,
                switch (token) {
                case Opt_scsi_host_id:
                        if (phv->phv_mode == PHV_LLD_SCSI_HOST_NO) {
-                               printk(KERN_ERR "PSCSI[%d]: Unable to accept"
+                               pr_err("PSCSI[%d]: Unable to accept"
                                        " scsi_host_id while phv_mode =="
                                        " PHV_LLD_SCSI_HOST_NO\n",
                                        phv->phv_host_id);
@@ -982,14 +924,14 @@ static ssize_t pscsi_set_configfs_dev_params(struct se_hba *hba,
                        }
                        match_int(args, &arg);
                        pdv->pdv_host_id = arg;
-                       printk(KERN_INFO "PSCSI[%d]: Referencing SCSI Host ID:"
+                       pr_debug("PSCSI[%d]: Referencing SCSI Host ID:"
                                " %d\n", phv->phv_host_id, pdv->pdv_host_id);
                        pdv->pdv_flags |= PDF_HAS_VIRT_HOST_ID;
                        break;
                case Opt_scsi_channel_id:
                        match_int(args, &arg);
                        pdv->pdv_channel_id = arg;
-                       printk(KERN_INFO "PSCSI[%d]: Referencing SCSI Channel"
+                       pr_debug("PSCSI[%d]: Referencing SCSI Channel"
                                " ID: %d\n",  phv->phv_host_id,
                                pdv->pdv_channel_id);
                        pdv->pdv_flags |= PDF_HAS_CHANNEL_ID;
@@ -997,7 +939,7 @@ static ssize_t pscsi_set_configfs_dev_params(struct se_hba *hba,
                case Opt_scsi_target_id:
                        match_int(args, &arg);
                        pdv->pdv_target_id = arg;
-                       printk(KERN_INFO "PSCSI[%d]: Referencing SCSI Target"
+                       pr_debug("PSCSI[%d]: Referencing SCSI Target"
                                " ID: %d\n", phv->phv_host_id,
                                pdv->pdv_target_id);
                        pdv->pdv_flags |= PDF_HAS_TARGET_ID;
@@ -1005,7 +947,7 @@ static ssize_t pscsi_set_configfs_dev_params(struct se_hba *hba,
                case Opt_scsi_lun_id:
                        match_int(args, &arg);
                        pdv->pdv_lun_id = arg;
-                       printk(KERN_INFO "PSCSI[%d]: Referencing SCSI LUN ID:"
+                       pr_debug("PSCSI[%d]: Referencing SCSI LUN ID:"
                                " %d\n", phv->phv_host_id, pdv->pdv_lun_id);
                        pdv->pdv_flags |= PDF_HAS_LUN_ID;
                        break;
@@ -1028,9 +970,9 @@ static ssize_t pscsi_check_configfs_dev_params(
        if (!(pdv->pdv_flags & PDF_HAS_CHANNEL_ID) ||
            !(pdv->pdv_flags & PDF_HAS_TARGET_ID) ||
            !(pdv->pdv_flags & PDF_HAS_LUN_ID)) {
-               printk(KERN_ERR "Missing scsi_channel_id=, scsi_target_id= and"
+               pr_err("Missing scsi_channel_id=, scsi_target_id= and"
                        " scsi_lun_id= parameters\n");
-               return -1;
+               return -EINVAL;
        }
 
        return 0;
@@ -1090,7 +1032,7 @@ static void pscsi_bi_endio(struct bio *bio, int error)
        bio_put(bio);
 }
 
-static inline struct bio *pscsi_get_bio(struct pscsi_dev_virt *pdv, int sg_num)
+static inline struct bio *pscsi_get_bio(int sg_num)
 {
        struct bio *bio;
        /*
@@ -1098,8 +1040,8 @@ static inline struct bio *pscsi_get_bio(struct pscsi_dev_virt *pdv, int sg_num)
         * in block/blk-core.c:blk_make_request()
         */
        bio = bio_kmalloc(GFP_KERNEL, sg_num);
-       if (!(bio)) {
-               printk(KERN_ERR "PSCSI: bio_kmalloc() failed\n");
+       if (!bio) {
+               pr_err("PSCSI: bio_kmalloc() failed\n");
                return NULL;
        }
        bio->bi_end_io = pscsi_bi_endio;
@@ -1107,13 +1049,7 @@ static inline struct bio *pscsi_get_bio(struct pscsi_dev_virt *pdv, int sg_num)
        return bio;
 }
 
-#if 0
-#define DEBUG_PSCSI(x...) printk(x)
-#else
-#define DEBUG_PSCSI(x...)
-#endif
-
-static int __pscsi_map_task_SG(
+static int __pscsi_map_SG(
        struct se_task *task,
        struct scatterlist *task_sg,
        u32 task_sg_num,
@@ -1134,7 +1070,7 @@ static int __pscsi_map_task_SG(
                return 0;
        /*
         * For SCF_SCSI_DATA_SG_IO_CDB, Use fs/bio.c:bio_add_page() to setup
-        * the bio_vec maplist from TC< struct se_mem -> task->task_sg ->
+        * the bio_vec maplist from task->task_sg ->
         * struct scatterlist memory.  The struct se_task->task_sg[] currently needs
         * to be attached to struct bios for submission to Linux/SCSI using
         * struct request to struct scsi_device->request_queue.
@@ -1143,34 +1079,34 @@ static int __pscsi_map_task_SG(
         * is ported to upstream SCSI passthrough functionality that accepts
         * struct scatterlist->page_link or struct page as a paraemeter.
         */
-       DEBUG_PSCSI("PSCSI: nr_pages: %d\n", nr_pages);
+       pr_debug("PSCSI: nr_pages: %d\n", nr_pages);
 
        for_each_sg(task_sg, sg, task_sg_num, i) {
                page = sg_page(sg);
                off = sg->offset;
                len = sg->length;
 
-               DEBUG_PSCSI("PSCSI: i: %d page: %p len: %d off: %d\n", i,
+               pr_debug("PSCSI: i: %d page: %p len: %d off: %d\n", i,
                        page, len, off);
 
                while (len > 0 && data_len > 0) {
                        bytes = min_t(unsigned int, len, PAGE_SIZE - off);
                        bytes = min(bytes, data_len);
 
-                       if (!(bio)) {
+                       if (!bio) {
                                nr_vecs = min_t(int, BIO_MAX_PAGES, nr_pages);
                                nr_pages -= nr_vecs;
                                /*
                                 * Calls bio_kmalloc() and sets bio->bi_end_io()
                                 */
-                               bio = pscsi_get_bio(pdv, nr_vecs);
-                               if (!(bio))
+                               bio = pscsi_get_bio(nr_vecs);
+                               if (!bio)
                                        goto fail;
 
                                if (rw)
                                        bio->bi_rw |= REQ_WRITE;
 
-                               DEBUG_PSCSI("PSCSI: Allocated bio: %p,"
+                               pr_debug("PSCSI: Allocated bio: %p,"
                                        " dir: %s nr_vecs: %d\n", bio,
                                        (rw) ? "rw" : "r", nr_vecs);
                                /*
@@ -1185,7 +1121,7 @@ static int __pscsi_map_task_SG(
                                        tbio = tbio->bi_next = bio;
                        }
 
-                       DEBUG_PSCSI("PSCSI: Calling bio_add_pc_page() i: %d"
+                       pr_debug("PSCSI: Calling bio_add_pc_page() i: %d"
                                " bio: %p page: %p len: %d off: %d\n", i, bio,
                                page, len, off);
 
@@ -1194,11 +1130,11 @@ static int __pscsi_map_task_SG(
                        if (rc != bytes)
                                goto fail;
 
-                       DEBUG_PSCSI("PSCSI: bio->bi_vcnt: %d nr_vecs: %d\n",
+                       pr_debug("PSCSI: bio->bi_vcnt: %d nr_vecs: %d\n",
                                bio->bi_vcnt, nr_vecs);
 
                        if (bio->bi_vcnt > nr_vecs) {
-                               DEBUG_PSCSI("PSCSI: Reached bio->bi_vcnt max:"
+                               pr_debug("PSCSI: Reached bio->bi_vcnt max:"
                                        " %d i: %d bio: %p, allocating another"
                                        " bio\n", bio->bi_vcnt, i, bio);
                                /*
@@ -1220,15 +1156,15 @@ static int __pscsi_map_task_SG(
         * Setup the primary pt->pscsi_req used for non BIDI and BIDI-COMMAND
         * primary SCSI WRITE poayload mapped for struct se_task->task_sg[]
         */
-       if (!(bidi_read)) {
+       if (!bidi_read) {
                /*
                 * Starting with v2.6.31, call blk_make_request() passing in *hbio to
                 * allocate the pSCSI task a struct request.
                 */
                pt->pscsi_req = blk_make_request(pdv->pdv_sd->request_queue,
                                        hbio, GFP_KERNEL);
-               if (!(pt->pscsi_req)) {
-                       printk(KERN_ERR "pSCSI: blk_make_request() failed\n");
+               if (!pt->pscsi_req) {
+                       pr_err("pSCSI: blk_make_request() failed\n");
                        goto fail;
                }
                /*
@@ -1237,7 +1173,7 @@ static int __pscsi_map_task_SG(
                 */
                pscsi_blk_init_request(task, pt, pt->pscsi_req, 0);
 
-               return task->task_sg_num;
+               return task->task_sg_nents;
        }
        /*
         * Setup the secondary pt->pscsi_req->next_rq used for the extra BIDI-COMMAND
@@ -1245,13 +1181,13 @@ static int __pscsi_map_task_SG(
         */
        pt->pscsi_req->next_rq = blk_make_request(pdv->pdv_sd->request_queue,
                                        hbio, GFP_KERNEL);
-       if (!(pt->pscsi_req->next_rq)) {
-               printk(KERN_ERR "pSCSI: blk_make_request() failed for BIDI\n");
+       if (!pt->pscsi_req->next_rq) {
+               pr_err("pSCSI: blk_make_request() failed for BIDI\n");
                goto fail;
        }
        pscsi_blk_init_request(task, pt, pt->pscsi_req->next_rq, 1);
 
-       return task->task_sg_num;
+       return task->task_sg_nents;
 fail:
        while (hbio) {
                bio = hbio;
@@ -1262,7 +1198,10 @@ fail:
        return ret;
 }
 
-static int pscsi_map_task_SG(struct se_task *task)
+/*
+ * pSCSI maps both ->map_control_SG() and ->map_data_SG() to a single call.
+ */
+static int pscsi_map_SG(struct se_task *task)
 {
        int ret;
 
@@ -1270,14 +1209,14 @@ static int pscsi_map_task_SG(struct se_task *task)
         * Setup the main struct request for the task->task_sg[] payload
         */
 
-       ret = __pscsi_map_task_SG(task, task->task_sg, task->task_sg_num, 0);
+       ret = __pscsi_map_SG(task, task->task_sg, task->task_sg_nents, 0);
        if (ret >= 0 && task->task_sg_bidi) {
                /*
                 * If present, set up the extra BIDI-COMMAND SCSI READ
                 * struct request and payload.
                 */
-               ret = __pscsi_map_task_SG(task, task->task_sg_bidi,
-                                       task->task_sg_num, 1);
+               ret = __pscsi_map_SG(task, task->task_sg_bidi,
+                                       task->task_sg_nents, 1);
        }
 
        if (ret < 0)
@@ -1285,33 +1224,6 @@ static int pscsi_map_task_SG(struct se_task *task)
        return 0;
 }
 
-/*     pscsi_map_task_non_SG():
- *
- *
- */
-static int pscsi_map_task_non_SG(struct se_task *task)
-{
-       struct se_cmd *cmd = TASK_CMD(task);
-       struct pscsi_plugin_task *pt = PSCSI_TASK(task);
-       struct pscsi_dev_virt *pdv = task->se_dev->dev_ptr;
-       int ret = 0;
-
-       if (pscsi_blk_get_request(task) < 0)
-               return PYX_TRANSPORT_LU_COMM_FAILURE;
-
-       if (!task->task_size)
-               return 0;
-
-       ret = blk_rq_map_kern(pdv->pdv_sd->request_queue,
-                       pt->pscsi_req, T_TASK(cmd)->t_task_buf,
-                       task->task_size, GFP_KERNEL);
-       if (ret < 0) {
-               printk(KERN_ERR "PSCSI: blk_rq_map_kern() failed: %d\n", ret);
-               return PYX_TRANSPORT_LU_COMM_FAILURE;
-       }
-       return 0;
-}
-
 static int pscsi_CDB_none(struct se_task *task)
 {
        return pscsi_blk_get_request(task);
@@ -1383,9 +1295,9 @@ static inline void pscsi_process_SAM_status(
        struct pscsi_plugin_task *pt)
 {
        task->task_scsi_status = status_byte(pt->pscsi_result);
-       if ((task->task_scsi_status)) {
+       if (task->task_scsi_status) {
                task->task_scsi_status <<= 1;
-               printk(KERN_INFO "PSCSI Status Byte exception at task: %p CDB:"
+               pr_debug("PSCSI Status Byte exception at task: %p CDB:"
                        " 0x%02x Result: 0x%08x\n", task, pt->pscsi_cdb[0],
                        pt->pscsi_result);
        }
@@ -1395,18 +1307,16 @@ static inline void pscsi_process_SAM_status(
                transport_complete_task(task, (!task->task_scsi_status));
                break;
        default:
-               printk(KERN_INFO "PSCSI Host Byte exception at task: %p CDB:"
+               pr_debug("PSCSI Host Byte exception at task: %p CDB:"
                        " 0x%02x Result: 0x%08x\n", task, pt->pscsi_cdb[0],
                        pt->pscsi_result);
                task->task_scsi_status = SAM_STAT_CHECK_CONDITION;
                task->task_error_status = PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
-               TASK_CMD(task)->transport_error_status =
+               task->task_se_cmd->transport_error_status =
                                        PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
                transport_complete_task(task, 0);
                break;
        }
-
-       return;
 }
 
 static void pscsi_req_done(struct request *req, int uptodate)
@@ -1433,8 +1343,8 @@ static struct se_subsystem_api pscsi_template = {
        .owner                  = THIS_MODULE,
        .transport_type         = TRANSPORT_PLUGIN_PHBA_PDEV,
        .cdb_none               = pscsi_CDB_none,
-       .map_task_non_SG        = pscsi_map_task_non_SG,
-       .map_task_SG            = pscsi_map_task_SG,
+       .map_control_SG         = pscsi_map_SG,
+       .map_data_SG            = pscsi_map_SG,
        .attach_hba             = pscsi_attach_hba,
        .detach_hba             = pscsi_detach_hba,
        .pmode_enable_hba       = pscsi_pmode_enable_hba,
index a4cd5d352c3aac585d6174fce04d246157407b75..ebf4f1ae2c83e78931bee09ae4c65ce64eb2e141 100644 (file)
@@ -2,7 +2,6 @@
 #define TARGET_CORE_PSCSI_H
 
 #define PSCSI_VERSION          "v4.0"
-#define PSCSI_VIRTUAL_HBA_DEPTH        2048
 
 /* used in pscsi_find_alloc_len() */
 #ifndef INQUIRY_DATA_SIZE
 
 struct pscsi_plugin_task {
        struct se_task pscsi_task;
-       unsigned char *pscsi_cdb;
-       unsigned char __pscsi_cdb[TCM_MAX_COMMAND_SIZE];
        unsigned char pscsi_sense[SCSI_SENSE_BUFFERSIZE];
        int     pscsi_direction;
        int     pscsi_result;
        u32     pscsi_resid;
        struct request *pscsi_req;
+       unsigned char pscsi_cdb[0];
 } ____cacheline_aligned;
 
 #define PDF_HAS_CHANNEL_ID     0x01
index 7837dd365a9d98be3ee9f606731a6e064b78acce..3dd81d24d9a914169f80372550fc6bf4f77edbff 100644 (file)
 
 #include "target_core_rd.h"
 
-static struct se_subsystem_api rd_dr_template;
 static struct se_subsystem_api rd_mcp_template;
 
-/* #define DEBUG_RAMDISK_MCP */
-/* #define DEBUG_RAMDISK_DR */
-
 /*     rd_attach_hba(): (Part of se_subsystem_api_t template)
  *
  *
@@ -59,24 +55,21 @@ static int rd_attach_hba(struct se_hba *hba, u32 host_id)
        struct rd_host *rd_host;
 
        rd_host = kzalloc(sizeof(struct rd_host), GFP_KERNEL);
-       if (!(rd_host)) {
-               printk(KERN_ERR "Unable to allocate memory for struct rd_host\n");
+       if (!rd_host) {
+               pr_err("Unable to allocate memory for struct rd_host\n");
                return -ENOMEM;
        }
 
        rd_host->rd_host_id = host_id;
 
-       atomic_set(&hba->left_queue_depth, RD_HBA_QUEUE_DEPTH);
-       atomic_set(&hba->max_queue_depth, RD_HBA_QUEUE_DEPTH);
-       hba->hba_ptr = (void *) rd_host;
+       hba->hba_ptr = rd_host;
 
-       printk(KERN_INFO "CORE_HBA[%d] - TCM Ramdisk HBA Driver %s on"
+       pr_debug("CORE_HBA[%d] - TCM Ramdisk HBA Driver %s on"
                " Generic Target Core Stack %s\n", hba->hba_id,
                RD_HBA_VERSION, TARGET_CORE_MOD_VERSION);
-       printk(KERN_INFO "CORE_HBA[%d] - Attached Ramdisk HBA: %u to Generic"
-               " Target Core TCQ Depth: %d MaxSectors: %u\n", hba->hba_id,
-               rd_host->rd_host_id, atomic_read(&hba->max_queue_depth),
-               RD_MAX_SECTORS);
+       pr_debug("CORE_HBA[%d] - Attached Ramdisk HBA: %u to Generic"
+               " MaxSectors: %u\n", hba->hba_id,
+               rd_host->rd_host_id, RD_MAX_SECTORS);
 
        return 0;
 }
@@ -85,7 +78,7 @@ static void rd_detach_hba(struct se_hba *hba)
 {
        struct rd_host *rd_host = hba->hba_ptr;
 
-       printk(KERN_INFO "CORE_HBA[%d] - Detached Ramdisk HBA: %u from"
+       pr_debug("CORE_HBA[%d] - Detached Ramdisk HBA: %u from"
                " Generic Target Core\n", hba->hba_id, rd_host->rd_host_id);
 
        kfree(rd_host);
@@ -114,7 +107,7 @@ static void rd_release_device_space(struct rd_dev *rd_dev)
 
                for (j = 0; j < sg_per_table; j++) {
                        pg = sg_page(&sg[j]);
-                       if ((pg)) {
+                       if (pg) {
                                __free_page(pg);
                                page_count++;
                        }
@@ -123,7 +116,7 @@ static void rd_release_device_space(struct rd_dev *rd_dev)
                kfree(sg);
        }
 
-       printk(KERN_INFO "CORE_RD[%u] - Released device space for Ramdisk"
+       pr_debug("CORE_RD[%u] - Released device space for Ramdisk"
                " Device ID: %u, pages %u in %u tables total bytes %lu\n",
                rd_dev->rd_host->rd_host_id, rd_dev->rd_dev_id, page_count,
                rd_dev->sg_table_count, (unsigned long)page_count * PAGE_SIZE);
@@ -148,7 +141,7 @@ static int rd_build_device_space(struct rd_dev *rd_dev)
        struct scatterlist *sg;
 
        if (rd_dev->rd_page_count <= 0) {
-               printk(KERN_ERR "Illegal page count: %u for Ramdisk device\n",
+               pr_err("Illegal page count: %u for Ramdisk device\n",
                        rd_dev->rd_page_count);
                return -EINVAL;
        }
@@ -157,8 +150,8 @@ static int rd_build_device_space(struct rd_dev *rd_dev)
        sg_tables = (total_sg_needed / max_sg_per_table) + 1;
 
        sg_table = kzalloc(sg_tables * sizeof(struct rd_dev_sg_table), GFP_KERNEL);
-       if (!(sg_table)) {
-               printk(KERN_ERR "Unable to allocate memory for Ramdisk"
+       if (!sg_table) {
+               pr_err("Unable to allocate memory for Ramdisk"
                        " scatterlist tables\n");
                return -ENOMEM;
        }
@@ -172,13 +165,13 @@ static int rd_build_device_space(struct rd_dev *rd_dev)
 
                sg = kzalloc(sg_per_table * sizeof(struct scatterlist),
                                GFP_KERNEL);
-               if (!(sg)) {
-                       printk(KERN_ERR "Unable to allocate scatterlist array"
+               if (!sg) {
+                       pr_err("Unable to allocate scatterlist array"
                                " for struct rd_dev\n");
                        return -ENOMEM;
                }
 
-               sg_init_table((struct scatterlist *)&sg[0], sg_per_table);
+               sg_init_table(sg, sg_per_table);
 
                sg_table[i].sg_table = sg;
                sg_table[i].rd_sg_count = sg_per_table;
@@ -188,8 +181,8 @@ static int rd_build_device_space(struct rd_dev *rd_dev)
 
                for (j = 0; j < sg_per_table; j++) {
                        pg = alloc_pages(GFP_KERNEL, 0);
-                       if (!(pg)) {
-                               printk(KERN_ERR "Unable to allocate scatterlist"
+                       if (!pg) {
+                               pr_err("Unable to allocate scatterlist"
                                        " pages for struct rd_dev_sg_table\n");
                                return -ENOMEM;
                        }
@@ -201,7 +194,7 @@ static int rd_build_device_space(struct rd_dev *rd_dev)
                total_sg_needed -= sg_per_table;
        }
 
-       printk(KERN_INFO "CORE_RD[%u] - Built Ramdisk Device ID: %u space of"
+       pr_debug("CORE_RD[%u] - Built Ramdisk Device ID: %u space of"
                " %u pages in %u tables\n", rd_dev->rd_host->rd_host_id,
                rd_dev->rd_dev_id, rd_dev->rd_page_count,
                rd_dev->sg_table_count);
@@ -218,8 +211,8 @@ static void *rd_allocate_virtdevice(
        struct rd_host *rd_host = hba->hba_ptr;
 
        rd_dev = kzalloc(sizeof(struct rd_dev), GFP_KERNEL);
-       if (!(rd_dev)) {
-               printk(KERN_ERR "Unable to allocate memory for struct rd_dev\n");
+       if (!rd_dev) {
+               pr_err("Unable to allocate memory for struct rd_dev\n");
                return NULL;
        }
 
@@ -229,11 +222,6 @@ static void *rd_allocate_virtdevice(
        return rd_dev;
 }
 
-static void *rd_DIRECT_allocate_virtdevice(struct se_hba *hba, const char *name)
-{
-       return rd_allocate_virtdevice(hba, name, 1);
-}
-
 static void *rd_MEMCPY_allocate_virtdevice(struct se_hba *hba, const char *name)
 {
        return rd_allocate_virtdevice(hba, name, 0);
@@ -273,16 +261,15 @@ static struct se_device *rd_create_virtdevice(
        dev_limits.queue_depth = RD_DEVICE_QUEUE_DEPTH;
 
        dev = transport_add_device_to_core_hba(hba,
-                       (rd_dev->rd_direct) ? &rd_dr_template :
-                       &rd_mcp_template, se_dev, dev_flags, (void *)rd_dev,
+                       &rd_mcp_template, se_dev, dev_flags, rd_dev,
                        &dev_limits, prod, rev);
-       if (!(dev))
+       if (!dev)
                goto fail;
 
        rd_dev->rd_dev_id = rd_host->rd_host_dev_id_count++;
        rd_dev->rd_queue_depth = dev->queue_depth;
 
-       printk(KERN_INFO "CORE_RD[%u] - Added TCM %s Ramdisk Device ID: %u of"
+       pr_debug("CORE_RD[%u] - Added TCM %s Ramdisk Device ID: %u of"
                " %u pages in %u tables, %lu total bytes\n",
                rd_host->rd_host_id, (!rd_dev->rd_direct) ? "MEMCPY" :
                "DIRECT", rd_dev->rd_dev_id, rd_dev->rd_page_count,
@@ -296,14 +283,6 @@ fail:
        return ERR_PTR(ret);
 }
 
-static struct se_device *rd_DIRECT_create_virtdevice(
-       struct se_hba *hba,
-       struct se_subsystem_dev *se_dev,
-       void *p)
-{
-       return rd_create_virtdevice(hba, se_dev, p, 1);
-}
-
 static struct se_device *rd_MEMCPY_create_virtdevice(
        struct se_hba *hba,
        struct se_subsystem_dev *se_dev,
@@ -330,16 +309,15 @@ static inline struct rd_request *RD_REQ(struct se_task *task)
 }
 
 static struct se_task *
-rd_alloc_task(struct se_cmd *cmd)
+rd_alloc_task(unsigned char *cdb)
 {
        struct rd_request *rd_req;
 
        rd_req = kzalloc(sizeof(struct rd_request), GFP_KERNEL);
        if (!rd_req) {
-               printk(KERN_ERR "Unable to allocate struct rd_request\n");
+               pr_err("Unable to allocate struct rd_request\n");
                return NULL;
        }
-       rd_req->rd_dev = SE_DEV(cmd)->dev_ptr;
 
        return &rd_req->rd_task;
 }
@@ -360,7 +338,7 @@ static struct rd_dev_sg_table *rd_get_sg_table(struct rd_dev *rd_dev, u32 page)
                        return sg_table;
        }
 
-       printk(KERN_ERR "Unable to locate struct rd_dev_sg_table for page: %u\n",
+       pr_err("Unable to locate struct rd_dev_sg_table for page: %u\n",
                        page);
 
        return NULL;
@@ -373,7 +351,7 @@ static struct rd_dev_sg_table *rd_get_sg_table(struct rd_dev *rd_dev, u32 page)
 static int rd_MEMCPY_read(struct rd_request *req)
 {
        struct se_task *task = &req->rd_task;
-       struct rd_dev *dev = req->rd_dev;
+       struct rd_dev *dev = req->rd_task.se_dev->dev_ptr;
        struct rd_dev_sg_table *table;
        struct scatterlist *sg_d, *sg_s;
        void *dst, *src;
@@ -382,32 +360,32 @@ static int rd_MEMCPY_read(struct rd_request *req)
        u32 rd_offset = req->rd_offset;
 
        table = rd_get_sg_table(dev, req->rd_page);
-       if (!(table))
-               return -1;
+       if (!table)
+               return -EINVAL;
 
        table_sg_end = (table->page_end_offset - req->rd_page);
        sg_d = task->task_sg;
        sg_s = &table->sg_table[req->rd_page - table->page_start_offset];
-#ifdef DEBUG_RAMDISK_MCP
-       printk(KERN_INFO "RD[%u]: Read LBA: %llu, Size: %u Page: %u, Offset:"
+
+       pr_debug("RD[%u]: Read LBA: %llu, Size: %u Page: %u, Offset:"
                " %u\n", dev->rd_dev_id, task->task_lba, req->rd_size,
                req->rd_page, req->rd_offset);
-#endif
+
        src_offset = rd_offset;
 
        while (req->rd_size) {
                if ((sg_d[i].length - dst_offset) <
                    (sg_s[j].length - src_offset)) {
                        length = (sg_d[i].length - dst_offset);
-#ifdef DEBUG_RAMDISK_MCP
-                       printk(KERN_INFO "Step 1 - sg_d[%d]: %p length: %d"
+
+                       pr_debug("Step 1 - sg_d[%d]: %p length: %d"
                                " offset: %u sg_s[%d].length: %u\n", i,
                                &sg_d[i], sg_d[i].length, sg_d[i].offset, j,
                                sg_s[j].length);
-                       printk(KERN_INFO "Step 1 - length: %u dst_offset: %u"
+                       pr_debug("Step 1 - length: %u dst_offset: %u"
                                " src_offset: %u\n", length, dst_offset,
                                src_offset);
-#endif
+
                        if (length > req->rd_size)
                                length = req->rd_size;
 
@@ -424,15 +402,15 @@ static int rd_MEMCPY_read(struct rd_request *req)
                        page_end = 0;
                } else {
                        length = (sg_s[j].length - src_offset);
-#ifdef DEBUG_RAMDISK_MCP
-                       printk(KERN_INFO "Step 2 - sg_d[%d]: %p length: %d"
+
+                       pr_debug("Step 2 - sg_d[%d]: %p length: %d"
                                " offset: %u sg_s[%d].length: %u\n", i,
                                &sg_d[i], sg_d[i].length, sg_d[i].offset,
                                j, sg_s[j].length);
-                       printk(KERN_INFO "Step 2 - length: %u dst_offset: %u"
+                       pr_debug("Step 2 - length: %u dst_offset: %u"
                                " src_offset: %u\n", length, dst_offset,
                                src_offset);
-#endif
+
                        if (length > req->rd_size)
                                length = req->rd_size;
 
@@ -456,32 +434,29 @@ static int rd_MEMCPY_read(struct rd_request *req)
 
                memcpy(dst, src, length);
 
-#ifdef DEBUG_RAMDISK_MCP
-               printk(KERN_INFO "page: %u, remaining size: %u, length: %u,"
+               pr_debug("page: %u, remaining size: %u, length: %u,"
                        " i: %u, j: %u\n", req->rd_page,
                        (req->rd_size - length), length, i, j);
-#endif
+
                req->rd_size -= length;
-               if (!(req->rd_size))
+               if (!req->rd_size)
                        return 0;
 
                if (!page_end)
                        continue;
 
                if (++req->rd_page <= table->page_end_offset) {
-#ifdef DEBUG_RAMDISK_MCP
-                       printk(KERN_INFO "page: %u in same page table\n",
+                       pr_debug("page: %u in same page table\n",
                                req->rd_page);
-#endif
                        continue;
                }
-#ifdef DEBUG_RAMDISK_MCP
-               printk(KERN_INFO "getting new page table for page: %u\n",
+
+               pr_debug("getting new page table for page: %u\n",
                                req->rd_page);
-#endif
+
                table = rd_get_sg_table(dev, req->rd_page);
-               if (!(table))
-                       return -1;
+               if (!table)
+                       return -EINVAL;
 
                sg_s = &table->sg_table[j = 0];
        }
@@ -496,7 +471,7 @@ static int rd_MEMCPY_read(struct rd_request *req)
 static int rd_MEMCPY_write(struct rd_request *req)
 {
        struct se_task *task = &req->rd_task;
-       struct rd_dev *dev = req->rd_dev;
+       struct rd_dev *dev = req->rd_task.se_dev->dev_ptr;
        struct rd_dev_sg_table *table;
        struct scatterlist *sg_d, *sg_s;
        void *dst, *src;
@@ -505,32 +480,32 @@ static int rd_MEMCPY_write(struct rd_request *req)
        u32 rd_offset = req->rd_offset;
 
        table = rd_get_sg_table(dev, req->rd_page);
-       if (!(table))
-               return -1;
+       if (!table)
+               return -EINVAL;
 
        table_sg_end = (table->page_end_offset - req->rd_page);
        sg_d = &table->sg_table[req->rd_page - table->page_start_offset];
        sg_s = task->task_sg;
-#ifdef DEBUG_RAMDISK_MCP
-       printk(KERN_INFO "RD[%d] Write LBA: %llu, Size: %u, Page: %u,"
+
+       pr_debug("RD[%d] Write LBA: %llu, Size: %u, Page: %u,"
                " Offset: %u\n", dev->rd_dev_id, task->task_lba, req->rd_size,
                req->rd_page, req->rd_offset);
-#endif
+
        dst_offset = rd_offset;
 
        while (req->rd_size) {
                if ((sg_s[i].length - src_offset) <
                    (sg_d[j].length - dst_offset)) {
                        length = (sg_s[i].length - src_offset);
-#ifdef DEBUG_RAMDISK_MCP
-                       printk(KERN_INFO "Step 1 - sg_s[%d]: %p length: %d"
+
+                       pr_debug("Step 1 - sg_s[%d]: %p length: %d"
                                " offset: %d sg_d[%d].length: %u\n", i,
                                &sg_s[i], sg_s[i].length, sg_s[i].offset,
                                j, sg_d[j].length);
-                       printk(KERN_INFO "Step 1 - length: %u src_offset: %u"
+                       pr_debug("Step 1 - length: %u src_offset: %u"
                                " dst_offset: %u\n", length, src_offset,
                                dst_offset);
-#endif
+
                        if (length > req->rd_size)
                                length = req->rd_size;
 
@@ -547,15 +522,15 @@ static int rd_MEMCPY_write(struct rd_request *req)
                        page_end = 0;
                } else {
                        length = (sg_d[j].length - dst_offset);
-#ifdef DEBUG_RAMDISK_MCP
-                       printk(KERN_INFO "Step 2 - sg_s[%d]: %p length: %d"
+
+                       pr_debug("Step 2 - sg_s[%d]: %p length: %d"
                                " offset: %d sg_d[%d].length: %u\n", i,
                                &sg_s[i], sg_s[i].length, sg_s[i].offset,
                                j, sg_d[j].length);
-                       printk(KERN_INFO "Step 2 - length: %u src_offset: %u"
+                       pr_debug("Step 2 - length: %u src_offset: %u"
                                " dst_offset: %u\n", length, src_offset,
                                dst_offset);
-#endif
+
                        if (length > req->rd_size)
                                length = req->rd_size;
 
@@ -579,32 +554,29 @@ static int rd_MEMCPY_write(struct rd_request *req)
 
                memcpy(dst, src, length);
 
-#ifdef DEBUG_RAMDISK_MCP
-               printk(KERN_INFO "page: %u, remaining size: %u, length: %u,"
+               pr_debug("page: %u, remaining size: %u, length: %u,"
                        " i: %u, j: %u\n", req->rd_page,
                        (req->rd_size - length), length, i, j);
-#endif
+
                req->rd_size -= length;
-               if (!(req->rd_size))
+               if (!req->rd_size)
                        return 0;
 
                if (!page_end)
                        continue;
 
                if (++req->rd_page <= table->page_end_offset) {
-#ifdef DEBUG_RAMDISK_MCP
-                       printk(KERN_INFO "page: %u in same page table\n",
+                       pr_debug("page: %u in same page table\n",
                                req->rd_page);
-#endif
                        continue;
                }
-#ifdef DEBUG_RAMDISK_MCP
-               printk(KERN_INFO "getting new page table for page: %u\n",
+
+               pr_debug("getting new page table for page: %u\n",
                                req->rd_page);
-#endif
+
                table = rd_get_sg_table(dev, req->rd_page);
-               if (!(table))
-                       return -1;
+               if (!table)
+                       return -EINVAL;
 
                sg_d = &table->sg_table[j = 0];
        }
@@ -623,11 +595,11 @@ static int rd_MEMCPY_do_task(struct se_task *task)
        unsigned long long lba;
        int ret;
 
-       req->rd_page = (task->task_lba * DEV_ATTRIB(dev)->block_size) / PAGE_SIZE;
+       req->rd_page = (task->task_lba * dev->se_sub_dev->se_dev_attrib.block_size) / PAGE_SIZE;
        lba = task->task_lba;
        req->rd_offset = (do_div(lba,
-                         (PAGE_SIZE / DEV_ATTRIB(dev)->block_size))) *
-                          DEV_ATTRIB(dev)->block_size;
+                         (PAGE_SIZE / dev->se_sub_dev->se_dev_attrib.block_size))) *
+                          dev->se_sub_dev->se_dev_attrib.block_size;
        req->rd_size = task->task_size;
 
        if (task->task_data_direction == DMA_FROM_DEVICE)
@@ -644,274 +616,6 @@ static int rd_MEMCPY_do_task(struct se_task *task)
        return PYX_TRANSPORT_SENT_TO_TRANSPORT;
 }
 
-/*     rd_DIRECT_with_offset():
- *
- *
- */
-static int rd_DIRECT_with_offset(
-       struct se_task *task,
-       struct list_head *se_mem_list,
-       u32 *se_mem_cnt,
-       u32 *task_offset)
-{
-       struct rd_request *req = RD_REQ(task);
-       struct rd_dev *dev = req->rd_dev;
-       struct rd_dev_sg_table *table;
-       struct se_mem *se_mem;
-       struct scatterlist *sg_s;
-       u32 j = 0, set_offset = 1;
-       u32 get_next_table = 0, offset_length, table_sg_end;
-
-       table = rd_get_sg_table(dev, req->rd_page);
-       if (!(table))
-               return -1;
-
-       table_sg_end = (table->page_end_offset - req->rd_page);
-       sg_s = &table->sg_table[req->rd_page - table->page_start_offset];
-#ifdef DEBUG_RAMDISK_DR
-       printk(KERN_INFO "%s DIRECT LBA: %llu, Size: %u Page: %u, Offset: %u\n",
-               (task->task_data_direction == DMA_TO_DEVICE) ?
-                       "Write" : "Read",
-               task->task_lba, req->rd_size, req->rd_page, req->rd_offset);
-#endif
-       while (req->rd_size) {
-               se_mem = kmem_cache_zalloc(se_mem_cache, GFP_KERNEL);
-               if (!(se_mem)) {
-                       printk(KERN_ERR "Unable to allocate struct se_mem\n");
-                       return -1;
-               }
-               INIT_LIST_HEAD(&se_mem->se_list);
-
-               if (set_offset) {
-                       offset_length = sg_s[j].length - req->rd_offset;
-                       if (offset_length > req->rd_size)
-                               offset_length = req->rd_size;
-
-                       se_mem->se_page = sg_page(&sg_s[j++]);
-                       se_mem->se_off = req->rd_offset;
-                       se_mem->se_len = offset_length;
-
-                       set_offset = 0;
-                       get_next_table = (j > table_sg_end);
-                       goto check_eot;
-               }
-
-               offset_length = (req->rd_size < req->rd_offset) ?
-                       req->rd_size : req->rd_offset;
-
-               se_mem->se_page = sg_page(&sg_s[j]);
-               se_mem->se_len = offset_length;
-
-               set_offset = 1;
-
-check_eot:
-#ifdef DEBUG_RAMDISK_DR
-               printk(KERN_INFO "page: %u, size: %u, offset_length: %u, j: %u"
-                       " se_mem: %p, se_page: %p se_off: %u se_len: %u\n",
-                       req->rd_page, req->rd_size, offset_length, j, se_mem,
-                       se_mem->se_page, se_mem->se_off, se_mem->se_len);
-#endif
-               list_add_tail(&se_mem->se_list, se_mem_list);
-               (*se_mem_cnt)++;
-
-               req->rd_size -= offset_length;
-               if (!(req->rd_size))
-                       goto out;
-
-               if (!set_offset && !get_next_table)
-                       continue;
-
-               if (++req->rd_page <= table->page_end_offset) {
-#ifdef DEBUG_RAMDISK_DR
-                       printk(KERN_INFO "page: %u in same page table\n",
-                                       req->rd_page);
-#endif
-                       continue;
-               }
-#ifdef DEBUG_RAMDISK_DR
-               printk(KERN_INFO "getting new page table for page: %u\n",
-                               req->rd_page);
-#endif
-               table = rd_get_sg_table(dev, req->rd_page);
-               if (!(table))
-                       return -1;
-
-               sg_s = &table->sg_table[j = 0];
-       }
-
-out:
-       T_TASK(task->task_se_cmd)->t_tasks_se_num += *se_mem_cnt;
-#ifdef DEBUG_RAMDISK_DR
-       printk(KERN_INFO "RD_DR - Allocated %u struct se_mem segments for task\n",
-                       *se_mem_cnt);
-#endif
-       return 0;
-}
-
-/*     rd_DIRECT_without_offset():
- *
- *
- */
-static int rd_DIRECT_without_offset(
-       struct se_task *task,
-       struct list_head *se_mem_list,
-       u32 *se_mem_cnt,
-       u32 *task_offset)
-{
-       struct rd_request *req = RD_REQ(task);
-       struct rd_dev *dev = req->rd_dev;
-       struct rd_dev_sg_table *table;
-       struct se_mem *se_mem;
-       struct scatterlist *sg_s;
-       u32 length, j = 0;
-
-       table = rd_get_sg_table(dev, req->rd_page);
-       if (!(table))
-               return -1;
-
-       sg_s = &table->sg_table[req->rd_page - table->page_start_offset];
-#ifdef DEBUG_RAMDISK_DR
-       printk(KERN_INFO "%s DIRECT LBA: %llu, Size: %u, Page: %u\n",
-               (task->task_data_direction == DMA_TO_DEVICE) ?
-                       "Write" : "Read",
-               task->task_lba, req->rd_size, req->rd_page);
-#endif
-       while (req->rd_size) {
-               se_mem = kmem_cache_zalloc(se_mem_cache, GFP_KERNEL);
-               if (!(se_mem)) {
-                       printk(KERN_ERR "Unable to allocate struct se_mem\n");
-                       return -1;
-               }
-               INIT_LIST_HEAD(&se_mem->se_list);
-
-               length = (req->rd_size < sg_s[j].length) ?
-                       req->rd_size : sg_s[j].length;
-
-               se_mem->se_page = sg_page(&sg_s[j++]);
-               se_mem->se_len = length;
-
-#ifdef DEBUG_RAMDISK_DR
-               printk(KERN_INFO "page: %u, size: %u, j: %u se_mem: %p,"
-                       " se_page: %p se_off: %u se_len: %u\n", req->rd_page,
-                       req->rd_size, j, se_mem, se_mem->se_page,
-                       se_mem->se_off, se_mem->se_len);
-#endif
-               list_add_tail(&se_mem->se_list, se_mem_list);
-               (*se_mem_cnt)++;
-
-               req->rd_size -= length;
-               if (!(req->rd_size))
-                       goto out;
-
-               if (++req->rd_page <= table->page_end_offset) {
-#ifdef DEBUG_RAMDISK_DR
-                       printk("page: %u in same page table\n",
-                               req->rd_page);
-#endif
-                       continue;
-               }
-#ifdef DEBUG_RAMDISK_DR
-               printk(KERN_INFO "getting new page table for page: %u\n",
-                               req->rd_page);
-#endif
-               table = rd_get_sg_table(dev, req->rd_page);
-               if (!(table))
-                       return -1;
-
-               sg_s = &table->sg_table[j = 0];
-       }
-
-out:
-       T_TASK(task->task_se_cmd)->t_tasks_se_num += *se_mem_cnt;
-#ifdef DEBUG_RAMDISK_DR
-       printk(KERN_INFO "RD_DR - Allocated %u struct se_mem segments for task\n",
-                       *se_mem_cnt);
-#endif
-       return 0;
-}
-
-/*     rd_DIRECT_do_se_mem_map():
- *
- *
- */
-static int rd_DIRECT_do_se_mem_map(
-       struct se_task *task,
-       struct list_head *se_mem_list,
-       void *in_mem,
-       struct se_mem *in_se_mem,
-       struct se_mem **out_se_mem,
-       u32 *se_mem_cnt,
-       u32 *task_offset_in)
-{
-       struct se_cmd *cmd = task->task_se_cmd;
-       struct rd_request *req = RD_REQ(task);
-       u32 task_offset = *task_offset_in;
-       unsigned long long lba;
-       int ret;
-
-       req->rd_page = ((task->task_lba * DEV_ATTRIB(task->se_dev)->block_size) /
-                       PAGE_SIZE);
-       lba = task->task_lba;
-       req->rd_offset = (do_div(lba,
-                         (PAGE_SIZE / DEV_ATTRIB(task->se_dev)->block_size))) *
-                          DEV_ATTRIB(task->se_dev)->block_size;
-       req->rd_size = task->task_size;
-
-       if (req->rd_offset)
-               ret = rd_DIRECT_with_offset(task, se_mem_list, se_mem_cnt,
-                               task_offset_in);
-       else
-               ret = rd_DIRECT_without_offset(task, se_mem_list, se_mem_cnt,
-                               task_offset_in);
-
-       if (ret < 0)
-               return ret;
-
-       if (CMD_TFO(cmd)->task_sg_chaining == 0)
-               return 0;
-       /*
-        * Currently prevent writers from multiple HW fabrics doing
-        * pci_map_sg() to RD_DR's internal scatterlist memory.
-        */
-       if (cmd->data_direction == DMA_TO_DEVICE) {
-               printk(KERN_ERR "DMA_TO_DEVICE not supported for"
-                               " RAMDISK_DR with task_sg_chaining=1\n");
-               return -1;
-       }
-       /*
-        * Special case for if task_sg_chaining is enabled, then
-        * we setup struct se_task->task_sg[], as it will be used by
-        * transport_do_task_sg_chain() for creating chainged SGLs
-        * across multiple struct se_task->task_sg[].
-        */
-       if (!(transport_calc_sg_num(task,
-                       list_entry(T_TASK(cmd)->t_mem_list->next,
-                                  struct se_mem, se_list),
-                       task_offset)))
-               return -1;
-
-       return transport_map_mem_to_sg(task, se_mem_list, task->task_sg,
-                       list_entry(T_TASK(cmd)->t_mem_list->next,
-                                  struct se_mem, se_list),
-                       out_se_mem, se_mem_cnt, task_offset_in);
-}
-
-/*     rd_DIRECT_do_task(): (Part of se_subsystem_api_t template)
- *
- *
- */
-static int rd_DIRECT_do_task(struct se_task *task)
-{
-       /*
-        * At this point the locally allocated RD tables have been mapped
-        * to struct se_mem elements in rd_DIRECT_do_se_mem_map().
-        */
-       task->task_scsi_status = GOOD;
-       transport_complete_task(task, 1);
-
-       return PYX_TRANSPORT_SENT_TO_TRANSPORT;
-}
-
 /*     rd_free_task(): (Part of se_subsystem_api_t template)
  *
  *
@@ -956,7 +660,7 @@ static ssize_t rd_set_configfs_dev_params(
                case Opt_rd_pages:
                        match_int(args, &arg);
                        rd_dev->rd_page_count = arg;
-                       printk(KERN_INFO "RAMDISK: Referencing Page"
+                       pr_debug("RAMDISK: Referencing Page"
                                " Count: %u\n", rd_dev->rd_page_count);
                        rd_dev->rd_flags |= RDF_HAS_PAGE_COUNT;
                        break;
@@ -974,8 +678,8 @@ static ssize_t rd_check_configfs_dev_params(struct se_hba *hba, struct se_subsys
        struct rd_dev *rd_dev = se_dev->se_dev_su_ptr;
 
        if (!(rd_dev->rd_flags & RDF_HAS_PAGE_COUNT)) {
-               printk(KERN_INFO "Missing rd_pages= parameter\n");
-               return -1;
+               pr_debug("Missing rd_pages= parameter\n");
+               return -EINVAL;
        }
 
        return 0;
@@ -1021,32 +725,11 @@ static sector_t rd_get_blocks(struct se_device *dev)
 {
        struct rd_dev *rd_dev = dev->dev_ptr;
        unsigned long long blocks_long = ((rd_dev->rd_page_count * PAGE_SIZE) /
-                       DEV_ATTRIB(dev)->block_size) - 1;
+                       dev->se_sub_dev->se_dev_attrib.block_size) - 1;
 
        return blocks_long;
 }
 
-static struct se_subsystem_api rd_dr_template = {
-       .name                   = "rd_dr",
-       .transport_type         = TRANSPORT_PLUGIN_VHBA_VDEV,
-       .attach_hba             = rd_attach_hba,
-       .detach_hba             = rd_detach_hba,
-       .allocate_virtdevice    = rd_DIRECT_allocate_virtdevice,
-       .create_virtdevice      = rd_DIRECT_create_virtdevice,
-       .free_device            = rd_free_device,
-       .alloc_task             = rd_alloc_task,
-       .do_task                = rd_DIRECT_do_task,
-       .free_task              = rd_free_task,
-       .check_configfs_dev_params = rd_check_configfs_dev_params,
-       .set_configfs_dev_params = rd_set_configfs_dev_params,
-       .show_configfs_dev_params = rd_show_configfs_dev_params,
-       .get_cdb                = rd_get_cdb,
-       .get_device_rev         = rd_get_device_rev,
-       .get_device_type        = rd_get_device_type,
-       .get_blocks             = rd_get_blocks,
-       .do_se_mem_map          = rd_DIRECT_do_se_mem_map,
-};
-
 static struct se_subsystem_api rd_mcp_template = {
        .name                   = "rd_mcp",
        .transport_type         = TRANSPORT_PLUGIN_VHBA_VDEV,
@@ -1071,13 +754,8 @@ int __init rd_module_init(void)
 {
        int ret;
 
-       ret = transport_subsystem_register(&rd_dr_template);
-       if (ret < 0)
-               return ret;
-
        ret = transport_subsystem_register(&rd_mcp_template);
        if (ret < 0) {
-               transport_subsystem_release(&rd_dr_template);
                return ret;
        }
 
@@ -1086,6 +764,5 @@ int __init rd_module_init(void)
 
 void rd_module_exit(void)
 {
-       transport_subsystem_release(&rd_dr_template);
        transport_subsystem_release(&rd_mcp_template);
 }
index 3ea19e29d8ece0fbcad14301aeef15b0b8413d75..0d027732cd00215836bb4529d3726382c2baff48 100644 (file)
@@ -7,8 +7,6 @@
 
 /* Largest piece of memory kmalloc can allocate */
 #define RD_MAX_ALLOCATION_SIZE 65536
-/* Maximum queuedepth for the Ramdisk HBA */
-#define RD_HBA_QUEUE_DEPTH     256
 #define RD_DEVICE_QUEUE_DEPTH  32
 #define RD_MAX_DEVICE_QUEUE_DEPTH 128
 #define RD_BLOCKSIZE           512
@@ -34,8 +32,6 @@ struct rd_request {
        u32             rd_page_count;
        /* Scatterlist count */
        u32             rd_size;
-       /* Ramdisk device */
-       struct rd_dev   *rd_dev;
 } ____cacheline_aligned;
 
 struct rd_dev_sg_table {
index dc6fed037ab3876b208ab8f5b8c08f69a2883eea..72843441d4fab7080c4dae47cdee9a07d16be4ed 100644 (file)
  */
 void split_cdb_XX_6(
        unsigned long long lba,
-       u32 *sectors,
+       u32 sectors,
        unsigned char *cdb)
 {
        cdb[1] = (lba >> 16) & 0x1f;
        cdb[2] = (lba >> 8) & 0xff;
        cdb[3] = lba & 0xff;
-       cdb[4] = *sectors & 0xff;
+       cdb[4] = sectors & 0xff;
 }
 
 /*     split_cdb_XX_10():
@@ -57,11 +57,11 @@ void split_cdb_XX_6(
  */
 void split_cdb_XX_10(
        unsigned long long lba,
-       u32 *sectors,
+       u32 sectors,
        unsigned char *cdb)
 {
        put_unaligned_be32(lba, &cdb[2]);
-       put_unaligned_be16(*sectors, &cdb[7]);
+       put_unaligned_be16(sectors, &cdb[7]);
 }
 
 /*     split_cdb_XX_12():
@@ -70,11 +70,11 @@ void split_cdb_XX_10(
  */
 void split_cdb_XX_12(
        unsigned long long lba,
-       u32 *sectors,
+       u32 sectors,
        unsigned char *cdb)
 {
        put_unaligned_be32(lba, &cdb[2]);
-       put_unaligned_be32(*sectors, &cdb[6]);
+       put_unaligned_be32(sectors, &cdb[6]);
 }
 
 /*     split_cdb_XX_16():
@@ -83,11 +83,11 @@ void split_cdb_XX_12(
  */
 void split_cdb_XX_16(
        unsigned long long lba,
-       u32 *sectors,
+       u32 sectors,
        unsigned char *cdb)
 {
        put_unaligned_be64(lba, &cdb[2]);
-       put_unaligned_be32(*sectors, &cdb[10]);
+       put_unaligned_be32(sectors, &cdb[10]);
 }
 
 /*
@@ -97,9 +97,9 @@ void split_cdb_XX_16(
  */
 void split_cdb_XX_32(
        unsigned long long lba,
-       u32 *sectors,
+       u32 sectors,
        unsigned char *cdb)
 {
        put_unaligned_be64(lba, &cdb[12]);
-       put_unaligned_be32(*sectors, &cdb[28]);
+       put_unaligned_be32(sectors, &cdb[28]);
 }
index 98cd1c01ed838843e4c4addf8a97453294cff488..48e9ccc9585e749245ba1fa841cfc84ecd1ad050 100644 (file)
@@ -1,10 +1,10 @@
 #ifndef TARGET_CORE_SCDB_H
 #define TARGET_CORE_SCDB_H
 
-extern void split_cdb_XX_6(unsigned long long, u32 *, unsigned char *);
-extern void split_cdb_XX_10(unsigned long long, u32 *, unsigned char *);
-extern void split_cdb_XX_12(unsigned long long, u32 *, unsigned char *);
-extern void split_cdb_XX_16(unsigned long long, u32 *, unsigned char *);
-extern void split_cdb_XX_32(unsigned long long, u32 *, unsigned char *);
+extern void split_cdb_XX_6(unsigned long long, u32, unsigned char *);
+extern void split_cdb_XX_10(unsigned long long, u32, unsigned char *);
+extern void split_cdb_XX_12(unsigned long long, u32, unsigned char *);
+extern void split_cdb_XX_16(unsigned long long, u32, unsigned char *);
+extern void split_cdb_XX_32(unsigned long long, u32, unsigned char *);
 
 #endif /* TARGET_CORE_SCDB_H */
index 5e3a067a7475ff25fce01b72fc67d96660997c00..a8d6e1dee938b418b9c1164ae979acb2a950fbc6 100644 (file)
@@ -402,8 +402,8 @@ static ssize_t target_stat_scsi_lu_show_attr_lu_name(
                return -ENODEV;
        /* scsiLuWwnName */
        return snprintf(page, PAGE_SIZE, "%s\n",
-                       (strlen(DEV_T10_WWN(dev)->unit_serial)) ?
-                       (char *)&DEV_T10_WWN(dev)->unit_serial[0] : "None");
+                       (strlen(dev->se_sub_dev->t10_wwn.unit_serial)) ?
+                       dev->se_sub_dev->t10_wwn.unit_serial : "None");
 }
 DEV_STAT_SCSI_LU_ATTR_RO(lu_name);
 
@@ -413,17 +413,17 @@ static ssize_t target_stat_scsi_lu_show_attr_vend(
        struct se_subsystem_dev *se_subdev = container_of(sgrps,
                        struct se_subsystem_dev, dev_stat_grps);
        struct se_device *dev = se_subdev->se_dev_ptr;
-       int j;
-       char str[28];
+       int i;
+       char str[sizeof(dev->se_sub_dev->t10_wwn.vendor)+1];
 
        if (!dev)
                return -ENODEV;
+
        /* scsiLuVendorId */
-       memcpy(&str[0], (void *)DEV_T10_WWN(dev), 28);
-       for (j = 0; j < 8; j++)
-               str[j] = ISPRINT(DEV_T10_WWN(dev)->vendor[j]) ?
-                               DEV_T10_WWN(dev)->vendor[j] : 0x20;
-       str[8] = 0;
+       for (i = 0; i < sizeof(dev->se_sub_dev->t10_wwn.vendor); i++)
+               str[i] = ISPRINT(dev->se_sub_dev->t10_wwn.vendor[i]) ?
+                       dev->se_sub_dev->t10_wwn.vendor[i] : ' ';
+       str[i] = '\0';
        return snprintf(page, PAGE_SIZE, "%s\n", str);
 }
 DEV_STAT_SCSI_LU_ATTR_RO(vend);
@@ -434,18 +434,17 @@ static ssize_t target_stat_scsi_lu_show_attr_prod(
        struct se_subsystem_dev *se_subdev = container_of(sgrps,
                        struct se_subsystem_dev, dev_stat_grps);
        struct se_device *dev = se_subdev->se_dev_ptr;
-       int j;
-       char str[28];
+       int i;
+       char str[sizeof(dev->se_sub_dev->t10_wwn.model)+1];
 
        if (!dev)
                return -ENODEV;
 
        /* scsiLuProductId */
-       memcpy(&str[0], (void *)DEV_T10_WWN(dev), 28);
-       for (j = 0; j < 16; j++)
-               str[j] = ISPRINT(DEV_T10_WWN(dev)->model[j]) ?
-                               DEV_T10_WWN(dev)->model[j] : 0x20;
-       str[16] = 0;
+       for (i = 0; i < sizeof(dev->se_sub_dev->t10_wwn.vendor); i++)
+               str[i] = ISPRINT(dev->se_sub_dev->t10_wwn.model[i]) ?
+                       dev->se_sub_dev->t10_wwn.model[i] : ' ';
+       str[i] = '\0';
        return snprintf(page, PAGE_SIZE, "%s\n", str);
 }
 DEV_STAT_SCSI_LU_ATTR_RO(prod);
@@ -456,18 +455,17 @@ static ssize_t target_stat_scsi_lu_show_attr_rev(
        struct se_subsystem_dev *se_subdev = container_of(sgrps,
                        struct se_subsystem_dev, dev_stat_grps);
        struct se_device *dev = se_subdev->se_dev_ptr;
-       int j;
-       char str[28];
+       int i;
+       char str[sizeof(dev->se_sub_dev->t10_wwn.revision)+1];
 
        if (!dev)
                return -ENODEV;
 
        /* scsiLuRevisionId */
-       memcpy(&str[0], (void *)DEV_T10_WWN(dev), 28);
-       for (j = 0; j < 4; j++)
-               str[j] = ISPRINT(DEV_T10_WWN(dev)->revision[j]) ?
-                               DEV_T10_WWN(dev)->revision[j] : 0x20;
-       str[4] = 0;
+       for (i = 0; i < sizeof(dev->se_sub_dev->t10_wwn.revision); i++)
+               str[i] = ISPRINT(dev->se_sub_dev->t10_wwn.revision[i]) ?
+                       dev->se_sub_dev->t10_wwn.revision[i] : ' ';
+       str[i] = '\0';
        return snprintf(page, PAGE_SIZE, "%s\n", str);
 }
 DEV_STAT_SCSI_LU_ATTR_RO(rev);
@@ -484,7 +482,7 @@ static ssize_t target_stat_scsi_lu_show_attr_dev_type(
 
        /* scsiLuPeripheralType */
        return snprintf(page, PAGE_SIZE, "%u\n",
-                       TRANSPORT(dev)->get_device_type(dev));
+                       dev->transport->get_device_type(dev));
 }
 DEV_STAT_SCSI_LU_ATTR_RO(dev_type);
 
@@ -668,18 +666,18 @@ static struct config_item_type target_stat_scsi_lu_cit = {
  */
 void target_stat_setup_dev_default_groups(struct se_subsystem_dev *se_subdev)
 {
-       struct config_group *dev_stat_grp = &DEV_STAT_GRP(se_subdev)->stat_group;
+       struct config_group *dev_stat_grp = &se_subdev->dev_stat_grps.stat_group;
 
-       config_group_init_type_name(&DEV_STAT_GRP(se_subdev)->scsi_dev_group,
+       config_group_init_type_name(&se_subdev->dev_stat_grps.scsi_dev_group,
                        "scsi_dev", &target_stat_scsi_dev_cit);
-       config_group_init_type_name(&DEV_STAT_GRP(se_subdev)->scsi_tgt_dev_group,
+       config_group_init_type_name(&se_subdev->dev_stat_grps.scsi_tgt_dev_group,
                        "scsi_tgt_dev", &target_stat_scsi_tgt_dev_cit);
-       config_group_init_type_name(&DEV_STAT_GRP(se_subdev)->scsi_lu_group,
+       config_group_init_type_name(&se_subdev->dev_stat_grps.scsi_lu_group,
                        "scsi_lu", &target_stat_scsi_lu_cit);
 
-       dev_stat_grp->default_groups[0] = &DEV_STAT_GRP(se_subdev)->scsi_dev_group;
-       dev_stat_grp->default_groups[1] = &DEV_STAT_GRP(se_subdev)->scsi_tgt_dev_group;
-       dev_stat_grp->default_groups[2] = &DEV_STAT_GRP(se_subdev)->scsi_lu_group;
+       dev_stat_grp->default_groups[0] = &se_subdev->dev_stat_grps.scsi_dev_group;
+       dev_stat_grp->default_groups[1] = &se_subdev->dev_stat_grps.scsi_tgt_dev_group;
+       dev_stat_grp->default_groups[2] = &se_subdev->dev_stat_grps.scsi_lu_group;
        dev_stat_grp->default_groups[3] = NULL;
 }
 
@@ -922,7 +920,7 @@ static ssize_t target_stat_scsi_tgt_port_show_attr_name(
        tpg = sep->sep_tpg;
 
        ret = snprintf(page, PAGE_SIZE, "%sPort#%u\n",
-               TPG_TFO(tpg)->get_fabric_name(), sep->sep_index);
+               tpg->se_tpg_tfo->get_fabric_name(), sep->sep_index);
        spin_unlock(&lun->lun_sep_lock);
        return ret;
 }
@@ -945,8 +943,8 @@ static ssize_t target_stat_scsi_tgt_port_show_attr_port_index(
        tpg = sep->sep_tpg;
 
        ret = snprintf(page, PAGE_SIZE, "%s%s%d\n",
-               TPG_TFO(tpg)->tpg_get_wwn(tpg), "+t+",
-               TPG_TFO(tpg)->tpg_get_tag(tpg));
+               tpg->se_tpg_tfo->tpg_get_wwn(tpg), "+t+",
+               tpg->se_tpg_tfo->tpg_get_tag(tpg));
        spin_unlock(&lun->lun_sep_lock);
        return ret;
 }
@@ -1128,7 +1126,7 @@ static ssize_t target_stat_scsi_transport_show_attr_device(
        tpg = sep->sep_tpg;
        /* scsiTransportType */
        ret = snprintf(page, PAGE_SIZE, "scsiTransport%s\n",
-                       TPG_TFO(tpg)->get_fabric_name());
+                       tpg->se_tpg_tfo->get_fabric_name());
        spin_unlock(&lun->lun_sep_lock);
        return ret;
 }
@@ -1150,7 +1148,7 @@ static ssize_t target_stat_scsi_transport_show_attr_indx(
        }
        tpg = sep->sep_tpg;
        ret = snprintf(page, PAGE_SIZE, "%u\n",
-                       TPG_TFO(tpg)->tpg_get_inst_index(tpg));
+                       tpg->se_tpg_tfo->tpg_get_inst_index(tpg));
        spin_unlock(&lun->lun_sep_lock);
        return ret;
 }
@@ -1173,10 +1171,10 @@ static ssize_t target_stat_scsi_transport_show_attr_dev_name(
                return -ENODEV;
        }
        tpg = sep->sep_tpg;
-       wwn = DEV_T10_WWN(dev);
+       wwn = &dev->se_sub_dev->t10_wwn;
        /* scsiTransportDevName */
        ret = snprintf(page, PAGE_SIZE, "%s+%s\n",
-                       TPG_TFO(tpg)->tpg_get_wwn(tpg),
+                       tpg->se_tpg_tfo->tpg_get_wwn(tpg),
                        (strlen(wwn->unit_serial)) ? wwn->unit_serial :
                        wwn->vendor);
        spin_unlock(&lun->lun_sep_lock);
@@ -1212,18 +1210,18 @@ static struct config_item_type target_stat_scsi_transport_cit = {
  */
 void target_stat_setup_port_default_groups(struct se_lun *lun)
 {
-       struct config_group *port_stat_grp = &PORT_STAT_GRP(lun)->stat_group;
+       struct config_group *port_stat_grp = &lun->port_stat_grps.stat_group;
 
-       config_group_init_type_name(&PORT_STAT_GRP(lun)->scsi_port_group,
+       config_group_init_type_name(&lun->port_stat_grps.scsi_port_group,
                        "scsi_port", &target_stat_scsi_port_cit);
-       config_group_init_type_name(&PORT_STAT_GRP(lun)->scsi_tgt_port_group,
+       config_group_init_type_name(&lun->port_stat_grps.scsi_tgt_port_group,
                        "scsi_tgt_port", &target_stat_scsi_tgt_port_cit);
-       config_group_init_type_name(&PORT_STAT_GRP(lun)->scsi_transport_group,
+       config_group_init_type_name(&lun->port_stat_grps.scsi_transport_group,
                        "scsi_transport", &target_stat_scsi_transport_cit);
 
-       port_stat_grp->default_groups[0] = &PORT_STAT_GRP(lun)->scsi_port_group;
-       port_stat_grp->default_groups[1] = &PORT_STAT_GRP(lun)->scsi_tgt_port_group;
-       port_stat_grp->default_groups[2] = &PORT_STAT_GRP(lun)->scsi_transport_group;
+       port_stat_grp->default_groups[0] = &lun->port_stat_grps.scsi_port_group;
+       port_stat_grp->default_groups[1] = &lun->port_stat_grps.scsi_tgt_port_group;
+       port_stat_grp->default_groups[2] = &lun->port_stat_grps.scsi_transport_group;
        port_stat_grp->default_groups[3] = NULL;
 }
 
@@ -1264,7 +1262,7 @@ static ssize_t target_stat_scsi_auth_intr_show_attr_inst(
        tpg = nacl->se_tpg;
        /* scsiInstIndex */
        ret = snprintf(page, PAGE_SIZE, "%u\n",
-                       TPG_TFO(tpg)->tpg_get_inst_index(tpg));
+                       tpg->se_tpg_tfo->tpg_get_inst_index(tpg));
        spin_unlock_irq(&nacl->device_list_lock);
        return ret;
 }
@@ -1314,7 +1312,7 @@ static ssize_t target_stat_scsi_auth_intr_show_attr_port(
        }
        tpg = nacl->se_tpg;
        /* scsiAuthIntrTgtPortIndex */
-       ret = snprintf(page, PAGE_SIZE, "%u\n", TPG_TFO(tpg)->tpg_get_tag(tpg));
+       ret = snprintf(page, PAGE_SIZE, "%u\n", tpg->se_tpg_tfo->tpg_get_tag(tpg));
        spin_unlock_irq(&nacl->device_list_lock);
        return ret;
 }
@@ -1632,7 +1630,7 @@ static ssize_t target_stat_scsi_att_intr_port_show_attr_inst(
        tpg = nacl->se_tpg;
        /* scsiInstIndex */
        ret = snprintf(page, PAGE_SIZE, "%u\n",
-                       TPG_TFO(tpg)->tpg_get_inst_index(tpg));
+                       tpg->se_tpg_tfo->tpg_get_inst_index(tpg));
        spin_unlock_irq(&nacl->device_list_lock);
        return ret;
 }
@@ -1682,7 +1680,7 @@ static ssize_t target_stat_scsi_att_intr_port_show_attr_port(
        }
        tpg = nacl->se_tpg;
        /* scsiPortIndex */
-       ret = snprintf(page, PAGE_SIZE, "%u\n", TPG_TFO(tpg)->tpg_get_tag(tpg));
+       ret = snprintf(page, PAGE_SIZE, "%u\n", tpg->se_tpg_tfo->tpg_get_tag(tpg));
        spin_unlock_irq(&nacl->device_list_lock);
        return ret;
 }
@@ -1708,7 +1706,7 @@ static ssize_t target_stat_scsi_att_intr_port_show_attr_indx(
        tpg = nacl->se_tpg;
        /* scsiAttIntrPortIndex */
        ret = snprintf(page, PAGE_SIZE, "%u\n",
-                       TPG_TFO(tpg)->sess_get_index(se_sess));
+                       tpg->se_tpg_tfo->sess_get_index(se_sess));
        spin_unlock_irq(&nacl->nacl_sess_lock);
        return ret;
 }
@@ -1757,8 +1755,8 @@ static ssize_t target_stat_scsi_att_intr_port_show_attr_port_ident(
        tpg = nacl->se_tpg;
        /* scsiAttIntrPortName+scsiAttIntrPortIdentifier */
        memset(buf, 0, 64);
-       if (TPG_TFO(tpg)->sess_get_initiator_sid != NULL)
-               TPG_TFO(tpg)->sess_get_initiator_sid(se_sess,
+       if (tpg->se_tpg_tfo->sess_get_initiator_sid != NULL)
+               tpg->se_tpg_tfo->sess_get_initiator_sid(se_sess,
                                (unsigned char *)&buf[0], 64);
 
        ret = snprintf(page, PAGE_SIZE, "%s+i+%s\n", nacl->initiatorname, buf);
@@ -1797,14 +1795,14 @@ static struct config_item_type target_stat_scsi_att_intr_port_cit = {
  */
 void target_stat_setup_mappedlun_default_groups(struct se_lun_acl *lacl)
 {
-       struct config_group *ml_stat_grp = &ML_STAT_GRPS(lacl)->stat_group;
+       struct config_group *ml_stat_grp = &lacl->ml_stat_grps.stat_group;
 
-       config_group_init_type_name(&ML_STAT_GRPS(lacl)->scsi_auth_intr_group,
+       config_group_init_type_name(&lacl->ml_stat_grps.scsi_auth_intr_group,
                        "scsi_auth_intr", &target_stat_scsi_auth_intr_cit);
-       config_group_init_type_name(&ML_STAT_GRPS(lacl)->scsi_att_intr_port_group,
+       config_group_init_type_name(&lacl->ml_stat_grps.scsi_att_intr_port_group,
                        "scsi_att_intr_port", &target_stat_scsi_att_intr_port_cit);
 
-       ml_stat_grp->default_groups[0] = &ML_STAT_GRPS(lacl)->scsi_auth_intr_group;
-       ml_stat_grp->default_groups[1] = &ML_STAT_GRPS(lacl)->scsi_att_intr_port_group;
+       ml_stat_grp->default_groups[0] = &lacl->ml_stat_grps.scsi_auth_intr_group;
+       ml_stat_grp->default_groups[1] = &lacl->ml_stat_grps.scsi_att_intr_port_group;
        ml_stat_grp->default_groups[2] = NULL;
 }
index 179063d81cdd6f30f1af19ef3cb20c2e12b923a4..27d4925e51c3f373914fe058630ccda7ccb121c1 100644 (file)
 #include "target_core_alua.h"
 #include "target_core_pr.h"
 
-#define DEBUG_LUN_RESET
-#ifdef DEBUG_LUN_RESET
-#define DEBUG_LR(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_LR(x...)
-#endif
-
 struct se_tmr_req *core_tmr_alloc_req(
        struct se_cmd *se_cmd,
        void *fabric_tmr_ptr,
@@ -57,8 +50,8 @@ struct se_tmr_req *core_tmr_alloc_req(
 
        tmr = kmem_cache_zalloc(se_tmr_req_cache, (in_interrupt()) ?
                                        GFP_ATOMIC : GFP_KERNEL);
-       if (!(tmr)) {
-               printk(KERN_ERR "Unable to allocate struct se_tmr_req\n");
+       if (!tmr) {
+               pr_err("Unable to allocate struct se_tmr_req\n");
                return ERR_PTR(-ENOMEM);
        }
        tmr->task_cmd = se_cmd;
@@ -80,9 +73,9 @@ void core_tmr_release_req(
                return;
        }
 
-       spin_lock(&dev->se_tmr_lock);
+       spin_lock_irq(&dev->se_tmr_lock);
        list_del(&tmr->tmr_list);
-       spin_unlock(&dev->se_tmr_lock);
+       spin_unlock_irq(&dev->se_tmr_lock);
 
        kmem_cache_free(se_tmr_req_cache, tmr);
 }
@@ -93,14 +86,14 @@ static void core_tmr_handle_tas_abort(
        int tas,
        int fe_count)
 {
-       if (!(fe_count)) {
+       if (!fe_count) {
                transport_cmd_finish_abort(cmd, 1);
                return;
        }
        /*
         * TASK ABORTED status (TAS) bit support
        */
-       if (((tmr_nacl != NULL) &&
+       if ((tmr_nacl &&
             (tmr_nacl == cmd->se_sess->se_node_acl)) || tas)
                transport_send_task_abort(cmd);
 
@@ -113,15 +106,14 @@ int core_tmr_lun_reset(
        struct list_head *preempt_and_abort_list,
        struct se_cmd *prout_cmd)
 {
-       struct se_cmd *cmd;
-       struct se_queue_req *qr, *qr_tmp;
+       struct se_cmd *cmd, *tcmd;
        struct se_node_acl *tmr_nacl = NULL;
        struct se_portal_group *tmr_tpg = NULL;
-       struct se_queue_obj *qobj = dev->dev_queue_obj;
+       struct se_queue_obj *qobj = &dev->dev_queue_obj;
        struct se_tmr_req *tmr_p, *tmr_pp;
        struct se_task *task, *task_tmp;
        unsigned long flags;
-       int fe_count, state, tas;
+       int fe_count, tas;
        /*
         * TASK_ABORTED status bit, this is configurable via ConfigFS
         * struct se_device attributes.  spc4r17 section 7.4.6 Control mode page
@@ -133,7 +125,7 @@ int core_tmr_lun_reset(
         * which the command was received shall be completed with TASK ABORTED
         * status (see SAM-4).
         */
-       tas = DEV_ATTRIB(dev)->emulate_tas;
+       tas = dev->se_sub_dev->se_dev_attrib.emulate_tas;
        /*
         * Determine if this se_tmr is coming from a $FABRIC_MOD
         * or struct se_device passthrough..
@@ -142,20 +134,20 @@ int core_tmr_lun_reset(
                tmr_nacl = tmr->task_cmd->se_sess->se_node_acl;
                tmr_tpg = tmr->task_cmd->se_sess->se_tpg;
                if (tmr_nacl && tmr_tpg) {
-                       DEBUG_LR("LUN_RESET: TMR caller fabric: %s"
+                       pr_debug("LUN_RESET: TMR caller fabric: %s"
                                " initiator port %s\n",
-                               TPG_TFO(tmr_tpg)->get_fabric_name(),
+                               tmr_tpg->se_tpg_tfo->get_fabric_name(),
                                tmr_nacl->initiatorname);
                }
        }
-       DEBUG_LR("LUN_RESET: %s starting for [%s], tas: %d\n",
+       pr_debug("LUN_RESET: %s starting for [%s], tas: %d\n",
                (preempt_and_abort_list) ? "Preempt" : "TMR",
-               TRANSPORT(dev)->name, tas);
+               dev->transport->name, tas);
        /*
         * Release all pending and outgoing TMRs aside from the received
         * LUN_RESET tmr..
         */
-       spin_lock(&dev->se_tmr_lock);
+       spin_lock_irq(&dev->se_tmr_lock);
        list_for_each_entry_safe(tmr_p, tmr_pp, &dev->dev_tmr_list, tmr_list) {
                /*
                 * Allow the received TMR to return with FUNCTION_COMPLETE.
@@ -164,8 +156,8 @@ int core_tmr_lun_reset(
                        continue;
 
                cmd = tmr_p->task_cmd;
-               if (!(cmd)) {
-                       printk(KERN_ERR "Unable to locate struct se_cmd for TMR\n");
+               if (!cmd) {
+                       pr_err("Unable to locate struct se_cmd for TMR\n");
                        continue;
                }
                /*
@@ -173,33 +165,33 @@ int core_tmr_lun_reset(
                 * parameter (eg: for PROUT PREEMPT_AND_ABORT service action
                 * skip non regisration key matching TMRs.
                 */
-               if ((preempt_and_abort_list != NULL) &&
+               if (preempt_and_abort_list &&
                    (core_scsi3_check_cdb_abort_and_preempt(
                                        preempt_and_abort_list, cmd) != 0))
                        continue;
-               spin_unlock(&dev->se_tmr_lock);
+               spin_unlock_irq(&dev->se_tmr_lock);
 
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-               if (!(atomic_read(&T_TASK(cmd)->t_transport_active))) {
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
-                       spin_lock(&dev->se_tmr_lock);
+               spin_lock_irqsave(&cmd->t_state_lock, flags);
+               if (!atomic_read(&cmd->t_transport_active)) {
+                       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
+                       spin_lock_irq(&dev->se_tmr_lock);
                        continue;
                }
                if (cmd->t_state == TRANSPORT_ISTATE_PROCESSING) {
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
-                       spin_lock(&dev->se_tmr_lock);
+                       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
+                       spin_lock_irq(&dev->se_tmr_lock);
                        continue;
                }
-               DEBUG_LR("LUN_RESET: %s releasing TMR %p Function: 0x%02x,"
+               pr_debug("LUN_RESET: %s releasing TMR %p Function: 0x%02x,"
                        " Response: 0x%02x, t_state: %d\n",
                        (preempt_and_abort_list) ? "Preempt" : "", tmr_p,
                        tmr_p->function, tmr_p->response, cmd->t_state);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
                transport_cmd_finish_abort_tmr(cmd);
-               spin_lock(&dev->se_tmr_lock);
+               spin_lock_irq(&dev->se_tmr_lock);
        }
-       spin_unlock(&dev->se_tmr_lock);
+       spin_unlock_irq(&dev->se_tmr_lock);
        /*
         * Complete outstanding struct se_task CDBs with TASK_ABORTED SAM status.
         * This is following sam4r17, section 5.6 Aborting commands, Table 38
@@ -224,23 +216,17 @@ int core_tmr_lun_reset(
        spin_lock_irqsave(&dev->execute_task_lock, flags);
        list_for_each_entry_safe(task, task_tmp, &dev->state_task_list,
                                t_state_list) {
-               if (!(TASK_CMD(task))) {
-                       printk(KERN_ERR "TASK_CMD(task) is NULL!\n");
+               if (!task->task_se_cmd) {
+                       pr_err("task->task_se_cmd is NULL!\n");
                        continue;
                }
-               cmd = TASK_CMD(task);
+               cmd = task->task_se_cmd;
 
-               if (!T_TASK(cmd)) {
-                       printk(KERN_ERR "T_TASK(cmd) is NULL for task: %p cmd:"
-                               " %p ITT: 0x%08x\n", task, cmd,
-                               CMD_TFO(cmd)->get_task_tag(cmd));
-                       continue;
-               }
                /*
                 * For PREEMPT_AND_ABORT usage, only process commands
                 * with a matching reservation key.
                 */
-               if ((preempt_and_abort_list != NULL) &&
+               if (preempt_and_abort_list &&
                    (core_scsi3_check_cdb_abort_and_preempt(
                                        preempt_and_abort_list, cmd) != 0))
                        continue;
@@ -254,38 +240,38 @@ int core_tmr_lun_reset(
                atomic_set(&task->task_state_active, 0);
                spin_unlock_irqrestore(&dev->execute_task_lock, flags);
 
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-               DEBUG_LR("LUN_RESET: %s cmd: %p task: %p"
+               spin_lock_irqsave(&cmd->t_state_lock, flags);
+               pr_debug("LUN_RESET: %s cmd: %p task: %p"
                        " ITT/CmdSN: 0x%08x/0x%08x, i_state: %d, t_state/"
                        "def_t_state: %d/%d cdb: 0x%02x\n",
                        (preempt_and_abort_list) ? "Preempt" : "", cmd, task,
-                       CMD_TFO(cmd)->get_task_tag(cmd), 0,
-                       CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state,
-                       cmd->deferred_t_state, T_TASK(cmd)->t_task_cdb[0]);
-               DEBUG_LR("LUN_RESET: ITT[0x%08x] - pr_res_key: 0x%016Lx"
+                       cmd->se_tfo->get_task_tag(cmd), 0,
+                       cmd->se_tfo->get_cmd_state(cmd), cmd->t_state,
+                       cmd->deferred_t_state, cmd->t_task_cdb[0]);
+               pr_debug("LUN_RESET: ITT[0x%08x] - pr_res_key: 0x%016Lx"
                        " t_task_cdbs: %d t_task_cdbs_left: %d"
                        " t_task_cdbs_sent: %d -- t_transport_active: %d"
                        " t_transport_stop: %d t_transport_sent: %d\n",
-                       CMD_TFO(cmd)->get_task_tag(cmd), cmd->pr_res_key,
-                       T_TASK(cmd)->t_task_cdbs,
-                       atomic_read(&T_TASK(cmd)->t_task_cdbs_left),
-                       atomic_read(&T_TASK(cmd)->t_task_cdbs_sent),
-                       atomic_read(&T_TASK(cmd)->t_transport_active),
-                       atomic_read(&T_TASK(cmd)->t_transport_stop),
-                       atomic_read(&T_TASK(cmd)->t_transport_sent));
+                       cmd->se_tfo->get_task_tag(cmd), cmd->pr_res_key,
+                       cmd->t_task_list_num,
+                       atomic_read(&cmd->t_task_cdbs_left),
+                       atomic_read(&cmd->t_task_cdbs_sent),
+                       atomic_read(&cmd->t_transport_active),
+                       atomic_read(&cmd->t_transport_stop),
+                       atomic_read(&cmd->t_transport_sent));
 
                if (atomic_read(&task->task_active)) {
                        atomic_set(&task->task_stop, 1);
                        spin_unlock_irqrestore(
-                               &T_TASK(cmd)->t_state_lock, flags);
+                               &cmd->t_state_lock, flags);
 
-                       DEBUG_LR("LUN_RESET: Waiting for task: %p to shutdown"
+                       pr_debug("LUN_RESET: Waiting for task: %p to shutdown"
                                " for dev: %p\n", task, dev);
                        wait_for_completion(&task->task_stop_comp);
-                       DEBUG_LR("LUN_RESET Completed task: %p shutdown for"
+                       pr_debug("LUN_RESET Completed task: %p shutdown for"
                                " dev: %p\n", task, dev);
-                       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-                       atomic_dec(&T_TASK(cmd)->t_task_cdbs_left);
+                       spin_lock_irqsave(&cmd->t_state_lock, flags);
+                       atomic_dec(&cmd->t_task_cdbs_left);
 
                        atomic_set(&task->task_active, 0);
                        atomic_set(&task->task_stop, 0);
@@ -295,34 +281,34 @@ int core_tmr_lun_reset(
                }
                __transport_stop_task_timer(task, &flags);
 
-               if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_ex_left))) {
+               if (!atomic_dec_and_test(&cmd->t_task_cdbs_ex_left)) {
                        spin_unlock_irqrestore(
-                                       &T_TASK(cmd)->t_state_lock, flags);
-                       DEBUG_LR("LUN_RESET: Skipping task: %p, dev: %p for"
+                                       &cmd->t_state_lock, flags);
+                       pr_debug("LUN_RESET: Skipping task: %p, dev: %p for"
                                " t_task_cdbs_ex_left: %d\n", task, dev,
-                               atomic_read(&T_TASK(cmd)->t_task_cdbs_ex_left));
+                               atomic_read(&cmd->t_task_cdbs_ex_left));
 
                        spin_lock_irqsave(&dev->execute_task_lock, flags);
                        continue;
                }
-               fe_count = atomic_read(&T_TASK(cmd)->t_fe_count);
+               fe_count = atomic_read(&cmd->t_fe_count);
 
-               if (atomic_read(&T_TASK(cmd)->t_transport_active)) {
-                       DEBUG_LR("LUN_RESET: got t_transport_active = 1 for"
+               if (atomic_read(&cmd->t_transport_active)) {
+                       pr_debug("LUN_RESET: got t_transport_active = 1 for"
                                " task: %p, t_fe_count: %d dev: %p\n", task,
                                fe_count, dev);
-                       atomic_set(&T_TASK(cmd)->t_transport_aborted, 1);
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+                       atomic_set(&cmd->t_transport_aborted, 1);
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                                flags);
                        core_tmr_handle_tas_abort(tmr_nacl, cmd, tas, fe_count);
 
                        spin_lock_irqsave(&dev->execute_task_lock, flags);
                        continue;
                }
-               DEBUG_LR("LUN_RESET: Got t_transport_active = 0 for task: %p,"
+               pr_debug("LUN_RESET: Got t_transport_active = 0 for task: %p,"
                        " t_fe_count: %d dev: %p\n", task, fe_count, dev);
-               atomic_set(&T_TASK(cmd)->t_transport_aborted, 1);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               atomic_set(&cmd->t_transport_aborted, 1);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                core_tmr_handle_tas_abort(tmr_nacl, cmd, tas, fe_count);
 
                spin_lock_irqsave(&dev->execute_task_lock, flags);
@@ -337,25 +323,12 @@ int core_tmr_lun_reset(
         * reference, otherwise the struct se_cmd is released.
         */
        spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
-       list_for_each_entry_safe(qr, qr_tmp, &qobj->qobj_list, qr_list) {
-               cmd = (struct se_cmd *)qr->cmd;
-               if (!(cmd)) {
-                       /*
-                        * Skip these for non PREEMPT_AND_ABORT usage..
-                        */
-                       if (preempt_and_abort_list != NULL)
-                               continue;
-
-                       atomic_dec(&qobj->queue_cnt);
-                       list_del(&qr->qr_list);
-                       kfree(qr);
-                       continue;
-               }
+       list_for_each_entry_safe(cmd, tcmd, &qobj->qobj_list, se_queue_node) {
                /*
                 * For PREEMPT_AND_ABORT usage, only process commands
                 * with a matching reservation key.
                 */
-               if ((preempt_and_abort_list != NULL) &&
+               if (preempt_and_abort_list &&
                    (core_scsi3_check_cdb_abort_and_preempt(
                                        preempt_and_abort_list, cmd) != 0))
                        continue;
@@ -365,30 +338,22 @@ int core_tmr_lun_reset(
                if (prout_cmd == cmd)
                        continue;
 
-               atomic_dec(&T_TASK(cmd)->t_transport_queue_active);
+               atomic_dec(&cmd->t_transport_queue_active);
                atomic_dec(&qobj->queue_cnt);
-               list_del(&qr->qr_list);
+               list_del(&cmd->se_queue_node);
                spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
 
-               state = qr->state;
-               kfree(qr);
-
-               DEBUG_LR("LUN_RESET: %s from Device Queue: cmd: %p t_state:"
+               pr_debug("LUN_RESET: %s from Device Queue: cmd: %p t_state:"
                        " %d t_fe_count: %d\n", (preempt_and_abort_list) ?
-                       "Preempt" : "", cmd, state,
-                       atomic_read(&T_TASK(cmd)->t_fe_count));
+                       "Preempt" : "", cmd, cmd->t_state,
+                       atomic_read(&cmd->t_fe_count));
                /*
                 * Signal that the command has failed via cmd->se_cmd_flags,
-                * and call TFO->new_cmd_failure() to wakeup any fabric
-                * dependent code used to wait for unsolicited data out
-                * allocation to complete.  The fabric module is expected
-                * to dump any remaining unsolicited data out for the aborted
-                * command at this point.
                 */
                transport_new_cmd_failure(cmd);
 
                core_tmr_handle_tas_abort(tmr_nacl, cmd, tas,
-                               atomic_read(&T_TASK(cmd)->t_fe_count));
+                               atomic_read(&cmd->t_fe_count));
                spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
        }
        spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
@@ -396,21 +361,21 @@ int core_tmr_lun_reset(
         * Clear any legacy SPC-2 reservation when called during
         * LOGICAL UNIT RESET
         */
-       if (!(preempt_and_abort_list) &&
+       if (!preempt_and_abort_list &&
             (dev->dev_flags & DF_SPC2_RESERVATIONS)) {
                spin_lock(&dev->dev_reservation_lock);
                dev->dev_reserved_node_acl = NULL;
                dev->dev_flags &= ~DF_SPC2_RESERVATIONS;
                spin_unlock(&dev->dev_reservation_lock);
-               printk(KERN_INFO "LUN_RESET: SCSI-2 Released reservation\n");
+               pr_debug("LUN_RESET: SCSI-2 Released reservation\n");
        }
 
        spin_lock_irq(&dev->stats_lock);
        dev->num_resets++;
        spin_unlock_irq(&dev->stats_lock);
 
-       DEBUG_LR("LUN_RESET: %s for [%s] Complete\n",
+       pr_debug("LUN_RESET: %s for [%s] Complete\n",
                        (preempt_and_abort_list) ? "Preempt" : "TMR",
-                       TRANSPORT(dev)->name);
+                       dev->transport->name);
        return 0;
 }
index 5ec745fed931d8c106f6bc70d5f8ac3cc844c8e3..4f1ba4c5ef1196e43e0d5ff4b52836c02e0dc9fe 100644 (file)
 #include <target/target_core_fabric_ops.h>
 
 #include "target_core_hba.h"
+#include "target_core_stat.h"
+
+extern struct se_device *g_lun0_dev;
+
+static DEFINE_SPINLOCK(tpg_lock);
+static LIST_HEAD(tpg_list);
 
 /*     core_clear_initiator_node_from_tpg():
  *
@@ -66,9 +72,9 @@ static void core_clear_initiator_node_from_tpg(
                        continue;
 
                if (!deve->se_lun) {
-                       printk(KERN_ERR "%s device entries device pointer is"
+                       pr_err("%s device entries device pointer is"
                                " NULL, but Initiator has access.\n",
-                               TPG_TFO(tpg)->get_fabric_name());
+                               tpg->se_tpg_tfo->get_fabric_name());
                        continue;
                }
 
@@ -80,14 +86,13 @@ static void core_clear_initiator_node_from_tpg(
                spin_lock(&lun->lun_acl_lock);
                list_for_each_entry_safe(acl, acl_tmp,
                                        &lun->lun_acl_list, lacl_list) {
-                       if (!(strcmp(acl->initiatorname,
-                                       nacl->initiatorname)) &&
-                            (acl->mapped_lun == deve->mapped_lun))
+                       if (!strcmp(acl->initiatorname, nacl->initiatorname) &&
+                           (acl->mapped_lun == deve->mapped_lun))
                                break;
                }
 
                if (!acl) {
-                       printk(KERN_ERR "Unable to locate struct se_lun_acl for %s,"
+                       pr_err("Unable to locate struct se_lun_acl for %s,"
                                " mapped_lun: %u\n", nacl->initiatorname,
                                deve->mapped_lun);
                        spin_unlock(&lun->lun_acl_lock);
@@ -115,7 +120,7 @@ struct se_node_acl *__core_tpg_get_initiator_node_acl(
        struct se_node_acl *acl;
 
        list_for_each_entry(acl, &tpg->acl_node_list, acl_list) {
-               if (!(strcmp(acl->initiatorname, initiatorname)))
+               if (!strcmp(acl->initiatorname, initiatorname))
                        return acl;
        }
 
@@ -134,8 +139,8 @@ struct se_node_acl *core_tpg_get_initiator_node_acl(
 
        spin_lock_bh(&tpg->acl_node_lock);
        list_for_each_entry(acl, &tpg->acl_node_list, acl_list) {
-               if (!(strcmp(acl->initiatorname, initiatorname)) &&
-                  (!(acl->dynamic_node_acl))) {
+               if (!strcmp(acl->initiatorname, initiatorname) &&
+                   !acl->dynamic_node_acl) {
                        spin_unlock_bh(&tpg->acl_node_lock);
                        return acl;
                }
@@ -171,7 +176,7 @@ void core_tpg_add_node_to_devs(
                 * By default in LIO-Target $FABRIC_MOD,
                 * demo_mode_write_protect is ON, or READ_ONLY;
                 */
-               if (!(TPG_TFO(tpg)->tpg_check_demo_mode_write_protect(tpg))) {
+               if (!tpg->se_tpg_tfo->tpg_check_demo_mode_write_protect(tpg)) {
                        if (dev->dev_flags & DF_READ_ONLY)
                                lun_access = TRANSPORT_LUNFLAGS_READ_ONLY;
                        else
@@ -181,16 +186,16 @@ void core_tpg_add_node_to_devs(
                         * Allow only optical drives to issue R/W in default RO
                         * demo mode.
                         */
-                       if (TRANSPORT(dev)->get_device_type(dev) == TYPE_DISK)
+                       if (dev->transport->get_device_type(dev) == TYPE_DISK)
                                lun_access = TRANSPORT_LUNFLAGS_READ_ONLY;
                        else
                                lun_access = TRANSPORT_LUNFLAGS_READ_WRITE;
                }
 
-               printk(KERN_INFO "TARGET_CORE[%s]->TPG[%u]_LUN[%u] - Adding %s"
+               pr_debug("TARGET_CORE[%s]->TPG[%u]_LUN[%u] - Adding %s"
                        " access for LUN in Demo Mode\n",
-                       TPG_TFO(tpg)->get_fabric_name(),
-                       TPG_TFO(tpg)->tpg_get_tag(tpg), lun->unpacked_lun,
+                       tpg->se_tpg_tfo->get_fabric_name(),
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg), lun->unpacked_lun,
                        (lun_access == TRANSPORT_LUNFLAGS_READ_WRITE) ?
                        "READ-WRITE" : "READ-ONLY");
 
@@ -210,8 +215,8 @@ static int core_set_queue_depth_for_node(
        struct se_node_acl *acl)
 {
        if (!acl->queue_depth) {
-               printk(KERN_ERR "Queue depth for %s Initiator Node: %s is 0,"
-                       "defaulting to 1.\n", TPG_TFO(tpg)->get_fabric_name(),
+               pr_err("Queue depth for %s Initiator Node: %s is 0,"
+                       "defaulting to 1.\n", tpg->se_tpg_tfo->get_fabric_name(),
                        acl->initiatorname);
                acl->queue_depth = 1;
        }
@@ -230,10 +235,10 @@ static int core_create_device_list_for_node(struct se_node_acl *nacl)
 
        nacl->device_list = kzalloc(sizeof(struct se_dev_entry) *
                                TRANSPORT_MAX_LUNS_PER_TPG, GFP_KERNEL);
-       if (!(nacl->device_list)) {
-               printk(KERN_ERR "Unable to allocate memory for"
+       if (!nacl->device_list) {
+               pr_err("Unable to allocate memory for"
                        " struct se_node_acl->device_list\n");
-               return -1;
+               return -ENOMEM;
        }
        for (i = 0; i < TRANSPORT_MAX_LUNS_PER_TPG; i++) {
                deve = &nacl->device_list[i];
@@ -259,14 +264,14 @@ struct se_node_acl *core_tpg_check_initiator_node_acl(
        struct se_node_acl *acl;
 
        acl = core_tpg_get_initiator_node_acl(tpg, initiatorname);
-       if ((acl))
+       if (acl)
                return acl;
 
-       if (!(TPG_TFO(tpg)->tpg_check_demo_mode(tpg)))
+       if (!tpg->se_tpg_tfo->tpg_check_demo_mode(tpg))
                return NULL;
 
-       acl =  TPG_TFO(tpg)->tpg_alloc_fabric_acl(tpg);
-       if (!(acl))
+       acl =  tpg->se_tpg_tfo->tpg_alloc_fabric_acl(tpg);
+       if (!acl)
                return NULL;
 
        INIT_LIST_HEAD(&acl->acl_list);
@@ -274,23 +279,23 @@ struct se_node_acl *core_tpg_check_initiator_node_acl(
        spin_lock_init(&acl->device_list_lock);
        spin_lock_init(&acl->nacl_sess_lock);
        atomic_set(&acl->acl_pr_ref_count, 0);
-       acl->queue_depth = TPG_TFO(tpg)->tpg_get_default_depth(tpg);
+       acl->queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg);
        snprintf(acl->initiatorname, TRANSPORT_IQN_LEN, "%s", initiatorname);
        acl->se_tpg = tpg;
        acl->acl_index = scsi_get_new_index(SCSI_AUTH_INTR_INDEX);
        spin_lock_init(&acl->stats_lock);
        acl->dynamic_node_acl = 1;
 
-       TPG_TFO(tpg)->set_default_node_attributes(acl);
+       tpg->se_tpg_tfo->set_default_node_attributes(acl);
 
        if (core_create_device_list_for_node(acl) < 0) {
-               TPG_TFO(tpg)->tpg_release_fabric_acl(tpg, acl);
+               tpg->se_tpg_tfo->tpg_release_fabric_acl(tpg, acl);
                return NULL;
        }
 
        if (core_set_queue_depth_for_node(tpg, acl) < 0) {
                core_free_device_list_for_node(acl, tpg);
-               TPG_TFO(tpg)->tpg_release_fabric_acl(tpg, acl);
+               tpg->se_tpg_tfo->tpg_release_fabric_acl(tpg, acl);
                return NULL;
        }
 
@@ -301,10 +306,10 @@ struct se_node_acl *core_tpg_check_initiator_node_acl(
        tpg->num_node_acls++;
        spin_unlock_bh(&tpg->acl_node_lock);
 
-       printk("%s_TPG[%u] - Added DYNAMIC ACL with TCQ Depth: %d for %s"
-               " Initiator Node: %s\n", TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg), acl->queue_depth,
-               TPG_TFO(tpg)->get_fabric_name(), initiatorname);
+       pr_debug("%s_TPG[%u] - Added DYNAMIC ACL with TCQ Depth: %d for %s"
+               " Initiator Node: %s\n", tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg), acl->queue_depth,
+               tpg->se_tpg_tfo->get_fabric_name(), initiatorname);
 
        return acl;
 }
@@ -351,12 +356,12 @@ struct se_node_acl *core_tpg_add_initiator_node_acl(
 
        spin_lock_bh(&tpg->acl_node_lock);
        acl = __core_tpg_get_initiator_node_acl(tpg, initiatorname);
-       if ((acl)) {
+       if (acl) {
                if (acl->dynamic_node_acl) {
                        acl->dynamic_node_acl = 0;
-                       printk(KERN_INFO "%s_TPG[%u] - Replacing dynamic ACL"
-                               " for %s\n", TPG_TFO(tpg)->get_fabric_name(),
-                               TPG_TFO(tpg)->tpg_get_tag(tpg), initiatorname);
+                       pr_debug("%s_TPG[%u] - Replacing dynamic ACL"
+                               " for %s\n", tpg->se_tpg_tfo->get_fabric_name(),
+                               tpg->se_tpg_tfo->tpg_get_tag(tpg), initiatorname);
                        spin_unlock_bh(&tpg->acl_node_lock);
                        /*
                         * Release the locally allocated struct se_node_acl
@@ -364,22 +369,22 @@ struct se_node_acl *core_tpg_add_initiator_node_acl(
                         * a pointer to an existing demo mode node ACL.
                         */
                        if (se_nacl)
-                               TPG_TFO(tpg)->tpg_release_fabric_acl(tpg,
+                               tpg->se_tpg_tfo->tpg_release_fabric_acl(tpg,
                                                        se_nacl);
                        goto done;
                }
 
-               printk(KERN_ERR "ACL entry for %s Initiator"
+               pr_err("ACL entry for %s Initiator"
                        " Node %s already exists for TPG %u, ignoring"
-                       " request.\n",  TPG_TFO(tpg)->get_fabric_name(),
-                       initiatorname, TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       " request.\n",  tpg->se_tpg_tfo->get_fabric_name(),
+                       initiatorname, tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock_bh(&tpg->acl_node_lock);
                return ERR_PTR(-EEXIST);
        }
        spin_unlock_bh(&tpg->acl_node_lock);
 
-       if (!(se_nacl)) {
-               printk("struct se_node_acl pointer is NULL\n");
+       if (!se_nacl) {
+               pr_err("struct se_node_acl pointer is NULL\n");
                return ERR_PTR(-EINVAL);
        }
        /*
@@ -400,16 +405,16 @@ struct se_node_acl *core_tpg_add_initiator_node_acl(
        acl->acl_index = scsi_get_new_index(SCSI_AUTH_INTR_INDEX);
        spin_lock_init(&acl->stats_lock);
 
-       TPG_TFO(tpg)->set_default_node_attributes(acl);
+       tpg->se_tpg_tfo->set_default_node_attributes(acl);
 
        if (core_create_device_list_for_node(acl) < 0) {
-               TPG_TFO(tpg)->tpg_release_fabric_acl(tpg, acl);
+               tpg->se_tpg_tfo->tpg_release_fabric_acl(tpg, acl);
                return ERR_PTR(-ENOMEM);
        }
 
        if (core_set_queue_depth_for_node(tpg, acl) < 0) {
                core_free_device_list_for_node(acl, tpg);
-               TPG_TFO(tpg)->tpg_release_fabric_acl(tpg, acl);
+               tpg->se_tpg_tfo->tpg_release_fabric_acl(tpg, acl);
                return ERR_PTR(-EINVAL);
        }
 
@@ -419,10 +424,10 @@ struct se_node_acl *core_tpg_add_initiator_node_acl(
        spin_unlock_bh(&tpg->acl_node_lock);
 
 done:
-       printk(KERN_INFO "%s_TPG[%hu] - Added ACL with TCQ Depth: %d for %s"
-               " Initiator Node: %s\n", TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg), acl->queue_depth,
-               TPG_TFO(tpg)->get_fabric_name(), initiatorname);
+       pr_debug("%s_TPG[%hu] - Added ACL with TCQ Depth: %d for %s"
+               " Initiator Node: %s\n", tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg), acl->queue_depth,
+               tpg->se_tpg_tfo->get_fabric_name(), initiatorname);
 
        return acl;
 }
@@ -457,7 +462,7 @@ int core_tpg_del_initiator_node_acl(
                /*
                 * Determine if the session needs to be closed by our context.
                 */
-               if (!(TPG_TFO(tpg)->shutdown_session(sess)))
+               if (!tpg->se_tpg_tfo->shutdown_session(sess))
                        continue;
 
                spin_unlock_bh(&tpg->session_lock);
@@ -465,7 +470,7 @@ int core_tpg_del_initiator_node_acl(
                 * If the $FABRIC_MOD session for the Initiator Node ACL exists,
                 * forcefully shutdown the $FABRIC_MOD session/nexus.
                 */
-               TPG_TFO(tpg)->close_session(sess);
+               tpg->se_tpg_tfo->close_session(sess);
 
                spin_lock_bh(&tpg->session_lock);
        }
@@ -475,10 +480,10 @@ int core_tpg_del_initiator_node_acl(
        core_clear_initiator_node_from_tpg(acl, tpg);
        core_free_device_list_for_node(acl, tpg);
 
-       printk(KERN_INFO "%s_TPG[%hu] - Deleted ACL with TCQ Depth: %d for %s"
-               " Initiator Node: %s\n", TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg), acl->queue_depth,
-               TPG_TFO(tpg)->get_fabric_name(), acl->initiatorname);
+       pr_debug("%s_TPG[%hu] - Deleted ACL with TCQ Depth: %d for %s"
+               " Initiator Node: %s\n", tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg), acl->queue_depth,
+               tpg->se_tpg_tfo->get_fabric_name(), acl->initiatorname);
 
        return 0;
 }
@@ -500,11 +505,11 @@ int core_tpg_set_initiator_node_queue_depth(
 
        spin_lock_bh(&tpg->acl_node_lock);
        acl = __core_tpg_get_initiator_node_acl(tpg, initiatorname);
-       if (!(acl)) {
-               printk(KERN_ERR "Access Control List entry for %s Initiator"
+       if (!acl) {
+               pr_err("Access Control List entry for %s Initiator"
                        " Node %s does not exists for TPG %hu, ignoring"
-                       " request.\n", TPG_TFO(tpg)->get_fabric_name(),
-                       initiatorname, TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       " request.\n", tpg->se_tpg_tfo->get_fabric_name(),
+                       initiatorname, tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock_bh(&tpg->acl_node_lock);
                return -ENODEV;
        }
@@ -520,12 +525,12 @@ int core_tpg_set_initiator_node_queue_depth(
                        continue;
 
                if (!force) {
-                       printk(KERN_ERR "Unable to change queue depth for %s"
+                       pr_err("Unable to change queue depth for %s"
                                " Initiator Node: %s while session is"
                                " operational.  To forcefully change the queue"
                                " depth and force session reinstatement"
                                " use the \"force=1\" parameter.\n",
-                               TPG_TFO(tpg)->get_fabric_name(), initiatorname);
+                               tpg->se_tpg_tfo->get_fabric_name(), initiatorname);
                        spin_unlock_bh(&tpg->session_lock);
 
                        spin_lock_bh(&tpg->acl_node_lock);
@@ -537,7 +542,7 @@ int core_tpg_set_initiator_node_queue_depth(
                /*
                 * Determine if the session needs to be closed by our context.
                 */
-               if (!(TPG_TFO(tpg)->shutdown_session(sess)))
+               if (!tpg->se_tpg_tfo->shutdown_session(sess))
                        continue;
 
                init_sess = sess;
@@ -549,7 +554,7 @@ int core_tpg_set_initiator_node_queue_depth(
         * Change the value in the Node's struct se_node_acl, and call
         * core_set_queue_depth_for_node() to add the requested queue depth.
         *
-        * Finally call  TPG_TFO(tpg)->close_session() to force session
+        * Finally call  tpg->se_tpg_tfo->close_session() to force session
         * reinstatement to occur if there is an active session for the
         * $FABRIC_MOD Initiator Node in question.
         */
@@ -561,10 +566,10 @@ int core_tpg_set_initiator_node_queue_depth(
                 * Force session reinstatement if
                 * core_set_queue_depth_for_node() failed, because we assume
                 * the $FABRIC_MOD has already the set session reinstatement
-                * bit from TPG_TFO(tpg)->shutdown_session() called above.
+                * bit from tpg->se_tpg_tfo->shutdown_session() called above.
                 */
                if (init_sess)
-                       TPG_TFO(tpg)->close_session(init_sess);
+                       tpg->se_tpg_tfo->close_session(init_sess);
 
                spin_lock_bh(&tpg->acl_node_lock);
                if (dynamic_acl)
@@ -578,12 +583,12 @@ int core_tpg_set_initiator_node_queue_depth(
         * forcefully shutdown the $FABRIC_MOD session/nexus.
         */
        if (init_sess)
-               TPG_TFO(tpg)->close_session(init_sess);
+               tpg->se_tpg_tfo->close_session(init_sess);
 
-       printk(KERN_INFO "Successfuly changed queue depth to: %d for Initiator"
+       pr_debug("Successfuly changed queue depth to: %d for Initiator"
                " Node: %s on %s Target Portal Group: %u\n", queue_depth,
-               initiatorname, TPG_TFO(tpg)->get_fabric_name(),
-               TPG_TFO(tpg)->tpg_get_tag(tpg));
+               initiatorname, tpg->se_tpg_tfo->get_fabric_name(),
+               tpg->se_tpg_tfo->tpg_get_tag(tpg));
 
        spin_lock_bh(&tpg->acl_node_lock);
        if (dynamic_acl)
@@ -597,7 +602,7 @@ EXPORT_SYMBOL(core_tpg_set_initiator_node_queue_depth);
 static int core_tpg_setup_virtual_lun0(struct se_portal_group *se_tpg)
 {
        /* Set in core_dev_setup_virtual_lun0() */
-       struct se_device *dev = se_global->g_lun0_dev;
+       struct se_device *dev = g_lun0_dev;
        struct se_lun *lun = &se_tpg->tpg_virt_lun0;
        u32 lun_access = TRANSPORT_LUNFLAGS_READ_ONLY;
        int ret;
@@ -614,7 +619,7 @@ static int core_tpg_setup_virtual_lun0(struct se_portal_group *se_tpg)
 
        ret = core_tpg_post_addlun(se_tpg, lun, lun_access, dev);
        if (ret < 0)
-               return -1;
+               return ret;
 
        return 0;
 }
@@ -638,8 +643,8 @@ int core_tpg_register(
 
        se_tpg->tpg_lun_list = kzalloc((sizeof(struct se_lun) *
                                TRANSPORT_MAX_LUNS_PER_TPG), GFP_KERNEL);
-       if (!(se_tpg->tpg_lun_list)) {
-               printk(KERN_ERR "Unable to allocate struct se_portal_group->"
+       if (!se_tpg->tpg_lun_list) {
+               pr_err("Unable to allocate struct se_portal_group->"
                                "tpg_lun_list\n");
                return -ENOMEM;
        }
@@ -663,7 +668,7 @@ int core_tpg_register(
        se_tpg->se_tpg_wwn = se_wwn;
        atomic_set(&se_tpg->tpg_pr_ref_count, 0);
        INIT_LIST_HEAD(&se_tpg->acl_node_list);
-       INIT_LIST_HEAD(&se_tpg->se_tpg_list);
+       INIT_LIST_HEAD(&se_tpg->se_tpg_node);
        INIT_LIST_HEAD(&se_tpg->tpg_sess_list);
        spin_lock_init(&se_tpg->acl_node_lock);
        spin_lock_init(&se_tpg->session_lock);
@@ -676,11 +681,11 @@ int core_tpg_register(
                }
        }
 
-       spin_lock_bh(&se_global->se_tpg_lock);
-       list_add_tail(&se_tpg->se_tpg_list, &se_global->g_se_tpg_list);
-       spin_unlock_bh(&se_global->se_tpg_lock);
+       spin_lock_bh(&tpg_lock);
+       list_add_tail(&se_tpg->se_tpg_node, &tpg_list);
+       spin_unlock_bh(&tpg_lock);
 
-       printk(KERN_INFO "TARGET_CORE[%s]: Allocated %s struct se_portal_group for"
+       pr_debug("TARGET_CORE[%s]: Allocated %s struct se_portal_group for"
                " endpoint: %s, Portal Tag: %u\n", tfo->get_fabric_name(),
                (se_tpg->se_tpg_type == TRANSPORT_TPG_TYPE_NORMAL) ?
                "Normal" : "Discovery", (tfo->tpg_get_wwn(se_tpg) == NULL) ?
@@ -694,16 +699,16 @@ int core_tpg_deregister(struct se_portal_group *se_tpg)
 {
        struct se_node_acl *nacl, *nacl_tmp;
 
-       printk(KERN_INFO "TARGET_CORE[%s]: Deallocating %s struct se_portal_group"
+       pr_debug("TARGET_CORE[%s]: Deallocating %s struct se_portal_group"
                " for endpoint: %s Portal Tag %u\n",
                (se_tpg->se_tpg_type == TRANSPORT_TPG_TYPE_NORMAL) ?
-               "Normal" : "Discovery", TPG_TFO(se_tpg)->get_fabric_name(),
-               TPG_TFO(se_tpg)->tpg_get_wwn(se_tpg),
-               TPG_TFO(se_tpg)->tpg_get_tag(se_tpg));
+               "Normal" : "Discovery", se_tpg->se_tpg_tfo->get_fabric_name(),
+               se_tpg->se_tpg_tfo->tpg_get_wwn(se_tpg),
+               se_tpg->se_tpg_tfo->tpg_get_tag(se_tpg));
 
-       spin_lock_bh(&se_global->se_tpg_lock);
-       list_del(&se_tpg->se_tpg_list);
-       spin_unlock_bh(&se_global->se_tpg_lock);
+       spin_lock_bh(&tpg_lock);
+       list_del(&se_tpg->se_tpg_node);
+       spin_unlock_bh(&tpg_lock);
 
        while (atomic_read(&se_tpg->tpg_pr_ref_count) != 0)
                cpu_relax();
@@ -721,7 +726,7 @@ int core_tpg_deregister(struct se_portal_group *se_tpg)
 
                core_tpg_wait_for_nacl_pr_ref(nacl);
                core_free_device_list_for_node(nacl, se_tpg);
-               TPG_TFO(se_tpg)->tpg_release_fabric_acl(se_tpg, nacl);
+               se_tpg->se_tpg_tfo->tpg_release_fabric_acl(se_tpg, nacl);
 
                spin_lock_bh(&se_tpg->acl_node_lock);
        }
@@ -743,21 +748,21 @@ struct se_lun *core_tpg_pre_addlun(
        struct se_lun *lun;
 
        if (unpacked_lun > (TRANSPORT_MAX_LUNS_PER_TPG-1)) {
-               printk(KERN_ERR "%s LUN: %u exceeds TRANSPORT_MAX_LUNS_PER_TPG"
+               pr_err("%s LUN: %u exceeds TRANSPORT_MAX_LUNS_PER_TPG"
                        "-1: %u for Target Portal Group: %u\n",
-                       TPG_TFO(tpg)->get_fabric_name(),
+                       tpg->se_tpg_tfo->get_fabric_name(),
                        unpacked_lun, TRANSPORT_MAX_LUNS_PER_TPG-1,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                return ERR_PTR(-EOVERFLOW);
        }
 
        spin_lock(&tpg->tpg_lun_lock);
        lun = &tpg->tpg_lun_list[unpacked_lun];
        if (lun->lun_status == TRANSPORT_LUN_STATUS_ACTIVE) {
-               printk(KERN_ERR "TPG Logical Unit Number: %u is already active"
+               pr_err("TPG Logical Unit Number: %u is already active"
                        " on %s Target Portal Group: %u, ignoring request.\n",
-                       unpacked_lun, TPG_TFO(tpg)->get_fabric_name(),
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       unpacked_lun, tpg->se_tpg_tfo->get_fabric_name(),
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock(&tpg->tpg_lun_lock);
                return ERR_PTR(-EINVAL);
        }
@@ -772,8 +777,11 @@ int core_tpg_post_addlun(
        u32 lun_access,
        void *lun_ptr)
 {
-       if (core_dev_export(lun_ptr, tpg, lun) < 0)
-               return -1;
+       int ret;
+
+       ret = core_dev_export(lun_ptr, tpg, lun);
+       if (ret < 0)
+               return ret;
 
        spin_lock(&tpg->tpg_lun_lock);
        lun->lun_access = lun_access;
@@ -799,21 +807,21 @@ struct se_lun *core_tpg_pre_dellun(
        struct se_lun *lun;
 
        if (unpacked_lun > (TRANSPORT_MAX_LUNS_PER_TPG-1)) {
-               printk(KERN_ERR "%s LUN: %u exceeds TRANSPORT_MAX_LUNS_PER_TPG"
+               pr_err("%s LUN: %u exceeds TRANSPORT_MAX_LUNS_PER_TPG"
                        "-1: %u for Target Portal Group: %u\n",
-                       TPG_TFO(tpg)->get_fabric_name(), unpacked_lun,
+                       tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
                        TRANSPORT_MAX_LUNS_PER_TPG-1,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                return ERR_PTR(-EOVERFLOW);
        }
 
        spin_lock(&tpg->tpg_lun_lock);
        lun = &tpg->tpg_lun_list[unpacked_lun];
        if (lun->lun_status != TRANSPORT_LUN_STATUS_ACTIVE) {
-               printk(KERN_ERR "%s Logical Unit Number: %u is not active on"
+               pr_err("%s Logical Unit Number: %u is not active on"
                        " Target Portal Group: %u, ignoring request.\n",
-                       TPG_TFO(tpg)->get_fabric_name(), unpacked_lun,
-                       TPG_TFO(tpg)->tpg_get_tag(tpg));
+                       tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
+                       tpg->se_tpg_tfo->tpg_get_tag(tpg));
                spin_unlock(&tpg->tpg_lun_lock);
                return ERR_PTR(-ENODEV);
        }
index 4b9b7169bdd96957d9a6054ea68ec99c897a2381..46352d658e35b5d2698dc95d75ce9d6eb1c1cd58 100644 (file)
 #include "target_core_scdb.h"
 #include "target_core_ua.h"
 
-/* #define DEBUG_CDB_HANDLER */
-#ifdef DEBUG_CDB_HANDLER
-#define DEBUG_CDB_H(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_CDB_H(x...)
-#endif
-
-/* #define DEBUG_CMD_MAP */
-#ifdef DEBUG_CMD_MAP
-#define DEBUG_CMD_M(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_CMD_M(x...)
-#endif
-
-/* #define DEBUG_MEM_ALLOC */
-#ifdef DEBUG_MEM_ALLOC
-#define DEBUG_MEM(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_MEM(x...)
-#endif
-
-/* #define DEBUG_MEM2_ALLOC */
-#ifdef DEBUG_MEM2_ALLOC
-#define DEBUG_MEM2(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_MEM2(x...)
-#endif
-
-/* #define DEBUG_SG_CALC */
-#ifdef DEBUG_SG_CALC
-#define DEBUG_SC(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_SC(x...)
-#endif
-
-/* #define DEBUG_SE_OBJ */
-#ifdef DEBUG_SE_OBJ
-#define DEBUG_SO(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_SO(x...)
-#endif
-
-/* #define DEBUG_CMD_VOL */
-#ifdef DEBUG_CMD_VOL
-#define DEBUG_VOL(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_VOL(x...)
-#endif
-
-/* #define DEBUG_CMD_STOP */
-#ifdef DEBUG_CMD_STOP
-#define DEBUG_CS(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_CS(x...)
-#endif
-
-/* #define DEBUG_PASSTHROUGH */
-#ifdef DEBUG_PASSTHROUGH
-#define DEBUG_PT(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_PT(x...)
-#endif
-
-/* #define DEBUG_TASK_STOP */
-#ifdef DEBUG_TASK_STOP
-#define DEBUG_TS(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_TS(x...)
-#endif
-
-/* #define DEBUG_TRANSPORT_STOP */
-#ifdef DEBUG_TRANSPORT_STOP
-#define DEBUG_TRANSPORT_S(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_TRANSPORT_S(x...)
-#endif
-
-/* #define DEBUG_TASK_FAILURE */
-#ifdef DEBUG_TASK_FAILURE
-#define DEBUG_TF(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_TF(x...)
-#endif
-
-/* #define DEBUG_DEV_OFFLINE */
-#ifdef DEBUG_DEV_OFFLINE
-#define DEBUG_DO(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_DO(x...)
-#endif
-
-/* #define DEBUG_TASK_STATE */
-#ifdef DEBUG_TASK_STATE
-#define DEBUG_TSTATE(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_TSTATE(x...)
-#endif
-
-/* #define DEBUG_STATUS_THR */
-#ifdef DEBUG_STATUS_THR
-#define DEBUG_ST(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_ST(x...)
-#endif
-
-/* #define DEBUG_TASK_TIMEOUT */
-#ifdef DEBUG_TASK_TIMEOUT
-#define DEBUG_TT(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_TT(x...)
-#endif
-
-/* #define DEBUG_GENERIC_REQUEST_FAILURE */
-#ifdef DEBUG_GENERIC_REQUEST_FAILURE
-#define DEBUG_GRF(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_GRF(x...)
-#endif
-
-/* #define DEBUG_SAM_TASK_ATTRS */
-#ifdef DEBUG_SAM_TASK_ATTRS
-#define DEBUG_STA(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_STA(x...)
-#endif
-
-struct se_global *se_global;
+static int sub_api_initialized;
 
 static struct kmem_cache *se_cmd_cache;
 static struct kmem_cache *se_sess_cache;
 struct kmem_cache *se_tmr_req_cache;
 struct kmem_cache *se_ua_cache;
-struct kmem_cache *se_mem_cache;
 struct kmem_cache *t10_pr_reg_cache;
 struct kmem_cache *t10_alua_lu_gp_cache;
 struct kmem_cache *t10_alua_lu_gp_mem_cache;
@@ -201,116 +74,87 @@ struct kmem_cache *t10_alua_tg_pt_gp_mem_cache;
 typedef int (*map_func_t)(struct se_task *, u32);
 
 static int transport_generic_write_pending(struct se_cmd *);
-static int transport_processing_thread(void *);
+static int transport_processing_thread(void *param);
 static int __transport_execute_tasks(struct se_device *dev);
 static void transport_complete_task_attr(struct se_cmd *cmd);
+static int transport_complete_qf(struct se_cmd *cmd);
+static void transport_handle_queue_full(struct se_cmd *cmd,
+               struct se_device *dev, int (*qf_callback)(struct se_cmd *));
 static void transport_direct_request_timeout(struct se_cmd *cmd);
 static void transport_free_dev_tasks(struct se_cmd *cmd);
-static u32 transport_generic_get_cdb_count(struct se_cmd *cmd,
-               unsigned long long starting_lba, u32 sectors,
+static u32 transport_allocate_tasks(struct se_cmd *cmd,
+               unsigned long long starting_lba,
                enum dma_data_direction data_direction,
-               struct list_head *mem_list, int set_counts);
-static int transport_generic_get_mem(struct se_cmd *cmd, u32 length,
-               u32 dma_size);
+               struct scatterlist *sgl, unsigned int nents);
+static int transport_generic_get_mem(struct se_cmd *cmd);
 static int transport_generic_remove(struct se_cmd *cmd,
-               int release_to_pool, int session_reinstatement);
-static int transport_get_sectors(struct se_cmd *cmd);
-static struct list_head *transport_init_se_mem_list(void);
-static int transport_map_sg_to_mem(struct se_cmd *cmd,
-               struct list_head *se_mem_list, void *in_mem,
-               u32 *se_mem_cnt);
-static void transport_memcpy_se_mem_read_contig(struct se_cmd *cmd,
-               unsigned char *dst, struct list_head *se_mem_list);
+               int session_reinstatement);
 static void transport_release_fe_cmd(struct se_cmd *cmd);
 static void transport_remove_cmd_from_queue(struct se_cmd *cmd,
                struct se_queue_obj *qobj);
 static int transport_set_sense_codes(struct se_cmd *cmd, u8 asc, u8 ascq);
 static void transport_stop_all_task_timers(struct se_cmd *cmd);
 
-int init_se_global(void)
+int init_se_kmem_caches(void)
 {
-       struct se_global *global;
-
-       global = kzalloc(sizeof(struct se_global), GFP_KERNEL);
-       if (!(global)) {
-               printk(KERN_ERR "Unable to allocate memory for struct se_global\n");
-               return -1;
-       }
-
-       INIT_LIST_HEAD(&global->g_lu_gps_list);
-       INIT_LIST_HEAD(&global->g_se_tpg_list);
-       INIT_LIST_HEAD(&global->g_hba_list);
-       INIT_LIST_HEAD(&global->g_se_dev_list);
-       spin_lock_init(&global->g_device_lock);
-       spin_lock_init(&global->hba_lock);
-       spin_lock_init(&global->se_tpg_lock);
-       spin_lock_init(&global->lu_gps_lock);
-       spin_lock_init(&global->plugin_class_lock);
-
        se_cmd_cache = kmem_cache_create("se_cmd_cache",
                        sizeof(struct se_cmd), __alignof__(struct se_cmd), 0, NULL);
-       if (!(se_cmd_cache)) {
-               printk(KERN_ERR "kmem_cache_create for struct se_cmd failed\n");
+       if (!se_cmd_cache) {
+               pr_err("kmem_cache_create for struct se_cmd failed\n");
                goto out;
        }
        se_tmr_req_cache = kmem_cache_create("se_tmr_cache",
                        sizeof(struct se_tmr_req), __alignof__(struct se_tmr_req),
                        0, NULL);
-       if (!(se_tmr_req_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for struct se_tmr_req"
+       if (!se_tmr_req_cache) {
+               pr_err("kmem_cache_create() for struct se_tmr_req"
                                " failed\n");
                goto out;
        }
        se_sess_cache = kmem_cache_create("se_sess_cache",
                        sizeof(struct se_session), __alignof__(struct se_session),
                        0, NULL);
-       if (!(se_sess_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for struct se_session"
+       if (!se_sess_cache) {
+               pr_err("kmem_cache_create() for struct se_session"
                                " failed\n");
                goto out;
        }
        se_ua_cache = kmem_cache_create("se_ua_cache",
                        sizeof(struct se_ua), __alignof__(struct se_ua),
                        0, NULL);
-       if (!(se_ua_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for struct se_ua failed\n");
-               goto out;
-       }
-       se_mem_cache = kmem_cache_create("se_mem_cache",
-                       sizeof(struct se_mem), __alignof__(struct se_mem), 0, NULL);
-       if (!(se_mem_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for struct se_mem failed\n");
+       if (!se_ua_cache) {
+               pr_err("kmem_cache_create() for struct se_ua failed\n");
                goto out;
        }
        t10_pr_reg_cache = kmem_cache_create("t10_pr_reg_cache",
                        sizeof(struct t10_pr_registration),
                        __alignof__(struct t10_pr_registration), 0, NULL);
-       if (!(t10_pr_reg_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for struct t10_pr_registration"
+       if (!t10_pr_reg_cache) {
+               pr_err("kmem_cache_create() for struct t10_pr_registration"
                                " failed\n");
                goto out;
        }
        t10_alua_lu_gp_cache = kmem_cache_create("t10_alua_lu_gp_cache",
                        sizeof(struct t10_alua_lu_gp), __alignof__(struct t10_alua_lu_gp),
                        0, NULL);
-       if (!(t10_alua_lu_gp_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for t10_alua_lu_gp_cache"
+       if (!t10_alua_lu_gp_cache) {
+               pr_err("kmem_cache_create() for t10_alua_lu_gp_cache"
                                " failed\n");
                goto out;
        }
        t10_alua_lu_gp_mem_cache = kmem_cache_create("t10_alua_lu_gp_mem_cache",
                        sizeof(struct t10_alua_lu_gp_member),
                        __alignof__(struct t10_alua_lu_gp_member), 0, NULL);
-       if (!(t10_alua_lu_gp_mem_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for t10_alua_lu_gp_mem_"
+       if (!t10_alua_lu_gp_mem_cache) {
+               pr_err("kmem_cache_create() for t10_alua_lu_gp_mem_"
                                "cache failed\n");
                goto out;
        }
        t10_alua_tg_pt_gp_cache = kmem_cache_create("t10_alua_tg_pt_gp_cache",
                        sizeof(struct t10_alua_tg_pt_gp),
                        __alignof__(struct t10_alua_tg_pt_gp), 0, NULL);
-       if (!(t10_alua_tg_pt_gp_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for t10_alua_tg_pt_gp_"
+       if (!t10_alua_tg_pt_gp_cache) {
+               pr_err("kmem_cache_create() for t10_alua_tg_pt_gp_"
                                "cache failed\n");
                goto out;
        }
@@ -319,14 +163,12 @@ int init_se_global(void)
                        sizeof(struct t10_alua_tg_pt_gp_member),
                        __alignof__(struct t10_alua_tg_pt_gp_member),
                        0, NULL);
-       if (!(t10_alua_tg_pt_gp_mem_cache)) {
-               printk(KERN_ERR "kmem_cache_create() for t10_alua_tg_pt_gp_"
+       if (!t10_alua_tg_pt_gp_mem_cache) {
+               pr_err("kmem_cache_create() for t10_alua_tg_pt_gp_"
                                "mem_t failed\n");
                goto out;
        }
 
-       se_global = global;
-
        return 0;
 out:
        if (se_cmd_cache)
@@ -337,8 +179,6 @@ out:
                kmem_cache_destroy(se_sess_cache);
        if (se_ua_cache)
                kmem_cache_destroy(se_ua_cache);
-       if (se_mem_cache)
-               kmem_cache_destroy(se_mem_cache);
        if (t10_pr_reg_cache)
                kmem_cache_destroy(t10_pr_reg_cache);
        if (t10_alua_lu_gp_cache)
@@ -349,45 +189,25 @@ out:
                kmem_cache_destroy(t10_alua_tg_pt_gp_cache);
        if (t10_alua_tg_pt_gp_mem_cache)
                kmem_cache_destroy(t10_alua_tg_pt_gp_mem_cache);
-       kfree(global);
-       return -1;
+       return -ENOMEM;
 }
 
-void release_se_global(void)
+void release_se_kmem_caches(void)
 {
-       struct se_global *global;
-
-       global = se_global;
-       if (!(global))
-               return;
-
        kmem_cache_destroy(se_cmd_cache);
        kmem_cache_destroy(se_tmr_req_cache);
        kmem_cache_destroy(se_sess_cache);
        kmem_cache_destroy(se_ua_cache);
-       kmem_cache_destroy(se_mem_cache);
        kmem_cache_destroy(t10_pr_reg_cache);
        kmem_cache_destroy(t10_alua_lu_gp_cache);
        kmem_cache_destroy(t10_alua_lu_gp_mem_cache);
        kmem_cache_destroy(t10_alua_tg_pt_gp_cache);
        kmem_cache_destroy(t10_alua_tg_pt_gp_mem_cache);
-       kfree(global);
-
-       se_global = NULL;
 }
 
-/* SCSI statistics table index */
-static struct scsi_index_table scsi_index_table;
-
-/*
- * Initialize the index table for allocating unique row indexes to various mib
- * tables.
- */
-void init_scsi_index_table(void)
-{
-       memset(&scsi_index_table, 0, sizeof(struct scsi_index_table));
-       spin_lock_init(&scsi_index_table.lock);
-}
+/* This code ensures unique mib indexes are handed out. */
+static DEFINE_SPINLOCK(scsi_mib_index_lock);
+static u32 scsi_mib_index[SCSI_INDEX_TYPE_MAX];
 
 /*
  * Allocate a new row index for the entry type specified
@@ -396,16 +216,11 @@ u32 scsi_get_new_index(scsi_index_t type)
 {
        u32 new_index;
 
-       if ((type < 0) || (type >= SCSI_INDEX_TYPE_MAX)) {
-               printk(KERN_ERR "Invalid index type %d\n", type);
-               return -EINVAL;
-       }
+       BUG_ON((type < 0) || (type >= SCSI_INDEX_TYPE_MAX));
 
-       spin_lock(&scsi_index_table.lock);
-       new_index = ++scsi_index_table.scsi_mib_index[type];
-       if (new_index == 0)
-               new_index = ++scsi_index_table.scsi_mib_index[type];
-       spin_unlock(&scsi_index_table.lock);
+       spin_lock(&scsi_mib_index_lock);
+       new_index = ++scsi_mib_index[type];
+       spin_unlock(&scsi_mib_index_lock);
 
        return new_index;
 }
@@ -425,34 +240,37 @@ static int transport_subsystem_reqmods(void)
 
        ret = request_module("target_core_iblock");
        if (ret != 0)
-               printk(KERN_ERR "Unable to load target_core_iblock\n");
+               pr_err("Unable to load target_core_iblock\n");
 
        ret = request_module("target_core_file");
        if (ret != 0)
-               printk(KERN_ERR "Unable to load target_core_file\n");
+               pr_err("Unable to load target_core_file\n");
 
        ret = request_module("target_core_pscsi");
        if (ret != 0)
-               printk(KERN_ERR "Unable to load target_core_pscsi\n");
+               pr_err("Unable to load target_core_pscsi\n");
 
        ret = request_module("target_core_stgt");
        if (ret != 0)
-               printk(KERN_ERR "Unable to load target_core_stgt\n");
+               pr_err("Unable to load target_core_stgt\n");
 
        return 0;
 }
 
 int transport_subsystem_check_init(void)
 {
-       if (se_global->g_sub_api_initialized)
+       int ret;
+
+       if (sub_api_initialized)
                return 0;
        /*
         * Request the loading of known TCM subsystem plugins..
         */
-       if (transport_subsystem_reqmods() < 0)
-               return -1;
+       ret = transport_subsystem_reqmods();
+       if (ret < 0)
+               return ret;
 
-       se_global->g_sub_api_initialized = 1;
+       sub_api_initialized = 1;
        return 0;
 }
 
@@ -461,8 +279,8 @@ struct se_session *transport_init_session(void)
        struct se_session *se_sess;
 
        se_sess = kmem_cache_zalloc(se_sess_cache, GFP_KERNEL);
-       if (!(se_sess)) {
-               printk(KERN_ERR "Unable to allocate struct se_session from"
+       if (!se_sess) {
+               pr_err("Unable to allocate struct se_session from"
                                " se_sess_cache\n");
                return ERR_PTR(-ENOMEM);
        }
@@ -497,9 +315,9 @@ void __transport_register_session(
                 * If the fabric module supports an ISID based TransportID,
                 * save this value in binary from the fabric I_T Nexus now.
                 */
-               if (TPG_TFO(se_tpg)->sess_get_initiator_sid != NULL) {
+               if (se_tpg->se_tpg_tfo->sess_get_initiator_sid != NULL) {
                        memset(&buf[0], 0, PR_REG_ISID_LEN);
-                       TPG_TFO(se_tpg)->sess_get_initiator_sid(se_sess,
+                       se_tpg->se_tpg_tfo->sess_get_initiator_sid(se_sess,
                                        &buf[0], PR_REG_ISID_LEN);
                        se_sess->sess_bin_isid = get_unaligned_be64(&buf[0]);
                }
@@ -516,8 +334,8 @@ void __transport_register_session(
        }
        list_add_tail(&se_sess->sess_list, &se_tpg->tpg_sess_list);
 
-       printk(KERN_INFO "TARGET_CORE[%s]: Registered fabric_sess_ptr: %p\n",
-               TPG_TFO(se_tpg)->get_fabric_name(), se_sess->fabric_sess_ptr);
+       pr_debug("TARGET_CORE[%s]: Registered fabric_sess_ptr: %p\n",
+               se_tpg->se_tpg_tfo->get_fabric_name(), se_sess->fabric_sess_ptr);
 }
 EXPORT_SYMBOL(__transport_register_session);
 
@@ -541,7 +359,7 @@ void transport_deregister_session_configfs(struct se_session *se_sess)
         * Used by struct se_node_acl's under ConfigFS to locate active struct se_session
         */
        se_nacl = se_sess->se_node_acl;
-       if ((se_nacl)) {
+       if (se_nacl) {
                spin_lock_irqsave(&se_nacl->nacl_sess_lock, flags);
                list_del(&se_sess->sess_acl_list);
                /*
@@ -572,7 +390,7 @@ void transport_deregister_session(struct se_session *se_sess)
        struct se_portal_group *se_tpg = se_sess->se_tpg;
        struct se_node_acl *se_nacl;
 
-       if (!(se_tpg)) {
+       if (!se_tpg) {
                transport_free_session(se_sess);
                return;
        }
@@ -588,18 +406,18 @@ void transport_deregister_session(struct se_session *se_sess)
         * struct se_node_acl if it had been previously dynamically generated.
         */
        se_nacl = se_sess->se_node_acl;
-       if ((se_nacl)) {
+       if (se_nacl) {
                spin_lock_bh(&se_tpg->acl_node_lock);
                if (se_nacl->dynamic_node_acl) {
-                       if (!(TPG_TFO(se_tpg)->tpg_check_demo_mode_cache(
-                                       se_tpg))) {
+                       if (!se_tpg->se_tpg_tfo->tpg_check_demo_mode_cache(
+                                       se_tpg)) {
                                list_del(&se_nacl->acl_list);
                                se_tpg->num_node_acls--;
                                spin_unlock_bh(&se_tpg->acl_node_lock);
 
                                core_tpg_wait_for_nacl_pr_ref(se_nacl);
                                core_free_device_list_for_node(se_nacl, se_tpg);
-                               TPG_TFO(se_tpg)->tpg_release_fabric_acl(se_tpg,
+                               se_tpg->se_tpg_tfo->tpg_release_fabric_acl(se_tpg,
                                                se_nacl);
                                spin_lock_bh(&se_tpg->acl_node_lock);
                        }
@@ -609,13 +427,13 @@ void transport_deregister_session(struct se_session *se_sess)
 
        transport_free_session(se_sess);
 
-       printk(KERN_INFO "TARGET_CORE[%s]: Deregistered fabric_sess\n",
-               TPG_TFO(se_tpg)->get_fabric_name());
+       pr_debug("TARGET_CORE[%s]: Deregistered fabric_sess\n",
+               se_tpg->se_tpg_tfo->get_fabric_name());
 }
 EXPORT_SYMBOL(transport_deregister_session);
 
 /*
- * Called with T_TASK(cmd)->t_state_lock held.
+ * Called with cmd->t_state_lock held.
  */
 static void transport_all_task_dev_remove_state(struct se_cmd *cmd)
 {
@@ -623,28 +441,25 @@ static void transport_all_task_dev_remove_state(struct se_cmd *cmd)
        struct se_task *task;
        unsigned long flags;
 
-       if (!T_TASK(cmd))
-               return;
-
-       list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+       list_for_each_entry(task, &cmd->t_task_list, t_list) {
                dev = task->se_dev;
-               if (!(dev))
+               if (!dev)
                        continue;
 
                if (atomic_read(&task->task_active))
                        continue;
 
-               if (!(atomic_read(&task->task_state_active)))
+               if (!atomic_read(&task->task_state_active))
                        continue;
 
                spin_lock_irqsave(&dev->execute_task_lock, flags);
                list_del(&task->t_state_list);
-               DEBUG_TSTATE("Removed ITT: 0x%08x dev: %p task[%p]\n",
-                       CMD_TFO(cmd)->tfo_get_task_tag(cmd), dev, task);
+               pr_debug("Removed ITT: 0x%08x dev: %p task[%p]\n",
+                       cmd->se_tfo->get_task_tag(cmd), dev, task);
                spin_unlock_irqrestore(&dev->execute_task_lock, flags);
 
                atomic_set(&task->task_state_active, 0);
-               atomic_dec(&T_TASK(cmd)->t_task_cdbs_ex_left);
+               atomic_dec(&cmd->t_task_cdbs_ex_left);
        }
 }
 
@@ -663,34 +478,34 @@ static int transport_cmd_check_stop(
 {
        unsigned long flags;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        /*
         * Determine if IOCTL context caller in requesting the stopping of this
         * command for LUN shutdown purposes.
         */
-       if (atomic_read(&T_TASK(cmd)->transport_lun_stop)) {
-               DEBUG_CS("%s:%d atomic_read(&T_TASK(cmd)->transport_lun_stop)"
+       if (atomic_read(&cmd->transport_lun_stop)) {
+               pr_debug("%s:%d atomic_read(&cmd->transport_lun_stop)"
                        " == TRUE for ITT: 0x%08x\n", __func__, __LINE__,
-                       CMD_TFO(cmd)->get_task_tag(cmd));
+                       cmd->se_tfo->get_task_tag(cmd));
 
                cmd->deferred_t_state = cmd->t_state;
                cmd->t_state = TRANSPORT_DEFERRED_CMD;
-               atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+               atomic_set(&cmd->t_transport_active, 0);
                if (transport_off == 2)
                        transport_all_task_dev_remove_state(cmd);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
-               complete(&T_TASK(cmd)->transport_lun_stop_comp);
+               complete(&cmd->transport_lun_stop_comp);
                return 1;
        }
        /*
         * Determine if frontend context caller is requesting the stopping of
-        * this command for frontend excpections.
+        * this command for frontend exceptions.
         */
-       if (atomic_read(&T_TASK(cmd)->t_transport_stop)) {
-               DEBUG_CS("%s:%d atomic_read(&T_TASK(cmd)->t_transport_stop) =="
+       if (atomic_read(&cmd->t_transport_stop)) {
+               pr_debug("%s:%d atomic_read(&cmd->t_transport_stop) =="
                        " TRUE for ITT: 0x%08x\n", __func__, __LINE__,
-                       CMD_TFO(cmd)->get_task_tag(cmd));
+                       cmd->se_tfo->get_task_tag(cmd));
 
                cmd->deferred_t_state = cmd->t_state;
                cmd->t_state = TRANSPORT_DEFERRED_CMD;
@@ -703,13 +518,13 @@ static int transport_cmd_check_stop(
                 */
                if (transport_off == 2)
                        cmd->se_lun = NULL;
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
-               complete(&T_TASK(cmd)->t_transport_stop_comp);
+               complete(&cmd->t_transport_stop_comp);
                return 1;
        }
        if (transport_off) {
-               atomic_set(&T_TASK(cmd)->t_transport_active, 0);
+               atomic_set(&cmd->t_transport_active, 0);
                if (transport_off == 2) {
                        transport_all_task_dev_remove_state(cmd);
                        /*
@@ -722,20 +537,20 @@ static int transport_cmd_check_stop(
                         * their internally allocated I/O reference now and
                         * struct se_cmd now.
                         */
-                       if (CMD_TFO(cmd)->check_stop_free != NULL) {
+                       if (cmd->se_tfo->check_stop_free != NULL) {
                                spin_unlock_irqrestore(
-                                       &T_TASK(cmd)->t_state_lock, flags);
+                                       &cmd->t_state_lock, flags);
 
-                               CMD_TFO(cmd)->check_stop_free(cmd);
+                               cmd->se_tfo->check_stop_free(cmd);
                                return 1;
                        }
                }
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
                return 0;
        } else if (t_state)
                cmd->t_state = t_state;
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        return 0;
 }
@@ -747,30 +562,30 @@ static int transport_cmd_check_stop_to_fabric(struct se_cmd *cmd)
 
 static void transport_lun_remove_cmd(struct se_cmd *cmd)
 {
-       struct se_lun *lun = SE_LUN(cmd);
+       struct se_lun *lun = cmd->se_lun;
        unsigned long flags;
 
        if (!lun)
                return;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       if (!(atomic_read(&T_TASK(cmd)->transport_dev_active))) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       if (!atomic_read(&cmd->transport_dev_active)) {
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                goto check_lun;
        }
-       atomic_set(&T_TASK(cmd)->transport_dev_active, 0);
+       atomic_set(&cmd->transport_dev_active, 0);
        transport_all_task_dev_remove_state(cmd);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
 
 check_lun:
        spin_lock_irqsave(&lun->lun_cmd_lock, flags);
-       if (atomic_read(&T_TASK(cmd)->transport_lun_active)) {
-               list_del(&cmd->se_lun_list);
-               atomic_set(&T_TASK(cmd)->transport_lun_active, 0);
+       if (atomic_read(&cmd->transport_lun_active)) {
+               list_del(&cmd->se_lun_node);
+               atomic_set(&cmd->transport_lun_active, 0);
 #if 0
-               printk(KERN_INFO "Removed ITT: 0x%08x from LUN LIST[%d]\n"
-                       CMD_TFO(cmd)->get_task_tag(cmd), lun->unpacked_lun);
+               pr_debug("Removed ITT: 0x%08x from LUN LIST[%d]\n"
+                       cmd->se_tfo->get_task_tag(cmd), lun->unpacked_lun);
 #endif
        }
        spin_unlock_irqrestore(&lun->lun_cmd_lock, flags);
@@ -778,92 +593,59 @@ check_lun:
 
 void transport_cmd_finish_abort(struct se_cmd *cmd, int remove)
 {
-       transport_remove_cmd_from_queue(cmd, SE_DEV(cmd)->dev_queue_obj);
+       transport_remove_cmd_from_queue(cmd, &cmd->se_dev->dev_queue_obj);
        transport_lun_remove_cmd(cmd);
 
        if (transport_cmd_check_stop_to_fabric(cmd))
                return;
        if (remove)
-               transport_generic_remove(cmd, 0, 0);
+               transport_generic_remove(cmd, 0);
 }
 
 void transport_cmd_finish_abort_tmr(struct se_cmd *cmd)
 {
-       transport_remove_cmd_from_queue(cmd, SE_DEV(cmd)->dev_queue_obj);
+       transport_remove_cmd_from_queue(cmd, &cmd->se_dev->dev_queue_obj);
 
        if (transport_cmd_check_stop_to_fabric(cmd))
                return;
 
-       transport_generic_remove(cmd, 0, 0);
+       transport_generic_remove(cmd, 0);
 }
 
-static int transport_add_cmd_to_queue(
+static void transport_add_cmd_to_queue(
        struct se_cmd *cmd,
        int t_state)
 {
        struct se_device *dev = cmd->se_dev;
-       struct se_queue_obj *qobj = dev->dev_queue_obj;
-       struct se_queue_req *qr;
+       struct se_queue_obj *qobj = &dev->dev_queue_obj;
        unsigned long flags;
 
-       qr = kzalloc(sizeof(struct se_queue_req), GFP_ATOMIC);
-       if (!(qr)) {
-               printk(KERN_ERR "Unable to allocate memory for"
-                               " struct se_queue_req\n");
-               return -1;
-       }
-       INIT_LIST_HEAD(&qr->qr_list);
-
-       qr->cmd = (void *)cmd;
-       qr->state = t_state;
+       INIT_LIST_HEAD(&cmd->se_queue_node);
 
        if (t_state) {
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+               spin_lock_irqsave(&cmd->t_state_lock, flags);
                cmd->t_state = t_state;
-               atomic_set(&T_TASK(cmd)->t_transport_active, 1);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               atomic_set(&cmd->t_transport_active, 1);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
        }
 
        spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
-       list_add_tail(&qr->qr_list, &qobj->qobj_list);
-       atomic_inc(&T_TASK(cmd)->t_transport_queue_active);
+       if (cmd->se_cmd_flags & SCF_EMULATE_QUEUE_FULL) {
+               cmd->se_cmd_flags &= ~SCF_EMULATE_QUEUE_FULL;
+               list_add(&cmd->se_queue_node, &qobj->qobj_list);
+       } else
+               list_add_tail(&cmd->se_queue_node, &qobj->qobj_list);
+       atomic_inc(&cmd->t_transport_queue_active);
        spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
 
        atomic_inc(&qobj->queue_cnt);
        wake_up_interruptible(&qobj->thread_wq);
-       return 0;
-}
-
-/*
- * Called with struct se_queue_obj->cmd_queue_lock held.
- */
-static struct se_queue_req *
-__transport_get_qr_from_queue(struct se_queue_obj *qobj)
-{
-       struct se_cmd *cmd;
-       struct se_queue_req *qr = NULL;
-
-       if (list_empty(&qobj->qobj_list))
-               return NULL;
-
-       list_for_each_entry(qr, &qobj->qobj_list, qr_list)
-               break;
-
-       if (qr->cmd) {
-               cmd = (struct se_cmd *)qr->cmd;
-               atomic_dec(&T_TASK(cmd)->t_transport_queue_active);
-       }
-       list_del(&qr->qr_list);
-       atomic_dec(&qobj->queue_cnt);
-
-       return qr;
 }
 
-static struct se_queue_req *
-transport_get_qr_from_queue(struct se_queue_obj *qobj)
+static struct se_cmd *
+transport_get_cmd_from_queue(struct se_queue_obj *qobj)
 {
        struct se_cmd *cmd;
-       struct se_queue_req *qr;
        unsigned long flags;
 
        spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
@@ -871,50 +653,42 @@ transport_get_qr_from_queue(struct se_queue_obj *qobj)
                spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
                return NULL;
        }
+       cmd = list_first_entry(&qobj->qobj_list, struct se_cmd, se_queue_node);
 
-       list_for_each_entry(qr, &qobj->qobj_list, qr_list)
-               break;
+       atomic_dec(&cmd->t_transport_queue_active);
 
-       if (qr->cmd) {
-               cmd = (struct se_cmd *)qr->cmd;
-               atomic_dec(&T_TASK(cmd)->t_transport_queue_active);
-       }
-       list_del(&qr->qr_list);
+       list_del(&cmd->se_queue_node);
        atomic_dec(&qobj->queue_cnt);
        spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
 
-       return qr;
+       return cmd;
 }
 
 static void transport_remove_cmd_from_queue(struct se_cmd *cmd,
                struct se_queue_obj *qobj)
 {
-       struct se_cmd *q_cmd;
-       struct se_queue_req *qr = NULL, *qr_p = NULL;
+       struct se_cmd *t;
        unsigned long flags;
 
        spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
-       if (!(atomic_read(&T_TASK(cmd)->t_transport_queue_active))) {
+       if (!atomic_read(&cmd->t_transport_queue_active)) {
                spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
                return;
        }
 
-       list_for_each_entry_safe(qr, qr_p, &qobj->qobj_list, qr_list) {
-               q_cmd = (struct se_cmd *)qr->cmd;
-               if (q_cmd != cmd)
-                       continue;
-
-               atomic_dec(&T_TASK(q_cmd)->t_transport_queue_active);
-               atomic_dec(&qobj->queue_cnt);
-               list_del(&qr->qr_list);
-               kfree(qr);
-       }
+       list_for_each_entry(t, &qobj->qobj_list, se_queue_node)
+               if (t == cmd) {
+                       atomic_dec(&cmd->t_transport_queue_active);
+                       atomic_dec(&qobj->queue_cnt);
+                       list_del(&cmd->se_queue_node);
+                       break;
+               }
        spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
 
-       if (atomic_read(&T_TASK(cmd)->t_transport_queue_active)) {
-               printk(KERN_ERR "ITT: 0x%08x t_transport_queue_active: %d\n",
-                       CMD_TFO(cmd)->get_task_tag(cmd),
-                       atomic_read(&T_TASK(cmd)->t_transport_queue_active));
+       if (atomic_read(&cmd->t_transport_queue_active)) {
+               pr_err("ITT: 0x%08x t_transport_queue_active: %d\n",
+                       cmd->se_tfo->get_task_tag(cmd),
+                       atomic_read(&cmd->t_transport_queue_active));
        }
 }
 
@@ -924,7 +698,7 @@ static void transport_remove_cmd_from_queue(struct se_cmd *cmd,
  */
 void transport_complete_sync_cache(struct se_cmd *cmd, int good)
 {
-       struct se_task *task = list_entry(T_TASK(cmd)->t_task_list.next,
+       struct se_task *task = list_entry(cmd->t_task_list.next,
                                struct se_task, t_list);
 
        if (good) {
@@ -933,7 +707,7 @@ void transport_complete_sync_cache(struct se_cmd *cmd, int good)
        } else {
                task->task_scsi_status = SAM_STAT_CHECK_CONDITION;
                task->task_error_status = PYX_TRANSPORT_ILLEGAL_REQUEST;
-               TASK_CMD(task)->transport_error_status =
+               task->task_se_cmd->transport_error_status =
                                        PYX_TRANSPORT_ILLEGAL_REQUEST;
        }
 
@@ -948,22 +722,18 @@ EXPORT_SYMBOL(transport_complete_sync_cache);
  */
 void transport_complete_task(struct se_task *task, int success)
 {
-       struct se_cmd *cmd = TASK_CMD(task);
+       struct se_cmd *cmd = task->task_se_cmd;
        struct se_device *dev = task->se_dev;
        int t_state;
        unsigned long flags;
 #if 0
-       printk(KERN_INFO "task: %p CDB: 0x%02x obj_ptr: %p\n", task,
-                       T_TASK(cmd)->t_task_cdb[0], dev);
+       pr_debug("task: %p CDB: 0x%02x obj_ptr: %p\n", task,
+                       cmd->t_task_cdb[0], dev);
 #endif
-       if (dev) {
-               spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);
+       if (dev)
                atomic_inc(&dev->depth_left);
-               atomic_inc(&SE_HBA(dev)->left_queue_depth);
-               spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
-       }
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        atomic_set(&task->task_active, 0);
 
        /*
@@ -985,14 +755,14 @@ void transport_complete_task(struct se_task *task, int success)
         */
        if (atomic_read(&task->task_stop)) {
                /*
-                * Decrement T_TASK(cmd)->t_se_count if this task had
+                * Decrement cmd->t_se_count if this task had
                 * previously thrown its timeout exception handler.
                 */
                if (atomic_read(&task->task_timeout)) {
-                       atomic_dec(&T_TASK(cmd)->t_se_count);
+                       atomic_dec(&cmd->t_se_count);
                        atomic_set(&task->task_timeout, 0);
                }
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
                complete(&task->task_stop_comp);
                return;
@@ -1003,34 +773,34 @@ void transport_complete_task(struct se_task *task, int success)
         * the processing thread.
         */
        if (atomic_read(&task->task_timeout)) {
-               if (!(atomic_dec_and_test(
-                               &T_TASK(cmd)->t_task_cdbs_timeout_left))) {
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+               if (!atomic_dec_and_test(
+                               &cmd->t_task_cdbs_timeout_left)) {
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                flags);
                        return;
                }
                t_state = TRANSPORT_COMPLETE_TIMEOUT;
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
                transport_add_cmd_to_queue(cmd, t_state);
                return;
        }
-       atomic_dec(&T_TASK(cmd)->t_task_cdbs_timeout_left);
+       atomic_dec(&cmd->t_task_cdbs_timeout_left);
 
        /*
         * Decrement the outstanding t_task_cdbs_left count.  The last
         * struct se_task from struct se_cmd will complete itself into the
         * device queue depending upon int success.
         */
-       if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_left))) {
+       if (!atomic_dec_and_test(&cmd->t_task_cdbs_left)) {
                if (!success)
-                       T_TASK(cmd)->t_tasks_failed = 1;
+                       cmd->t_tasks_failed = 1;
 
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return;
        }
 
-       if (!success || T_TASK(cmd)->t_tasks_failed) {
+       if (!success || cmd->t_tasks_failed) {
                t_state = TRANSPORT_COMPLETE_FAILURE;
                if (!task->task_error_status) {
                        task->task_error_status =
@@ -1039,10 +809,10 @@ void transport_complete_task(struct se_task *task, int success)
                                PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
                }
        } else {
-               atomic_set(&T_TASK(cmd)->t_transport_complete, 1);
+               atomic_set(&cmd->t_transport_complete, 1);
                t_state = TRANSPORT_COMPLETE_OK;
        }
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        transport_add_cmd_to_queue(cmd, t_state);
 }
@@ -1080,9 +850,9 @@ static inline int transport_add_task_check_sam_attr(
                                &task_prev->t_execute_list :
                                &dev->execute_task_list);
 
-               DEBUG_STA("Set HEAD_OF_QUEUE for task CDB: 0x%02x"
+               pr_debug("Set HEAD_OF_QUEUE for task CDB: 0x%02x"
                                " in execution queue\n",
-                               T_TASK(task->task_se_cmd)->t_task_cdb[0]);
+                               task->task_se_cmd->t_task_cdb[0]);
                return 1;
        }
        /*
@@ -1124,8 +894,8 @@ static void __transport_add_task_to_execute_queue(
 
        atomic_set(&task->task_state_active, 1);
 
-       DEBUG_TSTATE("Added ITT: 0x%08x task[%p] to dev: %p\n",
-               CMD_TFO(task->task_se_cmd)->get_task_tag(task->task_se_cmd),
+       pr_debug("Added ITT: 0x%08x task[%p] to dev: %p\n",
+               task->task_se_cmd->se_tfo->get_task_tag(task->task_se_cmd),
                task, dev);
 }
 
@@ -1135,8 +905,8 @@ static void transport_add_tasks_to_state_queue(struct se_cmd *cmd)
        struct se_task *task;
        unsigned long flags;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       list_for_each_entry(task, &cmd->t_task_list, t_list) {
                dev = task->se_dev;
 
                if (atomic_read(&task->task_state_active))
@@ -1146,23 +916,23 @@ static void transport_add_tasks_to_state_queue(struct se_cmd *cmd)
                list_add_tail(&task->t_state_list, &dev->state_task_list);
                atomic_set(&task->task_state_active, 1);
 
-               DEBUG_TSTATE("Added ITT: 0x%08x task[%p] to dev: %p\n",
-                       CMD_TFO(task->task_se_cmd)->get_task_tag(
+               pr_debug("Added ITT: 0x%08x task[%p] to dev: %p\n",
+                       task->task_se_cmd->se_tfo->get_task_tag(
                        task->task_se_cmd), task, dev);
 
                spin_unlock(&dev->execute_task_lock);
        }
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 }
 
 static void transport_add_tasks_from_cmd(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct se_task *task, *task_prev = NULL;
        unsigned long flags;
 
        spin_lock_irqsave(&dev->execute_task_lock, flags);
-       list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
+       list_for_each_entry(task, &cmd->t_task_list, t_list) {
                if (atomic_read(&task->task_execute_queue))
                        continue;
                /*
@@ -1174,30 +944,6 @@ static void transport_add_tasks_from_cmd(struct se_cmd *cmd)
                task_prev = task;
        }
        spin_unlock_irqrestore(&dev->execute_task_lock, flags);
-
-       return;
-}
-
-/*     transport_get_task_from_execute_queue():
- *
- *     Called with dev->execute_task_lock held.
- */
-static struct se_task *
-transport_get_task_from_execute_queue(struct se_device *dev)
-{
-       struct se_task *task;
-
-       if (list_empty(&dev->execute_task_list))
-               return NULL;
-
-       list_for_each_entry(task, &dev->execute_task_list, t_execute_list)
-               break;
-
-       list_del(&task->t_execute_list);
-       atomic_set(&task->task_execute_queue, 0);
-       atomic_dec(&dev->execute_tasks);
-
-       return task;
 }
 
 /*     transport_remove_task_from_execute_queue():
@@ -1222,6 +968,40 @@ void transport_remove_task_from_execute_queue(
        spin_unlock_irqrestore(&dev->execute_task_lock, flags);
 }
 
+/*
+ * Handle QUEUE_FULL / -EAGAIN status
+ */
+
+static void target_qf_do_work(struct work_struct *work)
+{
+       struct se_device *dev = container_of(work, struct se_device,
+                                       qf_work_queue);
+       struct se_cmd *cmd, *cmd_tmp;
+
+       spin_lock_irq(&dev->qf_cmd_lock);
+       list_for_each_entry_safe(cmd, cmd_tmp, &dev->qf_cmd_list, se_qf_node) {
+
+               list_del(&cmd->se_qf_node);
+               atomic_dec(&dev->dev_qf_count);
+               smp_mb__after_atomic_dec();
+               spin_unlock_irq(&dev->qf_cmd_lock);
+
+               pr_debug("Processing %s cmd: %p QUEUE_FULL in work queue"
+                       " context: %s\n", cmd->se_tfo->get_fabric_name(), cmd,
+                       (cmd->t_state == TRANSPORT_COMPLETE_OK) ? "COMPLETE_OK" :
+                       (cmd->t_state == TRANSPORT_COMPLETE_QF_WP) ? "WRITE_PENDING"
+                       : "UNKNOWN");
+               /*
+                * The SCF_EMULATE_QUEUE_FULL flag will be cleared once se_cmd
+                * has been added to head of queue
+                */
+               transport_add_cmd_to_queue(cmd, cmd->t_state);
+
+               spin_lock_irq(&dev->qf_cmd_lock);
+       }
+       spin_unlock_irq(&dev->qf_cmd_lock);
+}
+
 unsigned char *transport_dump_cmd_direction(struct se_cmd *cmd)
 {
        switch (cmd->data_direction) {
@@ -1269,7 +1049,7 @@ void transport_dump_dev_state(
                atomic_read(&dev->execute_tasks), atomic_read(&dev->depth_left),
                dev->queue_depth);
        *bl += sprintf(b + *bl, "  SectorSize: %u  MaxSectors: %u\n",
-               DEV_ATTRIB(dev)->block_size, DEV_ATTRIB(dev)->max_sectors);
+               dev->se_sub_dev->se_dev_attrib.block_size, dev->se_sub_dev->se_dev_attrib.max_sectors);
        *bl += sprintf(b + *bl, "        ");
 }
 
@@ -1279,33 +1059,29 @@ void transport_dump_dev_state(
  */
 static void transport_release_all_cmds(struct se_device *dev)
 {
-       struct se_cmd *cmd = NULL;
-       struct se_queue_req *qr = NULL, *qr_p = NULL;
+       struct se_cmd *cmd, *tcmd;
        int bug_out = 0, t_state;
        unsigned long flags;
 
-       spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
-       list_for_each_entry_safe(qr, qr_p, &dev->dev_queue_obj->qobj_list,
-                               qr_list) {
-
-               cmd = (struct se_cmd *)qr->cmd;
-               t_state = qr->state;
-               list_del(&qr->qr_list);
-               kfree(qr);
-               spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock,
+       spin_lock_irqsave(&dev->dev_queue_obj.cmd_queue_lock, flags);
+       list_for_each_entry_safe(cmd, tcmd, &dev->dev_queue_obj.qobj_list,
+                               se_queue_node) {
+               t_state = cmd->t_state;
+               list_del(&cmd->se_queue_node);
+               spin_unlock_irqrestore(&dev->dev_queue_obj.cmd_queue_lock,
                                flags);
 
-               printk(KERN_ERR "Releasing ITT: 0x%08x, i_state: %u,"
+               pr_err("Releasing ITT: 0x%08x, i_state: %u,"
                        " t_state: %u directly\n",
-                       CMD_TFO(cmd)->get_task_tag(cmd),
-                       CMD_TFO(cmd)->get_cmd_state(cmd), t_state);
+                       cmd->se_tfo->get_task_tag(cmd),
+                       cmd->se_tfo->get_cmd_state(cmd), t_state);
 
                transport_release_fe_cmd(cmd);
                bug_out = 1;
 
-               spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
+               spin_lock_irqsave(&dev->dev_queue_obj.cmd_queue_lock, flags);
        }
-       spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock, flags);
+       spin_unlock_irqrestore(&dev->dev_queue_obj.cmd_queue_lock, flags);
 #if 0
        if (bug_out)
                BUG();
@@ -1362,7 +1138,7 @@ void transport_dump_vpd_proto_id(
        if (p_buf)
                strncpy(p_buf, buf, p_buf_len);
        else
-               printk(KERN_INFO "%s", buf);
+               pr_debug("%s", buf);
 }
 
 void
@@ -1387,7 +1163,8 @@ int transport_dump_vpd_assoc(
        int p_buf_len)
 {
        unsigned char buf[VPD_TMP_BUF_SIZE];
-       int ret = 0, len;
+       int ret = 0;
+       int len;
 
        memset(buf, 0, VPD_TMP_BUF_SIZE);
        len = sprintf(buf, "T10 VPD Identifier Association: ");
@@ -1404,14 +1181,14 @@ int transport_dump_vpd_assoc(
                break;
        default:
                sprintf(buf+len, "Unknown 0x%02x\n", vpd->association);
-               ret = -1;
+               ret = -EINVAL;
                break;
        }
 
        if (p_buf)
                strncpy(p_buf, buf, p_buf_len);
        else
-               printk("%s", buf);
+               pr_debug("%s", buf);
 
        return ret;
 }
@@ -1434,7 +1211,8 @@ int transport_dump_vpd_ident_type(
        int p_buf_len)
 {
        unsigned char buf[VPD_TMP_BUF_SIZE];
-       int ret = 0, len;
+       int ret = 0;
+       int len;
 
        memset(buf, 0, VPD_TMP_BUF_SIZE);
        len = sprintf(buf, "T10 VPD Identifier Type: ");
@@ -1461,14 +1239,17 @@ int transport_dump_vpd_ident_type(
        default:
                sprintf(buf+len, "Unsupported: 0x%02x\n",
                                vpd->device_identifier_type);
-               ret = -1;
+               ret = -EINVAL;
                break;
        }
 
-       if (p_buf)
+       if (p_buf) {
+               if (p_buf_len < strlen(buf)+1)
+                       return -EINVAL;
                strncpy(p_buf, buf, p_buf_len);
-       else
-               printk("%s", buf);
+       } else {
+               pr_debug("%s", buf);
+       }
 
        return ret;
 }
@@ -1511,14 +1292,14 @@ int transport_dump_vpd_ident(
        default:
                sprintf(buf, "T10 VPD Device Identifier encoding unsupported:"
                        " 0x%02x", vpd->device_identifier_code_set);
-               ret = -1;
+               ret = -EINVAL;
                break;
        }
 
        if (p_buf)
                strncpy(p_buf, buf, p_buf_len);
        else
-               printk("%s", buf);
+               pr_debug("%s", buf);
 
        return ret;
 }
@@ -1569,51 +1350,51 @@ static void core_setup_task_attr_emulation(struct se_device *dev)
         * This is currently not available in upsream Linux/SCSI Target
         * mode code, and is assumed to be disabled while using TCM/pSCSI.
         */
-       if (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
+       if (dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV) {
                dev->dev_task_attr_type = SAM_TASK_ATTR_PASSTHROUGH;
                return;
        }
 
        dev->dev_task_attr_type = SAM_TASK_ATTR_EMULATED;
-       DEBUG_STA("%s: Using SAM_TASK_ATTR_EMULATED for SPC: 0x%02x"
-               " device\n", TRANSPORT(dev)->name,
-               TRANSPORT(dev)->get_device_rev(dev));
+       pr_debug("%s: Using SAM_TASK_ATTR_EMULATED for SPC: 0x%02x"
+               " device\n", dev->transport->name,
+               dev->transport->get_device_rev(dev));
 }
 
 static void scsi_dump_inquiry(struct se_device *dev)
 {
-       struct t10_wwn *wwn = DEV_T10_WWN(dev);
+       struct t10_wwn *wwn = &dev->se_sub_dev->t10_wwn;
        int i, device_type;
        /*
         * Print Linux/SCSI style INQUIRY formatting to the kernel ring buffer
         */
-       printk("  Vendor: ");
+       pr_debug("  Vendor: ");
        for (i = 0; i < 8; i++)
                if (wwn->vendor[i] >= 0x20)
-                       printk("%c", wwn->vendor[i]);
+                       pr_debug("%c", wwn->vendor[i]);
                else
-                       printk(" ");
+                       pr_debug(" ");
 
-       printk("  Model: ");
+       pr_debug("  Model: ");
        for (i = 0; i < 16; i++)
                if (wwn->model[i] >= 0x20)
-                       printk("%c", wwn->model[i]);
+                       pr_debug("%c", wwn->model[i]);
                else
-                       printk(" ");
+                       pr_debug(" ");
 
-       printk("  Revision: ");
+       pr_debug("  Revision: ");
        for (i = 0; i < 4; i++)
                if (wwn->revision[i] >= 0x20)
-                       printk("%c", wwn->revision[i]);
+                       pr_debug("%c", wwn->revision[i]);
                else
-                       printk(" ");
+                       pr_debug(" ");
 
-       printk("\n");
+       pr_debug("\n");
 
-       device_type = TRANSPORT(dev)->get_device_type(dev);
-       printk("  Type:   %s ", scsi_device_type(device_type));
-       printk("                 ANSI SCSI revision: %02x\n",
-                               TRANSPORT(dev)->get_device_rev(dev));
+       device_type = dev->transport->get_device_type(dev);
+       pr_debug("  Type:   %s ", scsi_device_type(device_type));
+       pr_debug("                 ANSI SCSI revision: %02x\n",
+                               dev->transport->get_device_rev(dev));
 }
 
 struct se_device *transport_add_device_to_core_hba(
@@ -1630,33 +1411,15 @@ struct se_device *transport_add_device_to_core_hba(
        struct se_device  *dev;
 
        dev = kzalloc(sizeof(struct se_device), GFP_KERNEL);
-       if (!(dev)) {
-               printk(KERN_ERR "Unable to allocate memory for se_dev_t\n");
-               return NULL;
-       }
-       dev->dev_queue_obj = kzalloc(sizeof(struct se_queue_obj), GFP_KERNEL);
-       if (!(dev->dev_queue_obj)) {
-               printk(KERN_ERR "Unable to allocate memory for"
-                               " dev->dev_queue_obj\n");
-               kfree(dev);
+       if (!dev) {
+               pr_err("Unable to allocate memory for se_dev_t\n");
                return NULL;
        }
-       transport_init_queue_obj(dev->dev_queue_obj);
-
-       dev->dev_status_queue_obj = kzalloc(sizeof(struct se_queue_obj),
-                                       GFP_KERNEL);
-       if (!(dev->dev_status_queue_obj)) {
-               printk(KERN_ERR "Unable to allocate memory for"
-                               " dev->dev_status_queue_obj\n");
-               kfree(dev->dev_queue_obj);
-               kfree(dev);
-               return NULL;
-       }
-       transport_init_queue_obj(dev->dev_status_queue_obj);
 
+       transport_init_queue_obj(&dev->dev_queue_obj);
        dev->dev_flags          = device_flags;
        dev->dev_status         |= TRANSPORT_DEVICE_DEACTIVATED;
-       dev->dev_ptr            = (void *) transport_dev;
+       dev->dev_ptr            = transport_dev;
        dev->se_hba             = hba;
        dev->se_sub_dev         = se_dev;
        dev->transport          = transport;
@@ -1668,6 +1431,7 @@ struct se_device *transport_add_device_to_core_hba(
        INIT_LIST_HEAD(&dev->delayed_cmd_list);
        INIT_LIST_HEAD(&dev->ordered_cmd_list);
        INIT_LIST_HEAD(&dev->state_task_list);
+       INIT_LIST_HEAD(&dev->qf_cmd_list);
        spin_lock_init(&dev->execute_task_lock);
        spin_lock_init(&dev->delayed_cmd_lock);
        spin_lock_init(&dev->ordered_cmd_lock);
@@ -1678,6 +1442,7 @@ struct se_device *transport_add_device_to_core_hba(
        spin_lock_init(&dev->dev_status_thr_lock);
        spin_lock_init(&dev->se_port_lock);
        spin_lock_init(&dev->se_tmr_lock);
+       spin_lock_init(&dev->qf_cmd_lock);
 
        dev->queue_depth        = dev_limits->queue_depth;
        atomic_set(&dev->depth_left, dev->queue_depth);
@@ -1715,13 +1480,16 @@ struct se_device *transport_add_device_to_core_hba(
         * Startup the struct se_device processing thread
         */
        dev->process_thread = kthread_run(transport_processing_thread, dev,
-                                         "LIO_%s", TRANSPORT(dev)->name);
+                                         "LIO_%s", dev->transport->name);
        if (IS_ERR(dev->process_thread)) {
-               printk(KERN_ERR "Unable to create kthread: LIO_%s\n",
-                       TRANSPORT(dev)->name);
+               pr_err("Unable to create kthread: LIO_%s\n",
+                       dev->transport->name);
                goto out;
        }
-
+       /*
+        * Setup work_queue for QUEUE_FULL
+        */
+       INIT_WORK(&dev->qf_work_queue, target_qf_do_work);
        /*
         * Preload the initial INQUIRY const values if we are doing
         * anything virtual (IBLOCK, FILEIO, RAMDISK), but not for TCM/pSCSI
@@ -1730,16 +1498,16 @@ struct se_device *transport_add_device_to_core_hba(
         * originals once back into DEV_T10_WWN(dev) for the virtual device
         * setup.
         */
-       if (TRANSPORT(dev)->transport_type != TRANSPORT_PLUGIN_PHBA_PDEV) {
-               if (!(inquiry_prod) || !(inquiry_prod)) {
-                       printk(KERN_ERR "All non TCM/pSCSI plugins require"
+       if (dev->transport->transport_type != TRANSPORT_PLUGIN_PHBA_PDEV) {
+               if (!inquiry_prod || !inquiry_rev) {
+                       pr_err("All non TCM/pSCSI plugins require"
                                " INQUIRY consts\n");
                        goto out;
                }
 
-               strncpy(&DEV_T10_WWN(dev)->vendor[0], "LIO-ORG", 8);
-               strncpy(&DEV_T10_WWN(dev)->model[0], inquiry_prod, 16);
-               strncpy(&DEV_T10_WWN(dev)->revision[0], inquiry_rev, 4);
+               strncpy(&dev->se_sub_dev->t10_wwn.vendor[0], "LIO-ORG", 8);
+               strncpy(&dev->se_sub_dev->t10_wwn.model[0], inquiry_prod, 16);
+               strncpy(&dev->se_sub_dev->t10_wwn.revision[0], inquiry_rev, 4);
        }
        scsi_dump_inquiry(dev);
 
@@ -1754,8 +1522,6 @@ out:
 
        se_release_vpd_for_dev(dev);
 
-       kfree(dev->dev_status_queue_obj);
-       kfree(dev->dev_queue_obj);
        kfree(dev);
 
        return NULL;
@@ -1794,12 +1560,11 @@ transport_generic_get_task(struct se_cmd *cmd,
                enum dma_data_direction data_direction)
 {
        struct se_task *task;
-       struct se_device *dev = SE_DEV(cmd);
-       unsigned long flags;
+       struct se_device *dev = cmd->se_dev;
 
-       task = dev->transport->alloc_task(cmd);
+       task = dev->transport->alloc_task(cmd->t_task_cdb);
        if (!task) {
-               printk(KERN_ERR "Unable to allocate struct se_task\n");
+               pr_err("Unable to allocate struct se_task\n");
                return NULL;
        }
 
@@ -1807,26 +1572,15 @@ transport_generic_get_task(struct se_cmd *cmd,
        INIT_LIST_HEAD(&task->t_execute_list);
        INIT_LIST_HEAD(&task->t_state_list);
        init_completion(&task->task_stop_comp);
-       task->task_no = T_TASK(cmd)->t_tasks_no++;
        task->task_se_cmd = cmd;
        task->se_dev = dev;
        task->task_data_direction = data_direction;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       list_add_tail(&task->t_list, &T_TASK(cmd)->t_task_list);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
-
        return task;
 }
 
 static int transport_generic_cmd_sequencer(struct se_cmd *, unsigned char *);
 
-void transport_device_setup_cmd(struct se_cmd *cmd)
-{
-       cmd->se_dev = SE_LUN(cmd)->lun_se_dev;
-}
-EXPORT_SYMBOL(transport_device_setup_cmd);
-
 /*
  * Used by fabric modules containing a local struct se_cmd within their
  * fabric dependent per I/O descriptor.
@@ -1840,20 +1594,17 @@ void transport_init_se_cmd(
        int task_attr,
        unsigned char *sense_buffer)
 {
-       INIT_LIST_HEAD(&cmd->se_lun_list);
-       INIT_LIST_HEAD(&cmd->se_delayed_list);
-       INIT_LIST_HEAD(&cmd->se_ordered_list);
-       /*
-        * Setup t_task pointer to t_task_backstore
-        */
-       cmd->t_task = &cmd->t_task_backstore;
+       INIT_LIST_HEAD(&cmd->se_lun_node);
+       INIT_LIST_HEAD(&cmd->se_delayed_node);
+       INIT_LIST_HEAD(&cmd->se_ordered_node);
+       INIT_LIST_HEAD(&cmd->se_qf_node);
 
-       INIT_LIST_HEAD(&T_TASK(cmd)->t_task_list);
-       init_completion(&T_TASK(cmd)->transport_lun_fe_stop_comp);
-       init_completion(&T_TASK(cmd)->transport_lun_stop_comp);
-       init_completion(&T_TASK(cmd)->t_transport_stop_comp);
-       spin_lock_init(&T_TASK(cmd)->t_state_lock);
-       atomic_set(&T_TASK(cmd)->transport_dev_active, 1);
+       INIT_LIST_HEAD(&cmd->t_task_list);
+       init_completion(&cmd->transport_lun_fe_stop_comp);
+       init_completion(&cmd->transport_lun_stop_comp);
+       init_completion(&cmd->t_transport_stop_comp);
+       spin_lock_init(&cmd->t_state_lock);
+       atomic_set(&cmd->transport_dev_active, 1);
 
        cmd->se_tfo = tfo;
        cmd->se_sess = se_sess;
@@ -1870,23 +1621,23 @@ static int transport_check_alloc_task_attr(struct se_cmd *cmd)
         * Check if SAM Task Attribute emulation is enabled for this
         * struct se_device storage object
         */
-       if (SE_DEV(cmd)->dev_task_attr_type != SAM_TASK_ATTR_EMULATED)
+       if (cmd->se_dev->dev_task_attr_type != SAM_TASK_ATTR_EMULATED)
                return 0;
 
        if (cmd->sam_task_attr == MSG_ACA_TAG) {
-               DEBUG_STA("SAM Task Attribute ACA"
+               pr_debug("SAM Task Attribute ACA"
                        " emulation is not supported\n");
-               return -1;
+               return -EINVAL;
        }
        /*
         * Used to determine when ORDERED commands should go from
         * Dormant to Active status.
         */
-       cmd->se_ordered_id = atomic_inc_return(&SE_DEV(cmd)->dev_ordered_id);
+       cmd->se_ordered_id = atomic_inc_return(&cmd->se_dev->dev_ordered_id);
        smp_mb__after_atomic_inc();
-       DEBUG_STA("Allocated se_ordered_id: %u for Task Attr: 0x%02x on %s\n",
+       pr_debug("Allocated se_ordered_id: %u for Task Attr: 0x%02x on %s\n",
                        cmd->se_ordered_id, cmd->sam_task_attr,
-                       TRANSPORT(cmd->se_dev)->name);
+                       cmd->se_dev->transport->name);
        return 0;
 }
 
@@ -1898,8 +1649,8 @@ void transport_free_se_cmd(
        /*
         * Check and free any extended CDB buffer that was allocated
         */
-       if (T_TASK(se_cmd)->t_task_cdb != T_TASK(se_cmd)->__t_task_cdb)
-               kfree(T_TASK(se_cmd)->t_task_cdb);
+       if (se_cmd->t_task_cdb != se_cmd->__t_task_cdb)
+               kfree(se_cmd->t_task_cdb);
 }
 EXPORT_SYMBOL(transport_free_se_cmd);
 
@@ -1922,42 +1673,41 @@ int transport_generic_allocate_tasks(
         */
        cmd->transport_wait_for_tasks = &transport_generic_wait_for_tasks;
 
-       transport_device_setup_cmd(cmd);
        /*
         * Ensure that the received CDB is less than the max (252 + 8) bytes
         * for VARIABLE_LENGTH_CMD
         */
        if (scsi_command_size(cdb) > SCSI_MAX_VARLEN_CDB_SIZE) {
-               printk(KERN_ERR "Received SCSI CDB with command_size: %d that"
+               pr_err("Received SCSI CDB with command_size: %d that"
                        " exceeds SCSI_MAX_VARLEN_CDB_SIZE: %d\n",
                        scsi_command_size(cdb), SCSI_MAX_VARLEN_CDB_SIZE);
-               return -1;
+               return -EINVAL;
        }
        /*
         * If the received CDB is larger than TCM_MAX_COMMAND_SIZE,
         * allocate the additional extended CDB buffer now..  Otherwise
         * setup the pointer from __t_task_cdb to t_task_cdb.
         */
-       if (scsi_command_size(cdb) > sizeof(T_TASK(cmd)->__t_task_cdb)) {
-               T_TASK(cmd)->t_task_cdb = kzalloc(scsi_command_size(cdb),
+       if (scsi_command_size(cdb) > sizeof(cmd->__t_task_cdb)) {
+               cmd->t_task_cdb = kzalloc(scsi_command_size(cdb),
                                                GFP_KERNEL);
-               if (!(T_TASK(cmd)->t_task_cdb)) {
-                       printk(KERN_ERR "Unable to allocate T_TASK(cmd)->t_task_cdb"
-                               " %u > sizeof(T_TASK(cmd)->__t_task_cdb): %lu ops\n",
+               if (!cmd->t_task_cdb) {
+                       pr_err("Unable to allocate cmd->t_task_cdb"
+                               " %u > sizeof(cmd->__t_task_cdb): %lu ops\n",
                                scsi_command_size(cdb),
-                               (unsigned long)sizeof(T_TASK(cmd)->__t_task_cdb));
-                       return -1;
+                               (unsigned long)sizeof(cmd->__t_task_cdb));
+                       return -ENOMEM;
                }
        } else
-               T_TASK(cmd)->t_task_cdb = &T_TASK(cmd)->__t_task_cdb[0];
+               cmd->t_task_cdb = &cmd->__t_task_cdb[0];
        /*
-        * Copy the original CDB into T_TASK(cmd).
+        * Copy the original CDB into cmd->
         */
-       memcpy(T_TASK(cmd)->t_task_cdb, cdb, scsi_command_size(cdb));
+       memcpy(cmd->t_task_cdb, cdb, scsi_command_size(cdb));
        /*
         * Setup the received CDB based on SCSI defined opcodes and
         * perform unit attention, persistent reservations and ALUA
-        * checks for virtual device backends.  The T_TASK(cmd)->t_task_cdb
+        * checks for virtual device backends.  The cmd->t_task_cdb
         * pointer is expected to be setup before we reach this point.
         */
        ret = transport_generic_cmd_sequencer(cmd, cdb);
@@ -1969,7 +1719,7 @@ int transport_generic_allocate_tasks(
        if (transport_check_alloc_task_attr(cmd) < 0) {
                cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
                cmd->scsi_sense_reason = TCM_INVALID_CDB_FIELD;
-               return -2;
+               return -EINVAL;
        }
        spin_lock(&cmd->se_lun->lun_sep_lock);
        if (cmd->se_lun->lun_sep)
@@ -1986,10 +1736,10 @@ EXPORT_SYMBOL(transport_generic_allocate_tasks);
 int transport_generic_handle_cdb(
        struct se_cmd *cmd)
 {
-       if (!SE_LUN(cmd)) {
+       if (!cmd->se_lun) {
                dump_stack();
-               printk(KERN_ERR "SE_LUN(cmd) is NULL\n");
-               return -1;
+               pr_err("cmd->se_lun is NULL\n");
+               return -EINVAL;
        }
 
        transport_add_cmd_to_queue(cmd, TRANSPORT_NEW_CMD);
@@ -1997,6 +1747,29 @@ int transport_generic_handle_cdb(
 }
 EXPORT_SYMBOL(transport_generic_handle_cdb);
 
+/*
+ * Used by fabric module frontends to queue tasks directly.
+ * Many only be used from process context only
+ */
+int transport_handle_cdb_direct(
+       struct se_cmd *cmd)
+{
+       if (!cmd->se_lun) {
+               dump_stack();
+               pr_err("cmd->se_lun is NULL\n");
+               return -EINVAL;
+       }
+       if (in_interrupt()) {
+               dump_stack();
+               pr_err("transport_generic_handle_cdb cannot be called"
+                               " from interrupt context\n");
+               return -EINVAL;
+       }
+
+       return transport_generic_new_cmd(cmd);
+}
+EXPORT_SYMBOL(transport_handle_cdb_direct);
+
 /*
  * Used by fabric module frontends defining a TFO->new_cmd_map() caller
  * to  queue up a newly setup se_cmd w/ TRANSPORT_NEW_CMD_MAP in order to
@@ -2005,10 +1778,10 @@ EXPORT_SYMBOL(transport_generic_handle_cdb);
 int transport_generic_handle_cdb_map(
        struct se_cmd *cmd)
 {
-       if (!SE_LUN(cmd)) {
+       if (!cmd->se_lun) {
                dump_stack();
-               printk(KERN_ERR "SE_LUN(cmd) is NULL\n");
-               return -1;
+               pr_err("cmd->se_lun is NULL\n");
+               return -EINVAL;
        }
 
        transport_add_cmd_to_queue(cmd, TRANSPORT_NEW_CMD_MAP);
@@ -2030,7 +1803,7 @@ int transport_generic_handle_data(
         * in interrupt code, the signal_pending() check is skipped.
         */
        if (!in_interrupt() && signal_pending(current))
-               return -1;
+               return -EPERM;
        /*
         * If the received CDB has aleady been ABORTED by the generic
         * target engine, we now call transport_check_aborted_status()
@@ -2057,7 +1830,6 @@ int transport_generic_handle_tmr(
         * This is needed for early exceptions.
         */
        cmd->transport_wait_for_tasks = &transport_generic_wait_for_tasks;
-       transport_device_setup_cmd(cmd);
 
        transport_add_cmd_to_queue(cmd, TRANSPORT_PROCESS_TMR);
        return 0;
@@ -2077,16 +1849,16 @@ static int transport_stop_tasks_for_cmd(struct se_cmd *cmd)
        unsigned long flags;
        int ret = 0;
 
-       DEBUG_TS("ITT[0x%08x] - Stopping tasks\n",
-               CMD_TFO(cmd)->get_task_tag(cmd));
+       pr_debug("ITT[0x%08x] - Stopping tasks\n",
+               cmd->se_tfo->get_task_tag(cmd));
 
        /*
         * No tasks remain in the execution queue
         */
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        list_for_each_entry_safe(task, task_tmp,
-                               &T_TASK(cmd)->t_task_list, t_list) {
-               DEBUG_TS("task_no[%d] - Processing task %p\n",
+                               &cmd->t_task_list, t_list) {
+               pr_debug("task_no[%d] - Processing task %p\n",
                                task->task_no, task);
                /*
                 * If the struct se_task has not been sent and is not active,
@@ -2094,14 +1866,14 @@ static int transport_stop_tasks_for_cmd(struct se_cmd *cmd)
                 */
                if (!atomic_read(&task->task_sent) &&
                    !atomic_read(&task->task_active)) {
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                        flags);
                        transport_remove_task_from_execute_queue(task,
                                        task->se_dev);
 
-                       DEBUG_TS("task_no[%d] - Removed from execute queue\n",
+                       pr_debug("task_no[%d] - Removed from execute queue\n",
                                task->task_no);
-                       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+                       spin_lock_irqsave(&cmd->t_state_lock, flags);
                        continue;
                }
 
@@ -2111,42 +1883,32 @@ static int transport_stop_tasks_for_cmd(struct se_cmd *cmd)
                 */
                if (atomic_read(&task->task_active)) {
                        atomic_set(&task->task_stop, 1);
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                        flags);
 
-                       DEBUG_TS("task_no[%d] - Waiting to complete\n",
+                       pr_debug("task_no[%d] - Waiting to complete\n",
                                task->task_no);
                        wait_for_completion(&task->task_stop_comp);
-                       DEBUG_TS("task_no[%d] - Stopped successfully\n",
+                       pr_debug("task_no[%d] - Stopped successfully\n",
                                task->task_no);
 
-                       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-                       atomic_dec(&T_TASK(cmd)->t_task_cdbs_left);
+                       spin_lock_irqsave(&cmd->t_state_lock, flags);
+                       atomic_dec(&cmd->t_task_cdbs_left);
 
                        atomic_set(&task->task_active, 0);
                        atomic_set(&task->task_stop, 0);
                } else {
-                       DEBUG_TS("task_no[%d] - Did nothing\n", task->task_no);
+                       pr_debug("task_no[%d] - Did nothing\n", task->task_no);
                        ret++;
                }
 
                __transport_stop_task_timer(task, &flags);
        }
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        return ret;
 }
 
-static void transport_failure_reset_queue_depth(struct se_device *dev)
-{
-       unsigned long flags;
-
-       spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);
-       atomic_inc(&dev->depth_left);
-       atomic_inc(&SE_HBA(dev)->left_queue_depth);
-       spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
-}
-
 /*
  * Handle SAM-esque emulation for generic transport request failures.
  */
@@ -2156,29 +1918,31 @@ static void transport_generic_request_failure(
        int complete,
        int sc)
 {
-       DEBUG_GRF("-----[ Storage Engine Exception for cmd: %p ITT: 0x%08x"
-               " CDB: 0x%02x\n", cmd, CMD_TFO(cmd)->get_task_tag(cmd),
-               T_TASK(cmd)->t_task_cdb[0]);
-       DEBUG_GRF("-----[ i_state: %d t_state/def_t_state:"
+       int ret = 0;
+
+       pr_debug("-----[ Storage Engine Exception for cmd: %p ITT: 0x%08x"
+               " CDB: 0x%02x\n", cmd, cmd->se_tfo->get_task_tag(cmd),
+               cmd->t_task_cdb[0]);
+       pr_debug("-----[ i_state: %d t_state/def_t_state:"
                " %d/%d transport_error_status: %d\n",
-               CMD_TFO(cmd)->get_cmd_state(cmd),
+               cmd->se_tfo->get_cmd_state(cmd),
                cmd->t_state, cmd->deferred_t_state,
                cmd->transport_error_status);
-       DEBUG_GRF("-----[ t_task_cdbs: %d t_task_cdbs_left: %d"
+       pr_debug("-----[ t_tasks: %d t_task_cdbs_left: %d"
                " t_task_cdbs_sent: %d t_task_cdbs_ex_left: %d --"
                " t_transport_active: %d t_transport_stop: %d"
-               " t_transport_sent: %d\n", T_TASK(cmd)->t_task_cdbs,
-               atomic_read(&T_TASK(cmd)->t_task_cdbs_left),
-               atomic_read(&T_TASK(cmd)->t_task_cdbs_sent),
-               atomic_read(&T_TASK(cmd)->t_task_cdbs_ex_left),
-               atomic_read(&T_TASK(cmd)->t_transport_active),
-               atomic_read(&T_TASK(cmd)->t_transport_stop),
-               atomic_read(&T_TASK(cmd)->t_transport_sent));
+               " t_transport_sent: %d\n", cmd->t_task_list_num,
+               atomic_read(&cmd->t_task_cdbs_left),
+               atomic_read(&cmd->t_task_cdbs_sent),
+               atomic_read(&cmd->t_task_cdbs_ex_left),
+               atomic_read(&cmd->t_transport_active),
+               atomic_read(&cmd->t_transport_stop),
+               atomic_read(&cmd->t_transport_sent));
 
        transport_stop_all_task_timers(cmd);
 
        if (dev)
-               transport_failure_reset_queue_depth(dev);
+               atomic_inc(&dev->depth_left);
        /*
         * For SAM Task Attribute emulation for failed struct se_cmd
         */
@@ -2211,8 +1975,8 @@ static void transport_generic_request_failure(
                 * we force this session to fall back to session
                 * recovery.
                 */
-               CMD_TFO(cmd)->fall_back_to_erl0(cmd->se_sess);
-               CMD_TFO(cmd)->stop_session(cmd->se_sess, 0, 0);
+               cmd->se_tfo->fall_back_to_erl0(cmd->se_sess);
+               cmd->se_tfo->stop_session(cmd->se_sess, 0, 0);
 
                goto check_stop;
        case PYX_TRANSPORT_LU_COMM_FAILURE:
@@ -2240,13 +2004,15 @@ static void transport_generic_request_failure(
                 *
                 * See spc4r17, section 7.4.6 Control Mode Page, Table 349
                 */
-               if (SE_SESS(cmd) &&
-                   DEV_ATTRIB(cmd->se_dev)->emulate_ua_intlck_ctrl == 2)
-                       core_scsi3_ua_allocate(SE_SESS(cmd)->se_node_acl,
+               if (cmd->se_sess &&
+                   cmd->se_dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl == 2)
+                       core_scsi3_ua_allocate(cmd->se_sess->se_node_acl,
                                cmd->orig_fe_lun, 0x2C,
                                ASCQ_2CH_PREVIOUS_RESERVATION_CONFLICT_STATUS);
 
-               CMD_TFO(cmd)->queue_status(cmd);
+               ret = cmd->se_tfo->queue_status(cmd);
+               if (ret == -EAGAIN)
+                       goto queue_full;
                goto check_stop;
        case PYX_TRANSPORT_USE_SENSE_REASON:
                /*
@@ -2254,8 +2020,8 @@ static void transport_generic_request_failure(
                 */
                break;
        default:
-               printk(KERN_ERR "Unknown transport error for CDB 0x%02x: %d\n",
-                       T_TASK(cmd)->t_task_cdb[0],
+               pr_err("Unknown transport error for CDB 0x%02x: %d\n",
+                       cmd->t_task_cdb[0],
                        cmd->transport_error_status);
                cmd->scsi_sense_reason = TCM_UNSUPPORTED_SCSI_OPCODE;
                break;
@@ -2263,32 +2029,41 @@ static void transport_generic_request_failure(
 
        if (!sc)
                transport_new_cmd_failure(cmd);
-       else
-               transport_send_check_condition_and_sense(cmd,
-                       cmd->scsi_sense_reason, 0);
+       else {
+               ret = transport_send_check_condition_and_sense(cmd,
+                               cmd->scsi_sense_reason, 0);
+               if (ret == -EAGAIN)
+                       goto queue_full;
+       }
+
 check_stop:
        transport_lun_remove_cmd(cmd);
-       if (!(transport_cmd_check_stop_to_fabric(cmd)))
+       if (!transport_cmd_check_stop_to_fabric(cmd))
                ;
+       return;
+
+queue_full:
+       cmd->t_state = TRANSPORT_COMPLETE_OK;
+       transport_handle_queue_full(cmd, cmd->se_dev, transport_complete_qf);
 }
 
 static void transport_direct_request_timeout(struct se_cmd *cmd)
 {
        unsigned long flags;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       if (!(atomic_read(&T_TASK(cmd)->t_transport_timeout))) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       if (!atomic_read(&cmd->t_transport_timeout)) {
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return;
        }
-       if (atomic_read(&T_TASK(cmd)->t_task_cdbs_timeout_left)) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       if (atomic_read(&cmd->t_task_cdbs_timeout_left)) {
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return;
        }
 
-       atomic_sub(atomic_read(&T_TASK(cmd)->t_transport_timeout),
-                  &T_TASK(cmd)->t_se_count);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       atomic_sub(atomic_read(&cmd->t_transport_timeout),
+                  &cmd->t_se_count);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 }
 
 static void transport_generic_request_timeout(struct se_cmd *cmd)
@@ -2296,35 +2071,18 @@ static void transport_generic_request_timeout(struct se_cmd *cmd)
        unsigned long flags;
 
        /*
-        * Reset T_TASK(cmd)->t_se_count to allow transport_generic_remove()
+        * Reset cmd->t_se_count to allow transport_generic_remove()
         * to allow last call to free memory resources.
         */
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       if (atomic_read(&T_TASK(cmd)->t_transport_timeout) > 1) {
-               int tmp = (atomic_read(&T_TASK(cmd)->t_transport_timeout) - 1);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       if (atomic_read(&cmd->t_transport_timeout) > 1) {
+               int tmp = (atomic_read(&cmd->t_transport_timeout) - 1);
 
-               atomic_sub(tmp, &T_TASK(cmd)->t_se_count);
+               atomic_sub(tmp, &cmd->t_se_count);
        }
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
-
-       transport_generic_remove(cmd, 0, 0);
-}
-
-static int
-transport_generic_allocate_buf(struct se_cmd *cmd, u32 data_length)
-{
-       unsigned char *buf;
-
-       buf = kzalloc(data_length, GFP_KERNEL);
-       if (!(buf)) {
-               printk(KERN_ERR "Unable to allocate memory for buffer\n");
-               return -1;
-       }
-
-       T_TASK(cmd)->t_tasks_se_num = 0;
-       T_TASK(cmd)->t_task_buf = buf;
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
-       return 0;
+       transport_generic_remove(cmd, 0);
 }
 
 static inline u32 transport_lba_21(unsigned char *cdb)
@@ -2364,9 +2122,9 @@ static void transport_set_supported_SAM_opcode(struct se_cmd *se_cmd)
 {
        unsigned long flags;
 
-       spin_lock_irqsave(&T_TASK(se_cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&se_cmd->t_state_lock, flags);
        se_cmd->se_cmd_flags |= SCF_SUPPORTED_SAM_OPCODE;
-       spin_unlock_irqrestore(&T_TASK(se_cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&se_cmd->t_state_lock, flags);
 }
 
 /*
@@ -2375,14 +2133,14 @@ static void transport_set_supported_SAM_opcode(struct se_cmd *se_cmd)
 static void transport_task_timeout_handler(unsigned long data)
 {
        struct se_task *task = (struct se_task *)data;
-       struct se_cmd *cmd = TASK_CMD(task);
+       struct se_cmd *cmd = task->task_se_cmd;
        unsigned long flags;
 
-       DEBUG_TT("transport task timeout fired! task: %p cmd: %p\n", task, cmd);
+       pr_debug("transport task timeout fired! task: %p cmd: %p\n", task, cmd);
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        if (task->task_flags & TF_STOP) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return;
        }
        task->task_flags &= ~TF_RUNNING;
@@ -2390,46 +2148,46 @@ static void transport_task_timeout_handler(unsigned long data)
        /*
         * Determine if transport_complete_task() has already been called.
         */
-       if (!(atomic_read(&task->task_active))) {
-               DEBUG_TT("transport task: %p cmd: %p timeout task_active"
+       if (!atomic_read(&task->task_active)) {
+               pr_debug("transport task: %p cmd: %p timeout task_active"
                                " == 0\n", task, cmd);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return;
        }
 
-       atomic_inc(&T_TASK(cmd)->t_se_count);
-       atomic_inc(&T_TASK(cmd)->t_transport_timeout);
-       T_TASK(cmd)->t_tasks_failed = 1;
+       atomic_inc(&cmd->t_se_count);
+       atomic_inc(&cmd->t_transport_timeout);
+       cmd->t_tasks_failed = 1;
 
        atomic_set(&task->task_timeout, 1);
        task->task_error_status = PYX_TRANSPORT_TASK_TIMEOUT;
        task->task_scsi_status = 1;
 
        if (atomic_read(&task->task_stop)) {
-               DEBUG_TT("transport task: %p cmd: %p timeout task_stop"
+               pr_debug("transport task: %p cmd: %p timeout task_stop"
                                " == 1\n", task, cmd);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                complete(&task->task_stop_comp);
                return;
        }
 
-       if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_left))) {
-               DEBUG_TT("transport task: %p cmd: %p timeout non zero"
+       if (!atomic_dec_and_test(&cmd->t_task_cdbs_left)) {
+               pr_debug("transport task: %p cmd: %p timeout non zero"
                                " t_task_cdbs_left\n", task, cmd);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return;
        }
-       DEBUG_TT("transport task: %p cmd: %p timeout ZERO t_task_cdbs_left\n",
+       pr_debug("transport task: %p cmd: %p timeout ZERO t_task_cdbs_left\n",
                        task, cmd);
 
        cmd->t_state = TRANSPORT_COMPLETE_FAILURE;
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        transport_add_cmd_to_queue(cmd, TRANSPORT_COMPLETE_FAILURE);
 }
 
 /*
- * Called with T_TASK(cmd)->t_state_lock held.
+ * Called with cmd->t_state_lock held.
  */
 static void transport_start_task_timer(struct se_task *task)
 {
@@ -2441,8 +2199,8 @@ static void transport_start_task_timer(struct se_task *task)
        /*
         * If the task_timeout is disabled, exit now.
         */
-       timeout = DEV_ATTRIB(dev)->task_timeout;
-       if (!(timeout))
+       timeout = dev->se_sub_dev->se_dev_attrib.task_timeout;
+       if (!timeout)
                return;
 
        init_timer(&task->task_timer);
@@ -2453,27 +2211,27 @@ static void transport_start_task_timer(struct se_task *task)
        task->task_flags |= TF_RUNNING;
        add_timer(&task->task_timer);
 #if 0
-       printk(KERN_INFO "Starting task timer for cmd: %p task: %p seconds:"
+       pr_debug("Starting task timer for cmd: %p task: %p seconds:"
                " %d\n", task->task_se_cmd, task, timeout);
 #endif
 }
 
 /*
- * Called with spin_lock_irq(&T_TASK(cmd)->t_state_lock) held.
+ * Called with spin_lock_irq(&cmd->t_state_lock) held.
  */
 void __transport_stop_task_timer(struct se_task *task, unsigned long *flags)
 {
-       struct se_cmd *cmd = TASK_CMD(task);
+       struct se_cmd *cmd = task->task_se_cmd;
 
-       if (!(task->task_flags & TF_RUNNING))
+       if (!task->task_flags & TF_RUNNING)
                return;
 
        task->task_flags |= TF_STOP;
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, *flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, *flags);
 
        del_timer_sync(&task->task_timer);
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, *flags);
+       spin_lock_irqsave(&cmd->t_state_lock, *flags);
        task->task_flags &= ~TF_RUNNING;
        task->task_flags &= ~TF_STOP;
 }
@@ -2483,11 +2241,11 @@ static void transport_stop_all_task_timers(struct se_cmd *cmd)
        struct se_task *task = NULL, *task_tmp;
        unsigned long flags;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        list_for_each_entry_safe(task, task_tmp,
-                               &T_TASK(cmd)->t_task_list, t_list)
+                               &cmd->t_task_list, t_list)
                __transport_stop_task_timer(task, &flags);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 }
 
 static inline int transport_tcq_window_closed(struct se_device *dev)
@@ -2498,7 +2256,7 @@ static inline int transport_tcq_window_closed(struct se_device *dev)
        } else
                msleep(PYX_TRANSPORT_WINDOW_CLOSED_WAIT_LONG);
 
-       wake_up_interruptible(&dev->dev_queue_obj->thread_wq);
+       wake_up_interruptible(&dev->dev_queue_obj.thread_wq);
        return 0;
 }
 
@@ -2511,45 +2269,45 @@ static inline int transport_tcq_window_closed(struct se_device *dev)
  */
 static inline int transport_execute_task_attr(struct se_cmd *cmd)
 {
-       if (SE_DEV(cmd)->dev_task_attr_type != SAM_TASK_ATTR_EMULATED)
+       if (cmd->se_dev->dev_task_attr_type != SAM_TASK_ATTR_EMULATED)
                return 1;
        /*
         * Check for the existence of HEAD_OF_QUEUE, and if true return 1
         * to allow the passed struct se_cmd list of tasks to the front of the list.
         */
         if (cmd->sam_task_attr == MSG_HEAD_TAG) {
-               atomic_inc(&SE_DEV(cmd)->dev_hoq_count);
+               atomic_inc(&cmd->se_dev->dev_hoq_count);
                smp_mb__after_atomic_inc();
-               DEBUG_STA("Added HEAD_OF_QUEUE for CDB:"
+               pr_debug("Added HEAD_OF_QUEUE for CDB:"
                        " 0x%02x, se_ordered_id: %u\n",
-                       T_TASK(cmd)->t_task_cdb[0],
+                       cmd->t_task_cdb[0],
                        cmd->se_ordered_id);
                return 1;
        } else if (cmd->sam_task_attr == MSG_ORDERED_TAG) {
-               spin_lock(&SE_DEV(cmd)->ordered_cmd_lock);
-               list_add_tail(&cmd->se_ordered_list,
-                               &SE_DEV(cmd)->ordered_cmd_list);
-               spin_unlock(&SE_DEV(cmd)->ordered_cmd_lock);
+               spin_lock(&cmd->se_dev->ordered_cmd_lock);
+               list_add_tail(&cmd->se_ordered_node,
+                               &cmd->se_dev->ordered_cmd_list);
+               spin_unlock(&cmd->se_dev->ordered_cmd_lock);
 
-               atomic_inc(&SE_DEV(cmd)->dev_ordered_sync);
+               atomic_inc(&cmd->se_dev->dev_ordered_sync);
                smp_mb__after_atomic_inc();
 
-               DEBUG_STA("Added ORDERED for CDB: 0x%02x to ordered"
+               pr_debug("Added ORDERED for CDB: 0x%02x to ordered"
                                " list, se_ordered_id: %u\n",
-                               T_TASK(cmd)->t_task_cdb[0],
+                               cmd->t_task_cdb[0],
                                cmd->se_ordered_id);
                /*
                 * Add ORDERED command to tail of execution queue if
                 * no other older commands exist that need to be
                 * completed first.
                 */
-               if (!(atomic_read(&SE_DEV(cmd)->simple_cmds)))
+               if (!atomic_read(&cmd->se_dev->simple_cmds))
                        return 1;
        } else {
                /*
                 * For SIMPLE and UNTAGGED Task Attribute commands
                 */
-               atomic_inc(&SE_DEV(cmd)->simple_cmds);
+               atomic_inc(&cmd->se_dev->simple_cmds);
                smp_mb__after_atomic_inc();
        }
        /*
@@ -2557,20 +2315,20 @@ static inline int transport_execute_task_attr(struct se_cmd *cmd)
         * add the dormant task(s) built for the passed struct se_cmd to the
         * execution queue and become in Active state for this struct se_device.
         */
-       if (atomic_read(&SE_DEV(cmd)->dev_ordered_sync) != 0) {
+       if (atomic_read(&cmd->se_dev->dev_ordered_sync) != 0) {
                /*
                 * Otherwise, add cmd w/ tasks to delayed cmd queue that
                 * will be drained upon completion of HEAD_OF_QUEUE task.
                 */
-               spin_lock(&SE_DEV(cmd)->delayed_cmd_lock);
+               spin_lock(&cmd->se_dev->delayed_cmd_lock);
                cmd->se_cmd_flags |= SCF_DELAYED_CMD_FROM_SAM_ATTR;
-               list_add_tail(&cmd->se_delayed_list,
-                               &SE_DEV(cmd)->delayed_cmd_list);
-               spin_unlock(&SE_DEV(cmd)->delayed_cmd_lock);
+               list_add_tail(&cmd->se_delayed_node,
+                               &cmd->se_dev->delayed_cmd_list);
+               spin_unlock(&cmd->se_dev->delayed_cmd_lock);
 
-               DEBUG_STA("Added CDB: 0x%02x Task Attr: 0x%02x to"
+               pr_debug("Added CDB: 0x%02x Task Attr: 0x%02x to"
                        " delayed CMD list, se_ordered_id: %u\n",
-                       T_TASK(cmd)->t_task_cdb[0], cmd->sam_task_attr,
+                       cmd->t_task_cdb[0], cmd->sam_task_attr,
                        cmd->se_ordered_id);
                /*
                 * Return zero to let transport_execute_tasks() know
@@ -2592,25 +2350,23 @@ static int transport_execute_tasks(struct se_cmd *cmd)
 {
        int add_tasks;
 
-       if (!(cmd->se_cmd_flags & SCF_SE_DISABLE_ONLINE_CHECK)) {
-               if (se_dev_check_online(cmd->se_orig_obj_ptr) != 0) {
-                       cmd->transport_error_status =
-                               PYX_TRANSPORT_LU_COMM_FAILURE;
-                       transport_generic_request_failure(cmd, NULL, 0, 1);
-                       return 0;
-               }
+       if (se_dev_check_online(cmd->se_orig_obj_ptr) != 0) {
+               cmd->transport_error_status = PYX_TRANSPORT_LU_COMM_FAILURE;
+               transport_generic_request_failure(cmd, NULL, 0, 1);
+               return 0;
        }
+
        /*
         * Call transport_cmd_check_stop() to see if a fabric exception
         * has occurred that prevents execution.
         */
-       if (!(transport_cmd_check_stop(cmd, 0, TRANSPORT_PROCESSING))) {
+       if (!transport_cmd_check_stop(cmd, 0, TRANSPORT_PROCESSING)) {
                /*
                 * Check for SAM Task Attribute emulation and HEAD_OF_QUEUE
                 * attribute for the tasks of the received struct se_cmd CDB
                 */
                add_tasks = transport_execute_task_attr(cmd);
-               if (add_tasks == 0)
+               if (!add_tasks)
                        goto execute_tasks;
                /*
                 * This calls transport_add_tasks_from_cmd() to handle
@@ -2625,7 +2381,7 @@ static int transport_execute_tasks(struct se_cmd *cmd)
         * storage object.
         */
 execute_tasks:
-       __transport_execute_tasks(SE_DEV(cmd));
+       __transport_execute_tasks(cmd->se_dev);
        return 0;
 }
 
@@ -2639,51 +2395,49 @@ static int __transport_execute_tasks(struct se_device *dev)
 {
        int error;
        struct se_cmd *cmd = NULL;
-       struct se_task *task;
+       struct se_task *task = NULL;
        unsigned long flags;
 
        /*
         * Check if there is enough room in the device and HBA queue to send
-        * struct se_transport_task's to the selected transport.
+        * struct se_tasks to the selected transport.
         */
 check_depth:
-       spin_lock_irqsave(&SE_HBA(dev)->hba_queue_lock, flags);
-       if (!(atomic_read(&dev->depth_left)) ||
-           !(atomic_read(&SE_HBA(dev)->left_queue_depth))) {
-               spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+       if (!atomic_read(&dev->depth_left))
                return transport_tcq_window_closed(dev);
-       }
-       dev->dev_tcq_window_closed = 0;
 
-       spin_lock(&dev->execute_task_lock);
-       task = transport_get_task_from_execute_queue(dev);
-       spin_unlock(&dev->execute_task_lock);
+       dev->dev_tcq_window_closed = 0;
 
-       if (!task) {
-               spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
+       spin_lock_irq(&dev->execute_task_lock);
+       if (list_empty(&dev->execute_task_list)) {
+               spin_unlock_irq(&dev->execute_task_lock);
                return 0;
        }
+       task = list_first_entry(&dev->execute_task_list,
+                               struct se_task, t_execute_list);
+       list_del(&task->t_execute_list);
+       atomic_set(&task->task_execute_queue, 0);
+       atomic_dec(&dev->execute_tasks);
+       spin_unlock_irq(&dev->execute_task_lock);
 
        atomic_dec(&dev->depth_left);
-       atomic_dec(&SE_HBA(dev)->left_queue_depth);
-       spin_unlock_irqrestore(&SE_HBA(dev)->hba_queue_lock, flags);
 
-       cmd = TASK_CMD(task);
+       cmd = task->task_se_cmd;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        atomic_set(&task->task_active, 1);
        atomic_set(&task->task_sent, 1);
-       atomic_inc(&T_TASK(cmd)->t_task_cdbs_sent);
+       atomic_inc(&cmd->t_task_cdbs_sent);
 
-       if (atomic_read(&T_TASK(cmd)->t_task_cdbs_sent) ==
-           T_TASK(cmd)->t_task_cdbs)
+       if (atomic_read(&cmd->t_task_cdbs_sent) ==
+           cmd->t_task_list_num)
                atomic_set(&cmd->transport_sent, 1);
 
        transport_start_task_timer(task);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
        /*
         * The struct se_cmd->transport_emulate_cdb() function pointer is used
-        * to grab REPORT_LUNS CDBs before they hit the
+        * to grab REPORT_LUNS and other CDBs we want to handle before they hit the
         * struct se_subsystem_api->do_task() caller below.
         */
        if (cmd->transport_emulate_cdb) {
@@ -2718,11 +2472,11 @@ check_depth:
                 * call ->do_task() directly and let the underlying TCM subsystem plugin
                 * code handle the CDB emulation.
                 */
-               if ((TRANSPORT(dev)->transport_type != TRANSPORT_PLUGIN_PHBA_PDEV) &&
-                   (!(TASK_CMD(task)->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB)))
+               if ((dev->transport->transport_type != TRANSPORT_PLUGIN_PHBA_PDEV) &&
+                   (!(task->task_se_cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB)))
                        error = transport_emulate_control_cdb(task);
                else
-                       error = TRANSPORT(dev)->do_task(task);
+                       error = dev->transport->do_task(task);
 
                if (error != 0) {
                        cmd->transport_error_status = error;
@@ -2745,12 +2499,10 @@ void transport_new_cmd_failure(struct se_cmd *se_cmd)
         * Any unsolicited data will get dumped for failed command inside of
         * the fabric plugin
         */
-       spin_lock_irqsave(&T_TASK(se_cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&se_cmd->t_state_lock, flags);
        se_cmd->se_cmd_flags |= SCF_SE_CMD_FAILED;
        se_cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-       spin_unlock_irqrestore(&T_TASK(se_cmd)->t_state_lock, flags);
-
-       CMD_TFO(se_cmd)->new_cmd_failure(se_cmd);
+       spin_unlock_irqrestore(&se_cmd->t_state_lock, flags);
 }
 
 static void transport_nop_wait_for_tasks(struct se_cmd *, int, int);
@@ -2760,7 +2512,7 @@ static inline u32 transport_get_sectors_6(
        struct se_cmd *cmd,
        int *ret)
 {
-       struct se_device *dev = SE_LUN(cmd)->lun_se_dev;
+       struct se_device *dev = cmd->se_dev;
 
        /*
         * Assume TYPE_DISK for non struct se_device objects.
@@ -2772,7 +2524,7 @@ static inline u32 transport_get_sectors_6(
        /*
         * Use 24-bit allocation length for TYPE_TAPE.
         */
-       if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+       if (dev->transport->get_device_type(dev) == TYPE_TAPE)
                return (u32)(cdb[2] << 16) + (cdb[3] << 8) + cdb[4];
 
        /*
@@ -2788,7 +2540,7 @@ static inline u32 transport_get_sectors_10(
        struct se_cmd *cmd,
        int *ret)
 {
-       struct se_device *dev = SE_LUN(cmd)->lun_se_dev;
+       struct se_device *dev = cmd->se_dev;
 
        /*
         * Assume TYPE_DISK for non struct se_device objects.
@@ -2800,8 +2552,8 @@ static inline u32 transport_get_sectors_10(
        /*
         * XXX_10 is not defined in SSC, throw an exception
         */
-       if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE) {
-               *ret = -1;
+       if (dev->transport->get_device_type(dev) == TYPE_TAPE) {
+               *ret = -EINVAL;
                return 0;
        }
 
@@ -2818,7 +2570,7 @@ static inline u32 transport_get_sectors_12(
        struct se_cmd *cmd,
        int *ret)
 {
-       struct se_device *dev = SE_LUN(cmd)->lun_se_dev;
+       struct se_device *dev = cmd->se_dev;
 
        /*
         * Assume TYPE_DISK for non struct se_device objects.
@@ -2830,8 +2582,8 @@ static inline u32 transport_get_sectors_12(
        /*
         * XXX_12 is not defined in SSC, throw an exception
         */
-       if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE) {
-               *ret = -1;
+       if (dev->transport->get_device_type(dev) == TYPE_TAPE) {
+               *ret = -EINVAL;
                return 0;
        }
 
@@ -2848,7 +2600,7 @@ static inline u32 transport_get_sectors_16(
        struct se_cmd *cmd,
        int *ret)
 {
-       struct se_device *dev = SE_LUN(cmd)->lun_se_dev;
+       struct se_device *dev = cmd->se_dev;
 
        /*
         * Assume TYPE_DISK for non struct se_device objects.
@@ -2860,7 +2612,7 @@ static inline u32 transport_get_sectors_16(
        /*
         * Use 24-bit allocation length for TYPE_TAPE.
         */
-       if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE)
+       if (dev->transport->get_device_type(dev) == TYPE_TAPE)
                return (u32)(cdb[12] << 16) + (cdb[13] << 8) + cdb[14];
 
 type_disk:
@@ -2890,57 +2642,30 @@ static inline u32 transport_get_size(
        unsigned char *cdb,
        struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
 
-       if (TRANSPORT(dev)->get_device_type(dev) == TYPE_TAPE) {
+       if (dev->transport->get_device_type(dev) == TYPE_TAPE) {
                if (cdb[1] & 1) { /* sectors */
-                       return DEV_ATTRIB(dev)->block_size * sectors;
+                       return dev->se_sub_dev->se_dev_attrib.block_size * sectors;
                } else /* bytes */
                        return sectors;
        }
 #if 0
-       printk(KERN_INFO "Returning block_size: %u, sectors: %u == %u for"
-                       " %s object\n", DEV_ATTRIB(dev)->block_size, sectors,
-                       DEV_ATTRIB(dev)->block_size * sectors,
-                       TRANSPORT(dev)->name);
+       pr_debug("Returning block_size: %u, sectors: %u == %u for"
+                       " %s object\n", dev->se_sub_dev->se_dev_attrib.block_size, sectors,
+                       dev->se_sub_dev->se_dev_attrib.block_size * sectors,
+                       dev->transport->name);
 #endif
-       return DEV_ATTRIB(dev)->block_size * sectors;
-}
-
-unsigned char transport_asciihex_to_binaryhex(unsigned char val[2])
-{
-       unsigned char result = 0;
-       /*
-        * MSB
-        */
-       if ((val[0] >= 'a') && (val[0] <= 'f'))
-               result = ((val[0] - 'a' + 10) & 0xf) << 4;
-       else
-               if ((val[0] >= 'A') && (val[0] <= 'F'))
-                       result = ((val[0] - 'A' + 10) & 0xf) << 4;
-               else /* digit */
-                       result = ((val[0] - '0') & 0xf) << 4;
-       /*
-        * LSB
-        */
-       if ((val[1] >= 'a') && (val[1] <= 'f'))
-               result |= ((val[1] - 'a' + 10) & 0xf);
-       else
-               if ((val[1] >= 'A') && (val[1] <= 'F'))
-                       result |= ((val[1] - 'A' + 10) & 0xf);
-               else /* digit */
-                       result |= ((val[1] - '0') & 0xf);
-
-       return result;
+       return dev->se_sub_dev->se_dev_attrib.block_size * sectors;
 }
-EXPORT_SYMBOL(transport_asciihex_to_binaryhex);
 
 static void transport_xor_callback(struct se_cmd *cmd)
 {
        unsigned char *buf, *addr;
-       struct se_mem *se_mem;
+       struct scatterlist *sg;
        unsigned int offset;
        int i;
+       int count;
        /*
         * From sbc3r22.pdf section 5.48 XDWRITEREAD (10) command
         *
@@ -2953,32 +2678,37 @@ static void transport_xor_callback(struct se_cmd *cmd)
         * 5) transfer the resulting XOR data to the data-in buffer.
         */
        buf = kmalloc(cmd->data_length, GFP_KERNEL);
-       if (!(buf)) {
-               printk(KERN_ERR "Unable to allocate xor_callback buf\n");
+       if (!buf) {
+               pr_err("Unable to allocate xor_callback buf\n");
                return;
        }
        /*
-        * Copy the scatterlist WRITE buffer located at T_TASK(cmd)->t_mem_list
+        * Copy the scatterlist WRITE buffer located at cmd->t_data_sg
         * into the locally allocated *buf
         */
-       transport_memcpy_se_mem_read_contig(cmd, buf, T_TASK(cmd)->t_mem_list);
+       sg_copy_to_buffer(cmd->t_data_sg,
+                         cmd->t_data_nents,
+                         buf,
+                         cmd->data_length);
+
        /*
         * Now perform the XOR against the BIDI read memory located at
-        * T_TASK(cmd)->t_mem_bidi_list
+        * cmd->t_mem_bidi_list
         */
 
        offset = 0;
-       list_for_each_entry(se_mem, T_TASK(cmd)->t_mem_bidi_list, se_list) {
-               addr = (unsigned char *)kmap_atomic(se_mem->se_page, KM_USER0);
-               if (!(addr))
+       for_each_sg(cmd->t_bidi_data_sg, sg, cmd->t_bidi_data_nents, count) {
+               addr = kmap_atomic(sg_page(sg), KM_USER0);
+               if (!addr)
                        goto out;
 
-               for (i = 0; i < se_mem->se_len; i++)
-                       *(addr + se_mem->se_off + i) ^= *(buf + offset + i);
+               for (i = 0; i < sg->length; i++)
+                       *(addr + sg->offset + i) ^= *(buf + offset + i);
 
-               offset += se_mem->se_len;
+               offset += sg->length;
                kunmap_atomic(addr, KM_USER0);
        }
+
 out:
        kfree(buf);
 }
@@ -2994,75 +2724,60 @@ static int transport_get_sense_data(struct se_cmd *cmd)
        unsigned long flags;
        u32 offset = 0;
 
-       if (!SE_LUN(cmd)) {
-               printk(KERN_ERR "SE_LUN(cmd) is NULL\n");
-               return -1;
-       }
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       WARN_ON(!cmd->se_lun);
+
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        if (cmd->se_cmd_flags & SCF_SENT_CHECK_CONDITION) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return 0;
        }
 
        list_for_each_entry_safe(task, task_tmp,
-                               &T_TASK(cmd)->t_task_list, t_list) {
+                               &cmd->t_task_list, t_list) {
 
                if (!task->task_sense)
                        continue;
 
                dev = task->se_dev;
-               if (!(dev))
+               if (!dev)
                        continue;
 
-               if (!TRANSPORT(dev)->get_sense_buffer) {
-                       printk(KERN_ERR "TRANSPORT(dev)->get_sense_buffer"
+               if (!dev->transport->get_sense_buffer) {
+                       pr_err("dev->transport->get_sense_buffer"
                                        " is NULL\n");
                        continue;
                }
 
-               sense_buffer = TRANSPORT(dev)->get_sense_buffer(task);
-               if (!(sense_buffer)) {
-                       printk(KERN_ERR "ITT[0x%08x]_TASK[%d]: Unable to locate"
+               sense_buffer = dev->transport->get_sense_buffer(task);
+               if (!sense_buffer) {
+                       pr_err("ITT[0x%08x]_TASK[%d]: Unable to locate"
                                " sense buffer for task with sense\n",
-                               CMD_TFO(cmd)->get_task_tag(cmd), task->task_no);
+                               cmd->se_tfo->get_task_tag(cmd), task->task_no);
                        continue;
                }
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
-               offset = CMD_TFO(cmd)->set_fabric_sense_len(cmd,
+               offset = cmd->se_tfo->set_fabric_sense_len(cmd,
                                TRANSPORT_SENSE_BUFFER);
 
-               memcpy((void *)&buffer[offset], (void *)sense_buffer,
+               memcpy(&buffer[offset], sense_buffer,
                                TRANSPORT_SENSE_BUFFER);
                cmd->scsi_status = task->task_scsi_status;
                /* Automatically padded */
                cmd->scsi_sense_length =
                                (TRANSPORT_SENSE_BUFFER + offset);
 
-               printk(KERN_INFO "HBA_[%u]_PLUG[%s]: Set SAM STATUS: 0x%02x"
+               pr_debug("HBA_[%u]_PLUG[%s]: Set SAM STATUS: 0x%02x"
                                " and sense\n",
-                       dev->se_hba->hba_id, TRANSPORT(dev)->name,
+                       dev->se_hba->hba_id, dev->transport->name,
                                cmd->scsi_status);
                return 0;
        }
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        return -1;
 }
 
-static int transport_allocate_resources(struct se_cmd *cmd)
-{
-       u32 length = cmd->data_length;
-
-       if ((cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB) ||
-           (cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB))
-               return transport_generic_get_mem(cmd, length, PAGE_SIZE);
-       else if (cmd->se_cmd_flags & SCF_SCSI_CONTROL_NONSG_IO_CDB)
-               return transport_generic_allocate_buf(cmd, length);
-       else
-               return 0;
-}
-
 static int
 transport_handle_reservation_conflict(struct se_cmd *cmd)
 {
@@ -3077,12 +2792,40 @@ transport_handle_reservation_conflict(struct se_cmd *cmd)
         *
         * See spc4r17, section 7.4.6 Control Mode Page, Table 349
         */
-       if (SE_SESS(cmd) &&
-           DEV_ATTRIB(cmd->se_dev)->emulate_ua_intlck_ctrl == 2)
-               core_scsi3_ua_allocate(SE_SESS(cmd)->se_node_acl,
+       if (cmd->se_sess &&
+           cmd->se_dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl == 2)
+               core_scsi3_ua_allocate(cmd->se_sess->se_node_acl,
                        cmd->orig_fe_lun, 0x2C,
                        ASCQ_2CH_PREVIOUS_RESERVATION_CONFLICT_STATUS);
-       return -2;
+       return -EINVAL;
+}
+
+static inline long long transport_dev_end_lba(struct se_device *dev)
+{
+       return dev->transport->get_blocks(dev) + 1;
+}
+
+static int transport_cmd_get_valid_sectors(struct se_cmd *cmd)
+{
+       struct se_device *dev = cmd->se_dev;
+       u32 sectors;
+
+       if (dev->transport->get_device_type(dev) != TYPE_DISK)
+               return 0;
+
+       sectors = (cmd->data_length / dev->se_sub_dev->se_dev_attrib.block_size);
+
+       if ((cmd->t_task_lba + sectors) > transport_dev_end_lba(dev)) {
+               pr_err("LBA: %llu Sectors: %u exceeds"
+                       " transport_dev_end_lba(): %llu\n",
+                       cmd->t_task_lba, sectors,
+                       transport_dev_end_lba(dev));
+               pr_err("  We should return CHECK_CONDITION"
+                      " but we don't yet\n");
+               return 0;
+       }
+
+       return sectors;
 }
 
 /*     transport_generic_cmd_sequencer():
@@ -3099,7 +2842,7 @@ static int transport_generic_cmd_sequencer(
        struct se_cmd *cmd,
        unsigned char *cdb)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct se_subsystem_dev *su_dev = dev->se_sub_dev;
        int ret = 0, sector_ret = 0, passthrough;
        u32 sectors = 0, size = 0, pr_reg_type = 0;
@@ -3113,12 +2856,12 @@ static int transport_generic_cmd_sequencer(
                                &transport_nop_wait_for_tasks;
                cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
                cmd->scsi_sense_reason = TCM_CHECK_CONDITION_UNIT_ATTENTION;
-               return -2;
+               return -EINVAL;
        }
        /*
         * Check status of Asymmetric Logical Unit Assignment port
         */
-       ret = T10_ALUA(su_dev)->alua_state_check(cmd, cdb, &alua_ascq);
+       ret = su_dev->t10_alua.alua_state_check(cmd, cdb, &alua_ascq);
        if (ret != 0) {
                cmd->transport_wait_for_tasks = &transport_nop_wait_for_tasks;
                /*
@@ -3128,22 +2871,22 @@ static int transport_generic_cmd_sequencer(
                 */
                if (ret > 0) {
 #if 0
-                       printk(KERN_INFO "[%s]: ALUA TG Port not available,"
+                       pr_debug("[%s]: ALUA TG Port not available,"
                                " SenseKey: NOT_READY, ASC/ASCQ: 0x04/0x%02x\n",
-                               CMD_TFO(cmd)->get_fabric_name(), alua_ascq);
+                               cmd->se_tfo->get_fabric_name(), alua_ascq);
 #endif
                        transport_set_sense_codes(cmd, 0x04, alua_ascq);
                        cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
                        cmd->scsi_sense_reason = TCM_CHECK_CONDITION_NOT_READY;
-                       return -2;
+                       return -EINVAL;
                }
                goto out_invalid_cdb_field;
        }
        /*
         * Check status for SPC-3 Persistent Reservations
         */
-       if (T10_PR_OPS(su_dev)->t10_reservation_check(cmd, &pr_reg_type) != 0) {
-               if (T10_PR_OPS(su_dev)->t10_seq_non_holder(
+       if (su_dev->t10_pr.pr_ops.t10_reservation_check(cmd, &pr_reg_type) != 0) {
+               if (su_dev->t10_pr.pr_ops.t10_seq_non_holder(
                                        cmd, cdb, pr_reg_type) != 0)
                        return transport_handle_reservation_conflict(cmd);
                /*
@@ -3160,7 +2903,7 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_6;
-               T_TASK(cmd)->t_task_lba = transport_lba_21(cdb);
+               cmd->t_task_lba = transport_lba_21(cdb);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case READ_10:
@@ -3169,7 +2912,7 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_10;
-               T_TASK(cmd)->t_task_lba = transport_lba_32(cdb);
+               cmd->t_task_lba = transport_lba_32(cdb);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case READ_12:
@@ -3178,7 +2921,7 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_12;
-               T_TASK(cmd)->t_task_lba = transport_lba_32(cdb);
+               cmd->t_task_lba = transport_lba_32(cdb);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case READ_16:
@@ -3187,7 +2930,7 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_16;
-               T_TASK(cmd)->t_task_lba = transport_lba_64(cdb);
+               cmd->t_task_lba = transport_lba_64(cdb);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case WRITE_6:
@@ -3196,7 +2939,7 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_6;
-               T_TASK(cmd)->t_task_lba = transport_lba_21(cdb);
+               cmd->t_task_lba = transport_lba_21(cdb);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case WRITE_10:
@@ -3205,8 +2948,8 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_10;
-               T_TASK(cmd)->t_task_lba = transport_lba_32(cdb);
-               T_TASK(cmd)->t_tasks_fua = (cdb[1] & 0x8);
+               cmd->t_task_lba = transport_lba_32(cdb);
+               cmd->t_tasks_fua = (cdb[1] & 0x8);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case WRITE_12:
@@ -3215,8 +2958,8 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_12;
-               T_TASK(cmd)->t_task_lba = transport_lba_32(cdb);
-               T_TASK(cmd)->t_tasks_fua = (cdb[1] & 0x8);
+               cmd->t_task_lba = transport_lba_32(cdb);
+               cmd->t_tasks_fua = (cdb[1] & 0x8);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case WRITE_16:
@@ -3225,22 +2968,22 @@ static int transport_generic_cmd_sequencer(
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_16;
-               T_TASK(cmd)->t_task_lba = transport_lba_64(cdb);
-               T_TASK(cmd)->t_tasks_fua = (cdb[1] & 0x8);
+               cmd->t_task_lba = transport_lba_64(cdb);
+               cmd->t_tasks_fua = (cdb[1] & 0x8);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
                break;
        case XDWRITEREAD_10:
                if ((cmd->data_direction != DMA_TO_DEVICE) ||
-                   !(T_TASK(cmd)->t_tasks_bidi))
+                   !(cmd->t_tasks_bidi))
                        goto out_invalid_cdb_field;
                sectors = transport_get_sectors_10(cdb, cmd, &sector_ret);
                if (sector_ret)
                        goto out_unsupported_cdb;
                size = transport_get_size(sectors, cdb, cmd);
                cmd->transport_split_cdb = &split_cdb_XX_10;
-               T_TASK(cmd)->t_task_lba = transport_lba_32(cdb);
+               cmd->t_task_lba = transport_lba_32(cdb);
                cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
-               passthrough = (TRANSPORT(dev)->transport_type ==
+               passthrough = (dev->transport->transport_type ==
                                TRANSPORT_PLUGIN_PHBA_PDEV);
                /*
                 * Skip the remaining assignments for TCM/PSCSI passthrough
@@ -3251,7 +2994,7 @@ static int transport_generic_cmd_sequencer(
                 * Setup BIDI XOR callback to be run during transport_generic_complete_ok()
                 */
                cmd->transport_complete_callback = &transport_xor_callback;
-               T_TASK(cmd)->t_tasks_fua = (cdb[1] & 0x8);
+               cmd->t_tasks_fua = (cdb[1] & 0x8);
                break;
        case VARIABLE_LENGTH_CMD:
                service_action = get_unaligned_be16(&cdb[8]);
@@ -3259,7 +3002,7 @@ static int transport_generic_cmd_sequencer(
                 * Determine if this is TCM/PSCSI device and we should disable
                 * internal emulation for this CDB.
                 */
-               passthrough = (TRANSPORT(dev)->transport_type ==
+               passthrough = (dev->transport->transport_type ==
                                        TRANSPORT_PLUGIN_PHBA_PDEV);
 
                switch (service_action) {
@@ -3273,7 +3016,7 @@ static int transport_generic_cmd_sequencer(
                         * XDWRITE_READ_32 logic.
                         */
                        cmd->transport_split_cdb = &split_cdb_XX_32;
-                       T_TASK(cmd)->t_task_lba = transport_lba_64_ext(cdb);
+                       cmd->t_task_lba = transport_lba_64_ext(cdb);
                        cmd->se_cmd_flags |= SCF_SCSI_DATA_SG_IO_CDB;
 
                        /*
@@ -3287,14 +3030,22 @@ static int transport_generic_cmd_sequencer(
                         * transport_generic_complete_ok()
                         */
                        cmd->transport_complete_callback = &transport_xor_callback;
-                       T_TASK(cmd)->t_tasks_fua = (cdb[10] & 0x8);
+                       cmd->t_tasks_fua = (cdb[10] & 0x8);
                        break;
                case WRITE_SAME_32:
                        sectors = transport_get_sectors_32(cdb, cmd, &sector_ret);
                        if (sector_ret)
                                goto out_unsupported_cdb;
-                       size = transport_get_size(sectors, cdb, cmd);
-                       T_TASK(cmd)->t_task_lba = get_unaligned_be64(&cdb[12]);
+
+                       if (sectors)
+                               size = transport_get_size(sectors, cdb, cmd);
+                       else {
+                               pr_err("WSNZ=1, WRITE_SAME w/sectors=0 not"
+                                      " supported\n");
+                               goto out_invalid_cdb_field;
+                       }
+
+                       cmd->t_task_lba = get_unaligned_be64(&cdb[12]);
                        cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
 
                        /*
@@ -3304,7 +3055,7 @@ static int transport_generic_cmd_sequencer(
                                break;
 
                        if ((cdb[10] & 0x04) || (cdb[10] & 0x02)) {
-                               printk(KERN_ERR "WRITE_SAME PBDATA and LBDATA"
+                               pr_err("WRITE_SAME PBDATA and LBDATA"
                                        " bits not supported for Block Discard"
                                        " Emulation\n");
                                goto out_invalid_cdb_field;
@@ -3314,28 +3065,28 @@ static int transport_generic_cmd_sequencer(
                         * tpws with the UNMAP=1 bit set.
                         */
                        if (!(cdb[10] & 0x08)) {
-                               printk(KERN_ERR "WRITE_SAME w/o UNMAP bit not"
+                               pr_err("WRITE_SAME w/o UNMAP bit not"
                                        " supported for Block Discard Emulation\n");
                                goto out_invalid_cdb_field;
                        }
                        break;
                default:
-                       printk(KERN_ERR "VARIABLE_LENGTH_CMD service action"
+                       pr_err("VARIABLE_LENGTH_CMD service action"
                                " 0x%04x not supported\n", service_action);
                        goto out_unsupported_cdb;
                }
                break;
-       case 0xa3:
-               if (TRANSPORT(dev)->get_device_type(dev) != TYPE_ROM) {
+       case MAINTENANCE_IN:
+               if (dev->transport->get_device_type(dev) != TYPE_ROM) {
                        /* MAINTENANCE_IN from SCC-2 */
                        /*
                         * Check for emulated MI_REPORT_TARGET_PGS.
                         */
                        if (cdb[1] == MI_REPORT_TARGET_PGS) {
                                cmd->transport_emulate_cdb =
-                               (T10_ALUA(su_dev)->alua_type ==
+                               (su_dev->t10_alua.alua_type ==
                                 SPC3_ALUA_EMULATED) ?
-                               &core_emulate_report_target_port_groups :
+                               core_emulate_report_target_port_groups :
                                NULL;
                        }
                        size = (cdb[6] << 24) | (cdb[7] << 16) |
@@ -3344,7 +3095,7 @@ static int transport_generic_cmd_sequencer(
                        /* GPCMD_SEND_KEY from multi media commands */
                        size = (cdb[8] << 8) + cdb[9];
                }
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case MODE_SELECT:
                size = cdb[4];
@@ -3356,7 +3107,7 @@ static int transport_generic_cmd_sequencer(
                break;
        case MODE_SENSE:
                size = cdb[4];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case MODE_SENSE_10:
        case GPCMD_READ_BUFFER_CAPACITY:
@@ -3364,11 +3115,11 @@ static int transport_generic_cmd_sequencer(
        case LOG_SELECT:
        case LOG_SENSE:
                size = (cdb[7] << 8) + cdb[8];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case READ_BLOCK_LIMITS:
                size = READ_BLOCK_LEN;
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case GPCMD_GET_CONFIGURATION:
        case GPCMD_READ_FORMAT_CAPACITIES:
@@ -3380,11 +3131,11 @@ static int transport_generic_cmd_sequencer(
        case PERSISTENT_RESERVE_IN:
        case PERSISTENT_RESERVE_OUT:
                cmd->transport_emulate_cdb =
-                       (T10_RES(su_dev)->res_type ==
+                       (su_dev->t10_pr.res_type ==
                         SPC3_PERSISTENT_RESERVATIONS) ?
-                       &core_scsi3_emulate_pr : NULL;
+                       core_scsi3_emulate_pr : NULL;
                size = (cdb[7] << 8) + cdb[8];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case GPCMD_MECHANISM_STATUS:
        case GPCMD_READ_DVD_STRUCTURE:
@@ -3393,19 +3144,19 @@ static int transport_generic_cmd_sequencer(
                break;
        case READ_POSITION:
                size = READ_POSITION_LEN;
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
-       case 0xa4:
-               if (TRANSPORT(dev)->get_device_type(dev) != TYPE_ROM) {
+       case MAINTENANCE_OUT:
+               if (dev->transport->get_device_type(dev) != TYPE_ROM) {
                        /* MAINTENANCE_OUT from SCC-2
                         *
                         * Check for emulated MO_SET_TARGET_PGS.
                         */
                        if (cdb[1] == MO_SET_TARGET_PGS) {
                                cmd->transport_emulate_cdb =
-                               (T10_ALUA(su_dev)->alua_type ==
+                               (su_dev->t10_alua.alua_type ==
                                        SPC3_ALUA_EMULATED) ?
-                               &core_emulate_set_target_port_groups :
+                               core_emulate_set_target_port_groups :
                                NULL;
                        }
 
@@ -3415,7 +3166,7 @@ static int transport_generic_cmd_sequencer(
                        /* GPCMD_REPORT_KEY from multi media commands */
                        size = (cdb[8] << 8) + cdb[9];
                }
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case INQUIRY:
                size = (cdb[3] << 8) + cdb[4];
@@ -3423,23 +3174,23 @@ static int transport_generic_cmd_sequencer(
                 * Do implict HEAD_OF_QUEUE processing for INQUIRY.
                 * See spc4r17 section 5.3
                 */
-               if (SE_DEV(cmd)->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
+               if (cmd->se_dev->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
                        cmd->sam_task_attr = MSG_HEAD_TAG;
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case READ_BUFFER:
                size = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case READ_CAPACITY:
                size = READ_CAP_LEN;
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case READ_MEDIA_SERIAL_NUMBER:
        case SECURITY_PROTOCOL_IN:
        case SECURITY_PROTOCOL_OUT:
                size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case SERVICE_ACTION_IN:
        case ACCESS_CONTROL_IN:
@@ -3450,36 +3201,36 @@ static int transport_generic_cmd_sequencer(
        case WRITE_ATTRIBUTE:
                size = (cdb[10] << 24) | (cdb[11] << 16) |
                       (cdb[12] << 8) | cdb[13];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case RECEIVE_DIAGNOSTIC:
        case SEND_DIAGNOSTIC:
                size = (cdb[3] << 8) | cdb[4];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
 /* #warning FIXME: Figure out correct GPCMD_READ_CD blocksize. */
 #if 0
        case GPCMD_READ_CD:
                sectors = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
                size = (2336 * sectors);
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
 #endif
        case READ_TOC:
                size = cdb[8];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case REQUEST_SENSE:
                size = cdb[4];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case READ_ELEMENT_STATUS:
                size = 65536 * cdb[7] + 256 * cdb[8] + cdb[9];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case WRITE_BUFFER:
                size = (cdb[6] << 16) + (cdb[7] << 8) + cdb[8];
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case RESERVE:
        case RESERVE_10:
@@ -3500,9 +3251,9 @@ static int transport_generic_cmd_sequencer(
                 * emulation disabled.
                 */
                cmd->transport_emulate_cdb =
-                               (T10_RES(su_dev)->res_type !=
+                               (su_dev->t10_pr.res_type !=
                                 SPC_PASSTHROUGH) ?
-                               &core_scsi2_emulate_crh : NULL;
+                               core_scsi2_emulate_crh : NULL;
                cmd->se_cmd_flags |= SCF_SCSI_NON_DATA_CDB;
                break;
        case RELEASE:
@@ -3517,9 +3268,9 @@ static int transport_generic_cmd_sequencer(
                        size = cmd->data_length;
 
                cmd->transport_emulate_cdb =
-                               (T10_RES(su_dev)->res_type !=
+                               (su_dev->t10_pr.res_type !=
                                 SPC_PASSTHROUGH) ?
-                               &core_scsi2_emulate_crh : NULL;
+                               core_scsi2_emulate_crh : NULL;
                cmd->se_cmd_flags |= SCF_SCSI_NON_DATA_CDB;
                break;
        case SYNCHRONIZE_CACHE:
@@ -3529,10 +3280,10 @@ static int transport_generic_cmd_sequencer(
                 */
                if (cdb[0] == SYNCHRONIZE_CACHE) {
                        sectors = transport_get_sectors_10(cdb, cmd, &sector_ret);
-                       T_TASK(cmd)->t_task_lba = transport_lba_32(cdb);
+                       cmd->t_task_lba = transport_lba_32(cdb);
                } else {
                        sectors = transport_get_sectors_16(cdb, cmd, &sector_ret);
-                       T_TASK(cmd)->t_task_lba = transport_lba_64(cdb);
+                       cmd->t_task_lba = transport_lba_64(cdb);
                }
                if (sector_ret)
                        goto out_unsupported_cdb;
@@ -3543,7 +3294,7 @@ static int transport_generic_cmd_sequencer(
                /*
                 * For TCM/pSCSI passthrough, skip cmd->transport_emulate_cdb()
                 */
-               if (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV)
+               if (dev->transport->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV)
                        break;
                /*
                 * Set SCF_EMULATE_CDB_ASYNC to ensure asynchronous operation
@@ -3554,32 +3305,27 @@ static int transport_generic_cmd_sequencer(
                 * Check to ensure that LBA + Range does not exceed past end of
                 * device.
                 */
-               if (transport_get_sectors(cmd) < 0)
+               if (!transport_cmd_get_valid_sectors(cmd))
                        goto out_invalid_cdb_field;
                break;
        case UNMAP:
                size = get_unaligned_be16(&cdb[7]);
-               passthrough = (TRANSPORT(dev)->transport_type ==
-                               TRANSPORT_PLUGIN_PHBA_PDEV);
-               /*
-                * Determine if the received UNMAP used to for direct passthrough
-                * into Linux/SCSI with struct request via TCM/pSCSI or we are
-                * signaling the use of internal transport_generic_unmap() emulation
-                * for UNMAP -> Linux/BLOCK disbard with TCM/IBLOCK and TCM/FILEIO
-                * subsystem plugin backstores.
-                */
-               if (!(passthrough))
-                       cmd->se_cmd_flags |= SCF_EMULATE_SYNC_UNMAP;
-
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        case WRITE_SAME_16:
                sectors = transport_get_sectors_16(cdb, cmd, &sector_ret);
                if (sector_ret)
                        goto out_unsupported_cdb;
-               size = transport_get_size(sectors, cdb, cmd);
-               T_TASK(cmd)->t_task_lba = get_unaligned_be16(&cdb[2]);
-               passthrough = (TRANSPORT(dev)->transport_type ==
+
+               if (sectors)
+                       size = transport_get_size(sectors, cdb, cmd);
+               else {
+                       pr_err("WSNZ=1, WRITE_SAME w/sectors=0 not supported\n");
+                       goto out_invalid_cdb_field;
+               }
+
+               cmd->t_task_lba = get_unaligned_be16(&cdb[2]);
+               passthrough = (dev->transport->transport_type ==
                                TRANSPORT_PLUGIN_PHBA_PDEV);
                /*
                 * Determine if the received WRITE_SAME_16 is used to for direct
@@ -3588,9 +3334,9 @@ static int transport_generic_cmd_sequencer(
                 * emulation for -> Linux/BLOCK disbard with TCM/IBLOCK and
                 * TCM/FILEIO subsystem plugin backstores.
                 */
-               if (!(passthrough)) {
+               if (!passthrough) {
                        if ((cdb[1] & 0x04) || (cdb[1] & 0x02)) {
-                               printk(KERN_ERR "WRITE_SAME PBDATA and LBDATA"
+                               pr_err("WRITE_SAME PBDATA and LBDATA"
                                        " bits not supported for Block Discard"
                                        " Emulation\n");
                                goto out_invalid_cdb_field;
@@ -3600,7 +3346,7 @@ static int transport_generic_cmd_sequencer(
                         * tpws with the UNMAP=1 bit set.
                         */
                        if (!(cdb[1] & 0x08)) {
-                               printk(KERN_ERR "WRITE_SAME w/o UNMAP bit not "
+                               pr_err("WRITE_SAME w/o UNMAP bit not "
                                        " supported for Block Discard Emulation\n");
                                goto out_invalid_cdb_field;
                        }
@@ -3625,34 +3371,34 @@ static int transport_generic_cmd_sequencer(
                break;
        case REPORT_LUNS:
                cmd->transport_emulate_cdb =
-                               &transport_core_report_lun_response;
+                               transport_core_report_lun_response;
                size = (cdb[6] << 24) | (cdb[7] << 16) | (cdb[8] << 8) | cdb[9];
                /*
                 * Do implict HEAD_OF_QUEUE processing for REPORT_LUNS
                 * See spc4r17 section 5.3
                 */
-               if (SE_DEV(cmd)->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
+               if (cmd->se_dev->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
                        cmd->sam_task_attr = MSG_HEAD_TAG;
-               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_NONSG_IO_CDB;
+               cmd->se_cmd_flags |= SCF_SCSI_CONTROL_SG_IO_CDB;
                break;
        default:
-               printk(KERN_WARNING "TARGET_CORE[%s]: Unsupported SCSI Opcode"
+               pr_warn("TARGET_CORE[%s]: Unsupported SCSI Opcode"
                        " 0x%02x, sending CHECK_CONDITION.\n",
-                       CMD_TFO(cmd)->get_fabric_name(), cdb[0]);
+                       cmd->se_tfo->get_fabric_name(), cdb[0]);
                cmd->transport_wait_for_tasks = &transport_nop_wait_for_tasks;
                goto out_unsupported_cdb;
        }
 
        if (size != cmd->data_length) {
-               printk(KERN_WARNING "TARGET_CORE[%s]: Expected Transfer Length:"
+               pr_warn("TARGET_CORE[%s]: Expected Transfer Length:"
                        " %u does not match SCSI CDB Length: %u for SAM Opcode:"
-                       " 0x%02x\n", CMD_TFO(cmd)->get_fabric_name(),
+                       " 0x%02x\n", cmd->se_tfo->get_fabric_name(),
                                cmd->data_length, size, cdb[0]);
 
                cmd->cmd_spdtl = size;
 
                if (cmd->data_direction == DMA_TO_DEVICE) {
-                       printk(KERN_ERR "Rejecting underflow/overflow"
+                       pr_err("Rejecting underflow/overflow"
                                        " WRITE data\n");
                        goto out_invalid_cdb_field;
                }
@@ -3660,10 +3406,10 @@ static int transport_generic_cmd_sequencer(
                 * Reject READ_* or WRITE_* with overflow/underflow for
                 * type SCF_SCSI_DATA_SG_IO_CDB.
                 */
-               if (!(ret) && (DEV_ATTRIB(dev)->block_size != 512))  {
-                       printk(KERN_ERR "Failing OVERFLOW/UNDERFLOW for LBA op"
+               if (!ret && (dev->se_sub_dev->se_dev_attrib.block_size != 512))  {
+                       pr_err("Failing OVERFLOW/UNDERFLOW for LBA op"
                                " CDB on non 512-byte sector setup subsystem"
-                               " plugin: %s\n", TRANSPORT(dev)->name);
+                               " plugin: %s\n", dev->transport->name);
                        /* Returns CHECK_CONDITION + INVALID_CDB_FIELD */
                        goto out_invalid_cdb_field;
                }
@@ -3678,105 +3424,22 @@ static int transport_generic_cmd_sequencer(
                cmd->data_length = size;
        }
 
+       /* Let's limit control cdbs to a page, for simplicity's sake. */
+       if ((cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB) &&
+           size > PAGE_SIZE)
+               goto out_invalid_cdb_field;
+
        transport_set_supported_SAM_opcode(cmd);
        return ret;
 
 out_unsupported_cdb:
        cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
        cmd->scsi_sense_reason = TCM_UNSUPPORTED_SCSI_OPCODE;
-       return -2;
+       return -EINVAL;
 out_invalid_cdb_field:
        cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
        cmd->scsi_sense_reason = TCM_INVALID_CDB_FIELD;
-       return -2;
-}
-
-static inline void transport_release_tasks(struct se_cmd *);
-
-/*
- * This function will copy a contiguous *src buffer into a destination
- * struct scatterlist array.
- */
-static void transport_memcpy_write_contig(
-       struct se_cmd *cmd,
-       struct scatterlist *sg_d,
-       unsigned char *src)
-{
-       u32 i = 0, length = 0, total_length = cmd->data_length;
-       void *dst;
-
-       while (total_length) {
-               length = sg_d[i].length;
-
-               if (length > total_length)
-                       length = total_length;
-
-               dst = sg_virt(&sg_d[i]);
-
-               memcpy(dst, src, length);
-
-               if (!(total_length -= length))
-                       return;
-
-               src += length;
-               i++;
-       }
-}
-
-/*
- * This function will copy a struct scatterlist array *sg_s into a destination
- * contiguous *dst buffer.
- */
-static void transport_memcpy_read_contig(
-       struct se_cmd *cmd,
-       unsigned char *dst,
-       struct scatterlist *sg_s)
-{
-       u32 i = 0, length = 0, total_length = cmd->data_length;
-       void *src;
-
-       while (total_length) {
-               length = sg_s[i].length;
-
-               if (length > total_length)
-                       length = total_length;
-
-               src = sg_virt(&sg_s[i]);
-
-               memcpy(dst, src, length);
-
-               if (!(total_length -= length))
-                       return;
-
-               dst += length;
-               i++;
-       }
-}
-
-static void transport_memcpy_se_mem_read_contig(
-       struct se_cmd *cmd,
-       unsigned char *dst,
-       struct list_head *se_mem_list)
-{
-       struct se_mem *se_mem;
-       void *src;
-       u32 length = 0, total_length = cmd->data_length;
-
-       list_for_each_entry(se_mem, se_mem_list, se_list) {
-               length = se_mem->se_len;
-
-               if (length > total_length)
-                       length = total_length;
-
-               src = page_address(se_mem->se_page) + se_mem->se_off;
-
-               memcpy(dst, src, length);
-
-               if (!(total_length -= length))
-                       return;
-
-               dst += length;
-       }
+       return -EINVAL;
 }
 
 /*
@@ -3786,7 +3449,7 @@ static void transport_memcpy_se_mem_read_contig(
  */
 static void transport_complete_task_attr(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct se_cmd *cmd_p, *cmd_tmp;
        int new_active_tasks = 0;
 
@@ -3794,25 +3457,25 @@ static void transport_complete_task_attr(struct se_cmd *cmd)
                atomic_dec(&dev->simple_cmds);
                smp_mb__after_atomic_dec();
                dev->dev_cur_ordered_id++;
-               DEBUG_STA("Incremented dev->dev_cur_ordered_id: %u for"
+               pr_debug("Incremented dev->dev_cur_ordered_id: %u for"
                        " SIMPLE: %u\n", dev->dev_cur_ordered_id,
                        cmd->se_ordered_id);
        } else if (cmd->sam_task_attr == MSG_HEAD_TAG) {
                atomic_dec(&dev->dev_hoq_count);
                smp_mb__after_atomic_dec();
                dev->dev_cur_ordered_id++;
-               DEBUG_STA("Incremented dev_cur_ordered_id: %u for"
+               pr_debug("Incremented dev_cur_ordered_id: %u for"
                        " HEAD_OF_QUEUE: %u\n", dev->dev_cur_ordered_id,
                        cmd->se_ordered_id);
        } else if (cmd->sam_task_attr == MSG_ORDERED_TAG) {
                spin_lock(&dev->ordered_cmd_lock);
-               list_del(&cmd->se_ordered_list);
+               list_del(&cmd->se_ordered_node);
                atomic_dec(&dev->dev_ordered_sync);
                smp_mb__after_atomic_dec();
                spin_unlock(&dev->ordered_cmd_lock);
 
                dev->dev_cur_ordered_id++;
-               DEBUG_STA("Incremented dev_cur_ordered_id: %u for ORDERED:"
+               pr_debug("Incremented dev_cur_ordered_id: %u for ORDERED:"
                        " %u\n", dev->dev_cur_ordered_id, cmd->se_ordered_id);
        }
        /*
@@ -3822,15 +3485,15 @@ static void transport_complete_task_attr(struct se_cmd *cmd)
         */
        spin_lock(&dev->delayed_cmd_lock);
        list_for_each_entry_safe(cmd_p, cmd_tmp,
-                       &dev->delayed_cmd_list, se_delayed_list) {
+                       &dev->delayed_cmd_list, se_delayed_node) {
 
-               list_del(&cmd_p->se_delayed_list);
+               list_del(&cmd_p->se_delayed_node);
                spin_unlock(&dev->delayed_cmd_lock);
 
-               DEBUG_STA("Calling add_tasks() for"
+               pr_debug("Calling add_tasks() for"
                        " cmd_p: 0x%02x Task Attr: 0x%02x"
                        " Dormant -> Active, se_ordered_id: %u\n",
-                       T_TASK(cmd_p)->t_task_cdb[0],
+                       cmd_p->t_task_cdb[0],
                        cmd_p->sam_task_attr, cmd_p->se_ordered_id);
 
                transport_add_tasks_from_cmd(cmd_p);
@@ -3846,19 +3509,78 @@ static void transport_complete_task_attr(struct se_cmd *cmd)
         * to do the processing of the Active tasks.
         */
        if (new_active_tasks != 0)
-               wake_up_interruptible(&dev->dev_queue_obj->thread_wq);
+               wake_up_interruptible(&dev->dev_queue_obj.thread_wq);
+}
+
+static int transport_complete_qf(struct se_cmd *cmd)
+{
+       int ret = 0;
+
+       if (cmd->se_cmd_flags & SCF_TRANSPORT_TASK_SENSE)
+               return cmd->se_tfo->queue_status(cmd);
+
+       switch (cmd->data_direction) {
+       case DMA_FROM_DEVICE:
+               ret = cmd->se_tfo->queue_data_in(cmd);
+               break;
+       case DMA_TO_DEVICE:
+               if (cmd->t_bidi_data_sg) {
+                       ret = cmd->se_tfo->queue_data_in(cmd);
+                       if (ret < 0)
+                               return ret;
+               }
+               /* Fall through for DMA_TO_DEVICE */
+       case DMA_NONE:
+               ret = cmd->se_tfo->queue_status(cmd);
+               break;
+       default:
+               break;
+       }
+
+       return ret;
+}
+
+static void transport_handle_queue_full(
+       struct se_cmd *cmd,
+       struct se_device *dev,
+       int (*qf_callback)(struct se_cmd *))
+{
+       spin_lock_irq(&dev->qf_cmd_lock);
+       cmd->se_cmd_flags |= SCF_EMULATE_QUEUE_FULL;
+       cmd->transport_qf_callback = qf_callback;
+       list_add_tail(&cmd->se_qf_node, &cmd->se_dev->qf_cmd_list);
+       atomic_inc(&dev->dev_qf_count);
+       smp_mb__after_atomic_inc();
+       spin_unlock_irq(&cmd->se_dev->qf_cmd_lock);
+
+       schedule_work(&cmd->se_dev->qf_work_queue);
 }
 
 static void transport_generic_complete_ok(struct se_cmd *cmd)
 {
-       int reason = 0;
+       int reason = 0, ret;
        /*
         * Check if we need to move delayed/dormant tasks from cmds on the
         * delayed execution list after a HEAD_OF_QUEUE or ORDERED Task
         * Attribute.
         */
-       if (SE_DEV(cmd)->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
+       if (cmd->se_dev->dev_task_attr_type == SAM_TASK_ATTR_EMULATED)
                transport_complete_task_attr(cmd);
+       /*
+        * Check to schedule QUEUE_FULL work, or execute an existing
+        * cmd->transport_qf_callback()
+        */
+       if (atomic_read(&cmd->se_dev->dev_qf_count) != 0)
+               schedule_work(&cmd->se_dev->qf_work_queue);
+
+       if (cmd->transport_qf_callback) {
+               ret = cmd->transport_qf_callback(cmd);
+               if (ret < 0)
+                       goto queue_full;
+
+               cmd->transport_qf_callback = NULL;
+               goto done;
+       }
        /*
         * Check if we need to retrieve a sense buffer from
         * the struct se_cmd in question.
@@ -3872,8 +3594,11 @@ static void transport_generic_complete_ok(struct se_cmd *cmd)
                 * a non GOOD status.
                 */
                if (cmd->scsi_status) {
-                       transport_send_check_condition_and_sense(
+                       ret = transport_send_check_condition_and_sense(
                                        cmd, reason, 1);
+                       if (ret == -EAGAIN)
+                               goto queue_full;
+
                        transport_lun_remove_cmd(cmd);
                        transport_cmd_check_stop_to_fabric(cmd);
                        return;
@@ -3889,53 +3614,57 @@ static void transport_generic_complete_ok(struct se_cmd *cmd)
        switch (cmd->data_direction) {
        case DMA_FROM_DEVICE:
                spin_lock(&cmd->se_lun->lun_sep_lock);
-               if (SE_LUN(cmd)->lun_sep) {
-                       SE_LUN(cmd)->lun_sep->sep_stats.tx_data_octets +=
+               if (cmd->se_lun->lun_sep) {
+                       cmd->se_lun->lun_sep->sep_stats.tx_data_octets +=
                                        cmd->data_length;
                }
                spin_unlock(&cmd->se_lun->lun_sep_lock);
-               /*
-                * If enabled by TCM fabirc module pre-registered SGL
-                * memory, perform the memcpy() from the TCM internal
-                * contigious buffer back to the original SGL.
-                */
-               if (cmd->se_cmd_flags & SCF_PASSTHROUGH_CONTIG_TO_SG)
-                       transport_memcpy_write_contig(cmd,
-                                T_TASK(cmd)->t_task_pt_sgl,
-                                T_TASK(cmd)->t_task_buf);
 
-               CMD_TFO(cmd)->queue_data_in(cmd);
+               ret = cmd->se_tfo->queue_data_in(cmd);
+               if (ret == -EAGAIN)
+                       goto queue_full;
                break;
        case DMA_TO_DEVICE:
                spin_lock(&cmd->se_lun->lun_sep_lock);
-               if (SE_LUN(cmd)->lun_sep) {
-                       SE_LUN(cmd)->lun_sep->sep_stats.rx_data_octets +=
+               if (cmd->se_lun->lun_sep) {
+                       cmd->se_lun->lun_sep->sep_stats.rx_data_octets +=
                                cmd->data_length;
                }
                spin_unlock(&cmd->se_lun->lun_sep_lock);
                /*
                 * Check if we need to send READ payload for BIDI-COMMAND
                 */
-               if (T_TASK(cmd)->t_mem_bidi_list != NULL) {
+               if (cmd->t_bidi_data_sg) {
                        spin_lock(&cmd->se_lun->lun_sep_lock);
-                       if (SE_LUN(cmd)->lun_sep) {
-                               SE_LUN(cmd)->lun_sep->sep_stats.tx_data_octets +=
+                       if (cmd->se_lun->lun_sep) {
+                               cmd->se_lun->lun_sep->sep_stats.tx_data_octets +=
                                        cmd->data_length;
                        }
                        spin_unlock(&cmd->se_lun->lun_sep_lock);
-                       CMD_TFO(cmd)->queue_data_in(cmd);
+                       ret = cmd->se_tfo->queue_data_in(cmd);
+                       if (ret == -EAGAIN)
+                               goto queue_full;
                        break;
                }
                /* Fall through for DMA_TO_DEVICE */
        case DMA_NONE:
-               CMD_TFO(cmd)->queue_status(cmd);
+               ret = cmd->se_tfo->queue_status(cmd);
+               if (ret == -EAGAIN)
+                       goto queue_full;
                break;
        default:
                break;
        }
 
+done:
        transport_lun_remove_cmd(cmd);
        transport_cmd_check_stop_to_fabric(cmd);
+       return;
+
+queue_full:
+       pr_debug("Handling complete_ok QUEUE_FULL: se_cmd: %p,"
+               " data_direction: %d\n", cmd, cmd->data_direction);
+       transport_handle_queue_full(cmd, cmd->se_dev, transport_complete_qf);
 }
 
 static void transport_free_dev_tasks(struct se_cmd *cmd)
@@ -3943,9 +3672,9 @@ static void transport_free_dev_tasks(struct se_cmd *cmd)
        struct se_task *task, *task_tmp;
        unsigned long flags;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        list_for_each_entry_safe(task, task_tmp,
-                               &T_TASK(cmd)->t_task_list, t_list) {
+                               &cmd->t_task_list, t_list) {
                if (atomic_read(&task->task_active))
                        continue;
 
@@ -3954,75 +3683,40 @@ static void transport_free_dev_tasks(struct se_cmd *cmd)
 
                list_del(&task->t_list);
 
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                if (task->se_dev)
-                       TRANSPORT(task->se_dev)->free_task(task);
+                       task->se_dev->transport->free_task(task);
                else
-                       printk(KERN_ERR "task[%u] - task->se_dev is NULL\n",
+                       pr_err("task[%u] - task->se_dev is NULL\n",
                                task->task_no);
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+               spin_lock_irqsave(&cmd->t_state_lock, flags);
        }
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 }
 
-static inline void transport_free_pages(struct se_cmd *cmd)
+static inline void transport_free_sgl(struct scatterlist *sgl, int nents)
 {
-       struct se_mem *se_mem, *se_mem_tmp;
-       int free_page = 1;
-
-       if (cmd->se_cmd_flags & SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC)
-               free_page = 0;
-       if (cmd->se_dev->transport->do_se_mem_map)
-               free_page = 0;
+       struct scatterlist *sg;
+       int count;
 
-       if (T_TASK(cmd)->t_task_buf) {
-               kfree(T_TASK(cmd)->t_task_buf);
-               T_TASK(cmd)->t_task_buf = NULL;
-               return;
-       }
+       for_each_sg(sgl, sg, nents, count)
+               __free_page(sg_page(sg));
 
-       /*
-        * Caller will handle releasing of struct se_mem.
-        */
-       if (cmd->se_cmd_flags & SCF_CMD_PASSTHROUGH_NOALLOC)
-               return;
+       kfree(sgl);
+}
 
-       if (!(T_TASK(cmd)->t_tasks_se_num))
+static inline void transport_free_pages(struct se_cmd *cmd)
+{
+       if (cmd->se_cmd_flags & SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC)
                return;
 
-       list_for_each_entry_safe(se_mem, se_mem_tmp,
-                       T_TASK(cmd)->t_mem_list, se_list) {
-               /*
-                * We only release call __free_page(struct se_mem->se_page) when
-                * SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC is NOT in use,
-                */
-               if (free_page)
-                       __free_page(se_mem->se_page);
-
-               list_del(&se_mem->se_list);
-               kmem_cache_free(se_mem_cache, se_mem);
-       }
-
-       if (T_TASK(cmd)->t_mem_bidi_list && T_TASK(cmd)->t_tasks_se_bidi_num) {
-               list_for_each_entry_safe(se_mem, se_mem_tmp,
-                               T_TASK(cmd)->t_mem_bidi_list, se_list) {
-                       /*
-                        * We only release call __free_page(struct se_mem->se_page) when
-                        * SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC is NOT in use,
-                        */
-                       if (free_page)
-                               __free_page(se_mem->se_page);
-
-                       list_del(&se_mem->se_list);
-                       kmem_cache_free(se_mem_cache, se_mem);
-               }
-       }
+       transport_free_sgl(cmd->t_data_sg, cmd->t_data_nents);
+       cmd->t_data_sg = NULL;
+       cmd->t_data_nents = 0;
 
-       kfree(T_TASK(cmd)->t_mem_bidi_list);
-       T_TASK(cmd)->t_mem_bidi_list = NULL;
-       kfree(T_TASK(cmd)->t_mem_list);
-       T_TASK(cmd)->t_mem_list = NULL;
-       T_TASK(cmd)->t_tasks_se_num = 0;
+       transport_free_sgl(cmd->t_bidi_data_sg, cmd->t_bidi_data_nents);
+       cmd->t_bidi_data_sg = NULL;
+       cmd->t_bidi_data_nents = 0;
 }
 
 static inline void transport_release_tasks(struct se_cmd *cmd)
@@ -4034,23 +3728,23 @@ static inline int transport_dec_and_check(struct se_cmd *cmd)
 {
        unsigned long flags;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
-               if (!(atomic_dec_and_test(&T_TASK(cmd)->t_fe_count))) {
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       if (atomic_read(&cmd->t_fe_count)) {
+               if (!atomic_dec_and_test(&cmd->t_fe_count)) {
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                        flags);
                        return 1;
                }
        }
 
-       if (atomic_read(&T_TASK(cmd)->t_se_count)) {
-               if (!(atomic_dec_and_test(&T_TASK(cmd)->t_se_count))) {
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+       if (atomic_read(&cmd->t_se_count)) {
+               if (!atomic_dec_and_test(&cmd->t_se_count)) {
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                        flags);
                        return 1;
                }
        }
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        return 0;
 }
@@ -4062,68 +3756,57 @@ static void transport_release_fe_cmd(struct se_cmd *cmd)
        if (transport_dec_and_check(cmd))
                return;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       if (!(atomic_read(&T_TASK(cmd)->transport_dev_active))) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       if (!atomic_read(&cmd->transport_dev_active)) {
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                goto free_pages;
        }
-       atomic_set(&T_TASK(cmd)->transport_dev_active, 0);
+       atomic_set(&cmd->transport_dev_active, 0);
        transport_all_task_dev_remove_state(cmd);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        transport_release_tasks(cmd);
 free_pages:
        transport_free_pages(cmd);
        transport_free_se_cmd(cmd);
-       CMD_TFO(cmd)->release_cmd_direct(cmd);
+       cmd->se_tfo->release_cmd(cmd);
 }
 
-static int transport_generic_remove(
-       struct se_cmd *cmd,
-       int release_to_pool,
-       int session_reinstatement)
+static int
+transport_generic_remove(struct se_cmd *cmd, int session_reinstatement)
 {
        unsigned long flags;
 
-       if (!(T_TASK(cmd)))
-               goto release_cmd;
-
        if (transport_dec_and_check(cmd)) {
                if (session_reinstatement) {
-                       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+                       spin_lock_irqsave(&cmd->t_state_lock, flags);
                        transport_all_task_dev_remove_state(cmd);
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                        flags);
                }
                return 1;
        }
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       if (!(atomic_read(&T_TASK(cmd)->transport_dev_active))) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       if (!atomic_read(&cmd->transport_dev_active)) {
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                goto free_pages;
        }
-       atomic_set(&T_TASK(cmd)->transport_dev_active, 0);
+       atomic_set(&cmd->transport_dev_active, 0);
        transport_all_task_dev_remove_state(cmd);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        transport_release_tasks(cmd);
+
 free_pages:
        transport_free_pages(cmd);
-
-release_cmd:
-       if (release_to_pool) {
-               transport_release_cmd_to_pool(cmd);
-       } else {
-               transport_free_se_cmd(cmd);
-               CMD_TFO(cmd)->release_cmd_direct(cmd);
-       }
-
+       transport_release_cmd(cmd);
        return 0;
 }
 
 /*
- * transport_generic_map_mem_to_cmd - Perform SGL -> struct se_mem map
+ * transport_generic_map_mem_to_cmd - Use fabric-alloced pages instead of
+ * allocating in the core.
  * @cmd:  Associated se_cmd descriptor
  * @mem:  SGL style memory for TCM WRITE / READ
  * @sg_mem_num: Number of SGL elements
@@ -4132,617 +3815,166 @@ release_cmd:
  *
  * Return: nonzero return cmd was rejected for -ENOMEM or inproper usage
  * of parameters.
- */
-int transport_generic_map_mem_to_cmd(
-       struct se_cmd *cmd,
-       struct scatterlist *mem,
-       u32 sg_mem_num,
-       struct scatterlist *mem_bidi_in,
-       u32 sg_mem_bidi_num)
-{
-       u32 se_mem_cnt_out = 0;
-       int ret;
-
-       if (!(mem) || !(sg_mem_num))
-               return 0;
-       /*
-        * Passed *mem will contain a list_head containing preformatted
-        * struct se_mem elements...
-        */
-       if (!(cmd->se_cmd_flags & SCF_PASSTHROUGH_SG_TO_MEM)) {
-               if ((mem_bidi_in) || (sg_mem_bidi_num)) {
-                       printk(KERN_ERR "SCF_CMD_PASSTHROUGH_NOALLOC not supported"
-                               " with BIDI-COMMAND\n");
-                       return -ENOSYS;
-               }
-
-               T_TASK(cmd)->t_mem_list = (struct list_head *)mem;
-               T_TASK(cmd)->t_tasks_se_num = sg_mem_num;
-               cmd->se_cmd_flags |= SCF_CMD_PASSTHROUGH_NOALLOC;
-               return 0;
-       }
-       /*
-        * Otherwise, assume the caller is passing a struct scatterlist
-        * array from include/linux/scatterlist.h
-        */
-       if ((cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB) ||
-           (cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB)) {
-               /*
-                * For CDB using TCM struct se_mem linked list scatterlist memory
-                * processed into a TCM struct se_subsystem_dev, we do the mapping
-                * from the passed physical memory to struct se_mem->se_page here.
-                */
-               T_TASK(cmd)->t_mem_list = transport_init_se_mem_list();
-               if (!(T_TASK(cmd)->t_mem_list))
-                       return -ENOMEM;
-
-               ret = transport_map_sg_to_mem(cmd,
-                       T_TASK(cmd)->t_mem_list, mem, &se_mem_cnt_out);
-               if (ret < 0)
-                       return -ENOMEM;
-
-               T_TASK(cmd)->t_tasks_se_num = se_mem_cnt_out;
-               /*
-                * Setup BIDI READ list of struct se_mem elements
-                */
-               if ((mem_bidi_in) && (sg_mem_bidi_num)) {
-                       T_TASK(cmd)->t_mem_bidi_list = transport_init_se_mem_list();
-                       if (!(T_TASK(cmd)->t_mem_bidi_list)) {
-                               kfree(T_TASK(cmd)->t_mem_list);
-                               return -ENOMEM;
-                       }
-                       se_mem_cnt_out = 0;
-
-                       ret = transport_map_sg_to_mem(cmd,
-                               T_TASK(cmd)->t_mem_bidi_list, mem_bidi_in,
-                               &se_mem_cnt_out);
-                       if (ret < 0) {
-                               kfree(T_TASK(cmd)->t_mem_list);
-                               return -ENOMEM;
-                       }
-
-                       T_TASK(cmd)->t_tasks_se_bidi_num = se_mem_cnt_out;
-               }
-               cmd->se_cmd_flags |= SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC;
-
-       } else if (cmd->se_cmd_flags & SCF_SCSI_CONTROL_NONSG_IO_CDB) {
-               if (mem_bidi_in || sg_mem_bidi_num) {
-                       printk(KERN_ERR "BIDI-Commands not supported using "
-                               "SCF_SCSI_CONTROL_NONSG_IO_CDB\n");
-                       return -ENOSYS;
-               }
-               /*
-                * For incoming CDBs using a contiguous buffer internall with TCM,
-                * save the passed struct scatterlist memory.  After TCM storage object
-                * processing has completed for this struct se_cmd, TCM core will call
-                * transport_memcpy_[write,read]_contig() as necessary from
-                * transport_generic_complete_ok() and transport_write_pending() in order
-                * to copy the TCM buffer to/from the original passed *mem in SGL ->
-                * struct scatterlist format.
-                */
-               cmd->se_cmd_flags |= SCF_PASSTHROUGH_CONTIG_TO_SG;
-               T_TASK(cmd)->t_task_pt_sgl = mem;
-       }
-
-       return 0;
-}
-EXPORT_SYMBOL(transport_generic_map_mem_to_cmd);
-
-
-static inline long long transport_dev_end_lba(struct se_device *dev)
-{
-       return dev->transport->get_blocks(dev) + 1;
-}
-
-static int transport_get_sectors(struct se_cmd *cmd)
-{
-       struct se_device *dev = SE_DEV(cmd);
-
-       T_TASK(cmd)->t_tasks_sectors =
-               (cmd->data_length / DEV_ATTRIB(dev)->block_size);
-       if (!(T_TASK(cmd)->t_tasks_sectors))
-               T_TASK(cmd)->t_tasks_sectors = 1;
-
-       if (TRANSPORT(dev)->get_device_type(dev) != TYPE_DISK)
-               return 0;
-
-       if ((T_TASK(cmd)->t_task_lba + T_TASK(cmd)->t_tasks_sectors) >
-            transport_dev_end_lba(dev)) {
-               printk(KERN_ERR "LBA: %llu Sectors: %u exceeds"
-                       " transport_dev_end_lba(): %llu\n",
-                       T_TASK(cmd)->t_task_lba, T_TASK(cmd)->t_tasks_sectors,
-                       transport_dev_end_lba(dev));
-               cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-               cmd->scsi_sense_reason = TCM_SECTOR_COUNT_TOO_MANY;
-               return PYX_TRANSPORT_REQ_TOO_MANY_SECTORS;
-       }
-
-       return 0;
-}
-
-static int transport_new_cmd_obj(struct se_cmd *cmd)
-{
-       struct se_device *dev = SE_DEV(cmd);
-       u32 task_cdbs = 0, rc;
-
-       if (!(cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB)) {
-               task_cdbs++;
-               T_TASK(cmd)->t_task_cdbs++;
-       } else {
-               int set_counts = 1;
-
-               /*
-                * Setup any BIDI READ tasks and memory from
-                * T_TASK(cmd)->t_mem_bidi_list so the READ struct se_tasks
-                * are queued first for the non pSCSI passthrough case.
-                */
-               if ((T_TASK(cmd)->t_mem_bidi_list != NULL) &&
-                   (TRANSPORT(dev)->transport_type != TRANSPORT_PLUGIN_PHBA_PDEV)) {
-                       rc = transport_generic_get_cdb_count(cmd,
-                               T_TASK(cmd)->t_task_lba,
-                               T_TASK(cmd)->t_tasks_sectors,
-                               DMA_FROM_DEVICE, T_TASK(cmd)->t_mem_bidi_list,
-                               set_counts);
-                       if (!(rc)) {
-                               cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-                               cmd->scsi_sense_reason =
-                                       TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
-                               return PYX_TRANSPORT_LU_COMM_FAILURE;
-                       }
-                       set_counts = 0;
-               }
-               /*
-                * Setup the tasks and memory from T_TASK(cmd)->t_mem_list
-                * Note for BIDI transfers this will contain the WRITE payload
-                */
-               task_cdbs = transport_generic_get_cdb_count(cmd,
-                               T_TASK(cmd)->t_task_lba,
-                               T_TASK(cmd)->t_tasks_sectors,
-                               cmd->data_direction, T_TASK(cmd)->t_mem_list,
-                               set_counts);
-               if (!(task_cdbs)) {
-                       cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
-                       cmd->scsi_sense_reason =
-                                       TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
-                       return PYX_TRANSPORT_LU_COMM_FAILURE;
-               }
-               T_TASK(cmd)->t_task_cdbs += task_cdbs;
-
-#if 0
-               printk(KERN_INFO "data_length: %u, LBA: %llu t_tasks_sectors:"
-                       " %u, t_task_cdbs: %u\n", obj_ptr, cmd->data_length,
-                       T_TASK(cmd)->t_task_lba, T_TASK(cmd)->t_tasks_sectors,
-                       T_TASK(cmd)->t_task_cdbs);
-#endif
-       }
-
-       atomic_set(&T_TASK(cmd)->t_task_cdbs_left, task_cdbs);
-       atomic_set(&T_TASK(cmd)->t_task_cdbs_ex_left, task_cdbs);
-       atomic_set(&T_TASK(cmd)->t_task_cdbs_timeout_left, task_cdbs);
-       return 0;
-}
-
-static struct list_head *transport_init_se_mem_list(void)
-{
-       struct list_head *se_mem_list;
-
-       se_mem_list = kzalloc(sizeof(struct list_head), GFP_KERNEL);
-       if (!(se_mem_list)) {
-               printk(KERN_ERR "Unable to allocate memory for se_mem_list\n");
-               return NULL;
-       }
-       INIT_LIST_HEAD(se_mem_list);
-
-       return se_mem_list;
-}
-
-static int
-transport_generic_get_mem(struct se_cmd *cmd, u32 length, u32 dma_size)
-{
-       unsigned char *buf;
-       struct se_mem *se_mem;
-
-       T_TASK(cmd)->t_mem_list = transport_init_se_mem_list();
-       if (!(T_TASK(cmd)->t_mem_list))
-               return -ENOMEM;
-
-       /*
-        * If the device uses memory mapping this is enough.
-        */
-       if (cmd->se_dev->transport->do_se_mem_map)
-               return 0;
-
-       /*
-        * Setup BIDI-COMMAND READ list of struct se_mem elements
-        */
-       if (T_TASK(cmd)->t_tasks_bidi) {
-               T_TASK(cmd)->t_mem_bidi_list = transport_init_se_mem_list();
-               if (!(T_TASK(cmd)->t_mem_bidi_list)) {
-                       kfree(T_TASK(cmd)->t_mem_list);
-                       return -ENOMEM;
-               }
-       }
-
-       while (length) {
-               se_mem = kmem_cache_zalloc(se_mem_cache, GFP_KERNEL);
-               if (!(se_mem)) {
-                       printk(KERN_ERR "Unable to allocate struct se_mem\n");
-                       goto out;
-               }
-
-/* #warning FIXME Allocate contigous pages for struct se_mem elements */
-               se_mem->se_page = alloc_pages(GFP_KERNEL, 0);
-               if (!(se_mem->se_page)) {
-                       printk(KERN_ERR "alloc_pages() failed\n");
-                       goto out;
-               }
-
-               buf = kmap_atomic(se_mem->se_page, KM_IRQ0);
-               if (!(buf)) {
-                       printk(KERN_ERR "kmap_atomic() failed\n");
-                       goto out;
-               }
-               INIT_LIST_HEAD(&se_mem->se_list);
-               se_mem->se_len = (length > dma_size) ? dma_size : length;
-               memset(buf, 0, se_mem->se_len);
-               kunmap_atomic(buf, KM_IRQ0);
-
-               list_add_tail(&se_mem->se_list, T_TASK(cmd)->t_mem_list);
-               T_TASK(cmd)->t_tasks_se_num++;
-
-               DEBUG_MEM("Allocated struct se_mem page(%p) Length(%u)"
-                       " Offset(%u)\n", se_mem->se_page, se_mem->se_len,
-                       se_mem->se_off);
-
-               length -= se_mem->se_len;
-       }
-
-       DEBUG_MEM("Allocated total struct se_mem elements(%u)\n",
-                       T_TASK(cmd)->t_tasks_se_num);
-
-       return 0;
-out:
-       if (se_mem)
-               __free_pages(se_mem->se_page, 0);
-       kmem_cache_free(se_mem_cache, se_mem);
-       return -1;
-}
-
-u32 transport_calc_sg_num(
-       struct se_task *task,
-       struct se_mem *in_se_mem,
-       u32 task_offset)
-{
-       struct se_cmd *se_cmd = task->task_se_cmd;
-       struct se_device *se_dev = SE_DEV(se_cmd);
-       struct se_mem *se_mem = in_se_mem;
-       struct target_core_fabric_ops *tfo = CMD_TFO(se_cmd);
-       u32 sg_length, task_size = task->task_size, task_sg_num_padded;
-
-       while (task_size != 0) {
-               DEBUG_SC("se_mem->se_page(%p) se_mem->se_len(%u)"
-                       " se_mem->se_off(%u) task_offset(%u)\n",
-                       se_mem->se_page, se_mem->se_len,
-                       se_mem->se_off, task_offset);
-
-               if (task_offset == 0) {
-                       if (task_size >= se_mem->se_len) {
-                               sg_length = se_mem->se_len;
-
-                               if (!(list_is_last(&se_mem->se_list,
-                                               T_TASK(se_cmd)->t_mem_list)))
-                                       se_mem = list_entry(se_mem->se_list.next,
-                                                       struct se_mem, se_list);
-                       } else {
-                               sg_length = task_size;
-                               task_size -= sg_length;
-                               goto next;
-                       }
-
-                       DEBUG_SC("sg_length(%u) task_size(%u)\n",
-                                       sg_length, task_size);
-               } else {
-                       if ((se_mem->se_len - task_offset) > task_size) {
-                               sg_length = task_size;
-                               task_size -= sg_length;
-                               goto next;
-                        } else {
-                               sg_length = (se_mem->se_len - task_offset);
-
-                               if (!(list_is_last(&se_mem->se_list,
-                                               T_TASK(se_cmd)->t_mem_list)))
-                                       se_mem = list_entry(se_mem->se_list.next,
-                                                       struct se_mem, se_list);
-                       }
-
-                       DEBUG_SC("sg_length(%u) task_size(%u)\n",
-                                       sg_length, task_size);
-
-                       task_offset = 0;
-               }
-               task_size -= sg_length;
-next:
-               DEBUG_SC("task[%u] - Reducing task_size to(%u)\n",
-                       task->task_no, task_size);
-
-               task->task_sg_num++;
-       }
-       /*
-        * Check if the fabric module driver is requesting that all
-        * struct se_task->task_sg[] be chained together..  If so,
-        * then allocate an extra padding SG entry for linking and
-        * marking the end of the chained SGL.
-        */
-       if (tfo->task_sg_chaining) {
-               task_sg_num_padded = (task->task_sg_num + 1);
-               task->task_padded_sg = 1;
-       } else
-               task_sg_num_padded = task->task_sg_num;
-
-       task->task_sg = kzalloc(task_sg_num_padded *
-                       sizeof(struct scatterlist), GFP_KERNEL);
-       if (!(task->task_sg)) {
-               printk(KERN_ERR "Unable to allocate memory for"
-                               " task->task_sg\n");
-               return 0;
-       }
-       sg_init_table(&task->task_sg[0], task_sg_num_padded);
-       /*
-        * Setup task->task_sg_bidi for SCSI READ payload for
-        * TCM/pSCSI passthrough if present for BIDI-COMMAND
-        */
-       if ((T_TASK(se_cmd)->t_mem_bidi_list != NULL) &&
-           (TRANSPORT(se_dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV)) {
-               task->task_sg_bidi = kzalloc(task_sg_num_padded *
-                               sizeof(struct scatterlist), GFP_KERNEL);
-               if (!(task->task_sg_bidi)) {
-                       printk(KERN_ERR "Unable to allocate memory for"
-                               " task->task_sg_bidi\n");
-                       return 0;
-               }
-               sg_init_table(&task->task_sg_bidi[0], task_sg_num_padded);
-       }
-       /*
-        * For the chaining case, setup the proper end of SGL for the
-        * initial submission struct task into struct se_subsystem_api.
-        * This will be cleared later by transport_do_task_sg_chain()
-        */
-       if (task->task_padded_sg) {
-               sg_mark_end(&task->task_sg[task->task_sg_num - 1]);
-               /*
-                * Added the 'if' check before marking end of bi-directional
-                * scatterlist (which gets created only in case of request
-                * (RD + WR).
-                */
-               if (task->task_sg_bidi)
-                       sg_mark_end(&task->task_sg_bidi[task->task_sg_num - 1]);
-       }
-
-       DEBUG_SC("Successfully allocated task->task_sg_num(%u),"
-               " task_sg_num_padded(%u)\n", task->task_sg_num,
-               task_sg_num_padded);
-
-       return task->task_sg_num;
-}
-
-static inline int transport_set_tasks_sectors_disk(
-       struct se_task *task,
-       struct se_device *dev,
-       unsigned long long lba,
-       u32 sectors,
-       int *max_sectors_set)
-{
-       if ((lba + sectors) > transport_dev_end_lba(dev)) {
-               task->task_sectors = ((transport_dev_end_lba(dev) - lba) + 1);
-
-               if (task->task_sectors > DEV_ATTRIB(dev)->max_sectors) {
-                       task->task_sectors = DEV_ATTRIB(dev)->max_sectors;
-                       *max_sectors_set = 1;
-               }
-       } else {
-               if (sectors > DEV_ATTRIB(dev)->max_sectors) {
-                       task->task_sectors = DEV_ATTRIB(dev)->max_sectors;
-                       *max_sectors_set = 1;
-               } else
-                       task->task_sectors = sectors;
-       }
-
-       return 0;
-}
-
-static inline int transport_set_tasks_sectors_non_disk(
-       struct se_task *task,
-       struct se_device *dev,
-       unsigned long long lba,
-       u32 sectors,
-       int *max_sectors_set)
-{
-       if (sectors > DEV_ATTRIB(dev)->max_sectors) {
-               task->task_sectors = DEV_ATTRIB(dev)->max_sectors;
-               *max_sectors_set = 1;
-       } else
-               task->task_sectors = sectors;
-
-       return 0;
-}
-
-static inline int transport_set_tasks_sectors(
-       struct se_task *task,
-       struct se_device *dev,
-       unsigned long long lba,
-       u32 sectors,
-       int *max_sectors_set)
-{
-       return (TRANSPORT(dev)->get_device_type(dev) == TYPE_DISK) ?
-               transport_set_tasks_sectors_disk(task, dev, lba, sectors,
-                               max_sectors_set) :
-               transport_set_tasks_sectors_non_disk(task, dev, lba, sectors,
-                               max_sectors_set);
-}
-
-static int transport_map_sg_to_mem(
+ */
+int transport_generic_map_mem_to_cmd(
        struct se_cmd *cmd,
-       struct list_head *se_mem_list,
-       void *in_mem,
-       u32 *se_mem_cnt)
+       struct scatterlist *sgl,
+       u32 sgl_count,
+       struct scatterlist *sgl_bidi,
+       u32 sgl_bidi_count)
 {
-       struct se_mem *se_mem;
-       struct scatterlist *sg;
-       u32 sg_count = 1, cmd_size = cmd->data_length;
-
-       if (!in_mem) {
-               printk(KERN_ERR "No source scatterlist\n");
-               return -1;
-       }
-       sg = (struct scatterlist *)in_mem;
+       if (!sgl || !sgl_count)
+               return 0;
 
-       while (cmd_size) {
-               se_mem = kmem_cache_zalloc(se_mem_cache, GFP_KERNEL);
-               if (!(se_mem)) {
-                       printk(KERN_ERR "Unable to allocate struct se_mem\n");
-                       return -1;
-               }
-               INIT_LIST_HEAD(&se_mem->se_list);
-               DEBUG_MEM("sg_to_mem: Starting loop with cmd_size: %u"
-                       " sg_page: %p offset: %d length: %d\n", cmd_size,
-                       sg_page(sg), sg->offset, sg->length);
+       if ((cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB) ||
+           (cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB)) {
 
-               se_mem->se_page = sg_page(sg);
-               se_mem->se_off = sg->offset;
+               cmd->t_data_sg = sgl;
+               cmd->t_data_nents = sgl_count;
 
-               if (cmd_size > sg->length) {
-                       se_mem->se_len = sg->length;
-                       sg = sg_next(sg);
-                       sg_count++;
-               } else
-                       se_mem->se_len = cmd_size;
+               if (sgl_bidi && sgl_bidi_count) {
+                       cmd->t_bidi_data_sg = sgl_bidi;
+                       cmd->t_bidi_data_nents = sgl_bidi_count;
+               }
+               cmd->se_cmd_flags |= SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC;
+       }
 
-               cmd_size -= se_mem->se_len;
+       return 0;
+}
+EXPORT_SYMBOL(transport_generic_map_mem_to_cmd);
 
-               DEBUG_MEM("sg_to_mem: *se_mem_cnt: %u cmd_size: %u\n",
-                               *se_mem_cnt, cmd_size);
-               DEBUG_MEM("sg_to_mem: Final se_page: %p se_off: %d se_len: %d\n",
-                               se_mem->se_page, se_mem->se_off, se_mem->se_len);
+static int transport_new_cmd_obj(struct se_cmd *cmd)
+{
+       struct se_device *dev = cmd->se_dev;
+       u32 task_cdbs;
+       u32 rc;
+       int set_counts = 1;
 
-               list_add_tail(&se_mem->se_list, se_mem_list);
-               (*se_mem_cnt)++;
+       /*
+        * Setup any BIDI READ tasks and memory from
+        * cmd->t_mem_bidi_list so the READ struct se_tasks
+        * are queued first for the non pSCSI passthrough case.
+        */
+       if (cmd->t_bidi_data_sg &&
+           (dev->transport->transport_type != TRANSPORT_PLUGIN_PHBA_PDEV)) {
+               rc = transport_allocate_tasks(cmd,
+                                             cmd->t_task_lba,
+                                             DMA_FROM_DEVICE,
+                                             cmd->t_bidi_data_sg,
+                                             cmd->t_bidi_data_nents);
+               if (rc <= 0) {
+                       cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+                       cmd->scsi_sense_reason =
+                               TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
+                       return PYX_TRANSPORT_LU_COMM_FAILURE;
+               }
+               atomic_inc(&cmd->t_fe_count);
+               atomic_inc(&cmd->t_se_count);
+               set_counts = 0;
+       }
+       /*
+        * Setup the tasks and memory from cmd->t_mem_list
+        * Note for BIDI transfers this will contain the WRITE payload
+        */
+       task_cdbs = transport_allocate_tasks(cmd,
+                                            cmd->t_task_lba,
+                                            cmd->data_direction,
+                                            cmd->t_data_sg,
+                                            cmd->t_data_nents);
+       if (task_cdbs <= 0) {
+               cmd->se_cmd_flags |= SCF_SCSI_CDB_EXCEPTION;
+               cmd->scsi_sense_reason =
+                       TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE;
+               return PYX_TRANSPORT_LU_COMM_FAILURE;
        }
 
-       DEBUG_MEM("task[0] - Mapped(%u) struct scatterlist segments to(%u)"
-               " struct se_mem\n", sg_count, *se_mem_cnt);
+       if (set_counts) {
+               atomic_inc(&cmd->t_fe_count);
+               atomic_inc(&cmd->t_se_count);
+       }
 
-       if (sg_count != *se_mem_cnt)
-               BUG();
+       cmd->t_task_list_num = task_cdbs;
 
+       atomic_set(&cmd->t_task_cdbs_left, task_cdbs);
+       atomic_set(&cmd->t_task_cdbs_ex_left, task_cdbs);
+       atomic_set(&cmd->t_task_cdbs_timeout_left, task_cdbs);
        return 0;
 }
 
-/*     transport_map_mem_to_sg():
- *
- *
- */
-int transport_map_mem_to_sg(
-       struct se_task *task,
-       struct list_head *se_mem_list,
-       void *in_mem,
-       struct se_mem *in_se_mem,
-       struct se_mem **out_se_mem,
-       u32 *se_mem_cnt,
-       u32 *task_offset)
+void *transport_kmap_first_data_page(struct se_cmd *cmd)
 {
-       struct se_cmd *se_cmd = task->task_se_cmd;
-       struct se_mem *se_mem = in_se_mem;
-       struct scatterlist *sg = (struct scatterlist *)in_mem;
-       u32 task_size = task->task_size, sg_no = 0;
-
-       if (!sg) {
-               printk(KERN_ERR "Unable to locate valid struct"
-                               " scatterlist pointer\n");
-               return -1;
-       }
+       struct scatterlist *sg = cmd->t_data_sg;
 
-       while (task_size != 0) {
-               /*
-                * Setup the contigious array of scatterlists for
-                * this struct se_task.
-                */
-               sg_assign_page(sg, se_mem->se_page);
+       BUG_ON(!sg);
+       /*
+        * We need to take into account a possible offset here for fabrics like
+        * tcm_loop who may be using a contig buffer from the SCSI midlayer for
+        * control CDBs passed as SGLs via transport_generic_map_mem_to_cmd()
+        */
+       return kmap(sg_page(sg)) + sg->offset;
+}
+EXPORT_SYMBOL(transport_kmap_first_data_page);
 
-               if (*task_offset == 0) {
-                       sg->offset = se_mem->se_off;
+void transport_kunmap_first_data_page(struct se_cmd *cmd)
+{
+       kunmap(sg_page(cmd->t_data_sg));
+}
+EXPORT_SYMBOL(transport_kunmap_first_data_page);
 
-                       if (task_size >= se_mem->se_len) {
-                               sg->length = se_mem->se_len;
+static int
+transport_generic_get_mem(struct se_cmd *cmd)
+{
+       u32 length = cmd->data_length;
+       unsigned int nents;
+       struct page *page;
+       int i = 0;
 
-                               if (!(list_is_last(&se_mem->se_list,
-                                               T_TASK(se_cmd)->t_mem_list))) {
-                                       se_mem = list_entry(se_mem->se_list.next,
-                                                       struct se_mem, se_list);
-                                       (*se_mem_cnt)++;
-                               }
-                       } else {
-                               sg->length = task_size;
-                               /*
-                                * Determine if we need to calculate an offset
-                                * into the struct se_mem on the next go around..
-                                */
-                               task_size -= sg->length;
-                               if (!(task_size))
-                                       *task_offset = sg->length;
-
-                               goto next;
-                       }
+       nents = DIV_ROUND_UP(length, PAGE_SIZE);
+       cmd->t_data_sg = kmalloc(sizeof(struct scatterlist) * nents, GFP_KERNEL);
+       if (!cmd->t_data_sg)
+               return -ENOMEM;
 
-               } else {
-                       sg->offset = (*task_offset + se_mem->se_off);
-
-                       if ((se_mem->se_len - *task_offset) > task_size) {
-                               sg->length = task_size;
-                               /*
-                                * Determine if we need to calculate an offset
-                                * into the struct se_mem on the next go around..
-                                */
-                               task_size -= sg->length;
-                               if (!(task_size))
-                                       *task_offset += sg->length;
-
-                               goto next;
-                       } else {
-                               sg->length = (se_mem->se_len - *task_offset);
-
-                               if (!(list_is_last(&se_mem->se_list,
-                                               T_TASK(se_cmd)->t_mem_list))) {
-                                       se_mem = list_entry(se_mem->se_list.next,
-                                                       struct se_mem, se_list);
-                                       (*se_mem_cnt)++;
-                               }
-                       }
+       cmd->t_data_nents = nents;
+       sg_init_table(cmd->t_data_sg, nents);
 
-                       *task_offset = 0;
-               }
-               task_size -= sg->length;
-next:
-               DEBUG_MEM("task[%u] mem_to_sg - sg[%u](%p)(%u)(%u) - Reducing"
-                       " task_size to(%u), task_offset: %u\n", task->task_no, sg_no,
-                       sg_page(sg), sg->length, sg->offset, task_size, *task_offset);
-
-               sg_no++;
-               if (!(task_size))
-                       break;
+       while (length) {
+               u32 page_len = min_t(u32, length, PAGE_SIZE);
+               page = alloc_page(GFP_KERNEL | __GFP_ZERO);
+               if (!page)
+                       goto out;
 
-               sg = sg_next(sg);
+               sg_set_page(&cmd->t_data_sg[i], page, page_len, 0);
+               length -= page_len;
+               i++;
+       }
+       return 0;
 
-               if (task_size > se_cmd->data_length)
-                       BUG();
+out:
+       while (i >= 0) {
+               __free_page(sg_page(&cmd->t_data_sg[i]));
+               i--;
        }
-       *out_se_mem = se_mem;
+       kfree(cmd->t_data_sg);
+       cmd->t_data_sg = NULL;
+       return -ENOMEM;
+}
+
+/* Reduce sectors if they are too long for the device */
+static inline sector_t transport_limit_task_sectors(
+       struct se_device *dev,
+       unsigned long long lba,
+       sector_t sectors)
+{
+       sectors = min_t(sector_t, sectors, dev->se_sub_dev->se_dev_attrib.max_sectors);
 
-       DEBUG_MEM("task[%u] - Mapped(%u) struct se_mem segments to total(%u)"
-               " SGs\n", task->task_no, *se_mem_cnt, sg_no);
+       if (dev->transport->get_device_type(dev) == TYPE_DISK)
+               if ((lba + sectors) > transport_dev_end_lba(dev))
+                       sectors = ((transport_dev_end_lba(dev) - lba) + 1);
 
-       return 0;
+       return sectors;
 }
 
+
 /*
  * This function can be used by HW target mode drivers to create a linked
  * scatterlist from all contiguously allocated struct se_task->task_sg[].
@@ -4751,334 +3983,236 @@ next:
  */
 void transport_do_task_sg_chain(struct se_cmd *cmd)
 {
-       struct scatterlist *sg_head = NULL, *sg_link = NULL, *sg_first = NULL;
-       struct scatterlist *sg_head_cur = NULL, *sg_link_cur = NULL;
-       struct scatterlist *sg, *sg_end = NULL, *sg_end_cur = NULL;
+       struct scatterlist *sg_first = NULL;
+       struct scatterlist *sg_prev = NULL;
+       int sg_prev_nents = 0;
+       struct scatterlist *sg;
        struct se_task *task;
-       struct target_core_fabric_ops *tfo = CMD_TFO(cmd);
-       u32 task_sg_num = 0, sg_count = 0;
+       u32 chained_nents = 0;
        int i;
 
-       if (tfo->task_sg_chaining == 0) {
-               printk(KERN_ERR "task_sg_chaining is diabled for fabric module:"
-                               " %s\n", tfo->get_fabric_name());
-               dump_stack();
-               return;
-       }
+       BUG_ON(!cmd->se_tfo->task_sg_chaining);
+
        /*
         * Walk the struct se_task list and setup scatterlist chains
-        * for each contiguosly allocated struct se_task->task_sg[].
+        * for each contiguously allocated struct se_task->task_sg[].
         */
-       list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
-               if (!(task->task_sg) || !(task->task_padded_sg))
+       list_for_each_entry(task, &cmd->t_task_list, t_list) {
+               if (!task->task_sg)
                        continue;
 
-               if (sg_head && sg_link) {
-                       sg_head_cur = &task->task_sg[0];
-                       sg_link_cur = &task->task_sg[task->task_sg_num];
-                       /*
-                        * Either add chain or mark end of scatterlist
-                        */
-                       if (!(list_is_last(&task->t_list,
-                                       &T_TASK(cmd)->t_task_list))) {
-                               /*
-                                * Clear existing SGL termination bit set in
-                                * transport_calc_sg_num(), see sg_mark_end()
-                                */
-                               sg_end_cur = &task->task_sg[task->task_sg_num - 1];
-                               sg_end_cur->page_link &= ~0x02;
-
-                               sg_chain(sg_head, task_sg_num, sg_head_cur);
-                               sg_count += task->task_sg_num;
-                               task_sg_num = (task->task_sg_num + 1);
-                       } else {
-                               sg_chain(sg_head, task_sg_num, sg_head_cur);
-                               sg_count += task->task_sg_num;
-                               task_sg_num = task->task_sg_num;
-                       }
+               BUG_ON(!task->task_padded_sg);
 
-                       sg_head = sg_head_cur;
-                       sg_link = sg_link_cur;
-                       continue;
-               }
-               sg_head = sg_first = &task->task_sg[0];
-               sg_link = &task->task_sg[task->task_sg_num];
-               /*
-                * Check for single task..
-                */
-               if (!(list_is_last(&task->t_list, &T_TASK(cmd)->t_task_list))) {
-                       /*
-                        * Clear existing SGL termination bit set in
-                        * transport_calc_sg_num(), see sg_mark_end()
-                        */
-                       sg_end = &task->task_sg[task->task_sg_num - 1];
-                       sg_end->page_link &= ~0x02;
-                       sg_count += task->task_sg_num;
-                       task_sg_num = (task->task_sg_num + 1);
+               if (!sg_first) {
+                       sg_first = task->task_sg;
+                       chained_nents = task->task_sg_nents;
                } else {
-                       sg_count += task->task_sg_num;
-                       task_sg_num = task->task_sg_num;
+                       sg_chain(sg_prev, sg_prev_nents, task->task_sg);
+                       chained_nents += task->task_sg_nents;
                }
+
+               sg_prev = task->task_sg;
+               sg_prev_nents = task->task_sg_nents;
        }
        /*
         * Setup the starting pointer and total t_tasks_sg_linked_no including
         * padding SGs for linking and to mark the end.
         */
-       T_TASK(cmd)->t_tasks_sg_chained = sg_first;
-       T_TASK(cmd)->t_tasks_sg_chained_no = sg_count;
+       cmd->t_tasks_sg_chained = sg_first;
+       cmd->t_tasks_sg_chained_no = chained_nents;
 
-       DEBUG_CMD_M("Setup cmd: %p T_TASK(cmd)->t_tasks_sg_chained: %p and"
-               " t_tasks_sg_chained_no: %u\n", cmd, T_TASK(cmd)->t_tasks_sg_chained,
-               T_TASK(cmd)->t_tasks_sg_chained_no);
+       pr_debug("Setup cmd: %p cmd->t_tasks_sg_chained: %p and"
+               " t_tasks_sg_chained_no: %u\n", cmd, cmd->t_tasks_sg_chained,
+               cmd->t_tasks_sg_chained_no);
 
-       for_each_sg(T_TASK(cmd)->t_tasks_sg_chained, sg,
-                       T_TASK(cmd)->t_tasks_sg_chained_no, i) {
+       for_each_sg(cmd->t_tasks_sg_chained, sg,
+                       cmd->t_tasks_sg_chained_no, i) {
 
-               DEBUG_CMD_M("SG[%d]: %p page: %p length: %d offset: %d, magic: 0x%08x\n",
-                       i, sg, sg_page(sg), sg->length, sg->offset, sg->sg_magic);
+               pr_debug("SG[%d]: %p page: %p length: %d offset: %d\n",
+                       i, sg, sg_page(sg), sg->length, sg->offset);
                if (sg_is_chain(sg))
-                       DEBUG_CMD_M("SG: %p sg_is_chain=1\n", sg);
+                       pr_debug("SG: %p sg_is_chain=1\n", sg);
                if (sg_is_last(sg))
-                       DEBUG_CMD_M("SG: %p sg_is_last=1\n", sg);
+                       pr_debug("SG: %p sg_is_last=1\n", sg);
        }
 }
 EXPORT_SYMBOL(transport_do_task_sg_chain);
 
-static int transport_do_se_mem_map(
-       struct se_device *dev,
-       struct se_task *task,
-       struct list_head *se_mem_list,
-       void *in_mem,
-       struct se_mem *in_se_mem,
-       struct se_mem **out_se_mem,
-       u32 *se_mem_cnt,
-       u32 *task_offset_in)
-{
-       u32 task_offset = *task_offset_in;
-       int ret = 0;
-       /*
-        * se_subsystem_api_t->do_se_mem_map is used when internal allocation
-        * has been done by the transport plugin.
-        */
-       if (TRANSPORT(dev)->do_se_mem_map) {
-               ret = TRANSPORT(dev)->do_se_mem_map(task, se_mem_list,
-                               in_mem, in_se_mem, out_se_mem, se_mem_cnt,
-                               task_offset_in);
-               if (ret == 0)
-                       T_TASK(task->task_se_cmd)->t_tasks_se_num += *se_mem_cnt;
-
-               return ret;
-       }
-
-       BUG_ON(list_empty(se_mem_list));
-       /*
-        * This is the normal path for all normal non BIDI and BIDI-COMMAND
-        * WRITE payloads..  If we need to do BIDI READ passthrough for
-        * TCM/pSCSI the first call to transport_do_se_mem_map ->
-        * transport_calc_sg_num() -> transport_map_mem_to_sg() will do the
-        * allocation for task->task_sg_bidi, and the subsequent call to
-        * transport_do_se_mem_map() from transport_generic_get_cdb_count()
-        */
-       if (!(task->task_sg_bidi)) {
-               /*
-                * Assume default that transport plugin speaks preallocated
-                * scatterlists.
-                */
-               if (!(transport_calc_sg_num(task, in_se_mem, task_offset)))
-                       return -1;
-               /*
-                * struct se_task->task_sg now contains the struct scatterlist array.
-                */
-               return transport_map_mem_to_sg(task, se_mem_list, task->task_sg,
-                                       in_se_mem, out_se_mem, se_mem_cnt,
-                                       task_offset_in);
-       }
-       /*
-        * Handle the se_mem_list -> struct task->task_sg_bidi
-        * memory map for the extra BIDI READ payload
-        */
-       return transport_map_mem_to_sg(task, se_mem_list, task->task_sg_bidi,
-                               in_se_mem, out_se_mem, se_mem_cnt,
-                               task_offset_in);
-}
-
-static u32 transport_generic_get_cdb_count(
+/*
+ * Break up cmd into chunks transport can handle
+ */
+static int transport_allocate_data_tasks(
        struct se_cmd *cmd,
        unsigned long long lba,
-       u32 sectors,
        enum dma_data_direction data_direction,
-       struct list_head *mem_list,
-       int set_counts)
+       struct scatterlist *sgl,
+       unsigned int sgl_nents)
 {
        unsigned char *cdb = NULL;
        struct se_task *task;
-       struct se_mem *se_mem = NULL, *se_mem_lout = NULL;
-       struct se_mem *se_mem_bidi = NULL, *se_mem_bidi_lout = NULL;
-       struct se_device *dev = SE_DEV(cmd);
-       int max_sectors_set = 0, ret;
-       u32 task_offset_in = 0, se_mem_cnt = 0, se_mem_bidi_cnt = 0, task_cdbs = 0;
-
-       if (!mem_list) {
-               printk(KERN_ERR "mem_list is NULL in transport_generic_get"
-                               "_cdb_count()\n");
-               return 0;
-       }
-       /*
-        * While using RAMDISK_DR backstores is the only case where
-        * mem_list will ever be empty at this point.
-        */
-       if (!(list_empty(mem_list)))
-               se_mem = list_entry(mem_list->next, struct se_mem, se_list);
-       /*
-        * Check for extra se_mem_bidi mapping for BIDI-COMMANDs to
-        * struct se_task->task_sg_bidi for TCM/pSCSI passthrough operation
-        */
-       if ((T_TASK(cmd)->t_mem_bidi_list != NULL) &&
-           !(list_empty(T_TASK(cmd)->t_mem_bidi_list)) &&
-           (TRANSPORT(dev)->transport_type == TRANSPORT_PLUGIN_PHBA_PDEV))
-               se_mem_bidi = list_entry(T_TASK(cmd)->t_mem_bidi_list->next,
-                                       struct se_mem, se_list);
-
-       while (sectors) {
-               DEBUG_VOL("ITT[0x%08x] LBA(%llu) SectorsLeft(%u) EOBJ(%llu)\n",
-                       CMD_TFO(cmd)->get_task_tag(cmd), lba, sectors,
-                       transport_dev_end_lba(dev));
+       struct se_device *dev = cmd->se_dev;
+       unsigned long flags;
+       sector_t sectors;
+       int task_count, i, ret;
+       sector_t dev_max_sectors = dev->se_sub_dev->se_dev_attrib.max_sectors;
+       u32 sector_size = dev->se_sub_dev->se_dev_attrib.block_size;
+       struct scatterlist *sg;
+       struct scatterlist *cmd_sg;
 
-               task = transport_generic_get_task(cmd, data_direction);
-               if (!(task))
-                       goto out;
+       WARN_ON(cmd->data_length % sector_size);
+       sectors = DIV_ROUND_UP(cmd->data_length, sector_size);
+       task_count = DIV_ROUND_UP(sectors, dev_max_sectors);
+
+       cmd_sg = sgl;
+       for (i = 0; i < task_count; i++) {
+               unsigned int task_size;
+               int count;
 
-               transport_set_tasks_sectors(task, dev, lba, sectors,
-                               &max_sectors_set);
+               task = transport_generic_get_task(cmd, data_direction);
+               if (!task)
+                       return -ENOMEM;
 
                task->task_lba = lba;
-               lba += task->task_sectors;
-               sectors -= task->task_sectors;
-               task->task_size = (task->task_sectors *
-                                  DEV_ATTRIB(dev)->block_size);
-
-               cdb = TRANSPORT(dev)->get_cdb(task);
-               if ((cdb)) {
-                       memcpy(cdb, T_TASK(cmd)->t_task_cdb,
-                               scsi_command_size(T_TASK(cmd)->t_task_cdb));
-                       cmd->transport_split_cdb(task->task_lba,
-                                       &task->task_sectors, cdb);
-               }
+               task->task_sectors = min(sectors, dev_max_sectors);
+               task->task_size = task->task_sectors * sector_size;
 
-               /*
-                * Perform the SE OBJ plugin and/or Transport plugin specific
-                * mapping for T_TASK(cmd)->t_mem_list. And setup the
-                * task->task_sg and if necessary task->task_sg_bidi
-                */
-               ret = transport_do_se_mem_map(dev, task, mem_list,
-                               NULL, se_mem, &se_mem_lout, &se_mem_cnt,
-                               &task_offset_in);
-               if (ret < 0)
-                       goto out;
+               cdb = dev->transport->get_cdb(task);
+               BUG_ON(!cdb);
+
+               memcpy(cdb, cmd->t_task_cdb,
+                      scsi_command_size(cmd->t_task_cdb));
+
+               /* Update new cdb with updated lba/sectors */
+               cmd->transport_split_cdb(task->task_lba, task->task_sectors, cdb);
 
-               se_mem = se_mem_lout;
                /*
-                * Setup the T_TASK(cmd)->t_mem_bidi_list -> task->task_sg_bidi
-                * mapping for SCSI READ for BIDI-COMMAND passthrough with TCM/pSCSI
-                *
-                * Note that the first call to transport_do_se_mem_map() above will
-                * allocate struct se_task->task_sg_bidi in transport_do_se_mem_map()
-                * -> transport_calc_sg_num(), and the second here will do the
-                * mapping for SCSI READ for BIDI-COMMAND passthrough with TCM/pSCSI.
+                * Check if the fabric module driver is requesting that all
+                * struct se_task->task_sg[] be chained together..  If so,
+                * then allocate an extra padding SG entry for linking and
+                * marking the end of the chained SGL.
+                * Possibly over-allocate task sgl size by using cmd sgl size.
+                * It's so much easier and only a waste when task_count > 1.
+                * That is extremely rare.
                 */
-               if (task->task_sg_bidi != NULL) {
-                       ret = transport_do_se_mem_map(dev, task,
-                               T_TASK(cmd)->t_mem_bidi_list, NULL,
-                               se_mem_bidi, &se_mem_bidi_lout, &se_mem_bidi_cnt,
-                               &task_offset_in);
-                       if (ret < 0)
-                               goto out;
+               task->task_sg_nents = sgl_nents;
+               if (cmd->se_tfo->task_sg_chaining) {
+                       task->task_sg_nents++;
+                       task->task_padded_sg = 1;
+               }
 
-                       se_mem_bidi = se_mem_bidi_lout;
+               task->task_sg = kmalloc(sizeof(struct scatterlist) *
+                                       task->task_sg_nents, GFP_KERNEL);
+               if (!task->task_sg) {
+                       cmd->se_dev->transport->free_task(task);
+                       return -ENOMEM;
                }
-               task_cdbs++;
 
-               DEBUG_VOL("Incremented task_cdbs(%u) task->task_sg_num(%u)\n",
-                               task_cdbs, task->task_sg_num);
+               sg_init_table(task->task_sg, task->task_sg_nents);
 
-               if (max_sectors_set) {
-                       max_sectors_set = 0;
-                       continue;
+               task_size = task->task_size;
+
+               /* Build new sgl, only up to task_size */
+               for_each_sg(task->task_sg, sg, task->task_sg_nents, count) {
+                       if (cmd_sg->length > task_size)
+                               break;
+
+                       *sg = *cmd_sg;
+                       task_size -= cmd_sg->length;
+                       cmd_sg = sg_next(cmd_sg);
                }
 
-               if (!sectors)
-                       break;
-       }
+               lba += task->task_sectors;
+               sectors -= task->task_sectors;
 
-       if (set_counts) {
-               atomic_inc(&T_TASK(cmd)->t_fe_count);
-               atomic_inc(&T_TASK(cmd)->t_se_count);
+               spin_lock_irqsave(&cmd->t_state_lock, flags);
+               list_add_tail(&task->t_list, &cmd->t_task_list);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
        }
+       /*
+        * Now perform the memory map of task->task_sg[] into backend
+        * subsystem memory..
+        */
+       list_for_each_entry(task, &cmd->t_task_list, t_list) {
+               if (atomic_read(&task->task_sent))
+                       continue;
+               if (!dev->transport->map_data_SG)
+                       continue;
 
-       DEBUG_VOL("ITT[0x%08x] total %s cdbs(%u)\n",
-               CMD_TFO(cmd)->get_task_tag(cmd), (data_direction == DMA_TO_DEVICE)
-               ? "DMA_TO_DEVICE" : "DMA_FROM_DEVICE", task_cdbs);
+               ret = dev->transport->map_data_SG(task);
+               if (ret < 0)
+                       return 0;
+       }
 
-       return task_cdbs;
-out:
-       return 0;
+       return task_count;
 }
 
 static int
-transport_map_control_cmd_to_task(struct se_cmd *cmd)
+transport_allocate_control_task(struct se_cmd *cmd)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        unsigned char *cdb;
        struct se_task *task;
-       int ret;
+       unsigned long flags;
+       int ret = 0;
 
        task = transport_generic_get_task(cmd, cmd->data_direction);
        if (!task)
-               return PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
+               return -ENOMEM;
 
-       cdb = TRANSPORT(dev)->get_cdb(task);
-       if (cdb)
-               memcpy(cdb, cmd->t_task->t_task_cdb,
-                       scsi_command_size(cmd->t_task->t_task_cdb));
+       cdb = dev->transport->get_cdb(task);
+       BUG_ON(!cdb);
+       memcpy(cdb, cmd->t_task_cdb,
+              scsi_command_size(cmd->t_task_cdb));
 
+       task->task_sg = kmalloc(sizeof(struct scatterlist) * cmd->t_data_nents,
+                               GFP_KERNEL);
+       if (!task->task_sg) {
+               cmd->se_dev->transport->free_task(task);
+               return -ENOMEM;
+       }
+
+       memcpy(task->task_sg, cmd->t_data_sg,
+              sizeof(struct scatterlist) * cmd->t_data_nents);
        task->task_size = cmd->data_length;
-       task->task_sg_num =
-               (cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB) ? 1 : 0;
+       task->task_sg_nents = cmd->t_data_nents;
 
-       atomic_inc(&cmd->t_task->t_fe_count);
-       atomic_inc(&cmd->t_task->t_se_count);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       list_add_tail(&task->t_list, &cmd->t_task_list);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        if (cmd->se_cmd_flags & SCF_SCSI_CONTROL_SG_IO_CDB) {
-               struct se_mem *se_mem = NULL, *se_mem_lout = NULL;
-               u32 se_mem_cnt = 0, task_offset = 0;
-
-               if (!list_empty(T_TASK(cmd)->t_mem_list))
-                       se_mem = list_entry(T_TASK(cmd)->t_mem_list->next,
-                                       struct se_mem, se_list);
-
-               ret = transport_do_se_mem_map(dev, task,
-                               cmd->t_task->t_mem_list, NULL, se_mem,
-                               &se_mem_lout, &se_mem_cnt, &task_offset);
-               if (ret < 0)
-                       return PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
-
-               if (dev->transport->map_task_SG)
-                       return dev->transport->map_task_SG(task);
-               return 0;
-       } else if (cmd->se_cmd_flags & SCF_SCSI_CONTROL_NONSG_IO_CDB) {
-               if (dev->transport->map_task_non_SG)
-                       return dev->transport->map_task_non_SG(task);
-               return 0;
+               if (dev->transport->map_control_SG)
+                       ret = dev->transport->map_control_SG(task);
        } else if (cmd->se_cmd_flags & SCF_SCSI_NON_DATA_CDB) {
                if (dev->transport->cdb_none)
-                       return dev->transport->cdb_none(task);
-               return 0;
+                       ret = dev->transport->cdb_none(task);
        } else {
+               pr_err("target: Unknown control cmd type!\n");
                BUG();
-               return PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
        }
+
+       /* Success! Return number of tasks allocated */
+       if (ret == 0)
+               return 1;
+       return ret;
+}
+
+static u32 transport_allocate_tasks(
+       struct se_cmd *cmd,
+       unsigned long long lba,
+       enum dma_data_direction data_direction,
+       struct scatterlist *sgl,
+       unsigned int sgl_nents)
+{
+       if (cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB)
+               return transport_allocate_data_tasks(cmd, lba, data_direction,
+                                                    sgl, sgl_nents);
+       else
+               return transport_allocate_control_task(cmd);
+
 }
 
+
 /*      transport_generic_new_cmd(): Called from transport_processing_thread()
  *
  *      Allocate storage transport resources from a set of values predefined
@@ -5088,64 +4222,33 @@ transport_map_control_cmd_to_task(struct se_cmd *cmd)
        /*
         * Generate struct se_task(s) and/or their payloads for this CDB.
         */
-static int transport_generic_new_cmd(struct se_cmd *cmd)
+int transport_generic_new_cmd(struct se_cmd *cmd)
 {
-       struct se_portal_group *se_tpg;
-       struct se_task *task;
-       struct se_device *dev = SE_DEV(cmd);
        int ret = 0;
 
        /*
         * Determine is the TCM fabric module has already allocated physical
         * memory, and is directly calling transport_generic_map_mem_to_cmd()
-        * to setup beforehand the linked list of physical memory at
-        * T_TASK(cmd)->t_mem_list of struct se_mem->se_page
+        * beforehand.
         */
-       if (!(cmd->se_cmd_flags & SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC)) {
-               ret = transport_allocate_resources(cmd);
+       if (!(cmd->se_cmd_flags & SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC) &&
+           cmd->data_length) {
+               ret = transport_generic_get_mem(cmd);
                if (ret < 0)
                        return ret;
        }
-
-       ret = transport_get_sectors(cmd);
-       if (ret < 0)
-               return ret;
-
+       /*
+        * Call transport_new_cmd_obj() to invoke transport_allocate_tasks() for
+        * control or data CDB types, and perform the map to backend subsystem
+        * code from SGL memory allocated here by transport_generic_get_mem(), or
+        * via pre-existing SGL memory setup explictly by fabric module code with
+        * transport_generic_map_mem_to_cmd().
+        */
        ret = transport_new_cmd_obj(cmd);
        if (ret < 0)
                return ret;
-
-       /*
-        * Determine if the calling TCM fabric module is talking to
-        * Linux/NET via kernel sockets and needs to allocate a
-        * struct iovec array to complete the struct se_cmd
-        */
-       se_tpg = SE_LUN(cmd)->lun_sep->sep_tpg;
-       if (TPG_TFO(se_tpg)->alloc_cmd_iovecs != NULL) {
-               ret = TPG_TFO(se_tpg)->alloc_cmd_iovecs(cmd);
-               if (ret < 0)
-                       return PYX_TRANSPORT_OUT_OF_MEMORY_RESOURCES;
-       }
-
-       if (cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB) {
-               list_for_each_entry(task, &T_TASK(cmd)->t_task_list, t_list) {
-                       if (atomic_read(&task->task_sent))
-                               continue;
-                       if (!dev->transport->map_task_SG)
-                               continue;
-
-                       ret = dev->transport->map_task_SG(task);
-                       if (ret < 0)
-                               return ret;
-               }
-       } else {
-               ret = transport_map_control_cmd_to_task(cmd);
-               if (ret < 0)
-                       return ret;
-       }
-
        /*
-        * For WRITEs, let the iSCSI Target RX Thread know its buffer is ready..
+        * For WRITEs, let the fabric know its buffer is ready..
         * This WRITE struct se_cmd (and all of its associated struct se_task's)
         * will be added to the struct se_device execution queue after its WRITE
         * data has arrived. (ie: It gets handled by the transport processing
@@ -5162,6 +4265,7 @@ static int transport_generic_new_cmd(struct se_cmd *cmd)
        transport_execute_tasks(cmd);
        return 0;
 }
+EXPORT_SYMBOL(transport_generic_new_cmd);
 
 /*     transport_generic_process_write():
  *
@@ -5169,68 +4273,15 @@ static int transport_generic_new_cmd(struct se_cmd *cmd)
  */
 void transport_generic_process_write(struct se_cmd *cmd)
 {
-#if 0
-       /*
-        * Copy SCSI Presented DTL sector(s) from received buffers allocated to
-        * original EDTL
-        */
-       if (cmd->se_cmd_flags & SCF_UNDERFLOW_BIT) {
-               if (!T_TASK(cmd)->t_tasks_se_num) {
-                       unsigned char *dst, *buf =
-                               (unsigned char *)T_TASK(cmd)->t_task_buf;
-
-                       dst = kzalloc(cmd->cmd_spdtl), GFP_KERNEL);
-                       if (!(dst)) {
-                               printk(KERN_ERR "Unable to allocate memory for"
-                                               " WRITE underflow\n");
-                               transport_generic_request_failure(cmd, NULL,
-                                       PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
-                               return;
-                       }
-                       memcpy(dst, buf, cmd->cmd_spdtl);
-
-                       kfree(T_TASK(cmd)->t_task_buf);
-                       T_TASK(cmd)->t_task_buf = dst;
-               } else {
-                       struct scatterlist *sg =
-                               (struct scatterlist *sg)T_TASK(cmd)->t_task_buf;
-                       struct scatterlist *orig_sg;
-
-                       orig_sg = kzalloc(sizeof(struct scatterlist) *
-                                       T_TASK(cmd)->t_tasks_se_num,
-                                       GFP_KERNEL))) {
-                       if (!(orig_sg)) {
-                               printk(KERN_ERR "Unable to allocate memory"
-                                               " for WRITE underflow\n");
-                               transport_generic_request_failure(cmd, NULL,
-                                       PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
-                               return;
-                       }
-
-                       memcpy(orig_sg, T_TASK(cmd)->t_task_buf,
-                                       sizeof(struct scatterlist) *
-                                       T_TASK(cmd)->t_tasks_se_num);
-
-                       cmd->data_length = cmd->cmd_spdtl;
-                       /*
-                        * FIXME, clear out original struct se_task and state
-                        * information.
-                        */
-                       if (transport_generic_new_cmd(cmd) < 0) {
-                               transport_generic_request_failure(cmd, NULL,
-                                       PYX_TRANSPORT_REQ_TOO_MANY_SECTORS, 1);
-                               kfree(orig_sg);
-                               return;
-                       }
-
-                       transport_memcpy_write_sg(cmd, orig_sg);
-               }
-       }
-#endif
        transport_execute_tasks(cmd);
 }
 EXPORT_SYMBOL(transport_generic_process_write);
 
+static int transport_write_pending_qf(struct se_cmd *cmd)
+{
+       return cmd->se_tfo->write_pending(cmd);
+}
+
 /*     transport_generic_write_pending():
  *
  *
@@ -5240,24 +4291,26 @@ static int transport_generic_write_pending(struct se_cmd *cmd)
        unsigned long flags;
        int ret;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        cmd->t_state = TRANSPORT_WRITE_PENDING;
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
-       /*
-        * For the TCM control CDBs using a contiguous buffer, do the memcpy
-        * from the passed Linux/SCSI struct scatterlist located at
-        * T_TASK(se_cmd)->t_task_pt_buf to the contiguous buffer at
-        * T_TASK(se_cmd)->t_task_buf.
-        */
-       if (cmd->se_cmd_flags & SCF_PASSTHROUGH_CONTIG_TO_SG)
-               transport_memcpy_read_contig(cmd,
-                               T_TASK(cmd)->t_task_buf,
-                               T_TASK(cmd)->t_task_pt_sgl);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
+
+       if (cmd->transport_qf_callback) {
+               ret = cmd->transport_qf_callback(cmd);
+               if (ret == -EAGAIN)
+                       goto queue_full;
+               else if (ret < 0)
+                       return ret;
+
+               cmd->transport_qf_callback = NULL;
+               return 0;
+       }
+
        /*
         * Clear the se_cmd for WRITE_PENDING status in order to set
-        * T_TASK(cmd)->t_transport_active=0 so that transport_generic_handle_data
+        * cmd->t_transport_active=0 so that transport_generic_handle_data
         * can be called from HW target mode interrupt code.  This is safe
-        * to be called with transport_off=1 before the CMD_TFO(cmd)->write_pending
+        * to be called with transport_off=1 before the cmd->se_tfo->write_pending
         * because the se_cmd->se_lun pointer is not being cleared.
         */
        transport_cmd_check_stop(cmd, 1, 0);
@@ -5266,26 +4319,30 @@ static int transport_generic_write_pending(struct se_cmd *cmd)
         * Call the fabric write_pending function here to let the
         * frontend know that WRITE buffers are ready.
         */
-       ret = CMD_TFO(cmd)->write_pending(cmd);
-       if (ret < 0)
+       ret = cmd->se_tfo->write_pending(cmd);
+       if (ret == -EAGAIN)
+               goto queue_full;
+       else if (ret < 0)
                return ret;
 
        return PYX_TRANSPORT_WRITE_PENDING;
+
+queue_full:
+       pr_debug("Handling write_pending QUEUE__FULL: se_cmd: %p\n", cmd);
+       cmd->t_state = TRANSPORT_COMPLETE_QF_WP;
+       transport_handle_queue_full(cmd, cmd->se_dev,
+                       transport_write_pending_qf);
+       return ret;
 }
 
-/*     transport_release_cmd_to_pool():
- *
- *
- */
-void transport_release_cmd_to_pool(struct se_cmd *cmd)
+void transport_release_cmd(struct se_cmd *cmd)
 {
-       BUG_ON(!T_TASK(cmd));
-       BUG_ON(!CMD_TFO(cmd));
+       BUG_ON(!cmd->se_tfo);
 
        transport_free_se_cmd(cmd);
-       CMD_TFO(cmd)->release_cmd_to_pool(cmd);
+       cmd->se_tfo->release_cmd(cmd);
 }
-EXPORT_SYMBOL(transport_release_cmd_to_pool);
+EXPORT_SYMBOL(transport_release_cmd);
 
 /*     transport_generic_free_cmd():
  *
@@ -5294,19 +4351,18 @@ EXPORT_SYMBOL(transport_release_cmd_to_pool);
 void transport_generic_free_cmd(
        struct se_cmd *cmd,
        int wait_for_tasks,
-       int release_to_pool,
        int session_reinstatement)
 {
-       if (!(cmd->se_cmd_flags & SCF_SE_LUN_CMD) || !T_TASK(cmd))
-               transport_release_cmd_to_pool(cmd);
+       if (!(cmd->se_cmd_flags & SCF_SE_LUN_CMD))
+               transport_release_cmd(cmd);
        else {
                core_dec_lacl_count(cmd->se_sess->se_node_acl, cmd);
 
-               if (SE_LUN(cmd)) {
+               if (cmd->se_lun) {
 #if 0
-                       printk(KERN_INFO "cmd: %p ITT: 0x%08x contains"
-                               " SE_LUN(cmd)\n", cmd,
-                               CMD_TFO(cmd)->get_task_tag(cmd));
+                       pr_debug("cmd: %p ITT: 0x%08x contains"
+                               " cmd->se_lun\n", cmd,
+                               cmd->se_tfo->get_task_tag(cmd));
 #endif
                        transport_lun_remove_cmd(cmd);
                }
@@ -5316,8 +4372,7 @@ void transport_generic_free_cmd(
 
                transport_free_dev_tasks(cmd);
 
-               transport_generic_remove(cmd, release_to_pool,
-                               session_reinstatement);
+               transport_generic_remove(cmd, session_reinstatement);
        }
 }
 EXPORT_SYMBOL(transport_generic_free_cmd);
@@ -5343,43 +4398,36 @@ static int transport_lun_wait_for_tasks(struct se_cmd *cmd, struct se_lun *lun)
         * If the frontend has already requested this struct se_cmd to
         * be stopped, we can safely ignore this struct se_cmd.
         */
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       if (atomic_read(&T_TASK(cmd)->t_transport_stop)) {
-               atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
-               DEBUG_TRANSPORT_S("ConfigFS ITT[0x%08x] - t_transport_stop =="
-                       " TRUE, skipping\n", CMD_TFO(cmd)->get_task_tag(cmd));
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       if (atomic_read(&cmd->t_transport_stop)) {
+               atomic_set(&cmd->transport_lun_stop, 0);
+               pr_debug("ConfigFS ITT[0x%08x] - t_transport_stop =="
+                       " TRUE, skipping\n", cmd->se_tfo->get_task_tag(cmd));
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                transport_cmd_check_stop(cmd, 1, 0);
-               return -1;
+               return -EPERM;
        }
-       atomic_set(&T_TASK(cmd)->transport_lun_fe_stop, 1);
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       atomic_set(&cmd->transport_lun_fe_stop, 1);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
-       wake_up_interruptible(&SE_DEV(cmd)->dev_queue_obj->thread_wq);
+       wake_up_interruptible(&cmd->se_dev->dev_queue_obj.thread_wq);
 
        ret = transport_stop_tasks_for_cmd(cmd);
 
-       DEBUG_TRANSPORT_S("ConfigFS: cmd: %p t_task_cdbs: %d stop tasks ret:"
-                       " %d\n", cmd, T_TASK(cmd)->t_task_cdbs, ret);
+       pr_debug("ConfigFS: cmd: %p t_tasks: %d stop tasks ret:"
+                       " %d\n", cmd, cmd->t_task_list_num, ret);
        if (!ret) {
-               DEBUG_TRANSPORT_S("ConfigFS: ITT[0x%08x] - stopping cmd....\n",
-                               CMD_TFO(cmd)->get_task_tag(cmd));
-               wait_for_completion(&T_TASK(cmd)->transport_lun_stop_comp);
-               DEBUG_TRANSPORT_S("ConfigFS: ITT[0x%08x] - stopped cmd....\n",
-                               CMD_TFO(cmd)->get_task_tag(cmd));
+               pr_debug("ConfigFS: ITT[0x%08x] - stopping cmd....\n",
+                               cmd->se_tfo->get_task_tag(cmd));
+               wait_for_completion(&cmd->transport_lun_stop_comp);
+               pr_debug("ConfigFS: ITT[0x%08x] - stopped cmd....\n",
+                               cmd->se_tfo->get_task_tag(cmd));
        }
-       transport_remove_cmd_from_queue(cmd, SE_DEV(cmd)->dev_queue_obj);
+       transport_remove_cmd_from_queue(cmd, &cmd->se_dev->dev_queue_obj);
 
        return 0;
 }
 
-/* #define DEBUG_CLEAR_LUN */
-#ifdef DEBUG_CLEAR_LUN
-#define DEBUG_CLEAR_L(x...) printk(KERN_INFO x)
-#else
-#define DEBUG_CLEAR_L(x...)
-#endif
-
 static void __transport_clear_lun_from_sessions(struct se_lun *lun)
 {
        struct se_cmd *cmd = NULL;
@@ -5389,66 +4437,59 @@ static void __transport_clear_lun_from_sessions(struct se_lun *lun)
         * Initiator Port.
         */
        spin_lock_irqsave(&lun->lun_cmd_lock, lun_flags);
-       while (!list_empty_careful(&lun->lun_cmd_list)) {
-               cmd = list_entry(lun->lun_cmd_list.next,
-                       struct se_cmd, se_lun_list);
-               list_del(&cmd->se_lun_list);
-
-               if (!(T_TASK(cmd))) {
-                       printk(KERN_ERR "ITT: 0x%08x, T_TASK(cmd) = NULL"
-                               "[i,t]_state: %u/%u\n",
-                               CMD_TFO(cmd)->get_task_tag(cmd),
-                               CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state);
-                       BUG();
-               }
-               atomic_set(&T_TASK(cmd)->transport_lun_active, 0);
+       while (!list_empty(&lun->lun_cmd_list)) {
+               cmd = list_first_entry(&lun->lun_cmd_list,
+                      struct se_cmd, se_lun_node);
+               list_del(&cmd->se_lun_node);
+
+               atomic_set(&cmd->transport_lun_active, 0);
                /*
                 * This will notify iscsi_target_transport.c:
                 * transport_cmd_check_stop() that a LUN shutdown is in
                 * progress for the iscsi_cmd_t.
                 */
-               spin_lock(&T_TASK(cmd)->t_state_lock);
-               DEBUG_CLEAR_L("SE_LUN[%d] - Setting T_TASK(cmd)->transport"
+               spin_lock(&cmd->t_state_lock);
+               pr_debug("SE_LUN[%d] - Setting cmd->transport"
                        "_lun_stop for  ITT: 0x%08x\n",
-                       SE_LUN(cmd)->unpacked_lun,
-                       CMD_TFO(cmd)->get_task_tag(cmd));
-               atomic_set(&T_TASK(cmd)->transport_lun_stop, 1);
-               spin_unlock(&T_TASK(cmd)->t_state_lock);
+                       cmd->se_lun->unpacked_lun,
+                       cmd->se_tfo->get_task_tag(cmd));
+               atomic_set(&cmd->transport_lun_stop, 1);
+               spin_unlock(&cmd->t_state_lock);
 
                spin_unlock_irqrestore(&lun->lun_cmd_lock, lun_flags);
 
-               if (!(SE_LUN(cmd))) {
-                       printk(KERN_ERR "ITT: 0x%08x, [i,t]_state: %u/%u\n",
-                               CMD_TFO(cmd)->get_task_tag(cmd),
-                               CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state);
+               if (!cmd->se_lun) {
+                       pr_err("ITT: 0x%08x, [i,t]_state: %u/%u\n",
+                               cmd->se_tfo->get_task_tag(cmd),
+                               cmd->se_tfo->get_cmd_state(cmd), cmd->t_state);
                        BUG();
                }
                /*
                 * If the Storage engine still owns the iscsi_cmd_t, determine
                 * and/or stop its context.
                 */
-               DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x before transport"
-                       "_lun_wait_for_tasks()\n", SE_LUN(cmd)->unpacked_lun,
-                       CMD_TFO(cmd)->get_task_tag(cmd));
+               pr_debug("SE_LUN[%d] - ITT: 0x%08x before transport"
+                       "_lun_wait_for_tasks()\n", cmd->se_lun->unpacked_lun,
+                       cmd->se_tfo->get_task_tag(cmd));
 
-               if (transport_lun_wait_for_tasks(cmd, SE_LUN(cmd)) < 0) {
+               if (transport_lun_wait_for_tasks(cmd, cmd->se_lun) < 0) {
                        spin_lock_irqsave(&lun->lun_cmd_lock, lun_flags);
                        continue;
                }
 
-               DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x after transport_lun"
+               pr_debug("SE_LUN[%d] - ITT: 0x%08x after transport_lun"
                        "_wait_for_tasks(): SUCCESS\n",
-                       SE_LUN(cmd)->unpacked_lun,
-                       CMD_TFO(cmd)->get_task_tag(cmd));
+                       cmd->se_lun->unpacked_lun,
+                       cmd->se_tfo->get_task_tag(cmd));
 
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, cmd_flags);
-               if (!(atomic_read(&T_TASK(cmd)->transport_dev_active))) {
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, cmd_flags);
+               spin_lock_irqsave(&cmd->t_state_lock, cmd_flags);
+               if (!atomic_read(&cmd->transport_dev_active)) {
+                       spin_unlock_irqrestore(&cmd->t_state_lock, cmd_flags);
                        goto check_cond;
                }
-               atomic_set(&T_TASK(cmd)->transport_dev_active, 0);
+               atomic_set(&cmd->transport_dev_active, 0);
                transport_all_task_dev_remove_state(cmd);
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, cmd_flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, cmd_flags);
 
                transport_free_dev_tasks(cmd);
                /*
@@ -5465,24 +4506,24 @@ check_cond:
                 * be released, notify the waiting thread now that LU has
                 * finished accessing it.
                 */
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, cmd_flags);
-               if (atomic_read(&T_TASK(cmd)->transport_lun_fe_stop)) {
-                       DEBUG_CLEAR_L("SE_LUN[%d] - Detected FE stop for"
+               spin_lock_irqsave(&cmd->t_state_lock, cmd_flags);
+               if (atomic_read(&cmd->transport_lun_fe_stop)) {
+                       pr_debug("SE_LUN[%d] - Detected FE stop for"
                                " struct se_cmd: %p ITT: 0x%08x\n",
                                lun->unpacked_lun,
-                               cmd, CMD_TFO(cmd)->get_task_tag(cmd));
+                               cmd, cmd->se_tfo->get_task_tag(cmd));
 
-                       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock,
+                       spin_unlock_irqrestore(&cmd->t_state_lock,
                                        cmd_flags);
                        transport_cmd_check_stop(cmd, 1, 0);
-                       complete(&T_TASK(cmd)->transport_lun_fe_stop_comp);
+                       complete(&cmd->transport_lun_fe_stop_comp);
                        spin_lock_irqsave(&lun->lun_cmd_lock, lun_flags);
                        continue;
                }
-               DEBUG_CLEAR_L("SE_LUN[%d] - ITT: 0x%08x finished processing\n",
-                       lun->unpacked_lun, CMD_TFO(cmd)->get_task_tag(cmd));
+               pr_debug("SE_LUN[%d] - ITT: 0x%08x finished processing\n",
+                       lun->unpacked_lun, cmd->se_tfo->get_task_tag(cmd));
 
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, cmd_flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, cmd_flags);
                spin_lock_irqsave(&lun->lun_cmd_lock, lun_flags);
        }
        spin_unlock_irqrestore(&lun->lun_cmd_lock, lun_flags);
@@ -5502,11 +4543,11 @@ int transport_clear_lun_from_sessions(struct se_lun *lun)
 {
        struct task_struct *kt;
 
-       kt = kthread_run(transport_clear_lun_thread, (void *)lun,
+       kt = kthread_run(transport_clear_lun_thread, lun,
                        "tcm_cl_%u", lun->unpacked_lun);
        if (IS_ERR(kt)) {
-               printk(KERN_ERR "Unable to start clear_lun thread\n");
-               return -1;
+               pr_err("Unable to start clear_lun thread\n");
+               return PTR_ERR(kt);
        }
        wait_for_completion(&lun->lun_shutdown_comp);
 
@@ -5528,20 +4569,20 @@ static void transport_generic_wait_for_tasks(
        if (!(cmd->se_cmd_flags & SCF_SE_LUN_CMD) && !(cmd->se_tmr_req))
                return;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        /*
         * If we are already stopped due to an external event (ie: LUN shutdown)
         * sleep until the connection can have the passed struct se_cmd back.
-        * The T_TASK(cmd)->transport_lun_stopped_sem will be upped by
+        * The cmd->transport_lun_stopped_sem will be upped by
         * transport_clear_lun_from_sessions() once the ConfigFS context caller
         * has completed its operation on the struct se_cmd.
         */
-       if (atomic_read(&T_TASK(cmd)->transport_lun_stop)) {
+       if (atomic_read(&cmd->transport_lun_stop)) {
 
-               DEBUG_TRANSPORT_S("wait_for_tasks: Stopping"
-                       " wait_for_completion(&T_TASK(cmd)transport_lun_fe"
+               pr_debug("wait_for_tasks: Stopping"
+                       " wait_for_completion(&cmd->t_tasktransport_lun_fe"
                        "_stop_comp); for ITT: 0x%08x\n",
-                       CMD_TFO(cmd)->get_task_tag(cmd));
+                       cmd->se_tfo->get_task_tag(cmd));
                /*
                 * There is a special case for WRITES where a FE exception +
                 * LUN shutdown means ConfigFS context is still sleeping on
@@ -5549,10 +4590,10 @@ static void transport_generic_wait_for_tasks(
                 * We go ahead and up transport_lun_stop_comp just to be sure
                 * here.
                 */
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
-               complete(&T_TASK(cmd)->transport_lun_stop_comp);
-               wait_for_completion(&T_TASK(cmd)->transport_lun_fe_stop_comp);
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
+               complete(&cmd->transport_lun_stop_comp);
+               wait_for_completion(&cmd->transport_lun_fe_stop_comp);
+               spin_lock_irqsave(&cmd->t_state_lock, flags);
 
                transport_all_task_dev_remove_state(cmd);
                /*
@@ -5560,44 +4601,44 @@ static void transport_generic_wait_for_tasks(
                 * struct se_cmd, now owns the structure and can be released through
                 * normal means below.
                 */
-               DEBUG_TRANSPORT_S("wait_for_tasks: Stopped"
-                       " wait_for_completion(&T_TASK(cmd)transport_lun_fe_"
+               pr_debug("wait_for_tasks: Stopped"
+                       " wait_for_completion(&cmd->t_tasktransport_lun_fe_"
                        "stop_comp); for ITT: 0x%08x\n",
-                       CMD_TFO(cmd)->get_task_tag(cmd));
+                       cmd->se_tfo->get_task_tag(cmd));
 
-               atomic_set(&T_TASK(cmd)->transport_lun_stop, 0);
+               atomic_set(&cmd->transport_lun_stop, 0);
        }
-       if (!atomic_read(&T_TASK(cmd)->t_transport_active) ||
-            atomic_read(&T_TASK(cmd)->t_transport_aborted))
+       if (!atomic_read(&cmd->t_transport_active) ||
+            atomic_read(&cmd->t_transport_aborted))
                goto remove;
 
-       atomic_set(&T_TASK(cmd)->t_transport_stop, 1);
+       atomic_set(&cmd->t_transport_stop, 1);
 
-       DEBUG_TRANSPORT_S("wait_for_tasks: Stopping %p ITT: 0x%08x"
+       pr_debug("wait_for_tasks: Stopping %p ITT: 0x%08x"
                " i_state: %d, t_state/def_t_state: %d/%d, t_transport_stop"
-               " = TRUE\n", cmd, CMD_TFO(cmd)->get_task_tag(cmd),
-               CMD_TFO(cmd)->get_cmd_state(cmd), cmd->t_state,
+               " = TRUE\n", cmd, cmd->se_tfo->get_task_tag(cmd),
+               cmd->se_tfo->get_cmd_state(cmd), cmd->t_state,
                cmd->deferred_t_state);
 
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
-       wake_up_interruptible(&SE_DEV(cmd)->dev_queue_obj->thread_wq);
+       wake_up_interruptible(&cmd->se_dev->dev_queue_obj.thread_wq);
 
-       wait_for_completion(&T_TASK(cmd)->t_transport_stop_comp);
+       wait_for_completion(&cmd->t_transport_stop_comp);
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-       atomic_set(&T_TASK(cmd)->t_transport_active, 0);
-       atomic_set(&T_TASK(cmd)->t_transport_stop, 0);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
+       atomic_set(&cmd->t_transport_active, 0);
+       atomic_set(&cmd->t_transport_stop, 0);
 
-       DEBUG_TRANSPORT_S("wait_for_tasks: Stopped wait_for_compltion("
-               "&T_TASK(cmd)->t_transport_stop_comp) for ITT: 0x%08x\n",
-               CMD_TFO(cmd)->get_task_tag(cmd));
+       pr_debug("wait_for_tasks: Stopped wait_for_compltion("
+               "&cmd->t_transport_stop_comp) for ITT: 0x%08x\n",
+               cmd->se_tfo->get_task_tag(cmd));
 remove:
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
        if (!remove_cmd)
                return;
 
-       transport_generic_free_cmd(cmd, 0, 0, session_reinstatement);
+       transport_generic_free_cmd(cmd, 0, session_reinstatement);
 }
 
 static int transport_get_sense_codes(
@@ -5632,13 +4673,13 @@ int transport_send_check_condition_and_sense(
        int offset;
        u8 asc = 0, ascq = 0;
 
-       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+       spin_lock_irqsave(&cmd->t_state_lock, flags);
        if (cmd->se_cmd_flags & SCF_SENT_CHECK_CONDITION) {
-               spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+               spin_unlock_irqrestore(&cmd->t_state_lock, flags);
                return 0;
        }
        cmd->se_cmd_flags |= SCF_SENT_CHECK_CONDITION;
-       spin_unlock_irqrestore(&T_TASK(cmd)->t_state_lock, flags);
+       spin_unlock_irqrestore(&cmd->t_state_lock, flags);
 
        if (!reason && from_transport)
                goto after_reason;
@@ -5651,7 +4692,7 @@ int transport_send_check_condition_and_sense(
         * TRANSPORT_SENSE_BUFFER is now set to SCSI_SENSE_BUFFERSIZE
         * from include/scsi/scsi_cmnd.h
         */
-       offset = CMD_TFO(cmd)->set_fabric_sense_len(cmd,
+       offset = cmd->se_tfo->set_fabric_sense_len(cmd,
                                TRANSPORT_SENSE_BUFFER);
        /*
         * Actual SENSE DATA, see SPC-3 7.23.2  SPC_SENSE_KEY_OFFSET uses
@@ -5788,8 +4829,7 @@ int transport_send_check_condition_and_sense(
        cmd->scsi_sense_length  = TRANSPORT_SENSE_BUFFER + offset;
 
 after_reason:
-       CMD_TFO(cmd)->queue_status(cmd);
-       return 0;
+       return cmd->se_tfo->queue_status(cmd);
 }
 EXPORT_SYMBOL(transport_send_check_condition_and_sense);
 
@@ -5797,18 +4837,18 @@ int transport_check_aborted_status(struct se_cmd *cmd, int send_status)
 {
        int ret = 0;
 
-       if (atomic_read(&T_TASK(cmd)->t_transport_aborted) != 0) {
-               if (!(send_status) ||
+       if (atomic_read(&cmd->t_transport_aborted) != 0) {
+               if (!send_status ||
                     (cmd->se_cmd_flags & SCF_SENT_DELAYED_TAS))
                        return 1;
 #if 0
-               printk(KERN_INFO "Sending delayed SAM_STAT_TASK_ABORTED"
+               pr_debug("Sending delayed SAM_STAT_TASK_ABORTED"
                        " status for CDB: 0x%02x ITT: 0x%08x\n",
-                       T_TASK(cmd)->t_task_cdb[0],
-                       CMD_TFO(cmd)->get_task_tag(cmd));
+                       cmd->t_task_cdb[0],
+                       cmd->se_tfo->get_task_tag(cmd));
 #endif
                cmd->se_cmd_flags |= SCF_SENT_DELAYED_TAS;
-               CMD_TFO(cmd)->queue_status(cmd);
+               cmd->se_tfo->queue_status(cmd);
                ret = 1;
        }
        return ret;
@@ -5824,8 +4864,8 @@ void transport_send_task_abort(struct se_cmd *cmd)
         * queued back to fabric module by transport_check_aborted_status().
         */
        if (cmd->data_direction == DMA_TO_DEVICE) {
-               if (CMD_TFO(cmd)->write_pending_status(cmd) != 0) {
-                       atomic_inc(&T_TASK(cmd)->t_transport_aborted);
+               if (cmd->se_tfo->write_pending_status(cmd) != 0) {
+                       atomic_inc(&cmd->t_transport_aborted);
                        smp_mb__after_atomic_inc();
                        cmd->scsi_status = SAM_STAT_TASK_ABORTED;
                        transport_new_cmd_failure(cmd);
@@ -5834,11 +4874,11 @@ void transport_send_task_abort(struct se_cmd *cmd)
        }
        cmd->scsi_status = SAM_STAT_TASK_ABORTED;
 #if 0
-       printk(KERN_INFO "Setting SAM_STAT_TASK_ABORTED status for CDB: 0x%02x,"
-               " ITT: 0x%08x\n", T_TASK(cmd)->t_task_cdb[0],
-               CMD_TFO(cmd)->get_task_tag(cmd));
+       pr_debug("Setting SAM_STAT_TASK_ABORTED status for CDB: 0x%02x,"
+               " ITT: 0x%08x\n", cmd->t_task_cdb[0],
+               cmd->se_tfo->get_task_tag(cmd));
 #endif
-       CMD_TFO(cmd)->queue_status(cmd);
+       cmd->se_tfo->queue_status(cmd);
 }
 
 /*     transport_generic_do_tmr():
@@ -5847,14 +4887,12 @@ void transport_send_task_abort(struct se_cmd *cmd)
  */
 int transport_generic_do_tmr(struct se_cmd *cmd)
 {
-       struct se_cmd *ref_cmd;
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct se_tmr_req *tmr = cmd->se_tmr_req;
        int ret;
 
        switch (tmr->function) {
        case TMR_ABORT_TASK:
-               ref_cmd = tmr->ref_cmd;
                tmr->response = TMR_FUNCTION_REJECTED;
                break;
        case TMR_ABORT_TASK_SET:
@@ -5874,14 +4912,14 @@ int transport_generic_do_tmr(struct se_cmd *cmd)
                tmr->response = TMR_FUNCTION_REJECTED;
                break;
        default:
-               printk(KERN_ERR "Uknown TMR function: 0x%02x.\n",
+               pr_err("Uknown TMR function: 0x%02x.\n",
                                tmr->function);
                tmr->response = TMR_FUNCTION_REJECTED;
                break;
        }
 
        cmd->t_state = TRANSPORT_ISTATE_PROCESSING;
-       CMD_TFO(cmd)->queue_tm_rsp(cmd);
+       cmd->se_tfo->queue_tm_rsp(cmd);
 
        transport_cmd_check_stop(cmd, 2, 0);
        return 0;
@@ -5911,62 +4949,54 @@ transport_get_task_from_state_list(struct se_device *dev)
 static void transport_processing_shutdown(struct se_device *dev)
 {
        struct se_cmd *cmd;
-       struct se_queue_req *qr;
        struct se_task *task;
-       u8 state;
        unsigned long flags;
        /*
         * Empty the struct se_device's struct se_task state list.
         */
        spin_lock_irqsave(&dev->execute_task_lock, flags);
        while ((task = transport_get_task_from_state_list(dev))) {
-               if (!(TASK_CMD(task))) {
-                       printk(KERN_ERR "TASK_CMD(task) is NULL!\n");
+               if (!task->task_se_cmd) {
+                       pr_err("task->task_se_cmd is NULL!\n");
                        continue;
                }
-               cmd = TASK_CMD(task);
+               cmd = task->task_se_cmd;
 
-               if (!T_TASK(cmd)) {
-                       printk(KERN_ERR "T_TASK(cmd) is NULL for task: %p cmd:"
-                               " %p ITT: 0x%08x\n", task, cmd,
-                               CMD_TFO(cmd)->get_task_tag(cmd));
-                       continue;
-               }
                spin_unlock_irqrestore(&dev->execute_task_lock, flags);
 
-               spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
+               spin_lock_irqsave(&cmd->t_state_lock, flags);
 
-               DEBUG_DO("PT: cmd: %p task: %p ITT/CmdSN: 0x%08x/0x%08x,"
-                       " i_state/def_i_state: %d/%d, t_state/def_t_state:"
+               pr_debug("PT: cmd: %p task: %p ITT: 0x%08x,"
+                       " i_state%d, t_state/def_t_state:"
                        " %d/%d cdb: 0x%02x\n", cmd, task,
-                       CMD_TFO(cmd)->get_task_tag(cmd), cmd->cmd_sn,
-                       CMD_TFO(cmd)->get_cmd_state(cmd), cmd->deferred_i_state,
+                       cmd->se_tfo->get_task_tag(cmd),
+                       cmd->se_tfo->get_cmd_state(cmd),
                        cmd->t_state, cmd->deferred_t_state,
-                       T_TASK(cmd)->t_task_cdb[0]);
-               DEBUG_DO("PT: ITT[0x%08x] - t_task_cdbs: %d t_task_cdbs_left:"
+                       cmd->t_task_cdb[0]);
+               pr_debug("PT: ITT[0x%08x] - t_tasks: %d t_task_cdbs_left:"
                        " %d t_task_cdbs_sent: %d -- t_transport_active: %d"
                        " t_transport_stop: %d t_transport_sent: %d\n",
-                       CMD_TFO(cmd)->get_task_tag(cmd),
-                       T_TASK(cmd)->t_task_cdbs,
-                       atomic_read(&T_TASK(cmd)->t_task_cdbs_left),
-                       atomic_read(&T_TASK(cmd)->t_task_cdbs_sent),
-                       atomic_read(&T_TASK(cmd)->t_transport_active),
-                       atomic_read(&T_TASK(cmd)->t_transport_stop),
-                       atomic_read(&T_TASK(cmd)->t_transport_sent));
+                       cmd->se_tfo->get_task_tag(cmd),
+                       cmd->t_task_list_num,
+                       atomic_read(&cmd->t_task_cdbs_left),
+                       atomic_read(&cmd->t_task_cdbs_sent),
+                       atomic_read(&cmd->t_transport_active),
+                       atomic_read(&cmd->t_transport_stop),
+                       atomic_read(&cmd->t_transport_sent));
 
                if (atomic_read(&task->task_active)) {
                        atomic_set(&task->task_stop, 1);
                        spin_unlock_irqrestore(
-                               &T_TASK(cmd)->t_state_lock, flags);
+                               &cmd->t_state_lock, flags);
 
-                       DEBUG_DO("Waiting for task: %p to shutdown for dev:"
+                       pr_debug("Waiting for task: %p to shutdown for dev:"
                                " %p\n", task, dev);
                        wait_for_completion(&task->task_stop_comp);
-                       DEBUG_DO("Completed task: %p shutdown for dev: %p\n",
+                       pr_debug("Completed task: %p shutdown for dev: %p\n",
                                task, dev);
 
-                       spin_lock_irqsave(&T_TASK(cmd)->t_state_lock, flags);
-                       atomic_dec(&T_TASK(cmd)->t_task_cdbs_left);
+                       spin_lock_irqsave(&cmd->t_state_lock, flags);
+                       atomic_dec(&cmd->t_task_cdbs_left);
 
                        atomic_set(&task->task_active, 0);
                        atomic_set(&task->task_stop, 0);
@@ -5976,72 +5006,72 @@ static void transport_processing_shutdown(struct se_device *dev)
                }
                __transport_stop_task_timer(task, &flags);
 
-               if (!(atomic_dec_and_test(&T_TASK(cmd)->t_task_cdbs_ex_left))) {
+               if (!atomic_dec_and_test(&cmd->t_task_cdbs_ex_left)) {
                        spin_unlock_irqrestore(
-                                       &T_TASK(cmd)->t_state_lock, flags);
+                                       &cmd->t_state_lock, flags);
 
-                       DEBUG_DO("Skipping task: %p, dev: %p for"
+                       pr_debug("Skipping task: %p, dev: %p for"
                                " t_task_cdbs_ex_left: %d\n", task, dev,
-                               atomic_read(&T_TASK(cmd)->t_task_cdbs_ex_left));
+                               atomic_read(&cmd->t_task_cdbs_ex_left));
 
                        spin_lock_irqsave(&dev->execute_task_lock, flags);
                        continue;
                }
 
-               if (atomic_read(&T_TASK(cmd)->t_transport_active)) {
-                       DEBUG_DO("got t_transport_active = 1 for task: %p, dev:"
+               if (atomic_read(&cmd->t_transport_active)) {
+                       pr_debug("got t_transport_active = 1 for task: %p, dev:"
                                        " %p\n", task, dev);
 
-                       if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+                       if (atomic_read(&cmd->t_fe_count)) {
                                spin_unlock_irqrestore(
-                                       &T_TASK(cmd)->t_state_lock, flags);
+                                       &cmd->t_state_lock, flags);
                                transport_send_check_condition_and_sense(
                                        cmd, TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE,
                                        0);
                                transport_remove_cmd_from_queue(cmd,
-                                       SE_DEV(cmd)->dev_queue_obj);
+                                       &cmd->se_dev->dev_queue_obj);
 
                                transport_lun_remove_cmd(cmd);
                                transport_cmd_check_stop(cmd, 1, 0);
                        } else {
                                spin_unlock_irqrestore(
-                                       &T_TASK(cmd)->t_state_lock, flags);
+                                       &cmd->t_state_lock, flags);
 
                                transport_remove_cmd_from_queue(cmd,
-                                       SE_DEV(cmd)->dev_queue_obj);
+                                       &cmd->se_dev->dev_queue_obj);
 
                                transport_lun_remove_cmd(cmd);
 
                                if (transport_cmd_check_stop(cmd, 1, 0))
-                                       transport_generic_remove(cmd, 0, 0);
+                                       transport_generic_remove(cmd, 0);
                        }
 
                        spin_lock_irqsave(&dev->execute_task_lock, flags);
                        continue;
                }
-               DEBUG_DO("Got t_transport_active = 0 for task: %p, dev: %p\n",
+               pr_debug("Got t_transport_active = 0 for task: %p, dev: %p\n",
                                task, dev);
 
-               if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+               if (atomic_read(&cmd->t_fe_count)) {
                        spin_unlock_irqrestore(
-                               &T_TASK(cmd)->t_state_lock, flags);
+                               &cmd->t_state_lock, flags);
                        transport_send_check_condition_and_sense(cmd,
                                TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
                        transport_remove_cmd_from_queue(cmd,
-                               SE_DEV(cmd)->dev_queue_obj);
+                               &cmd->se_dev->dev_queue_obj);
 
                        transport_lun_remove_cmd(cmd);
                        transport_cmd_check_stop(cmd, 1, 0);
                } else {
                        spin_unlock_irqrestore(
-                               &T_TASK(cmd)->t_state_lock, flags);
+                               &cmd->t_state_lock, flags);
 
                        transport_remove_cmd_from_queue(cmd,
-                               SE_DEV(cmd)->dev_queue_obj);
+                               &cmd->se_dev->dev_queue_obj);
                        transport_lun_remove_cmd(cmd);
 
                        if (transport_cmd_check_stop(cmd, 1, 0))
-                               transport_generic_remove(cmd, 0, 0);
+                               transport_generic_remove(cmd, 0);
                }
 
                spin_lock_irqsave(&dev->execute_task_lock, flags);
@@ -6050,18 +5080,12 @@ static void transport_processing_shutdown(struct se_device *dev)
        /*
         * Empty the struct se_device's struct se_cmd list.
         */
-       spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
-       while ((qr = __transport_get_qr_from_queue(dev->dev_queue_obj))) {
-               spin_unlock_irqrestore(
-                               &dev->dev_queue_obj->cmd_queue_lock, flags);
-               cmd = (struct se_cmd *)qr->cmd;
-               state = qr->state;
-               kfree(qr);
-
-               DEBUG_DO("From Device Queue: cmd: %p t_state: %d\n",
-                               cmd, state);
-
-               if (atomic_read(&T_TASK(cmd)->t_fe_count)) {
+       while ((cmd = transport_get_cmd_from_queue(&dev->dev_queue_obj))) {
+
+               pr_debug("From Device Queue: cmd: %p t_state: %d\n",
+                               cmd, cmd->t_state);
+
+               if (atomic_read(&cmd->t_fe_count)) {
                        transport_send_check_condition_and_sense(cmd,
                                TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
 
@@ -6070,11 +5094,9 @@ static void transport_processing_shutdown(struct se_device *dev)
                } else {
                        transport_lun_remove_cmd(cmd);
                        if (transport_cmd_check_stop(cmd, 1, 0))
-                               transport_generic_remove(cmd, 0, 0);
+                               transport_generic_remove(cmd, 0);
                }
-               spin_lock_irqsave(&dev->dev_queue_obj->cmd_queue_lock, flags);
        }
-       spin_unlock_irqrestore(&dev->dev_queue_obj->cmd_queue_lock, flags);
 }
 
 /*     transport_processing_thread():
@@ -6083,16 +5105,15 @@ static void transport_processing_shutdown(struct se_device *dev)
  */
 static int transport_processing_thread(void *param)
 {
-       int ret, t_state;
+       int ret;
        struct se_cmd *cmd;
        struct se_device *dev = (struct se_device *) param;
-       struct se_queue_req *qr;
 
        set_user_nice(current, -20);
 
        while (!kthread_should_stop()) {
-               ret = wait_event_interruptible(dev->dev_queue_obj->thread_wq,
-                               atomic_read(&dev->dev_queue_obj->queue_cnt) ||
+               ret = wait_event_interruptible(dev->dev_queue_obj.thread_wq,
+                               atomic_read(&dev->dev_queue_obj.queue_cnt) ||
                                kthread_should_stop());
                if (ret < 0)
                        goto out;
@@ -6108,22 +5129,18 @@ static int transport_processing_thread(void *param)
 get_cmd:
                __transport_execute_tasks(dev);
 
-               qr = transport_get_qr_from_queue(dev->dev_queue_obj);
-               if (!(qr))
+               cmd = transport_get_cmd_from_queue(&dev->dev_queue_obj);
+               if (!cmd)
                        continue;
 
-               cmd = (struct se_cmd *)qr->cmd;
-               t_state = qr->state;
-               kfree(qr);
-
-               switch (t_state) {
+               switch (cmd->t_state) {
                case TRANSPORT_NEW_CMD_MAP:
-                       if (!(CMD_TFO(cmd)->new_cmd_map)) {
-                               printk(KERN_ERR "CMD_TFO(cmd)->new_cmd_map is"
+                       if (!cmd->se_tfo->new_cmd_map) {
+                               pr_err("cmd->se_tfo->new_cmd_map is"
                                        " NULL for TRANSPORT_NEW_CMD_MAP\n");
                                BUG();
                        }
-                       ret = CMD_TFO(cmd)->new_cmd_map(cmd);
+                       ret = cmd->se_tfo->new_cmd_map(cmd);
                        if (ret < 0) {
                                cmd->transport_error_status = ret;
                                transport_generic_request_failure(cmd, NULL,
@@ -6134,7 +5151,9 @@ get_cmd:
                        /* Fall through */
                case TRANSPORT_NEW_CMD:
                        ret = transport_generic_new_cmd(cmd);
-                       if (ret < 0) {
+                       if (ret == -EAGAIN)
+                               break;
+                       else if (ret < 0) {
                                cmd->transport_error_status = ret;
                                transport_generic_request_failure(cmd, NULL,
                                        0, (cmd->data_direction !=
@@ -6149,10 +5168,10 @@ get_cmd:
                        transport_generic_complete_ok(cmd);
                        break;
                case TRANSPORT_REMOVE:
-                       transport_generic_remove(cmd, 1, 0);
+                       transport_generic_remove(cmd, 0);
                        break;
                case TRANSPORT_FREE_CMD_INTR:
-                       transport_generic_free_cmd(cmd, 0, 1, 0);
+                       transport_generic_free_cmd(cmd, 0, 0);
                        break;
                case TRANSPORT_PROCESS_TMR:
                        transport_generic_do_tmr(cmd);
@@ -6164,13 +5183,16 @@ get_cmd:
                        transport_stop_all_task_timers(cmd);
                        transport_generic_request_timeout(cmd);
                        break;
+               case TRANSPORT_COMPLETE_QF_WP:
+                       transport_generic_write_pending(cmd);
+                       break;
                default:
-                       printk(KERN_ERR "Unknown t_state: %d deferred_t_state:"
+                       pr_err("Unknown t_state: %d deferred_t_state:"
                                " %d for ITT: 0x%08x i_state: %d on SE LUN:"
-                               " %u\n", t_state, cmd->deferred_t_state,
-                               CMD_TFO(cmd)->get_task_tag(cmd),
-                               CMD_TFO(cmd)->get_cmd_state(cmd),
-                               SE_LUN(cmd)->unpacked_lun);
+                               " %u\n", cmd->t_state, cmd->deferred_t_state,
+                               cmd->se_tfo->get_task_tag(cmd),
+                               cmd->se_tfo->get_cmd_state(cmd),
+                               cmd->se_lun->unpacked_lun);
                        BUG();
                }
 
index df355176a377432b8784bc7085d397019ca9978d..31e3c652527e0d3999983f022fd1a5627a99ed63 100644 (file)
@@ -49,15 +49,15 @@ int core_scsi3_ua_check(
        struct se_session *sess = cmd->se_sess;
        struct se_node_acl *nacl;
 
-       if (!(sess))
+       if (!sess)
                return 0;
 
        nacl = sess->se_node_acl;
-       if (!(nacl))
+       if (!nacl)
                return 0;
 
        deve = &nacl->device_list[cmd->orig_fe_lun];
-       if (!(atomic_read(&deve->ua_count)))
+       if (!atomic_read(&deve->ua_count))
                return 0;
        /*
         * From sam4r14, section 5.14 Unit attention condition:
@@ -80,10 +80,10 @@ int core_scsi3_ua_check(
        case REQUEST_SENSE:
                return 0;
        default:
-               return -1;
+               return -EINVAL;
        }
 
-       return -1;
+       return -EINVAL;
 }
 
 int core_scsi3_ua_allocate(
@@ -97,13 +97,13 @@ int core_scsi3_ua_allocate(
        /*
         * PASSTHROUGH OPS
         */
-       if (!(nacl))
-               return -1;
+       if (!nacl)
+               return -EINVAL;
 
        ua = kmem_cache_zalloc(se_ua_cache, GFP_ATOMIC);
-       if (!(ua)) {
-               printk(KERN_ERR "Unable to allocate struct se_ua\n");
-               return -1;
+       if (!ua) {
+               pr_err("Unable to allocate struct se_ua\n");
+               return -ENOMEM;
        }
        INIT_LIST_HEAD(&ua->ua_dev_list);
        INIT_LIST_HEAD(&ua->ua_nacl_list);
@@ -177,9 +177,9 @@ int core_scsi3_ua_allocate(
        spin_unlock(&deve->ua_lock);
        spin_unlock_irq(&nacl->device_list_lock);
 
-       printk(KERN_INFO "[%s]: Allocated UNIT ATTENTION, mapped LUN: %u, ASC:"
+       pr_debug("[%s]: Allocated UNIT ATTENTION, mapped LUN: %u, ASC:"
                " 0x%02x, ASCQ: 0x%02x\n",
-               TPG_TFO(nacl->se_tpg)->get_fabric_name(), unpacked_lun,
+               nacl->se_tpg->se_tpg_tfo->get_fabric_name(), unpacked_lun,
                asc, ascq);
 
        atomic_inc(&deve->ua_count);
@@ -208,23 +208,23 @@ void core_scsi3_ua_for_check_condition(
        u8 *asc,
        u8 *ascq)
 {
-       struct se_device *dev = SE_DEV(cmd);
+       struct se_device *dev = cmd->se_dev;
        struct se_dev_entry *deve;
        struct se_session *sess = cmd->se_sess;
        struct se_node_acl *nacl;
        struct se_ua *ua = NULL, *ua_p;
        int head = 1;
 
-       if (!(sess))
+       if (!sess)
                return;
 
        nacl = sess->se_node_acl;
-       if (!(nacl))
+       if (!nacl)
                return;
 
        spin_lock_irq(&nacl->device_list_lock);
        deve = &nacl->device_list[cmd->orig_fe_lun];
-       if (!(atomic_read(&deve->ua_count))) {
+       if (!atomic_read(&deve->ua_count)) {
                spin_unlock_irq(&nacl->device_list_lock);
                return;
        }
@@ -240,7 +240,7 @@ void core_scsi3_ua_for_check_condition(
                 * highest priority UNIT_ATTENTION and ASC/ASCQ without
                 * clearing it.
                 */
-               if (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl != 0) {
+               if (dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl != 0) {
                        *asc = ua->ua_asc;
                        *ascq = ua->ua_ascq;
                        break;
@@ -264,13 +264,13 @@ void core_scsi3_ua_for_check_condition(
        spin_unlock(&deve->ua_lock);
        spin_unlock_irq(&nacl->device_list_lock);
 
-       printk(KERN_INFO "[%s]: %s UNIT ATTENTION condition with"
+       pr_debug("[%s]: %s UNIT ATTENTION condition with"
                " INTLCK_CTRL: %d, mapped LUN: %u, got CDB: 0x%02x"
                " reported ASC: 0x%02x, ASCQ: 0x%02x\n",
-               TPG_TFO(nacl->se_tpg)->get_fabric_name(),
-               (DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl != 0) ? "Reporting" :
-               "Releasing", DEV_ATTRIB(dev)->emulate_ua_intlck_ctrl,
-               cmd->orig_fe_lun, T_TASK(cmd)->t_task_cdb[0], *asc, *ascq);
+               nacl->se_tpg->se_tpg_tfo->get_fabric_name(),
+               (dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl != 0) ? "Reporting" :
+               "Releasing", dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl,
+               cmd->orig_fe_lun, cmd->t_task_cdb[0], *asc, *ascq);
 }
 
 int core_scsi3_ua_clear_for_request_sense(
@@ -284,18 +284,18 @@ int core_scsi3_ua_clear_for_request_sense(
        struct se_ua *ua = NULL, *ua_p;
        int head = 1;
 
-       if (!(sess))
-               return -1;
+       if (!sess)
+               return -EINVAL;
 
        nacl = sess->se_node_acl;
-       if (!(nacl))
-               return -1;
+       if (!nacl)
+               return -EINVAL;
 
        spin_lock_irq(&nacl->device_list_lock);
        deve = &nacl->device_list[cmd->orig_fe_lun];
-       if (!(atomic_read(&deve->ua_count))) {
+       if (!atomic_read(&deve->ua_count)) {
                spin_unlock_irq(&nacl->device_list_lock);
-               return -1;
+               return -EPERM;
        }
        /*
         * The highest priority Unit Attentions are placed at the head of the
@@ -323,10 +323,10 @@ int core_scsi3_ua_clear_for_request_sense(
        spin_unlock(&deve->ua_lock);
        spin_unlock_irq(&nacl->device_list_lock);
 
-       printk(KERN_INFO "[%s]: Released UNIT ATTENTION condition, mapped"
+       pr_debug("[%s]: Released UNIT ATTENTION condition, mapped"
                " LUN: %u, got REQUEST_SENSE reported ASC: 0x%02x,"
-               " ASCQ: 0x%02x\n", TPG_TFO(nacl->se_tpg)->get_fabric_name(),
+               " ASCQ: 0x%02x\n", nacl->se_tpg->se_tpg_tfo->get_fabric_name(),
                cmd->orig_fe_lun, *asc, *ascq);
 
-       return (head) ? -1 : 0;
+       return (head) ? -EPERM : 0;
 }
index 7a5c2b64cf65e208cf6c9061e75891f8e101bf33..20b14bb087c95f9924db6bab384cc48f6d5dfa8a 100644 (file)
@@ -1,15 +1,6 @@
-EXTRA_CFLAGS += -I$(srctree)/drivers/target/ \
-               -I$(srctree)/drivers/scsi/ \
-               -I$(srctree)/include/scsi/ \
-               -I$(srctree)/drivers/target/tcm_fc/
-
-tcm_fc-y +=    tfc_cmd.o \
-               tfc_conf.o \
-               tfc_io.o \
-               tfc_sess.o
+tcm_fc-y +=            tfc_cmd.o \
+                       tfc_conf.o \
+                       tfc_io.o \
+                       tfc_sess.o
 
 obj-$(CONFIG_TCM_FC)   += tcm_fc.o
-
-ifdef CONFIGFS_TCM_FC_DEBUG
-EXTRA_CFLAGS   += -DTCM_FC_DEBUG
-endif
index 7b82f1b7fef814b595f8cc266f182baa3f031642..f7fff7ed63c36ac98fa5eaa01439a3d20d1afb37 100644 (file)
 #define FT_TPG_NAMELEN 32      /* max length of TPG name */
 #define FT_LUN_NAMELEN 32      /* max length of LUN name */
 
-/*
- * Debug options.
- */
-#define FT_DEBUG_CONF  0x01    /* configuration messages */
-#define        FT_DEBUG_SESS   0x02    /* session messages */
-#define        FT_DEBUG_TM     0x04    /* TM operations */
-#define        FT_DEBUG_IO     0x08    /* I/O commands */
-#define        FT_DEBUG_DATA   0x10    /* Data transfer */
-
-extern unsigned int ft_debug_logging;  /* debug options */
-
-#define FT_DEBUG(mask, fmt, args...)                                   \
-       do {                                                            \
-               if (ft_debug_logging & (mask))                          \
-                       printk(KERN_INFO "tcm_fc: %s: " fmt,            \
-                               __func__, ##args);                      \
-       } while (0)
-
-#define        FT_CONF_DBG(fmt, args...)       FT_DEBUG(FT_DEBUG_CONF, fmt, ##args)
-#define        FT_SESS_DBG(fmt, args...)       FT_DEBUG(FT_DEBUG_SESS, fmt, ##args)
-#define        FT_TM_DBG(fmt, args...)         FT_DEBUG(FT_DEBUG_TM, fmt, ##args)
-#define        FT_IO_DBG(fmt, args...)         FT_DEBUG(FT_DEBUG_IO, fmt, ##args)
-#define        FT_DATA_DBG(fmt, args...)       FT_DEBUG(FT_DEBUG_DATA, fmt, ##args)
-
 struct ft_transport_id {
        __u8    format;
        __u8    __resvd1[7];
@@ -195,7 +171,6 @@ int ft_write_pending(struct se_cmd *);
 int ft_write_pending_status(struct se_cmd *);
 u32 ft_get_task_tag(struct se_cmd *);
 int ft_get_cmd_state(struct se_cmd *);
-void ft_new_cmd_failure(struct se_cmd *);
 int ft_queue_tm_resp(struct se_cmd *);
 int ft_is_state_remove(struct se_cmd *);
 
index b2a106729d4914eca2c275adb956a944939d9b99..a9e9a31da11d7cd433a6b12d68628f910642cc84 100644 (file)
@@ -59,33 +59,30 @@ void ft_dump_cmd(struct ft_cmd *cmd, const char *caller)
        struct fc_exch *ep;
        struct fc_seq *sp;
        struct se_cmd *se_cmd;
-       struct se_mem *mem;
-       struct se_transport_task *task;
-
-       if (!(ft_debug_logging & FT_DEBUG_IO))
-               return;
+       struct scatterlist *sg;
+       int count;
 
        se_cmd = &cmd->se_cmd;
-       printk(KERN_INFO "%s: cmd %p state %d sess %p seq %p se_cmd %p\n",
+       pr_debug("%s: cmd %p state %d sess %p seq %p se_cmd %p\n",
                caller, cmd, cmd->state, cmd->sess, cmd->seq, se_cmd);
-       printk(KERN_INFO "%s: cmd %p cdb %p\n",
+       pr_debug("%s: cmd %p cdb %p\n",
                caller, cmd, cmd->cdb);
-       printk(KERN_INFO "%s: cmd %p lun %d\n", caller, cmd, cmd->lun);
-
-       task = T_TASK(se_cmd);
-       printk(KERN_INFO "%s: cmd %p task %p se_num %u buf %p len %u se_cmd_flags <0x%x>\n",
-              caller, cmd, task, task->t_tasks_se_num,
-              task->t_task_buf, se_cmd->data_length, se_cmd->se_cmd_flags);
-       if (task->t_mem_list)
-               list_for_each_entry(mem, task->t_mem_list, se_list)
-                       printk(KERN_INFO "%s: cmd %p mem %p page %p "
-                              "len 0x%x off 0x%x\n",
-                              caller, cmd, mem,
-                              mem->se_page, mem->se_len, mem->se_off);
+       pr_debug("%s: cmd %p lun %d\n", caller, cmd, cmd->lun);
+
+       pr_debug("%s: cmd %p data_nents %u len %u se_cmd_flags <0x%x>\n",
+               caller, cmd, se_cmd->t_data_nents,
+              se_cmd->data_length, se_cmd->se_cmd_flags);
+
+       for_each_sg(se_cmd->t_data_sg, sg, se_cmd->t_data_nents, count)
+               pr_debug("%s: cmd %p sg %p page %p "
+                       "len 0x%x off 0x%x\n",
+                       caller, cmd, sg,
+                       sg_page(sg), sg->length, sg->offset);
+
        sp = cmd->seq;
        if (sp) {
                ep = fc_seq_exch(sp);
-               printk(KERN_INFO "%s: cmd %p sid %x did %x "
+               pr_debug("%s: cmd %p sid %x did %x "
                        "ox_id %x rx_id %x seq_id %x e_stat %x\n",
                        caller, cmd, ep->sid, ep->did, ep->oxid, ep->rxid,
                        sp->id, ep->esb_stat);
@@ -96,15 +93,17 @@ void ft_dump_cmd(struct ft_cmd *cmd, const char *caller)
 
 static void ft_queue_cmd(struct ft_sess *sess, struct ft_cmd *cmd)
 {
-       struct se_queue_obj *qobj;
+       struct ft_tpg *tpg = sess->tport->tpg;
+       struct se_queue_obj *qobj = &tpg->qobj;
        unsigned long flags;
 
        qobj = &sess->tport->tpg->qobj;
        spin_lock_irqsave(&qobj->cmd_queue_lock, flags);
        list_add_tail(&cmd->se_req.qr_list, &qobj->qobj_list);
-       spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
        atomic_inc(&qobj->queue_cnt);
-       wake_up_interruptible(&qobj->thread_wq);
+       spin_unlock_irqrestore(&qobj->cmd_queue_lock, flags);
+
+       wake_up_process(tpg->thread);
 }
 
 static struct ft_cmd *ft_dequeue_cmd(struct se_queue_obj *qobj)
@@ -149,7 +148,7 @@ void ft_release_cmd(struct se_cmd *se_cmd)
 
 void ft_check_stop_free(struct se_cmd *se_cmd)
 {
-       transport_generic_free_cmd(se_cmd, 0, 1, 0);
+       transport_generic_free_cmd(se_cmd, 0, 0);
 }
 
 /*
@@ -256,15 +255,14 @@ int ft_write_pending(struct se_cmd *se_cmd)
                    (fh->fh_r_ctl == FC_RCTL_DD_DATA_DESC)) {
                        if (se_cmd->se_cmd_flags & SCF_SCSI_DATA_SG_IO_CDB) {
                                /*
-                                * Map se_mem list to scatterlist, so that
-                                * DDP can be setup. DDP setup function require
-                                * scatterlist. se_mem_list is internal to
-                                * TCM/LIO target
+                                * cmd may have been broken up into multiple
+                                * tasks. Link their sgs together so we can
+                                * operate on them all at once.
                                 */
                                transport_do_task_sg_chain(se_cmd);
-                               cmd->sg = T_TASK(se_cmd)->t_tasks_sg_chained;
+                               cmd->sg = se_cmd->t_tasks_sg_chained;
                                cmd->sg_cnt =
-                                       T_TASK(se_cmd)->t_tasks_sg_chained_no;
+                                       se_cmd->t_tasks_sg_chained_no;
                        }
                        if (cmd->sg && lport->tt.ddp_setup(lport, ep->xid,
                                                    cmd->sg, cmd->sg_cnt))
@@ -294,12 +292,6 @@ int ft_is_state_remove(struct se_cmd *se_cmd)
        return 0;       /* XXX TBD */
 }
 
-void ft_new_cmd_failure(struct se_cmd *se_cmd)
-{
-       /* XXX TBD */
-       printk(KERN_INFO "%s: se_cmd %p\n", __func__, se_cmd);
-}
-
 /*
  * FC sequence response handler for follow-on sequences (data) and aborts.
  */
@@ -312,7 +304,7 @@ static void ft_recv_seq(struct fc_seq *sp, struct fc_frame *fp, void *arg)
                /* XXX need to find cmd if queued */
                cmd->se_cmd.t_state = TRANSPORT_REMOVE;
                cmd->seq = NULL;
-               transport_generic_free_cmd(&cmd->se_cmd, 0, 1, 0);
+               transport_generic_free_cmd(&cmd->se_cmd, 0, 0);
                return;
        }
 
@@ -326,10 +318,10 @@ static void ft_recv_seq(struct fc_seq *sp, struct fc_frame *fp, void *arg)
        case FC_RCTL_DD_SOL_CTL:        /* transfer ready */
        case FC_RCTL_DD_DATA_DESC:      /* transfer ready */
        default:
-               printk(KERN_INFO "%s: unhandled frame r_ctl %x\n",
+               pr_debug("%s: unhandled frame r_ctl %x\n",
                       __func__, fh->fh_r_ctl);
                fc_frame_free(fp);
-               transport_generic_free_cmd(&cmd->se_cmd, 0, 1, 0);
+               transport_generic_free_cmd(&cmd->se_cmd, 0, 0);
                break;
        }
 }
@@ -351,7 +343,7 @@ static void ft_send_resp_status(struct fc_lport *lport,
        struct fcp_resp_rsp_info *info;
 
        fh = fc_frame_header_get(rx_fp);
-       FT_IO_DBG("FCP error response: did %x oxid %x status %x code %x\n",
+       pr_debug("FCP error response: did %x oxid %x status %x code %x\n",
                  ntoh24(fh->fh_s_id), ntohs(fh->fh_ox_id), status, code);
        len = sizeof(*fcp);
        if (status == SAM_STAT_GOOD)
@@ -421,15 +413,15 @@ static void ft_send_tm(struct ft_cmd *cmd)
                 * FCP4r01 indicates having a combination of
                 * tm_flags set is invalid.
                 */
-               FT_TM_DBG("invalid FCP tm_flags %x\n", fcp->fc_tm_flags);
+               pr_debug("invalid FCP tm_flags %x\n", fcp->fc_tm_flags);
                ft_send_resp_code(cmd, FCP_CMND_FIELDS_INVALID);
                return;
        }
 
-       FT_TM_DBG("alloc tm cmd fn %d\n", tm_func);
+       pr_debug("alloc tm cmd fn %d\n", tm_func);
        tmr = core_tmr_alloc_req(&cmd->se_cmd, cmd, tm_func);
        if (!tmr) {
-               FT_TM_DBG("alloc failed\n");
+               pr_debug("alloc failed\n");
                ft_send_resp_code(cmd, FCP_TMF_FAILED);
                return;
        }
@@ -438,20 +430,20 @@ static void ft_send_tm(struct ft_cmd *cmd)
        switch (fcp->fc_tm_flags) {
        case FCP_TMF_LUN_RESET:
                cmd->lun = scsilun_to_int((struct scsi_lun *)fcp->fc_lun);
-               if (transport_get_lun_for_tmr(&cmd->se_cmd, cmd->lun) < 0) {
+               if (transport_lookup_tmr_lun(&cmd->se_cmd, cmd->lun) < 0) {
                        /*
                         * Make sure to clean up newly allocated TMR request
                         * since "unable to  handle TMR request because failed
                         * to get to LUN"
                         */
-                       FT_TM_DBG("Failed to get LUN for TMR func %d, "
+                       pr_debug("Failed to get LUN for TMR func %d, "
                                  "se_cmd %p, unpacked_lun %d\n",
                                  tm_func, &cmd->se_cmd, cmd->lun);
                        ft_dump_cmd(cmd, __func__);
                        sess = cmd->sess;
                        transport_send_check_condition_and_sense(&cmd->se_cmd,
                                cmd->se_cmd.scsi_sense_reason, 0);
-                       transport_generic_free_cmd(&cmd->se_cmd, 0, 1, 0);
+                       transport_generic_free_cmd(&cmd->se_cmd, 0, 0);
                        ft_sess_put(sess);
                        return;
                }
@@ -495,7 +487,7 @@ int ft_queue_tm_resp(struct se_cmd *se_cmd)
                code = FCP_TMF_FAILED;
                break;
        }
-       FT_TM_DBG("tmr fn %d resp %d fcp code %d\n",
+       pr_debug("tmr fn %d resp %d fcp code %d\n",
                  tmr->function, tmr->response, code);
        ft_send_resp_code(cmd, code);
        return 0;
@@ -523,7 +515,7 @@ static void ft_recv_cmd(struct ft_sess *sess, struct fc_frame *fp)
        return;
 
 busy:
-       FT_IO_DBG("cmd or seq allocation failure - sending BUSY\n");
+       pr_debug("cmd or seq allocation failure - sending BUSY\n");
        ft_send_resp_status(lport, fp, SAM_STAT_BUSY, 0);
        fc_frame_free(fp);
        ft_sess_put(sess);              /* undo get from lookup */
@@ -548,7 +540,7 @@ void ft_recv_req(struct ft_sess *sess, struct fc_frame *fp)
        case FC_RCTL_DD_DATA_DESC:      /* transfer ready */
        case FC_RCTL_ELS4_REQ:          /* SRR, perhaps */
        default:
-               printk(KERN_INFO "%s: unhandled frame r_ctl %x\n",
+               pr_debug("%s: unhandled frame r_ctl %x\n",
                       __func__, fh->fh_r_ctl);
                fc_frame_free(fp);
                ft_sess_put(sess);      /* undo get from lookup */
@@ -637,7 +629,7 @@ static void ft_send_cmd(struct ft_cmd *cmd)
        fc_seq_exch(cmd->seq)->lp->tt.seq_set_resp(cmd->seq, ft_recv_seq, cmd);
 
        cmd->lun = scsilun_to_int((struct scsi_lun *)fcp->fc_lun);
-       ret = transport_get_lun_for_cmd(&cmd->se_cmd, NULL, cmd->lun);
+       ret = transport_lookup_cmd_lun(&cmd->se_cmd, cmd->lun);
        if (ret < 0) {
                ft_dump_cmd(cmd, __func__);
                transport_send_check_condition_and_sense(&cmd->se_cmd,
@@ -647,22 +639,22 @@ static void ft_send_cmd(struct ft_cmd *cmd)
 
        ret = transport_generic_allocate_tasks(se_cmd, cmd->cdb);
 
-       FT_IO_DBG("r_ctl %x alloc task ret %d\n", fh->fh_r_ctl, ret);
+       pr_debug("r_ctl %x alloc task ret %d\n", fh->fh_r_ctl, ret);
        ft_dump_cmd(cmd, __func__);
 
-       if (ret == -1) {
+       if (ret == -ENOMEM) {
                transport_send_check_condition_and_sense(se_cmd,
                                TCM_LOGICAL_UNIT_COMMUNICATION_FAILURE, 0);
-               transport_generic_free_cmd(se_cmd, 0, 1, 0);
+               transport_generic_free_cmd(se_cmd, 0, 0);
                return;
        }
-       if (ret == -2) {
+       if (ret == -EINVAL) {
                if (se_cmd->se_cmd_flags & SCF_SCSI_RESERVATION_CONFLICT)
                        ft_queue_status(se_cmd);
                else
                        transport_send_check_condition_and_sense(se_cmd,
                                        se_cmd->scsi_sense_reason, 0);
-               transport_generic_free_cmd(se_cmd, 0, 1, 0);
+               transport_generic_free_cmd(se_cmd, 0, 0);
                return;
        }
        transport_generic_handle_cdb(se_cmd);
@@ -670,7 +662,6 @@ static void ft_send_cmd(struct ft_cmd *cmd)
 
 err:
        ft_send_resp_code(cmd, FCP_CMND_FIELDS_INVALID);
-       return;
 }
 
 /*
@@ -678,7 +669,7 @@ err:
  */
 static void ft_exec_req(struct ft_cmd *cmd)
 {
-       FT_IO_DBG("cmd state %x\n", cmd->state);
+       pr_debug("cmd state %x\n", cmd->state);
        switch (cmd->state) {
        case FC_CMD_ST_NEW:
                ft_send_cmd(cmd);
@@ -697,15 +688,12 @@ int ft_thread(void *arg)
        struct ft_tpg *tpg = arg;
        struct se_queue_obj *qobj = &tpg->qobj;
        struct ft_cmd *cmd;
-       int ret;
-
-       set_user_nice(current, -20);
 
        while (!kthread_should_stop()) {
-               ret = wait_event_interruptible(qobj->thread_wq,
-                       atomic_read(&qobj->queue_cnt) || kthread_should_stop());
-               if (ret < 0 || kthread_should_stop())
+               schedule_timeout_interruptible(MAX_SCHEDULE_TIMEOUT);
+               if (kthread_should_stop())
                        goto out;
+
                cmd = ft_dequeue_cmd(qobj);
                if (cmd)
                        ft_exec_req(cmd);
index 84e868c255dd1d24177dfcda58f3276cc5bfe7f6..d63e3dd3b180d093ddf542d23326a297e666b6e8 100644 (file)
@@ -106,7 +106,7 @@ static ssize_t ft_parse_wwn(const char *name, u64 *wwn, int strict)
        }
        err = 4;
 fail:
-       FT_CONF_DBG("err %u len %zu pos %u byte %u\n",
+       pr_debug("err %u len %zu pos %u byte %u\n",
                    err, cp - name, pos, byte);
        return -1;
 }
@@ -216,14 +216,14 @@ static struct se_node_acl *ft_add_acl(
        u64 wwpn;
        u32 q_depth;
 
-       FT_CONF_DBG("add acl %s\n", name);
+       pr_debug("add acl %s\n", name);
        tpg = container_of(se_tpg, struct ft_tpg, se_tpg);
 
        if (ft_parse_wwn(name, &wwpn, 1) < 0)
                return ERR_PTR(-EINVAL);
 
        acl = kzalloc(sizeof(struct ft_node_acl), GFP_KERNEL);
-       if (!(acl))
+       if (!acl)
                return ERR_PTR(-ENOMEM);
        acl->node_auth.port_name = wwpn;
 
@@ -239,11 +239,11 @@ static void ft_del_acl(struct se_node_acl *se_acl)
        struct ft_node_acl *acl = container_of(se_acl,
                                struct ft_node_acl, se_node_acl);
 
-       FT_CONF_DBG("del acl %s\n",
+       pr_debug("del acl %s\n",
                config_item_name(&se_acl->acl_group.cg_item));
 
        tpg = container_of(se_tpg, struct ft_tpg, se_tpg);
-       FT_CONF_DBG("del acl %p se_acl %p tpg %p se_tpg %p\n",
+       pr_debug("del acl %p se_acl %p tpg %p se_tpg %p\n",
                    acl, se_acl, tpg, &tpg->se_tpg);
 
        core_tpg_del_initiator_node_acl(&tpg->se_tpg, se_acl, 1);
@@ -260,11 +260,11 @@ struct ft_node_acl *ft_acl_get(struct ft_tpg *tpg, struct fc_rport_priv *rdata)
        spin_lock_bh(&se_tpg->acl_node_lock);
        list_for_each_entry(se_acl, &se_tpg->acl_node_list, acl_list) {
                acl = container_of(se_acl, struct ft_node_acl, se_node_acl);
-               FT_CONF_DBG("acl %p port_name %llx\n",
+               pr_debug("acl %p port_name %llx\n",
                        acl, (unsigned long long)acl->node_auth.port_name);
                if (acl->node_auth.port_name == rdata->ids.port_name ||
                    acl->node_auth.node_name == rdata->ids.node_name) {
-                       FT_CONF_DBG("acl %p port_name %llx matched\n", acl,
+                       pr_debug("acl %p port_name %llx matched\n", acl,
                                    (unsigned long long)rdata->ids.port_name);
                        found = acl;
                        /* XXX need to hold onto ACL */
@@ -280,11 +280,11 @@ struct se_node_acl *ft_tpg_alloc_fabric_acl(struct se_portal_group *se_tpg)
        struct ft_node_acl *acl;
 
        acl = kzalloc(sizeof(*acl), GFP_KERNEL);
-       if (!(acl)) {
-               printk(KERN_ERR "Unable to allocate struct ft_node_acl\n");
+       if (!acl) {
+               pr_err("Unable to allocate struct ft_node_acl\n");
                return NULL;
        }
-       FT_CONF_DBG("acl %p\n", acl);
+       pr_debug("acl %p\n", acl);
        return &acl->se_node_acl;
 }
 
@@ -294,7 +294,7 @@ static void ft_tpg_release_fabric_acl(struct se_portal_group *se_tpg,
        struct ft_node_acl *acl = container_of(se_acl,
                                struct ft_node_acl, se_node_acl);
 
-       FT_CONF_DBG(KERN_INFO "acl %p\n", acl);
+       pr_debug("acl %p\n", acl);
        kfree(acl);
 }
 
@@ -311,7 +311,7 @@ static struct se_portal_group *ft_add_tpg(
        unsigned long index;
        int ret;
 
-       FT_CONF_DBG("tcm_fc: add tpg %s\n", name);
+       pr_debug("tcm_fc: add tpg %s\n", name);
 
        /*
         * Name must be "tpgt_" followed by the index.
@@ -331,7 +331,7 @@ static struct se_portal_group *ft_add_tpg(
        transport_init_queue_obj(&tpg->qobj);
 
        ret = core_tpg_register(&ft_configfs->tf_ops, wwn, &tpg->se_tpg,
-                               (void *)tpg, TRANSPORT_TPG_TYPE_NORMAL);
+                               tpg, TRANSPORT_TPG_TYPE_NORMAL);
        if (ret < 0) {
                kfree(tpg);
                return NULL;
@@ -354,7 +354,7 @@ static void ft_del_tpg(struct se_portal_group *se_tpg)
 {
        struct ft_tpg *tpg = container_of(se_tpg, struct ft_tpg, se_tpg);
 
-       FT_CONF_DBG("del tpg %s\n",
+       pr_debug("del tpg %s\n",
                    config_item_name(&tpg->se_tpg.tpg_group.cg_item));
 
        kthread_stop(tpg->thread);
@@ -412,7 +412,7 @@ static struct se_wwn *ft_add_lport(
        struct ft_lport_acl *old_lacl;
        u64 wwpn;
 
-       FT_CONF_DBG("add lport %s\n", name);
+       pr_debug("add lport %s\n", name);
        if (ft_parse_wwn(name, &wwpn, 1) < 0)
                return NULL;
        lacl = kzalloc(sizeof(*lacl), GFP_KERNEL);
@@ -441,7 +441,7 @@ static void ft_del_lport(struct se_wwn *wwn)
        struct ft_lport_acl *lacl = container_of(wwn,
                                struct ft_lport_acl, fc_lport_wwn);
 
-       FT_CONF_DBG("del lport %s\n",
+       pr_debug("del lport %s\n",
                        config_item_name(&wwn->wwn_group.cg_item));
        mutex_lock(&ft_lport_lock);
        list_del(&lacl->list);
@@ -536,8 +536,7 @@ static struct target_core_fabric_ops ft_fabric_ops = {
        .tpg_release_fabric_acl =       ft_tpg_release_fabric_acl,
        .tpg_get_inst_index =           ft_tpg_get_inst_index,
        .check_stop_free =              ft_check_stop_free,
-       .release_cmd_to_pool =          ft_release_cmd,
-       .release_cmd_direct =           ft_release_cmd,
+       .release_cmd =                  ft_release_cmd,
        .shutdown_session =             ft_sess_shutdown,
        .close_session =                ft_sess_close,
        .stop_session =                 ft_sess_stop,
@@ -550,7 +549,6 @@ static struct target_core_fabric_ops ft_fabric_ops = {
        .set_default_node_attributes =  ft_set_default_node_attr,
        .get_task_tag =                 ft_get_task_tag,
        .get_cmd_state =                ft_get_cmd_state,
-       .new_cmd_failure =              ft_new_cmd_failure,
        .queue_data_in =                ft_queue_data_in,
        .queue_status =                 ft_queue_status,
        .queue_tm_rsp =                 ft_queue_tm_resp,
@@ -582,10 +580,10 @@ int ft_register_configfs(void)
         * Register the top level struct config_item_type with TCM core
         */
        fabric = target_fabric_configfs_init(THIS_MODULE, "fc");
-       if (!fabric) {
-               printk(KERN_INFO "%s: target_fabric_configfs_init() failed!\n",
+       if (IS_ERR(fabric)) {
+               pr_err("%s: target_fabric_configfs_init() failed!\n",
                       __func__);
-               return -1;
+               return PTR_ERR(fabric);
        }
        fabric->tf_ops = ft_fabric_ops;
 
@@ -610,11 +608,8 @@ int ft_register_configfs(void)
         */
        ret = target_fabric_configfs_register(fabric);
        if (ret < 0) {
-               FT_CONF_DBG("target_fabric_configfs_register() for"
+               pr_debug("target_fabric_configfs_register() for"
                            " FC Target failed!\n");
-               printk(KERN_INFO
-                      "%s: target_fabric_configfs_register() failed!\n",
-                      __func__);
                target_fabric_configfs_free(fabric);
                return -1;
        }
index 8c4a24077d9d701fc6902501f9ad9f53775c24b2..11e6483fc1276013e92234cb1c19538d6b09994a 100644 (file)
@@ -39,6 +39,7 @@
 #include <linux/configfs.h>
 #include <linux/ctype.h>
 #include <linux/hash.h>
+#include <linux/ratelimit.h>
 #include <asm/unaligned.h>
 #include <scsi/scsi.h>
 #include <scsi/scsi_host.h>
 int ft_queue_data_in(struct se_cmd *se_cmd)
 {
        struct ft_cmd *cmd = container_of(se_cmd, struct ft_cmd, se_cmd);
-       struct se_transport_task *task;
        struct fc_frame *fp = NULL;
        struct fc_exch *ep;
        struct fc_lport *lport;
-       struct se_mem *mem;
+       struct scatterlist *sg = NULL;
        size_t remaining;
        u32 f_ctl = FC_FC_EX_CTX | FC_FC_REL_OFF;
-       u32 mem_off;
+       u32 mem_off = 0;
        u32 fh_off = 0;
        u32 frame_off = 0;
        size_t frame_len = 0;
-       size_t mem_len;
+       size_t mem_len = 0;
        size_t tlen;
        size_t off_in_page;
-       struct page *page;
+       struct page *page = NULL;
        int use_sg;
        int error;
        void *page_addr;
@@ -90,24 +90,17 @@ int ft_queue_data_in(struct se_cmd *se_cmd)
        lport = ep->lp;
        cmd->seq = lport->tt.seq_start_next(cmd->seq);
 
-       task = T_TASK(se_cmd);
-       BUG_ON(!task);
        remaining = se_cmd->data_length;
 
        /*
-        * Setup to use first mem list entry if any.
+        * Setup to use first mem list entry, unless no data.
         */
-       if (task->t_tasks_se_num) {
-               mem = list_first_entry(task->t_mem_list,
-                        struct se_mem, se_list);
-               mem_len = mem->se_len;
-               mem_off = mem->se_off;
-               page = mem->se_page;
-       } else {
-               mem = NULL;
-               mem_len = remaining;
-               mem_off = 0;
-               page = NULL;
+       BUG_ON(remaining && !se_cmd->t_data_sg);
+       if (remaining) {
+               sg = se_cmd->t_data_sg;
+               mem_len = sg->length;
+               mem_off = sg->offset;
+               page = sg_page(sg);
        }
 
        /* no scatter/gather in skb for odd word length due to fc_seq_send() */
@@ -115,12 +108,10 @@ int ft_queue_data_in(struct se_cmd *se_cmd)
 
        while (remaining) {
                if (!mem_len) {
-                       BUG_ON(!mem);
-                       mem = list_entry(mem->se_list.next,
-                               struct se_mem, se_list);
-                       mem_len = min((size_t)mem->se_len, remaining);
-                       mem_off = mem->se_off;
-                       page = mem->se_page;
+                       sg = sg_next(sg);
+                       mem_len = min((size_t)sg->length, remaining);
+                       mem_off = sg->offset;
+                       page = sg_page(sg);
                }
                if (!frame_len) {
                        /*
@@ -148,18 +139,7 @@ int ft_queue_data_in(struct se_cmd *se_cmd)
                tlen = min(mem_len, frame_len);
 
                if (use_sg) {
-                       if (!mem) {
-                               BUG_ON(!task->t_task_buf);
-                               page_addr = task->t_task_buf + mem_off;
-                               /*
-                                * In this case, offset is 'offset_in_page' of
-                                * (t_task_buf + mem_off) instead of 'mem_off'.
-                                */
-                               off_in_page = offset_in_page(page_addr);
-                               page = virt_to_page(page_addr);
-                               tlen = min(tlen, PAGE_SIZE - off_in_page);
-                       } else
-                               off_in_page = mem_off;
+                       off_in_page = mem_off;
                        BUG_ON(!page);
                        get_page(page);
                        skb_fill_page_desc(fp_skb(fp),
@@ -169,7 +149,7 @@ int ft_queue_data_in(struct se_cmd *se_cmd)
                        fp_skb(fp)->data_len += tlen;
                        fp_skb(fp)->truesize +=
                                        PAGE_SIZE << compound_order(page);
-               } else if (mem) {
+               } else {
                        BUG_ON(!page);
                        from = kmap_atomic(page + (mem_off >> PAGE_SHIFT),
                                           KM_SOFTIRQ0);
@@ -180,10 +160,6 @@ int ft_queue_data_in(struct se_cmd *se_cmd)
                        memcpy(to, from, tlen);
                        kunmap_atomic(page_addr, KM_SOFTIRQ0);
                        to += tlen;
-               } else {
-                       from = task->t_task_buf + mem_off;
-                       memcpy(to, from, tlen);
-                       to += tlen;
                }
 
                mem_off += tlen;
@@ -201,8 +177,7 @@ int ft_queue_data_in(struct se_cmd *se_cmd)
                error = lport->tt.seq_send(lport, cmd->seq, fp);
                if (error) {
                        /* XXX For now, initiator will retry */
-                       if (printk_ratelimit())
-                               printk(KERN_ERR "%s: Failed to send frame %p, "
+                       pr_err_ratelimited("%s: Failed to send frame %p, "
                                                "xid <0x%x>, remaining %zu, "
                                                "lso_max <0x%x>\n",
                                                __func__, fp, ep->xid,
@@ -221,24 +196,20 @@ void ft_recv_write_data(struct ft_cmd *cmd, struct fc_frame *fp)
        struct fc_seq *seq = cmd->seq;
        struct fc_exch *ep;
        struct fc_lport *lport;
-       struct se_transport_task *task;
        struct fc_frame_header *fh;
-       struct se_mem *mem;
-       u32 mem_off;
+       struct scatterlist *sg = NULL;
+       u32 mem_off = 0;
        u32 rel_off;
        size_t frame_len;
-       size_t mem_len;
+       size_t mem_len = 0;
        size_t tlen;
-       struct page *page;
+       struct page *page = NULL;
        void *page_addr;
        void *from;
        void *to;
        u32 f_ctl;
        void *buf;
 
-       task = T_TASK(se_cmd);
-       BUG_ON(!task);
-
        fh = fc_frame_header_get(fp);
        if (!(ntoh24(fh->fh_f_ctl) & FC_FC_REL_OFF))
                goto drop;
@@ -251,7 +222,7 @@ void ft_recv_write_data(struct ft_cmd *cmd, struct fc_frame *fp)
         */
        buf = fc_frame_payload_get(fp, 1);
        if (cmd->was_ddp_setup && buf) {
-               printk(KERN_INFO "%s: When DDP was setup, not expected to"
+               pr_debug("%s: When DDP was setup, not expected to"
                                 "receive frame with payload, Payload shall be"
                                 "copied directly to buffer instead of coming "
                                 "via. legacy receive queues\n", __func__);
@@ -289,7 +260,7 @@ void ft_recv_write_data(struct ft_cmd *cmd, struct fc_frame *fp)
                         * this point, but just in case if required in future
                         * for debugging or any other purpose
                         */
-                       printk(KERN_ERR "%s: Received frame with TSI bit not"
+                       pr_err("%s: Received frame with TSI bit not"
                                        " being SET, dropping the frame, "
                                        "cmd->sg <%p>, cmd->sg_cnt <0x%x>\n",
                                        __func__, cmd->sg, cmd->sg_cnt);
@@ -312,29 +283,22 @@ void ft_recv_write_data(struct ft_cmd *cmd, struct fc_frame *fp)
                frame_len = se_cmd->data_length - rel_off;
 
        /*
-        * Setup to use first mem list entry if any.
+        * Setup to use first mem list entry, unless no data.
         */
-       if (task->t_tasks_se_num) {
-               mem = list_first_entry(task->t_mem_list,
-                                      struct se_mem, se_list);
-               mem_len = mem->se_len;
-               mem_off = mem->se_off;
-               page = mem->se_page;
-       } else {
-               mem = NULL;
-               page = NULL;
-               mem_off = 0;
-               mem_len = frame_len;
+       BUG_ON(frame_len && !se_cmd->t_data_sg);
+       if (frame_len) {
+               sg = se_cmd->t_data_sg;
+               mem_len = sg->length;
+               mem_off = sg->offset;
+               page = sg_page(sg);
        }
 
        while (frame_len) {
                if (!mem_len) {
-                       BUG_ON(!mem);
-                       mem = list_entry(mem->se_list.next,
-                                        struct se_mem, se_list);
-                       mem_len = mem->se_len;
-                       mem_off = mem->se_off;
-                       page = mem->se_page;
+                       sg = sg_next(sg);
+                       mem_len = sg->length;
+                       mem_off = sg->offset;
+                       page = sg_page(sg);
                }
                if (rel_off >= mem_len) {
                        rel_off -= mem_len;
@@ -347,19 +311,15 @@ void ft_recv_write_data(struct ft_cmd *cmd, struct fc_frame *fp)
 
                tlen = min(mem_len, frame_len);
 
-               if (mem) {
-                       to = kmap_atomic(page + (mem_off >> PAGE_SHIFT),
-                                        KM_SOFTIRQ0);
-                       page_addr = to;
-                       to += mem_off & ~PAGE_MASK;
-                       tlen = min(tlen, (size_t)(PAGE_SIZE -
-                                               (mem_off & ~PAGE_MASK)));
-                       memcpy(to, from, tlen);
-                       kunmap_atomic(page_addr, KM_SOFTIRQ0);
-               } else {
-                       to = task->t_task_buf + mem_off;
-                       memcpy(to, from, tlen);
-               }
+               to = kmap_atomic(page + (mem_off >> PAGE_SHIFT),
+                                KM_SOFTIRQ0);
+               page_addr = to;
+               to += mem_off & ~PAGE_MASK;
+               tlen = min(tlen, (size_t)(PAGE_SIZE -
+                                         (mem_off & ~PAGE_MASK)));
+               memcpy(to, from, tlen);
+               kunmap_atomic(page_addr, KM_SOFTIRQ0);
+
                from += tlen;
                frame_len -= tlen;
                mem_off += tlen;
index 7491e21cc6ae4c35629614b687a424ac2917d87a..fbcbb3d1d06b64f06742ced9d5073764678d0676 100644 (file)
@@ -198,13 +198,13 @@ static struct ft_sess *ft_sess_get(struct fc_lport *lport, u32 port_id)
                if (sess->port_id == port_id) {
                        kref_get(&sess->kref);
                        rcu_read_unlock();
-                       FT_SESS_DBG("port_id %x found %p\n", port_id, sess);
+                       pr_debug("port_id %x found %p\n", port_id, sess);
                        return sess;
                }
        }
 out:
        rcu_read_unlock();
-       FT_SESS_DBG("port_id %x not found\n", port_id);
+       pr_debug("port_id %x not found\n", port_id);
        return NULL;
 }
 
@@ -240,7 +240,7 @@ static struct ft_sess *ft_sess_create(struct ft_tport *tport, u32 port_id,
        hlist_add_head_rcu(&sess->hash, head);
        tport->sess_count++;
 
-       FT_SESS_DBG("port_id %x sess %p\n", port_id, sess);
+       pr_debug("port_id %x sess %p\n", port_id, sess);
 
        transport_register_session(&tport->tpg->se_tpg, &acl->se_node_acl,
                                   sess->se_sess, sess);
@@ -314,7 +314,7 @@ int ft_sess_shutdown(struct se_session *se_sess)
 {
        struct ft_sess *sess = se_sess->fabric_sess_ptr;
 
-       FT_SESS_DBG("port_id %x\n", sess->port_id);
+       pr_debug("port_id %x\n", sess->port_id);
        return 1;
 }
 
@@ -335,7 +335,7 @@ void ft_sess_close(struct se_session *se_sess)
                mutex_unlock(&ft_lport_lock);
                return;
        }
-       FT_SESS_DBG("port_id %x\n", port_id);
+       pr_debug("port_id %x\n", port_id);
        ft_sess_unhash(sess);
        mutex_unlock(&ft_lport_lock);
        transport_deregister_session_configfs(se_sess);
@@ -348,7 +348,7 @@ void ft_sess_stop(struct se_session *se_sess, int sess_sleep, int conn_sleep)
 {
        struct ft_sess *sess = se_sess->fabric_sess_ptr;
 
-       FT_SESS_DBG("port_id %x\n", sess->port_id);
+       pr_debug("port_id %x\n", sess->port_id);
 }
 
 int ft_sess_logged_in(struct se_session *se_sess)
@@ -458,7 +458,7 @@ static int ft_prli(struct fc_rport_priv *rdata, u32 spp_len,
        mutex_lock(&ft_lport_lock);
        ret = ft_prli_locked(rdata, spp_len, rspp, spp);
        mutex_unlock(&ft_lport_lock);
-       FT_SESS_DBG("port_id %x flags %x ret %x\n",
+       pr_debug("port_id %x flags %x ret %x\n",
               rdata->ids.port_id, rspp ? rspp->spp_flags : 0, ret);
        return ret;
 }
@@ -518,11 +518,11 @@ static void ft_recv(struct fc_lport *lport, struct fc_frame *fp)
        struct ft_sess *sess;
        u32 sid = fc_frame_sid(fp);
 
-       FT_SESS_DBG("sid %x\n", sid);
+       pr_debug("sid %x\n", sid);
 
        sess = ft_sess_get(lport, sid);
        if (!sess) {
-               FT_SESS_DBG("sid %x sess lookup failed\n", sid);
+               pr_debug("sid %x sess lookup failed\n", sid);
                /* TBD XXX - if FCP_CMND, send PRLO */
                fc_frame_free(fp);
                return;
index 561ac99def5af8b51d9db422625dcf1712907267..27040653005e42f8d4babb506b7badd209cca014 100644 (file)
@@ -9,12 +9,13 @@
 #include <net/sock.h>
 #include <net/tcp.h>
 
-#define TARGET_CORE_MOD_VERSION                "v4.0.0-rc7-ml"
+#define TARGET_CORE_MOD_VERSION                "v4.1.0-rc1-ml"
 #define SHUTDOWN_SIGS  (sigmask(SIGKILL)|sigmask(SIGINT)|sigmask(SIGABRT))
 
 /* Used by transport_generic_allocate_iovecs() */
 #define TRANSPORT_IOV_DATA_BUFFER              5
 /* Maximum Number of LUNs per Target Portal Group */
+/* Don't raise above 511 or REPORT_LUNS needs to handle >1 page */
 #define TRANSPORT_MAX_LUNS_PER_TPG             256
 /*
  * By default we use 32-byte CDBs in TCM Core and subsystem plugin code.
@@ -99,6 +100,7 @@ enum transport_state_table {
        TRANSPORT_FREE          = 15,
        TRANSPORT_NEW_CMD_MAP   = 16,
        TRANSPORT_FREE_CMD_INTR = 17,
+       TRANSPORT_COMPLETE_QF_WP = 18,
 };
 
 /* Used for struct se_cmd->se_cmd_flags */
@@ -108,27 +110,22 @@ enum se_cmd_flags_table {
        SCF_EMULATED_TASK_SENSE         = 0x00000004,
        SCF_SCSI_DATA_SG_IO_CDB         = 0x00000008,
        SCF_SCSI_CONTROL_SG_IO_CDB      = 0x00000010,
-       SCF_SCSI_CONTROL_NONSG_IO_CDB   = 0x00000020,
        SCF_SCSI_NON_DATA_CDB           = 0x00000040,
        SCF_SCSI_CDB_EXCEPTION          = 0x00000080,
        SCF_SCSI_RESERVATION_CONFLICT   = 0x00000100,
-       SCF_CMD_PASSTHROUGH_NOALLOC     = 0x00000200,
        SCF_SE_CMD_FAILED               = 0x00000400,
        SCF_SE_LUN_CMD                  = 0x00000800,
        SCF_SE_ALLOW_EOO                = 0x00001000,
-       SCF_SE_DISABLE_ONLINE_CHECK     = 0x00002000,
        SCF_SENT_CHECK_CONDITION        = 0x00004000,
        SCF_OVERFLOW_BIT                = 0x00008000,
        SCF_UNDERFLOW_BIT               = 0x00010000,
        SCF_SENT_DELAYED_TAS            = 0x00020000,
        SCF_ALUA_NON_OPTIMIZED          = 0x00040000,
        SCF_DELAYED_CMD_FROM_SAM_ATTR   = 0x00080000,
-       SCF_PASSTHROUGH_SG_TO_MEM       = 0x00100000,
-       SCF_PASSTHROUGH_CONTIG_TO_SG    = 0x00200000,
+       SCF_UNUSED                      = 0x00100000,
        SCF_PASSTHROUGH_SG_TO_MEM_NOALLOC = 0x00400000,
-       SCF_EMULATE_SYNC_CACHE          = 0x00800000,
        SCF_EMULATE_CDB_ASYNC           = 0x01000000,
-       SCF_EMULATE_SYNC_UNMAP          = 0x02000000
+       SCF_EMULATE_QUEUE_FULL          = 0x02000000,
 };
 
 /* struct se_dev_entry->lun_flags and struct se_lun->lun_access */
@@ -205,11 +202,6 @@ typedef enum {
        SCSI_INDEX_TYPE_MAX
 } scsi_index_t;
 
-struct scsi_index_table {
-       spinlock_t      lock;
-       u32             scsi_mib_index[SCSI_INDEX_TYPE_MAX];
-} ____cacheline_aligned;
-
 struct se_cmd;
 
 struct t10_alua {
@@ -235,7 +227,7 @@ struct t10_alua_lu_gp {
        atomic_t lu_gp_ref_cnt;
        spinlock_t lu_gp_lock;
        struct config_group lu_gp_group;
-       struct list_head lu_gp_list;
+       struct list_head lu_gp_node;
        struct list_head lu_gp_mem_list;
 } ____cacheline_aligned;
 
@@ -291,10 +283,10 @@ struct t10_vpd {
 } ____cacheline_aligned;
 
 struct t10_wwn {
-       unsigned char vendor[8];
-       unsigned char model[16];
-       unsigned char revision[4];
-       unsigned char unit_serial[INQUIRY_VPD_SERIAL_LEN];
+       char vendor[8];
+       char model[16];
+       char revision[4];
+       char unit_serial[INQUIRY_VPD_SERIAL_LEN];
        spinlock_t t10_vpd_lock;
        struct se_subsystem_dev *t10_sub_dev;
        struct config_group t10_wwn_group;
@@ -366,13 +358,13 @@ struct t10_reservation_ops {
        int (*t10_pr_clear)(struct se_cmd *);
 };
 
-struct t10_reservation_template {
+struct t10_reservation {
        /* Reservation effects all target ports */
        int pr_all_tg_pt;
        /* Activate Persistence across Target Power Loss enabled
         * for SCSI device */
        int pr_aptpl_active;
-       /* Used by struct t10_reservation_template->pr_aptpl_buf_len */
+       /* Used by struct t10_reservation->pr_aptpl_buf_len */
 #define PR_APTPL_BUF_LEN                       8192
        u32 pr_aptpl_buf_len;
        u32 pr_generation;
@@ -397,7 +389,7 @@ struct t10_reservation_template {
 
 struct se_queue_req {
        int                     state;
-       void                    *cmd;
+       struct se_cmd           *cmd;
        struct list_head        qr_list;
 } ____cacheline_aligned;
 
@@ -408,64 +400,10 @@ struct se_queue_obj {
        wait_queue_head_t       thread_wq;
 } ____cacheline_aligned;
 
-/*
- * Used one per struct se_cmd to hold all extra struct se_task
- * metadata.  This structure is setup and allocated in
- * drivers/target/target_core_transport.c:__transport_alloc_se_cmd()
- */
-struct se_transport_task {
-       unsigned char           *t_task_cdb;
-       unsigned char           __t_task_cdb[TCM_MAX_COMMAND_SIZE];
-       unsigned long long      t_task_lba;
-       int                     t_tasks_failed;
-       int                     t_tasks_fua;
-       bool                    t_tasks_bidi;
-       u32                     t_task_cdbs;
-       u32                     t_tasks_check;
-       u32                     t_tasks_no;
-       u32                     t_tasks_sectors;
-       u32                     t_tasks_se_num;
-       u32                     t_tasks_se_bidi_num;
-       u32                     t_tasks_sg_chained_no;
-       atomic_t                t_fe_count;
-       atomic_t                t_se_count;
-       atomic_t                t_task_cdbs_left;
-       atomic_t                t_task_cdbs_ex_left;
-       atomic_t                t_task_cdbs_timeout_left;
-       atomic_t                t_task_cdbs_sent;
-       atomic_t                t_transport_aborted;
-       atomic_t                t_transport_active;
-       atomic_t                t_transport_complete;
-       atomic_t                t_transport_queue_active;
-       atomic_t                t_transport_sent;
-       atomic_t                t_transport_stop;
-       atomic_t                t_transport_timeout;
-       atomic_t                transport_dev_active;
-       atomic_t                transport_lun_active;
-       atomic_t                transport_lun_fe_stop;
-       atomic_t                transport_lun_stop;
-       spinlock_t              t_state_lock;
-       struct completion       t_transport_stop_comp;
-       struct completion       transport_lun_fe_stop_comp;
-       struct completion       transport_lun_stop_comp;
-       struct scatterlist      *t_tasks_sg_chained;
-       struct scatterlist      t_tasks_sg_bounce;
-       void                    *t_task_buf;
-       /*
-        * Used for pre-registered fabric SGL passthrough WRITE and READ
-        * with the special SCF_PASSTHROUGH_CONTIG_TO_SG case for TCM_Loop
-        * and other HW target mode fabric modules.
-        */
-       struct scatterlist      *t_task_pt_sgl;
-       struct list_head        *t_mem_list;
-       /* Used for BIDI READ */
-       struct list_head        *t_mem_bidi_list;
-       struct list_head        t_task_list;
-} ____cacheline_aligned;
-
 struct se_task {
        unsigned char   task_sense;
        struct scatterlist *task_sg;
+       u32             task_sg_nents;
        struct scatterlist *task_sg_bidi;
        u8              task_scsi_status;
        u8              task_flags;
@@ -476,8 +414,6 @@ struct se_task {
        u32             task_no;
        u32             task_sectors;
        u32             task_size;
-       u32             task_sg_num;
-       u32             task_sg_offset;
        enum dma_data_direction task_data_direction;
        struct se_cmd *task_se_cmd;
        struct se_device        *se_dev;
@@ -495,9 +431,6 @@ struct se_task {
        struct list_head t_state_list;
 } ____cacheline_aligned;
 
-#define TASK_CMD(task) ((task)->task_se_cmd)
-#define TASK_DEV(task) ((task)->se_dev)
-
 struct se_cmd {
        /* SAM response code being sent to initiator */
        u8                      scsi_status;
@@ -531,9 +464,10 @@ struct se_cmd {
        atomic_t                transport_sent;
        /* Used for sense data */
        void                    *sense_buffer;
-       struct list_head        se_delayed_list;
-       struct list_head        se_ordered_list;
-       struct list_head        se_lun_list;
+       struct list_head        se_delayed_node;
+       struct list_head        se_ordered_node;
+       struct list_head        se_lun_node;
+       struct list_head        se_qf_node;
        struct se_device      *se_dev;
        struct se_dev_entry   *se_deve;
        struct se_device        *se_obj_ptr;
@@ -542,18 +476,62 @@ struct se_cmd {
        /* Only used for internal passthrough and legacy TCM fabric modules */
        struct se_session       *se_sess;
        struct se_tmr_req       *se_tmr_req;
-       /* t_task is setup to t_task_backstore in transport_init_se_cmd() */
-       struct se_transport_task *t_task;
-       struct se_transport_task t_task_backstore;
+       struct list_head        se_queue_node;
        struct target_core_fabric_ops *se_tfo;
        int (*transport_emulate_cdb)(struct se_cmd *);
-       void (*transport_split_cdb)(unsigned long long, u32 *, unsigned char *);
+       void (*transport_split_cdb)(unsigned long long, u32, unsigned char *);
        void (*transport_wait_for_tasks)(struct se_cmd *, int, int);
        void (*transport_complete_callback)(struct se_cmd *);
-} ____cacheline_aligned;
+       int (*transport_qf_callback)(struct se_cmd *);
 
-#define T_TASK(cmd)     ((cmd)->t_task)
-#define CMD_TFO(cmd)   ((cmd)->se_tfo)
+       unsigned char           *t_task_cdb;
+       unsigned char           __t_task_cdb[TCM_MAX_COMMAND_SIZE];
+       unsigned long long      t_task_lba;
+       int                     t_tasks_failed;
+       int                     t_tasks_fua;
+       bool                    t_tasks_bidi;
+       u32                     t_tasks_sg_chained_no;
+       atomic_t                t_fe_count;
+       atomic_t                t_se_count;
+       atomic_t                t_task_cdbs_left;
+       atomic_t                t_task_cdbs_ex_left;
+       atomic_t                t_task_cdbs_timeout_left;
+       atomic_t                t_task_cdbs_sent;
+       atomic_t                t_transport_aborted;
+       atomic_t                t_transport_active;
+       atomic_t                t_transport_complete;
+       atomic_t                t_transport_queue_active;
+       atomic_t                t_transport_sent;
+       atomic_t                t_transport_stop;
+       atomic_t                t_transport_timeout;
+       atomic_t                transport_dev_active;
+       atomic_t                transport_lun_active;
+       atomic_t                transport_lun_fe_stop;
+       atomic_t                transport_lun_stop;
+       spinlock_t              t_state_lock;
+       struct completion       t_transport_stop_comp;
+       struct completion       transport_lun_fe_stop_comp;
+       struct completion       transport_lun_stop_comp;
+       struct scatterlist      *t_tasks_sg_chained;
+
+       /*
+        * Used for pre-registered fabric SGL passthrough WRITE and READ
+        * with the special SCF_PASSTHROUGH_CONTIG_TO_SG case for TCM_Loop
+        * and other HW target mode fabric modules.
+        */
+       struct scatterlist      *t_task_pt_sgl;
+       u32                     t_task_pt_sgl_num;
+
+       struct scatterlist      *t_data_sg;
+       unsigned int            t_data_nents;
+       struct scatterlist      *t_bidi_data_sg;
+       unsigned int            t_bidi_data_nents;
+
+       /* Used for BIDI READ */
+       struct list_head        t_task_list;
+       u32                     t_task_list_num;
+
+} ____cacheline_aligned;
 
 struct se_tmr_req {
        /* Task Management function to be preformed */
@@ -617,9 +595,6 @@ struct se_session {
        struct list_head        sess_acl_list;
 } ____cacheline_aligned;
 
-#define SE_SESS(cmd)           ((cmd)->se_sess)
-#define SE_NODE_ACL(sess)      ((sess)->se_node_acl)
-
 struct se_device;
 struct se_transform_info;
 struct scatterlist;
@@ -640,8 +615,6 @@ struct se_lun_acl {
        struct se_ml_stat_grps  ml_stat_grps;
 }  ____cacheline_aligned;
 
-#define ML_STAT_GRPS(lacl)     (&(lacl)->ml_stat_grps)
-
 struct se_dev_entry {
        bool                    def_pr_registered;
        /* See transport_lunflags_table */
@@ -688,6 +661,8 @@ struct se_dev_attrib {
        int             emulate_reservations;
        int             emulate_alua;
        int             enforce_pr_isids;
+       int             is_nonrot;
+       int             emulate_rest_reord;
        u32             hw_block_size;
        u32             block_size;
        u32             hw_max_sectors;
@@ -727,10 +702,10 @@ struct se_subsystem_dev {
        /* T10 Inquiry and VPD WWN Information */
        struct t10_wwn  t10_wwn;
        /* T10 SPC-2 + SPC-3 Reservations */
-       struct t10_reservation_template t10_reservation;
+       struct t10_reservation t10_pr;
        spinlock_t      se_dev_lock;
        void            *se_dev_su_ptr;
-       struct list_head g_se_dev_list;
+       struct list_head se_dev_node;
        struct config_group se_dev_group;
        /* For T10 Reservations */
        struct config_group se_dev_pr_group;
@@ -738,11 +713,6 @@ struct se_subsystem_dev {
        struct se_dev_stat_grps dev_stat_grps;
 } ____cacheline_aligned;
 
-#define T10_ALUA(su_dev)       (&(su_dev)->t10_alua)
-#define T10_RES(su_dev)                (&(su_dev)->t10_reservation)
-#define T10_PR_OPS(su_dev)     (&(su_dev)->t10_reservation.pr_ops)
-#define DEV_STAT_GRP(dev)      (&(dev)->dev_stat_grps)
-
 struct se_device {
        /* Set to 1 if thread is NOT sleeping on thread_sem */
        u8                      thread_active;
@@ -780,11 +750,11 @@ struct se_device {
        atomic_t                dev_status_thr_count;
        atomic_t                dev_hoq_count;
        atomic_t                dev_ordered_sync;
+       atomic_t                dev_qf_count;
        struct se_obj           dev_obj;
        struct se_obj           dev_access_obj;
        struct se_obj           dev_export_obj;
-       struct se_queue_obj     *dev_queue_obj;
-       struct se_queue_obj     *dev_status_queue_obj;
+       struct se_queue_obj     dev_queue_obj;
        spinlock_t              delayed_cmd_lock;
        spinlock_t              ordered_cmd_lock;
        spinlock_t              execute_task_lock;
@@ -796,6 +766,7 @@ struct se_device {
        spinlock_t              dev_status_thr_lock;
        spinlock_t              se_port_lock;
        spinlock_t              se_tmr_lock;
+       spinlock_t              qf_cmd_lock;
        /* Used for legacy SPC-2 reservationsa */
        struct se_node_acl      *dev_reserved_node_acl;
        /* Used for ALUA Logical Unit Group membership */
@@ -809,10 +780,12 @@ struct se_device {
        struct task_struct      *process_thread;
        pid_t                   process_thread_pid;
        struct task_struct              *dev_mgmt_thread;
+       struct work_struct      qf_work_queue;
        struct list_head        delayed_cmd_list;
        struct list_head        ordered_cmd_list;
        struct list_head        execute_task_list;
        struct list_head        state_task_list;
+       struct list_head        qf_cmd_list;
        /* Pointer to associated SE HBA */
        struct se_hba           *se_hba;
        struct se_subsystem_dev *se_sub_dev;
@@ -824,11 +797,6 @@ struct se_device {
        struct list_head        g_se_dev_list;
 }  ____cacheline_aligned;
 
-#define SE_DEV(cmd)            ((cmd)->se_lun->lun_se_dev)
-#define SU_DEV(dev)            ((dev)->se_sub_dev)
-#define DEV_ATTRIB(dev)                (&(dev)->se_sub_dev->se_dev_attrib)
-#define DEV_T10_WWN(dev)       (&(dev)->se_sub_dev->t10_wwn)
-
 struct se_hba {
        u16                     hba_tpgt;
        u32                     hba_id;
@@ -837,24 +805,17 @@ struct se_hba {
        /* Virtual iSCSI devices attached. */
        u32                     dev_count;
        u32                     hba_index;
-       atomic_t                load_balance_queue;
-       atomic_t                left_queue_depth;
-       /* Maximum queue depth the HBA can handle. */
-       atomic_t                max_queue_depth;
        /* Pointer to transport specific host structure. */
        void                    *hba_ptr;
        /* Linked list for struct se_device */
        struct list_head        hba_dev_list;
-       struct list_head        hba_list;
+       struct list_head        hba_node;
        spinlock_t              device_lock;
-       spinlock_t              hba_queue_lock;
        struct config_group     hba_group;
        struct mutex            hba_access_mutex;
        struct se_subsystem_api *transport;
 }  ____cacheline_aligned;
 
-#define SE_HBA(dev)            ((dev)->se_hba)
-
 struct se_port_stat_grps {
        struct config_group stat_group;
        struct config_group scsi_port_group;
@@ -881,9 +842,6 @@ struct se_lun {
        struct se_port_stat_grps port_stat_grps;
 } ____cacheline_aligned;
 
-#define SE_LUN(cmd)            ((cmd)->se_lun)
-#define PORT_STAT_GRP(lun)     (&(lun)->port_stat_grps)
-
 struct scsi_port_stats {
        u64     cmd_pdus;
        u64     tx_data_octets;
@@ -930,7 +888,7 @@ struct se_portal_group {
        spinlock_t              tpg_lun_lock;
        /* Pointer to $FABRIC_MOD portal group */
        void                    *se_tpg_fabric_ptr;
-       struct list_head        se_tpg_list;
+       struct list_head        se_tpg_node;
        /* linked list for initiator ACL list */
        struct list_head        acl_node_list;
        struct se_lun           *tpg_lun_list;
@@ -949,8 +907,6 @@ struct se_portal_group {
        struct config_group     tpg_param_group;
 } ____cacheline_aligned;
 
-#define TPG_TFO(se_tpg)        ((se_tpg)->se_tpg_tfo)
-
 struct se_wwn {
        struct target_fabric_configfs *wwn_tf;
        struct config_group     wwn_group;
@@ -958,28 +914,4 @@ struct se_wwn {
        struct config_group     fabric_stat_group;
 } ____cacheline_aligned;
 
-struct se_global {
-       u16                     alua_lu_gps_counter;
-       int                     g_sub_api_initialized;
-       u32                     in_shutdown;
-       u32                     alua_lu_gps_count;
-       u32                     g_hba_id_counter;
-       struct config_group     target_core_hbagroup;
-       struct config_group     alua_group;
-       struct config_group     alua_lu_gps_group;
-       struct list_head        g_lu_gps_list;
-       struct list_head        g_se_tpg_list;
-       struct list_head        g_hba_list;
-       struct list_head        g_se_dev_list;
-       struct se_hba           *g_lun0_hba;
-       struct se_subsystem_dev *g_lun0_su_dev;
-       struct se_device        *g_lun0_dev;
-       struct t10_alua_lu_gp   *default_lu_gp;
-       spinlock_t              g_device_lock;
-       spinlock_t              hba_lock;
-       spinlock_t              se_tpg_lock;
-       spinlock_t              lu_gps_lock;
-       spinlock_t              plugin_class_lock;
-} ____cacheline_aligned;
-
 #endif /* TARGET_CORE_BASE_H */
index 52b18a5752c93d7aec425817bffcfb782f7a1a36..46571912086c1b5c7cc5592e91a4ec4ace9fd50d 100644 (file)
@@ -1,8 +1,8 @@
 #ifndef TARGET_CORE_DEVICE_H
 #define TARGET_CORE_DEVICE_H
 
-extern int transport_get_lun_for_cmd(struct se_cmd *, unsigned char *, u32);
-extern int transport_get_lun_for_tmr(struct se_cmd *, u32);
+extern int transport_lookup_cmd_lun(struct se_cmd *, u32);
+extern int transport_lookup_tmr_lun(struct se_cmd *, u32);
 extern struct se_dev_entry *core_get_se_deve_from_rtpi(
                                        struct se_node_acl *, u16);
 extern int core_free_device_list_for_node(struct se_node_acl *,
@@ -39,6 +39,8 @@ extern int se_dev_set_emulate_tas(struct se_device *, int);
 extern int se_dev_set_emulate_tpu(struct se_device *, int);
 extern int se_dev_set_emulate_tpws(struct se_device *, int);
 extern int se_dev_set_enforce_pr_isids(struct se_device *, int);
+extern int se_dev_set_is_nonrot(struct se_device *, int);
+extern int se_dev_set_emulate_rest_reord(struct se_device *dev, int);
 extern int se_dev_set_queue_depth(struct se_device *, u32);
 extern int se_dev_set_max_sectors(struct se_device *, u32);
 extern int se_dev_set_optimal_sectors(struct se_device *, u32);
index 747e1404dca0db5c43041e6521eb4ba95ee5ed14..2de8fe9075963350f8159759c118dedc70db1b57 100644 (file)
@@ -38,18 +38,12 @@ struct target_core_fabric_ops {
         * initially allocated in interrupt context.
         */
        int (*new_cmd_map)(struct se_cmd *);
-       /*
-        * Optional function pointer for TCM fabric modules that use
-        * Linux/NET sockets to allocate struct iovec array to struct se_cmd
-        */
-       int (*alloc_cmd_iovecs)(struct se_cmd *);
        /*
         * Optional to release struct se_cmd and fabric dependent allocated
         * I/O descriptor in transport_cmd_check_stop()
         */
        void (*check_stop_free)(struct se_cmd *);
-       void (*release_cmd_to_pool)(struct se_cmd *);
-       void (*release_cmd_direct)(struct se_cmd *);
+       void (*release_cmd)(struct se_cmd *);
        /*
         * Called with spin_lock_bh(struct se_portal_group->session_lock held.
         */
@@ -70,7 +64,6 @@ struct target_core_fabric_ops {
        void (*set_default_node_attributes)(struct se_node_acl *);
        u32 (*get_task_tag)(struct se_cmd *);
        int (*get_cmd_state)(struct se_cmd *);
-       void (*new_cmd_failure)(struct se_cmd *);
        int (*queue_data_in)(struct se_cmd *);
        int (*queue_status)(struct se_cmd *);
        int (*queue_tm_rsp)(struct se_cmd *);
index 24a1c6cb83c348eeeb4e575e460b4e5ac7c36882..46aae4f94ede427814b6f6cdef7974c1f89d5fd3 100644 (file)
 #define DA_ENFORCE_PR_ISIDS                    1
 #define DA_STATUS_MAX_SECTORS_MIN              16
 #define DA_STATUS_MAX_SECTORS_MAX              8192
+/* By default don't report non-rotating (solid state) medium */
+#define DA_IS_NONROT                           0
+/* Queue Algorithm Modifier default for restricted reordering in control mode page */
+#define DA_EMULATE_REST_REORD                  0
 
 #define SE_MODE_PAGE_BUF                       512
 
@@ -111,9 +115,8 @@ struct se_subsystem_api;
 
 extern struct kmem_cache *se_mem_cache;
 
-extern int init_se_global(void);
-extern void release_se_global(void);
-extern void init_scsi_index_table(void);
+extern int init_se_kmem_caches(void);
+extern void release_se_kmem_caches(void);
 extern u32 scsi_get_new_index(scsi_index_t);
 extern void transport_init_queue_obj(struct se_queue_obj *);
 extern int transport_subsystem_check_init(void);
@@ -160,36 +163,38 @@ extern struct se_device *transport_add_device_to_core_hba(struct se_hba *,
                                        struct se_subsystem_dev *, u32,
                                        void *, struct se_dev_limits *,
                                        const char *, const char *);
-extern void transport_device_setup_cmd(struct se_cmd *);
 extern void transport_init_se_cmd(struct se_cmd *,
                                        struct target_core_fabric_ops *,
                                        struct se_session *, u32, int, int,
                                        unsigned char *);
+void *transport_kmap_first_data_page(struct se_cmd *cmd);
+void transport_kunmap_first_data_page(struct se_cmd *cmd);
 extern void transport_free_se_cmd(struct se_cmd *);
 extern int transport_generic_allocate_tasks(struct se_cmd *, unsigned char *);
 extern int transport_generic_handle_cdb(struct se_cmd *);
+extern int transport_handle_cdb_direct(struct se_cmd *);
 extern int transport_generic_handle_cdb_map(struct se_cmd *);
 extern int transport_generic_handle_data(struct se_cmd *);
 extern void transport_new_cmd_failure(struct se_cmd *);
 extern int transport_generic_handle_tmr(struct se_cmd *);
 extern void transport_generic_free_cmd_intr(struct se_cmd *);
 extern void __transport_stop_task_timer(struct se_task *, unsigned long *);
-extern unsigned char transport_asciihex_to_binaryhex(unsigned char val[2]);
 extern int transport_generic_map_mem_to_cmd(struct se_cmd *cmd, struct scatterlist *, u32,
                                struct scatterlist *, u32);
 extern int transport_clear_lun_from_sessions(struct se_lun *);
 extern int transport_check_aborted_status(struct se_cmd *, int);
 extern int transport_send_check_condition_and_sense(struct se_cmd *, u8, int);
 extern void transport_send_task_abort(struct se_cmd *);
-extern void transport_release_cmd_to_pool(struct se_cmd *);
-extern void transport_generic_free_cmd(struct se_cmd *, int, int, int);
+extern void transport_release_cmd(struct se_cmd *);
+extern void transport_generic_free_cmd(struct se_cmd *, int, int);
 extern void transport_generic_wait_for_cmds(struct se_cmd *, int);
-extern u32 transport_calc_sg_num(struct se_task *, struct se_mem *, u32);
+extern int transport_init_task_sg(struct se_task *, struct se_mem *, u32);
 extern int transport_map_mem_to_sg(struct se_task *, struct list_head *,
-                                       void *, struct se_mem *,
+                                       struct scatterlist *, struct se_mem *,
                                        struct se_mem **, u32 *, u32 *);
 extern void transport_do_task_sg_chain(struct se_cmd *);
 extern void transport_generic_process_write(struct se_cmd *);
+extern int transport_generic_new_cmd(struct se_cmd *);
 extern int transport_generic_do_tmr(struct se_cmd *);
 /* From target_core_alua.c */
 extern int core_alua_check_nonop_delay(struct se_cmd *);
@@ -235,13 +240,13 @@ struct se_subsystem_api {
         */
        int (*cdb_none)(struct se_task *);
        /*
-        * For SCF_SCSI_CONTROL_NONSG_IO_CDB
+        * For SCF_SCSI_DATA_SG_IO_CDB
         */
-       int (*map_task_non_SG)(struct se_task *);
+       int (*map_data_SG)(struct se_task *);
        /*
-        * For SCF_SCSI_DATA_SG_IO_CDB and SCF_SCSI_CONTROL_SG_IO_CDB
+        * For SCF_SCSI_CONTROL_SG_IO_CDB
         */
-       int (*map_task_SG)(struct se_task *);
+       int (*map_control_SG)(struct se_task *);
        /*
         * attach_hba():
         */
@@ -292,7 +297,7 @@ struct se_subsystem_api {
         * drivers.  Provided out of convenience.
         */
        int (*transport_complete)(struct se_task *task);
-       struct se_task *(*alloc_task)(struct se_cmd *);
+       struct se_task *(*alloc_task)(unsigned char *cdb);
        /*
         * do_task():
         */
@@ -341,20 +346,10 @@ struct se_subsystem_api {
         * Get the sector_t from a subsystem backstore..
         */
        sector_t (*get_blocks)(struct se_device *);
-       /*
-        * do_se_mem_map():
-        */
-       int (*do_se_mem_map)(struct se_task *, struct list_head *, void *,
-                               struct se_mem *, struct se_mem **, u32 *, u32 *);
        /*
         * get_sense_buffer():
         */
        unsigned char *(*get_sense_buffer)(struct se_task *);
 } ____cacheline_aligned;
 
-#define TRANSPORT(dev)         ((dev)->transport)
-#define HBA_TRANSPORT(hba)     ((hba)->transport)
-
-extern struct se_global *se_global;
-
 #endif /* TARGET_CORE_TRANSPORT_H */