]> git.kernelconcepts.de Git - karo-tx-linux.git/commitdiff
target: simplify backend attribute implementation
authorChristoph Hellwig <hch@lst.de>
Sun, 10 May 2015 16:14:58 +0000 (18:14 +0200)
committerNicholas Bellinger <nab@linux-iscsi.org>
Mon, 1 Jun 2015 07:25:38 +0000 (00:25 -0700)
Consolidate the implementation of the backend attributes in a single file
and single function per attribute show/store function instead of splitting
it into multiple functions in multiple files.

Also use the proper strto* helpers for exposed data types, add macros to
implement the store methods for the most common data types and share the
show methods between the two different attribute implementations.

(Fix bogus store_pi_prot_format flag=0 return value - nab)

Signed-off-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Nicholas Bellinger <nab@linux-iscsi.org>
drivers/target/target_core_configfs.c
drivers/target/target_core_device.c
include/target/target_core_backend.h

index dbf91f02ee5ceedcc9e8b9679b4b832ee229cd4d..4313eea060edbb1e462647c580e2a4f6334bfc8a 100644 (file)
@@ -471,82 +471,507 @@ EXPORT_SYMBOL(target_unregister_template);
 //############################################################################*/
 
 /* Start functions for struct config_item_type tb_dev_attrib_cit */
-#define DEF_TB_DEV_ATTRIB_SHOW(_backend, _name)                                \
-static ssize_t _backend##_dev_show_attr_##_name(                       \
-       struct se_dev_attrib *da,                                       \
-       char *page)                                                     \
+#define DEF_TB_DEV_ATTRIB_SHOW(_name)                                  \
+static ssize_t show_##_name(struct se_dev_attrib *da, char *page)      \
 {                                                                      \
-       return snprintf(page, PAGE_SIZE, "%u\n",                        \
-                       (u32)da->da_dev->dev_attrib._name);             \
-}
-
-#define DEF_TB_DEV_ATTRIB_STORE(_backend, _name)                       \
-static ssize_t _backend##_dev_store_attr_##_name(                      \
-       struct se_dev_attrib *da,                                       \
-       const char *page,                                               \
-       size_t count)                                                   \
+       return snprintf(page, PAGE_SIZE, "%u\n", da->_name);            \
+}
+
+DEF_TB_DEV_ATTRIB_SHOW(emulate_model_alias);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_dpo);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_fua_write);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_fua_read);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_write_cache);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_ua_intlck_ctrl);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_tas);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_tpu);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_tpws);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_caw);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_3pc);
+DEF_TB_DEV_ATTRIB_SHOW(pi_prot_type);
+DEF_TB_DEV_ATTRIB_SHOW(hw_pi_prot_type);
+DEF_TB_DEV_ATTRIB_SHOW(pi_prot_format);
+DEF_TB_DEV_ATTRIB_SHOW(enforce_pr_isids);
+DEF_TB_DEV_ATTRIB_SHOW(is_nonrot);
+DEF_TB_DEV_ATTRIB_SHOW(emulate_rest_reord);
+DEF_TB_DEV_ATTRIB_SHOW(force_pr_aptpl);
+DEF_TB_DEV_ATTRIB_SHOW(hw_block_size);
+DEF_TB_DEV_ATTRIB_SHOW(block_size);
+DEF_TB_DEV_ATTRIB_SHOW(hw_max_sectors);
+DEF_TB_DEV_ATTRIB_SHOW(optimal_sectors);
+DEF_TB_DEV_ATTRIB_SHOW(hw_queue_depth);
+DEF_TB_DEV_ATTRIB_SHOW(queue_depth);
+DEF_TB_DEV_ATTRIB_SHOW(max_unmap_lba_count);
+DEF_TB_DEV_ATTRIB_SHOW(max_unmap_block_desc_count);
+DEF_TB_DEV_ATTRIB_SHOW(unmap_granularity);
+DEF_TB_DEV_ATTRIB_SHOW(unmap_granularity_alignment);
+DEF_TB_DEV_ATTRIB_SHOW(max_write_same_len);
+
+#define DEF_TB_DEV_ATTRIB_STORE_U32(_name)                             \
+static ssize_t store_##_name(struct se_dev_attrib *da, const char *page,\
+               size_t count)                                           \
 {                                                                      \
-       unsigned long val;                                              \
+       u32 val;                                                        \
        int ret;                                                        \
                                                                        \
-       ret = kstrtoul(page, 0, &val);                                  \
-       if (ret < 0) {                                                  \
-               pr_err("kstrtoul() failed with ret: %d\n", ret);        \
-               return -EINVAL;                                         \
-       }                                                               \
-       ret = se_dev_set_##_name(da->da_dev, (u32)val);                 \
+       ret = kstrtou32(page, 0, &val);                                 \
+       if (ret < 0)                                                    \
+               return ret;                                             \
+       da->_name = val;                                                \
+       return count;                                                   \
+}
+
+DEF_TB_DEV_ATTRIB_STORE_U32(max_unmap_lba_count);
+DEF_TB_DEV_ATTRIB_STORE_U32(max_unmap_block_desc_count);
+DEF_TB_DEV_ATTRIB_STORE_U32(unmap_granularity);
+DEF_TB_DEV_ATTRIB_STORE_U32(unmap_granularity_alignment);
+DEF_TB_DEV_ATTRIB_STORE_U32(max_write_same_len);
+
+#define DEF_TB_DEV_ATTRIB_STORE_BOOL(_name)                            \
+static ssize_t store_##_name(struct se_dev_attrib *da, const char *page,\
+               size_t count)                                           \
+{                                                                      \
+       bool flag;                                                      \
+       int ret;                                                        \
                                                                        \
-       return (!ret) ? count : -EINVAL;                                \
+       ret = strtobool(page, &flag);                                   \
+       if (ret < 0)                                                    \
+               return ret;                                             \
+       da->_name = flag;                                               \
+       return count;                                                   \
+}
+
+DEF_TB_DEV_ATTRIB_STORE_BOOL(emulate_fua_write);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(emulate_caw);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(emulate_3pc);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(enforce_pr_isids);
+DEF_TB_DEV_ATTRIB_STORE_BOOL(is_nonrot);
+
+#define DEF_TB_DEV_ATTRIB_STORE_STUB(_name)                            \
+static ssize_t store_##_name(struct se_dev_attrib *da, const char *page,\
+               size_t count)                                           \
+{                                                                      \
+       printk_once(KERN_WARNING                                        \
+               "ignoring deprecated ##_name## attribute\n");   \
+       return count;                                                   \
+}
+
+DEF_TB_DEV_ATTRIB_STORE_STUB(emulate_dpo);
+DEF_TB_DEV_ATTRIB_STORE_STUB(emulate_fua_read);
+
+static void dev_set_t10_wwn_model_alias(struct se_device *dev)
+{
+       const char *configname;
+
+       configname = config_item_name(&dev->dev_group.cg_item);
+       if (strlen(configname) >= 16) {
+               pr_warn("dev[%p]: Backstore name '%s' is too long for "
+                       "INQUIRY_MODEL, truncating to 16 bytes\n", dev,
+                       configname);
+       }
+       snprintf(&dev->t10_wwn.model[0], 16, "%s", configname);
+}
+
+static ssize_t store_emulate_model_alias(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       struct se_device *dev = da->da_dev;
+       bool flag;
+       int ret;
+
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to change model alias"
+                       " while export_count is %d\n",
+                       dev, dev->export_count);
+               return -EINVAL;
+       }
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag) {
+               dev_set_t10_wwn_model_alias(dev);
+       } else {
+               strncpy(&dev->t10_wwn.model[0],
+                       dev->transport->inquiry_prod, 16);
+       }
+       da->emulate_model_alias = flag;
+       return count;
+}
+
+static ssize_t store_emulate_write_cache(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag && da->da_dev->transport->get_write_cache) {
+               pr_err("emulate_write_cache not supported for this device\n");
+               return -EINVAL;
+       }
+
+       da->emulate_write_cache = flag;
+       pr_debug("dev[%p]: SE Device WRITE_CACHE_EMULATION flag: %d\n",
+                       da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_emulate_ua_intlck_ctrl(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       u32 val;
+       int ret;
+
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (val != 0 && val != 1 && val != 2) {
+               pr_err("Illegal value %d\n", val);
+               return -EINVAL;
+       }
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device"
+                       " UA_INTRLCK_CTRL while export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+       da->emulate_ua_intlck_ctrl = val;
+       pr_debug("dev[%p]: SE Device UA_INTRLCK_CTRL flag: %d\n",
+               da->da_dev, val);
+       return count;
+}
+
+static ssize_t store_emulate_tas(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device TAS while"
+                       " export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+       da->emulate_tas = flag;
+       pr_debug("dev[%p]: SE Device TASK_ABORTED status bit: %s\n",
+               da->da_dev, flag ? "Enabled" : "Disabled");
+
+       return count;
+}
+
+static ssize_t store_emulate_tpu(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       /*
+        * We expect this value to be non-zero when generic Block Layer
+        * Discard supported is detected iblock_create_virtdevice().
+        */
+       if (flag && !da->max_unmap_block_desc_count) {
+               pr_err("Generic Block Discard not supported\n");
+               return -ENOSYS;
+       }
+
+       da->emulate_tpu = flag;
+       pr_debug("dev[%p]: SE Device Thin Provisioning UNMAP bit: %d\n",
+               da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_emulate_tpws(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       /*
+        * We expect this value to be non-zero when generic Block Layer
+        * Discard supported is detected iblock_create_virtdevice().
+        */
+       if (flag && !da->max_unmap_block_desc_count) {
+               pr_err("Generic Block Discard not supported\n");
+               return -ENOSYS;
+       }
+
+       da->emulate_tpws = flag;
+       pr_debug("dev[%p]: SE Device Thin Provisioning WRITE_SAME: %d\n",
+                               da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_pi_prot_type(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       int old_prot = da->pi_prot_type, ret;
+       struct se_device *dev = da->da_dev;
+       u32 flag;
+
+       ret = kstrtou32(page, 0, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag != 0 && flag != 1 && flag != 2 && flag != 3) {
+               pr_err("Illegal value %d for pi_prot_type\n", flag);
+               return -EINVAL;
+       }
+       if (flag == 2) {
+               pr_err("DIF TYPE2 protection currently not supported\n");
+               return -ENOSYS;
+       }
+       if (da->hw_pi_prot_type) {
+               pr_warn("DIF protection enabled on underlying hardware,"
+                       " ignoring\n");
+               return count;
+       }
+       if (!dev->transport->init_prot || !dev->transport->free_prot) {
+               /* 0 is only allowed value for non-supporting backends */
+               if (flag == 0)
+                       return 0;
+
+               pr_err("DIF protection not supported by backend: %s\n",
+                      dev->transport->name);
+               return -ENOSYS;
+       }
+       if (!(dev->dev_flags & DF_CONFIGURED)) {
+               pr_err("DIF protection requires device to be configured\n");
+               return -ENODEV;
+       }
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device PROT type while"
+                      " export_count is %d\n", dev, dev->export_count);
+               return -EINVAL;
+       }
+
+       da->pi_prot_type = flag;
+
+       if (flag && !old_prot) {
+               ret = dev->transport->init_prot(dev);
+               if (ret) {
+                       da->pi_prot_type = old_prot;
+                       return ret;
+               }
+
+       } else if (!flag && old_prot) {
+               dev->transport->free_prot(dev);
+       }
+
+       pr_debug("dev[%p]: SE Device Protection Type: %d\n", dev, flag);
+       return count;
+}
+
+static ssize_t store_pi_prot_format(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       struct se_device *dev = da->da_dev;
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (!flag)
+               return count;
+
+       if (!dev->transport->format_prot) {
+               pr_err("DIF protection format not supported by backend %s\n",
+                      dev->transport->name);
+               return -ENOSYS;
+       }
+       if (!(dev->dev_flags & DF_CONFIGURED)) {
+               pr_err("DIF protection format requires device to be configured\n");
+               return -ENODEV;
+       }
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to format SE Device PROT type while"
+                      " export_count is %d\n", dev, dev->export_count);
+               return -EINVAL;
+       }
+
+       ret = dev->transport->format_prot(dev);
+       if (ret)
+               return ret;
+
+       pr_debug("dev[%p]: SE Device Protection Format complete\n", dev);
+       return count;
+}
+
+static ssize_t store_force_pr_aptpl(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to set force_pr_aptpl while"
+                      " export_count is %d\n",
+                      da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+
+       da->force_pr_aptpl = flag;
+       pr_debug("dev[%p]: SE Device force_pr_aptpl: %d\n", da->da_dev, flag);
+       return count;
+}
+
+static ssize_t store_emulate_rest_reord(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       bool flag;
+       int ret;
+
+       ret = strtobool(page, &flag);
+       if (ret < 0)
+               return ret;
+
+       if (flag != 0) {
+               printk(KERN_ERR "dev[%p]: SE Device emulation of restricted"
+                       " reordering not implemented\n", da->da_dev);
+               return -ENOSYS;
+       }
+       da->emulate_rest_reord = flag;
+       pr_debug("dev[%p]: SE Device emulate_rest_reord: %d\n",
+               da->da_dev, flag);
+       return count;
 }
 
-#define DEF_TB_DEV_ATTRIB(_backend, _name)                             \
-DEF_TB_DEV_ATTRIB_SHOW(_backend, _name);                               \
-DEF_TB_DEV_ATTRIB_STORE(_backend, _name);
+/*
+ * Note, this can only be called on unexported SE Device Object.
+ */
+static ssize_t store_queue_depth(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       struct se_device *dev = da->da_dev;
+       u32 val;
+       int ret;
 
-#define DEF_TB_DEV_ATTRIB_RO(_backend, name)                           \
-DEF_TB_DEV_ATTRIB_SHOW(_backend, name);
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device TCQ while"
+                       " export_count is %d\n",
+                       dev, dev->export_count);
+               return -EINVAL;
+       }
+       if (!val) {
+               pr_err("dev[%p]: Illegal ZERO value for queue_depth\n", dev);
+               return -EINVAL;
+       }
+
+       if (val > dev->dev_attrib.queue_depth) {
+               if (val > dev->dev_attrib.hw_queue_depth) {
+                       pr_err("dev[%p]: Passed queue_depth:"
+                               " %u exceeds TCM/SE_Device MAX"
+                               " TCQ: %u\n", dev, val,
+                               dev->dev_attrib.hw_queue_depth);
+                       return -EINVAL;
+               }
+       }
+       da->queue_depth = dev->queue_depth = val;
+       pr_debug("dev[%p]: SE Device TCQ Depth changed to: %u\n", dev, val);
+       return count;
+}
+
+static ssize_t store_optimal_sectors(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       u32 val;
+       int ret;
+
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device"
+                       " optimal_sectors while export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+       if (val > da->hw_max_sectors) {
+               pr_err("dev[%p]: Passed optimal_sectors %u cannot be"
+                       " greater than hw_max_sectors: %u\n",
+                       da->da_dev, val, da->hw_max_sectors);
+               return -EINVAL;
+       }
+
+       da->optimal_sectors = val;
+       pr_debug("dev[%p]: SE Device optimal_sectors changed to %u\n",
+                       da->da_dev, val);
+       return count;
+}
+
+static ssize_t store_block_size(struct se_dev_attrib *da,
+               const char *page, size_t count)
+{
+       u32 val;
+       int ret;
+
+       ret = kstrtou32(page, 0, &val);
+       if (ret < 0)
+               return ret;
+
+       if (da->da_dev->export_count) {
+               pr_err("dev[%p]: Unable to change SE Device block_size"
+                       " while export_count is %d\n",
+                       da->da_dev, da->da_dev->export_count);
+               return -EINVAL;
+       }
+
+       if (val != 512 && val != 1024 && val != 2048 && val != 4096) {
+               pr_err("dev[%p]: Illegal value for block_device: %u"
+                       " for SE device, must be 512, 1024, 2048 or 4096\n",
+                       da->da_dev, val);
+               return -EINVAL;
+       }
+
+       da->block_size = val;
+       if (da->max_bytes_per_io)
+               da->hw_max_sectors = da->max_bytes_per_io / val;
+
+       pr_debug("dev[%p]: SE Device block_size changed to %u\n",
+                       da->da_dev, val);
+       return count;
+}
 
 CONFIGFS_EATTR_STRUCT(target_backend_dev_attrib, se_dev_attrib);
 #define TB_DEV_ATTR(_backend, _name, _mode)                            \
 static struct target_backend_dev_attrib_attribute _backend##_dev_attrib_##_name = \
-               __CONFIGFS_EATTR(_name, _mode,                          \
-               _backend##_dev_show_attr_##_name,                       \
-               _backend##_dev_store_attr_##_name);
+       __CONFIGFS_EATTR(_name, _mode,                                  \
+       show_##_name,                                                   \
+       store_##_name);
 
-#define TB_DEV_ATTR_RO(_backend, _name)                                                \
+#define TB_DEV_ATTR_RO(_backend, _name)                                        \
 static struct target_backend_dev_attrib_attribute _backend##_dev_attrib_##_name = \
        __CONFIGFS_EATTR_RO(_name,                                      \
-       _backend##_dev_show_attr_##_name);
-
-DEF_TB_DEV_ATTRIB(target_core, emulate_model_alias);
-DEF_TB_DEV_ATTRIB(target_core, emulate_dpo);
-DEF_TB_DEV_ATTRIB(target_core, emulate_fua_write);
-DEF_TB_DEV_ATTRIB(target_core, emulate_fua_read);
-DEF_TB_DEV_ATTRIB(target_core, emulate_write_cache);
-DEF_TB_DEV_ATTRIB(target_core, emulate_ua_intlck_ctrl);
-DEF_TB_DEV_ATTRIB(target_core, emulate_tas);
-DEF_TB_DEV_ATTRIB(target_core, emulate_tpu);
-DEF_TB_DEV_ATTRIB(target_core, emulate_tpws);
-DEF_TB_DEV_ATTRIB(target_core, emulate_caw);
-DEF_TB_DEV_ATTRIB(target_core, emulate_3pc);
-DEF_TB_DEV_ATTRIB(target_core, pi_prot_type);
-DEF_TB_DEV_ATTRIB_RO(target_core, hw_pi_prot_type);
-DEF_TB_DEV_ATTRIB(target_core, pi_prot_format);
-DEF_TB_DEV_ATTRIB(target_core, enforce_pr_isids);
-DEF_TB_DEV_ATTRIB(target_core, is_nonrot);
-DEF_TB_DEV_ATTRIB(target_core, emulate_rest_reord);
-DEF_TB_DEV_ATTRIB(target_core, force_pr_aptpl);
-DEF_TB_DEV_ATTRIB_RO(target_core, hw_block_size);
-DEF_TB_DEV_ATTRIB(target_core, block_size);
-DEF_TB_DEV_ATTRIB_RO(target_core, hw_max_sectors);
-DEF_TB_DEV_ATTRIB(target_core, optimal_sectors);
-DEF_TB_DEV_ATTRIB_RO(target_core, hw_queue_depth);
-DEF_TB_DEV_ATTRIB(target_core, queue_depth);
-DEF_TB_DEV_ATTRIB(target_core, max_unmap_lba_count);
-DEF_TB_DEV_ATTRIB(target_core, max_unmap_block_desc_count);
-DEF_TB_DEV_ATTRIB(target_core, unmap_granularity);
-DEF_TB_DEV_ATTRIB(target_core, unmap_granularity_alignment);
-DEF_TB_DEV_ATTRIB(target_core, max_write_same_len);
+       show_##_name);
 
 TB_DEV_ATTR(target_core, emulate_model_alias, S_IRUGO | S_IWUSR);
 TB_DEV_ATTR(target_core, emulate_dpo, S_IRUGO | S_IWUSR);
@@ -620,11 +1045,6 @@ struct configfs_attribute *sbc_attrib_attrs[] = {
 };
 EXPORT_SYMBOL(sbc_attrib_attrs);
 
-DEF_TB_DEV_ATTRIB_RO(target_pt, hw_pi_prot_type);
-DEF_TB_DEV_ATTRIB_RO(target_pt, hw_block_size);
-DEF_TB_DEV_ATTRIB_RO(target_pt, hw_max_sectors);
-DEF_TB_DEV_ATTRIB_RO(target_pt, hw_queue_depth);
-
 TB_DEV_ATTR_RO(target_pt, hw_pi_prot_type);
 TB_DEV_ATTR_RO(target_pt, hw_block_size);
 TB_DEV_ATTR_RO(target_pt, hw_max_sectors);
index af906f1caaa78af6cc852b1fd7748ecb2ec5b5e5..22f1e180d8ee64a0b2f13a66560ee9e96a018915 100644 (file)
@@ -665,510 +665,6 @@ bool se_dev_check_wce(struct se_device *dev)
        return wce;
 }
 
-int se_dev_set_max_unmap_lba_count(
-       struct se_device *dev,
-       u32 max_unmap_lba_count)
-{
-       dev->dev_attrib.max_unmap_lba_count = max_unmap_lba_count;
-       pr_debug("dev[%p]: Set max_unmap_lba_count: %u\n",
-                       dev, dev->dev_attrib.max_unmap_lba_count);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_max_unmap_lba_count);
-
-int se_dev_set_max_unmap_block_desc_count(
-       struct se_device *dev,
-       u32 max_unmap_block_desc_count)
-{
-       dev->dev_attrib.max_unmap_block_desc_count =
-               max_unmap_block_desc_count;
-       pr_debug("dev[%p]: Set max_unmap_block_desc_count: %u\n",
-                       dev, dev->dev_attrib.max_unmap_block_desc_count);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_max_unmap_block_desc_count);
-
-int se_dev_set_unmap_granularity(
-       struct se_device *dev,
-       u32 unmap_granularity)
-{
-       dev->dev_attrib.unmap_granularity = unmap_granularity;
-       pr_debug("dev[%p]: Set unmap_granularity: %u\n",
-                       dev, dev->dev_attrib.unmap_granularity);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_unmap_granularity);
-
-int se_dev_set_unmap_granularity_alignment(
-       struct se_device *dev,
-       u32 unmap_granularity_alignment)
-{
-       dev->dev_attrib.unmap_granularity_alignment = unmap_granularity_alignment;
-       pr_debug("dev[%p]: Set unmap_granularity_alignment: %u\n",
-                       dev, dev->dev_attrib.unmap_granularity_alignment);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_unmap_granularity_alignment);
-
-int se_dev_set_max_write_same_len(
-       struct se_device *dev,
-       u32 max_write_same_len)
-{
-       dev->dev_attrib.max_write_same_len = max_write_same_len;
-       pr_debug("dev[%p]: Set max_write_same_len: %u\n",
-                       dev, dev->dev_attrib.max_write_same_len);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_max_write_same_len);
-
-static void dev_set_t10_wwn_model_alias(struct se_device *dev)
-{
-       const char *configname;
-
-       configname = config_item_name(&dev->dev_group.cg_item);
-       if (strlen(configname) >= 16) {
-               pr_warn("dev[%p]: Backstore name '%s' is too long for "
-                       "INQUIRY_MODEL, truncating to 16 bytes\n", dev,
-                       configname);
-       }
-       snprintf(&dev->t10_wwn.model[0], 16, "%s", configname);
-}
-
-int se_dev_set_emulate_model_alias(struct se_device *dev, int flag)
-{
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to change model alias"
-                       " while export_count is %d\n",
-                       dev, dev->export_count);
-                       return -EINVAL;
-       }
-
-       if (flag != 0 && flag != 1) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-
-       if (flag) {
-               dev_set_t10_wwn_model_alias(dev);
-       } else {
-               strncpy(&dev->t10_wwn.model[0],
-                       dev->transport->inquiry_prod, 16);
-       }
-       dev->dev_attrib.emulate_model_alias = flag;
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_model_alias);
-
-int se_dev_set_emulate_dpo(struct se_device *dev, int flag)
-{
-       printk_once(KERN_WARNING
-               "ignoring deprecated emulate_dpo attribute\n");
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_dpo);
-
-int se_dev_set_emulate_fua_write(struct se_device *dev, int flag)
-{
-       if (flag != 0 && flag != 1) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       if (flag &&
-           dev->transport->get_write_cache) {
-               pr_warn("emulate_fua_write not supported for this device, ignoring\n");
-               return 0;
-       }
-       if (dev->export_count) {
-               pr_err("emulate_fua_write cannot be changed with active"
-                      " exports: %d\n", dev->export_count);
-               return -EINVAL;
-       }
-       dev->dev_attrib.emulate_fua_write = flag;
-       pr_debug("dev[%p]: SE Device Forced Unit Access WRITEs: %d\n",
-                       dev, dev->dev_attrib.emulate_fua_write);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_fua_write);
-
-int se_dev_set_emulate_fua_read(struct se_device *dev, int flag)
-{
-       printk_once(KERN_WARNING
-               "ignoring deprecated emulate_fua_read attribute\n");
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_fua_read);
-
-int se_dev_set_emulate_write_cache(struct se_device *dev, int flag)
-{
-       if (flag != 0 && flag != 1) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       if (flag &&
-           dev->transport->get_write_cache) {
-               pr_err("emulate_write_cache not supported for this device\n");
-               return -EINVAL;
-       }
-       if (dev->export_count) {
-               pr_err("emulate_write_cache cannot be changed with active"
-                      " exports: %d\n", dev->export_count);
-               return -EINVAL;
-       }
-       dev->dev_attrib.emulate_write_cache = flag;
-       pr_debug("dev[%p]: SE Device WRITE_CACHE_EMULATION flag: %d\n",
-                       dev, dev->dev_attrib.emulate_write_cache);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_write_cache);
-
-int se_dev_set_emulate_ua_intlck_ctrl(struct se_device *dev, int flag)
-{
-       if ((flag != 0) && (flag != 1) && (flag != 2)) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to change SE Device"
-                       " UA_INTRLCK_CTRL while export_count is %d\n",
-                       dev, dev->export_count);
-               return -EINVAL;
-       }
-       dev->dev_attrib.emulate_ua_intlck_ctrl = flag;
-       pr_debug("dev[%p]: SE Device UA_INTRLCK_CTRL flag: %d\n",
-               dev, dev->dev_attrib.emulate_ua_intlck_ctrl);
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_ua_intlck_ctrl);
-
-int se_dev_set_emulate_tas(struct se_device *dev, int flag)
-{
-       if ((flag != 0) && (flag != 1)) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to change SE Device TAS while"
-                       " export_count is %d\n",
-                       dev, dev->export_count);
-               return -EINVAL;
-       }
-       dev->dev_attrib.emulate_tas = flag;
-       pr_debug("dev[%p]: SE Device TASK_ABORTED status bit: %s\n",
-               dev, (dev->dev_attrib.emulate_tas) ? "Enabled" : "Disabled");
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_tas);
-
-int se_dev_set_emulate_tpu(struct se_device *dev, int flag)
-{
-       if ((flag != 0) && (flag != 1)) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       /*
-        * We expect this value to be non-zero when generic Block Layer
-        * Discard supported is detected iblock_create_virtdevice().
-        */
-       if (flag && !dev->dev_attrib.max_unmap_block_desc_count) {
-               pr_err("Generic Block Discard not supported\n");
-               return -ENOSYS;
-       }
-
-       dev->dev_attrib.emulate_tpu = flag;
-       pr_debug("dev[%p]: SE Device Thin Provisioning UNMAP bit: %d\n",
-                               dev, flag);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_tpu);
-
-int se_dev_set_emulate_tpws(struct se_device *dev, int flag)
-{
-       if ((flag != 0) && (flag != 1)) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       /*
-        * We expect this value to be non-zero when generic Block Layer
-        * Discard supported is detected iblock_create_virtdevice().
-        */
-       if (flag && !dev->dev_attrib.max_unmap_block_desc_count) {
-               pr_err("Generic Block Discard not supported\n");
-               return -ENOSYS;
-       }
-
-       dev->dev_attrib.emulate_tpws = flag;
-       pr_debug("dev[%p]: SE Device Thin Provisioning WRITE_SAME: %d\n",
-                               dev, flag);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_tpws);
-
-int se_dev_set_emulate_caw(struct se_device *dev, int flag)
-{
-       if (flag != 0 && flag != 1) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       dev->dev_attrib.emulate_caw = flag;
-       pr_debug("dev[%p]: SE Device CompareAndWrite (AtomicTestandSet): %d\n",
-                dev, flag);
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_caw);
-
-int se_dev_set_emulate_3pc(struct se_device *dev, int flag)
-{
-       if (flag != 0 && flag != 1) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       dev->dev_attrib.emulate_3pc = flag;
-       pr_debug("dev[%p]: SE Device 3rd Party Copy (EXTENDED_COPY): %d\n",
-               dev, flag);
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_3pc);
-
-int se_dev_set_pi_prot_type(struct se_device *dev, int flag)
-{
-       int rc, old_prot = dev->dev_attrib.pi_prot_type;
-
-       if (flag != 0 && flag != 1 && flag != 2 && flag != 3) {
-               pr_err("Illegal value %d for pi_prot_type\n", flag);
-               return -EINVAL;
-       }
-       if (flag == 2) {
-               pr_err("DIF TYPE2 protection currently not supported\n");
-               return -ENOSYS;
-       }
-       if (dev->dev_attrib.hw_pi_prot_type) {
-               pr_warn("DIF protection enabled on underlying hardware,"
-                       " ignoring\n");
-               return 0;
-       }
-       if (!dev->transport->init_prot || !dev->transport->free_prot) {
-               /* 0 is only allowed value for non-supporting backends */
-               if (flag == 0)
-                       return 0;
-
-               pr_err("DIF protection not supported by backend: %s\n",
-                      dev->transport->name);
-               return -ENOSYS;
-       }
-       if (!(dev->dev_flags & DF_CONFIGURED)) {
-               pr_err("DIF protection requires device to be configured\n");
-               return -ENODEV;
-       }
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to change SE Device PROT type while"
-                      " export_count is %d\n", dev, dev->export_count);
-               return -EINVAL;
-       }
-
-       dev->dev_attrib.pi_prot_type = flag;
-
-       if (flag && !old_prot) {
-               rc = dev->transport->init_prot(dev);
-               if (rc) {
-                       dev->dev_attrib.pi_prot_type = old_prot;
-                       return rc;
-               }
-
-       } else if (!flag && old_prot) {
-               dev->transport->free_prot(dev);
-       }
-       pr_debug("dev[%p]: SE Device Protection Type: %d\n", dev, flag);
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_pi_prot_type);
-
-int se_dev_set_pi_prot_format(struct se_device *dev, int flag)
-{
-       int rc;
-
-       if (!flag)
-               return 0;
-
-       if (flag != 1) {
-               pr_err("Illegal value %d for pi_prot_format\n", flag);
-               return -EINVAL;
-       }
-       if (!dev->transport->format_prot) {
-               pr_err("DIF protection format not supported by backend %s\n",
-                      dev->transport->name);
-               return -ENOSYS;
-       }
-       if (!(dev->dev_flags & DF_CONFIGURED)) {
-               pr_err("DIF protection format requires device to be configured\n");
-               return -ENODEV;
-       }
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to format SE Device PROT type while"
-                      " export_count is %d\n", dev, dev->export_count);
-               return -EINVAL;
-       }
-
-       rc = dev->transport->format_prot(dev);
-       if (rc)
-               return rc;
-
-       pr_debug("dev[%p]: SE Device Protection Format complete\n", dev);
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_pi_prot_format);
-
-int se_dev_set_enforce_pr_isids(struct se_device *dev, int flag)
-{
-       if ((flag != 0) && (flag != 1)) {
-               pr_err("Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       dev->dev_attrib.enforce_pr_isids = flag;
-       pr_debug("dev[%p]: SE Device enforce_pr_isids bit: %s\n", dev,
-               (dev->dev_attrib.enforce_pr_isids) ? "Enabled" : "Disabled");
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_enforce_pr_isids);
-
-int se_dev_set_force_pr_aptpl(struct se_device *dev, int flag)
-{
-       if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to set force_pr_aptpl while"
-                      " export_count is %d\n", dev, dev->export_count);
-               return -EINVAL;
-       }
-
-       dev->dev_attrib.force_pr_aptpl = flag;
-       pr_debug("dev[%p]: SE Device force_pr_aptpl: %d\n", dev, flag);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_force_pr_aptpl);
-
-int se_dev_set_is_nonrot(struct se_device *dev, int flag)
-{
-       if ((flag != 0) && (flag != 1)) {
-               printk(KERN_ERR "Illegal value %d\n", flag);
-               return -EINVAL;
-       }
-       dev->dev_attrib.is_nonrot = flag;
-       pr_debug("dev[%p]: SE Device is_nonrot bit: %d\n",
-              dev, flag);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_is_nonrot);
-
-int se_dev_set_emulate_rest_reord(struct se_device *dev, int flag)
-{
-       if (flag != 0) {
-               printk(KERN_ERR "dev[%p]: SE Device emulation of restricted"
-                       " reordering not implemented\n", dev);
-               return -ENOSYS;
-       }
-       dev->dev_attrib.emulate_rest_reord = flag;
-       pr_debug("dev[%p]: SE Device emulate_rest_reord: %d\n", dev, flag);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_emulate_rest_reord);
-
-/*
- * Note, this can only be called on unexported SE Device Object.
- */
-int se_dev_set_queue_depth(struct se_device *dev, u32 queue_depth)
-{
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to change SE Device TCQ while"
-                       " export_count is %d\n",
-                       dev, dev->export_count);
-               return -EINVAL;
-       }
-       if (!queue_depth) {
-               pr_err("dev[%p]: Illegal ZERO value for queue"
-                       "_depth\n", dev);
-               return -EINVAL;
-       }
-
-       if (queue_depth > dev->dev_attrib.queue_depth) {
-               if (queue_depth > dev->dev_attrib.hw_queue_depth) {
-                       pr_err("dev[%p]: Passed queue_depth:"
-                               " %u exceeds TCM/SE_Device MAX"
-                               " TCQ: %u\n", dev, queue_depth,
-                               dev->dev_attrib.hw_queue_depth);
-                       return -EINVAL;
-               }
-       }
-       dev->dev_attrib.queue_depth = dev->queue_depth = queue_depth;
-       pr_debug("dev[%p]: SE Device TCQ Depth changed to: %u\n",
-                       dev, queue_depth);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_queue_depth);
-
-int se_dev_set_optimal_sectors(struct se_device *dev, u32 optimal_sectors)
-{
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to change SE Device"
-                       " optimal_sectors while export_count is %d\n",
-                       dev, dev->export_count);
-               return -EINVAL;
-       }
-       if (optimal_sectors > dev->dev_attrib.hw_max_sectors) {
-               pr_err("dev[%p]: Passed optimal_sectors %u cannot be"
-                       " greater than hw_max_sectors: %u\n", dev,
-                       optimal_sectors, dev->dev_attrib.hw_max_sectors);
-               return -EINVAL;
-       }
-
-       dev->dev_attrib.optimal_sectors = optimal_sectors;
-       pr_debug("dev[%p]: SE Device optimal_sectors changed to %u\n",
-                       dev, optimal_sectors);
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_optimal_sectors);
-
-int se_dev_set_block_size(struct se_device *dev, u32 block_size)
-{
-       if (dev->export_count) {
-               pr_err("dev[%p]: Unable to change SE Device block_size"
-                       " while export_count is %d\n",
-                       dev, dev->export_count);
-               return -EINVAL;
-       }
-
-       if ((block_size != 512) &&
-           (block_size != 1024) &&
-           (block_size != 2048) &&
-           (block_size != 4096)) {
-               pr_err("dev[%p]: Illegal value for block_device: %u"
-                       " for SE device, must be 512, 1024, 2048 or 4096\n",
-                       dev, block_size);
-               return -EINVAL;
-       }
-
-       dev->dev_attrib.block_size = block_size;
-       pr_debug("dev[%p]: SE Device block_size changed to %u\n",
-                       dev, block_size);
-
-       if (dev->dev_attrib.max_bytes_per_io)
-               dev->dev_attrib.hw_max_sectors =
-                       dev->dev_attrib.max_bytes_per_io / block_size;
-
-       return 0;
-}
-EXPORT_SYMBOL(se_dev_set_block_size);
-
 int core_dev_add_lun(
        struct se_portal_group *tpg,
        struct se_device *dev,
index 2d561051649464aacf995522b4f4512b28adb684..566c7d27f6d0e1049589297d2969351a123db795 100644 (file)
@@ -94,35 +94,6 @@ sense_reason_t       transport_generic_map_mem_to_cmd(struct se_cmd *,
                struct scatterlist *, u32, struct scatterlist *, u32);
 
 bool   target_lun_is_rdonly(struct se_cmd *);
-
-/* attribute helpers from target_core_device.c for backend drivers */
-bool   se_dev_check_wce(struct se_device *);
-int    se_dev_set_max_unmap_lba_count(struct se_device *, u32);
-int    se_dev_set_max_unmap_block_desc_count(struct se_device *, u32);
-int    se_dev_set_unmap_granularity(struct se_device *, u32);
-int    se_dev_set_unmap_granularity_alignment(struct se_device *, u32);
-int    se_dev_set_max_write_same_len(struct se_device *, u32);
-int    se_dev_set_emulate_model_alias(struct se_device *, int);
-int    se_dev_set_emulate_dpo(struct se_device *, int);
-int    se_dev_set_emulate_fua_write(struct se_device *, int);
-int    se_dev_set_emulate_fua_read(struct se_device *, int);
-int    se_dev_set_emulate_write_cache(struct se_device *, int);
-int    se_dev_set_emulate_ua_intlck_ctrl(struct se_device *, int);
-int    se_dev_set_emulate_tas(struct se_device *, int);
-int    se_dev_set_emulate_tpu(struct se_device *, int);
-int    se_dev_set_emulate_tpws(struct se_device *, int);
-int    se_dev_set_emulate_caw(struct se_device *, int);
-int    se_dev_set_emulate_3pc(struct se_device *, int);
-int    se_dev_set_pi_prot_type(struct se_device *, int);
-int    se_dev_set_pi_prot_format(struct se_device *, int);
-int    se_dev_set_enforce_pr_isids(struct se_device *, int);
-int    se_dev_set_force_pr_aptpl(struct se_device *, int);
-int    se_dev_set_is_nonrot(struct se_device *, int);
-int    se_dev_set_emulate_rest_reord(struct se_device *dev, int);
-int    se_dev_set_queue_depth(struct se_device *, u32);
-int    se_dev_set_max_sectors(struct se_device *, u32);
-int    se_dev_set_optimal_sectors(struct se_device *, u32);
-int    se_dev_set_block_size(struct se_device *, u32);
 sense_reason_t passthrough_parse_cdb(struct se_cmd *cmd,
        sense_reason_t (*exec_cmd)(struct se_cmd *cmd));