aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMartin K. Petersen <martin.petersen@oracle.com>2014-03-25 12:35:44 -0400
committerMartin K. Petersen <martin.petersen@oracle.com>2014-03-25 12:35:44 -0400
commit0bdeed274e16b3038a851552188512071974eea8 (patch)
tree123e80a1a364d4bd99db9f80c44ea533f673eb07
parentcd2aff22daff66dc883a11db55505c2d0cfc2cb5 (diff)
downloadlinux-0bdeed274e16b3038a851552188512071974eea8.tar.gz
patch xcopy.patchxcopy
-rw-r--r--Documentation/ABI/testing/sysfs-block9
-rw-r--r--block/blk-core.c5
-rw-r--r--block/blk-lib.c86
-rw-r--r--block/blk-merge.c7
-rw-r--r--block/blk-settings.c15
-rw-r--r--block/blk-sysfs.c10
-rw-r--r--block/ioctl.c35
-rw-r--r--drivers/scsi/scsi.c57
-rw-r--r--drivers/scsi/sd.c257
-rw-r--r--drivers/scsi/sd.h4
-rw-r--r--include/linux/bio.h7
-rw-r--r--include/linux/blk_types.h15
-rw-r--r--include/linux/blkdev.h15
-rw-r--r--include/scsi/scsi_device.h3
-rw-r--r--include/uapi/linux/fs.h1
15 files changed, 514 insertions, 12 deletions
diff --git a/Documentation/ABI/testing/sysfs-block b/Documentation/ABI/testing/sysfs-block
index 279da08f75419..d1304cc305f72 100644
--- a/Documentation/ABI/testing/sysfs-block
+++ b/Documentation/ABI/testing/sysfs-block
@@ -220,3 +220,12 @@ Description:
write_same_max_bytes is 0, write same is not supported
by the device.
+
+What: /sys/block/<disk>/queue/copy_max_bytes
+Date: January 2014
+Contact: Martin K. Petersen <martin.petersen@oracle.com>
+Description:
+ Devices that support copy offloading will set this value
+ to indicate the maximum buffer size in bytes that can be
+ copied in one operation. If the copy_max_bytes is 0 the
+ device does not support copy offload.
diff --git a/block/blk-core.c b/block/blk-core.c
index 853f92749202c..f512bc961ea35 100644
--- a/block/blk-core.c
+++ b/block/blk-core.c
@@ -1772,6 +1772,11 @@ generic_make_request_checks(struct bio *bio)
goto end_io;
}
+ if (bio->bi_rw & REQ_COPY && !bdev_copy_offload(bio->bi_bdev)) {
+ err = -EOPNOTSUPP;
+ goto end_io;
+ }
+
/*
* Various block parts want %current->io_context and lazy ioc
* allocation ends up trading a lot of pain for a small amount of
diff --git a/block/blk-lib.c b/block/blk-lib.c
index 97a733cf3d5f9..720d67839a1ae 100644
--- a/block/blk-lib.c
+++ b/block/blk-lib.c
@@ -305,3 +305,89 @@ int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
return __blkdev_issue_zeroout(bdev, sector, nr_sects, gfp_mask);
}
EXPORT_SYMBOL(blkdev_issue_zeroout);
+
+/**
+ * blkdev_issue_copy - queue a copy same operation
+ * @src_bdev: source blockdev
+ * @src_sector: source sector
+ * @dst_bdev: destination blockdev
+ * @dst_sector: destination sector
+ * @nr_sects: number of sectors to copy
+ * @gfp_mask: memory allocation flags (for bio_alloc)
+ *
+ * Description:
+ * Copy a block range from source device to target device.
+ */
+int blkdev_issue_copy(struct block_device *src_bdev, sector_t src_sector,
+ struct block_device *dst_bdev, sector_t dst_sector,
+ unsigned int nr_sects, gfp_t gfp_mask)
+{
+ DECLARE_COMPLETION_ONSTACK(wait);
+ struct request_queue *sq = bdev_get_queue(src_bdev);
+ struct request_queue *dq = bdev_get_queue(dst_bdev);
+ unsigned int max_copy_sectors;
+ struct bio_batch bb;
+ int ret = 0;
+
+ if (!sq || !dq)
+ return -ENXIO;
+
+ max_copy_sectors = min(sq->limits.max_copy_sectors,
+ dq->limits.max_copy_sectors);
+
+ if (max_copy_sectors == 0)
+ return -EOPNOTSUPP;
+
+ atomic_set(&bb.done, 1);
+ bb.flags = 1 << BIO_UPTODATE;
+ bb.wait = &wait;
+
+ while (nr_sects) {
+ struct bio *bio;
+ struct bio_copy *bc;
+ unsigned int chunk;
+
+ bc = kmalloc(sizeof(struct bio_copy), gfp_mask);
+ if (!bc) {
+ ret = -ENOMEM;
+ break;
+ }
+
+ bio = bio_alloc(gfp_mask, 1);
+ if (!bio) {
+ kfree(bc);
+ ret = -ENOMEM;
+ break;
+ }
+
+ chunk = min(nr_sects, max_copy_sectors);
+
+ bio->bi_iter.bi_sector = dst_sector;
+ bio->bi_iter.bi_size = chunk << 9;
+ bio->bi_end_io = bio_batch_end_io;
+ bio->bi_bdev = dst_bdev;
+ bio->bi_private = &bb;
+ bio->bi_copy = bc;
+
+ bc->bic_bdev = src_bdev;
+ bc->bic_sector = src_sector;
+
+ atomic_inc(&bb.done);
+ submit_bio(REQ_WRITE | REQ_COPY, bio);
+
+ src_sector += chunk;
+ dst_sector += chunk;
+ nr_sects -= chunk;
+ }
+
+ /* Wait for bios in-flight */
+ if (!atomic_dec_and_test(&bb.done))
+ wait_for_completion_io(&wait);
+
+ if (!test_bit(BIO_UPTODATE, &bb.flags))
+ ret = -ENOTSUPP;
+
+ return ret;
+}
+EXPORT_SYMBOL(blkdev_issue_copy);
+
diff --git a/block/blk-merge.c b/block/blk-merge.c
index 6c583f9c5b65d..0e1027e2e32b1 100644
--- a/block/blk-merge.c
+++ b/block/blk-merge.c
@@ -25,10 +25,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
* This should probably be returning 0, but blk_add_request_payload()
* (Christoph!!!!)
*/
- if (bio->bi_rw & REQ_DISCARD)
- return 1;
-
- if (bio->bi_rw & REQ_WRITE_SAME)
+ if (bio->bi_rw & (REQ_DISCARD | REQ_WRITE_SAME | REQ_COPY))
return 1;
fbio = bio;
@@ -182,7 +179,7 @@ static int __blk_bios_map_sg(struct request_queue *q, struct bio *bio,
nsegs = 0;
cluster = blk_queue_cluster(q);
- if (bio->bi_rw & REQ_DISCARD) {
+ if (bio->bi_rw & (REQ_DISCARD | REQ_COPY)) {
/*
* This is a hack - drivers should be neither modifying the
* biovec, nor relying on bi_vcnt - but because of
diff --git a/block/blk-settings.c b/block/blk-settings.c
index 5d21239bc8599..98801bcc02b02 100644
--- a/block/blk-settings.c
+++ b/block/blk-settings.c
@@ -114,6 +114,7 @@ void blk_set_default_limits(struct queue_limits *lim)
lim->max_segment_size = BLK_MAX_SEGMENT_SIZE;
lim->max_sectors = lim->max_hw_sectors = BLK_SAFE_MAX_SECTORS;
lim->max_write_same_sectors = 0;
+ lim->max_copy_sectors = 0;
lim->max_discard_sectors = 0;
lim->discard_granularity = 0;
lim->discard_alignment = 0;
@@ -147,6 +148,7 @@ void blk_set_stacking_limits(struct queue_limits *lim)
lim->max_segment_size = UINT_MAX;
lim->max_sectors = UINT_MAX;
lim->max_write_same_sectors = UINT_MAX;
+ lim->max_copy_sectors = UINT_MAX;
}
EXPORT_SYMBOL(blk_set_stacking_limits);
@@ -301,6 +303,18 @@ void blk_queue_max_write_same_sectors(struct request_queue *q,
EXPORT_SYMBOL(blk_queue_max_write_same_sectors);
/**
+ * blk_queue_max_copy_sectors - set max sectors for a single copy operation
+ * @q: the request queue for the device
+ * @max_copy_sectors: maximum number of sectors per copy operation
+ **/
+void blk_queue_max_copy_sectors(struct request_queue *q,
+ unsigned int max_copy_sectors)
+{
+ q->limits.max_copy_sectors = max_copy_sectors;
+}
+EXPORT_SYMBOL(blk_queue_max_copy_sectors);
+
+/**
* blk_queue_max_segments - set max hw segments for a request for this queue
* @q: the request queue for the device
* @max_segments: max number of segments
@@ -527,6 +541,7 @@ int blk_stack_limits(struct queue_limits *t, struct queue_limits *b,
t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors);
t->max_write_same_sectors = min(t->max_write_same_sectors,
b->max_write_same_sectors);
+ t->max_copy_sectors = min(t->max_copy_sectors, b->max_copy_sectors);
t->bounce_pfn = min_not_zero(t->bounce_pfn, b->bounce_pfn);
t->seg_boundary_mask = min_not_zero(t->seg_boundary_mask,
diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c
index 7500f876dae40..4ecdb7d325933 100644
--- a/block/blk-sysfs.c
+++ b/block/blk-sysfs.c
@@ -188,6 +188,11 @@ static ssize_t queue_write_same_max_show(struct request_queue *q, char *page)
(unsigned long long)q->limits.max_write_same_sectors << 9);
}
+static ssize_t queue_copy_max_show(struct request_queue *q, char *page)
+{
+ return sprintf(page, "%llu\n",
+ (unsigned long long)q->limits.max_copy_sectors << 9);
+}
static ssize_t
queue_max_sectors_store(struct request_queue *q, const char *page, size_t count)
@@ -401,6 +406,10 @@ static struct queue_sysfs_entry queue_write_same_max_entry = {
.show = queue_write_same_max_show,
};
+static struct queue_sysfs_entry queue_copy_max_entry = {
+ .attr = {.name = "copy_max_bytes", .mode = S_IRUGO },
+ .show = queue_copy_max_show,
+};
static struct queue_sysfs_entry queue_nonrot_entry = {
.attr = {.name = "rotational", .mode = S_IRUGO | S_IWUSR },
.show = queue_show_nonrot,
@@ -449,6 +458,7 @@ static struct attribute *default_attrs[] = {
&queue_discard_max_entry.attr,
&queue_discard_zeroes_data_entry.attr,
&queue_write_same_max_entry.attr,
+ &queue_copy_max_entry.attr,
&queue_nonrot_entry.attr,
&queue_nomerges_entry.attr,
&queue_rq_affinity_entry.attr,
diff --git a/block/ioctl.c b/block/ioctl.c
index 7d5c3b20af451..5efb6e666f186 100644
--- a/block/ioctl.c
+++ b/block/ioctl.c
@@ -201,6 +201,29 @@ static int blk_ioctl_zeroout(struct block_device *bdev, uint64_t start,
return blkdev_issue_zeroout(bdev, start, len, GFP_KERNEL);
}
+static int blk_ioctl_copy(struct block_device *bdev, uint64_t src_offset,
+ uint64_t dst_offset, uint64_t len)
+{
+ if (src_offset & 511)
+ return -EINVAL;
+ if (dst_offset & 511)
+ return -EINVAL;
+ if (len & 511)
+ return -EINVAL;
+ src_offset >>= 9;
+ dst_offset >>= 9;
+ len >>= 9;
+
+ if (src_offset + len > (i_size_read(bdev->bd_inode) >> 9))
+ return -EINVAL;
+
+ if (dst_offset + len > (i_size_read(bdev->bd_inode) >> 9))
+ return -EINVAL;
+
+ return blkdev_issue_copy(bdev, src_offset, bdev, dst_offset, len,
+ GFP_KERNEL);
+}
+
static int put_ushort(unsigned long arg, unsigned short val)
{
return put_user(val, (unsigned short __user *)arg);
@@ -328,6 +351,18 @@ int blkdev_ioctl(struct block_device *bdev, fmode_t mode, unsigned cmd,
return blk_ioctl_zeroout(bdev, range[0], range[1]);
}
+ case BLKCOPY: {
+ uint64_t range[3];
+
+ if (!(mode & FMODE_WRITE))
+ return -EBADF;
+
+ if (copy_from_user(range, (void __user *)arg, sizeof(range)))
+ return -EFAULT;
+
+ return blk_ioctl_copy(bdev, range[0], range[1], range[2]);
+ }
+
case HDIO_GETGEO: {
struct hd_geometry geo;
diff --git a/drivers/scsi/scsi.c b/drivers/scsi/scsi.c
index 91c0595d9a503..ed6831fa7987f 100644
--- a/drivers/scsi/scsi.c
+++ b/drivers/scsi/scsi.c
@@ -1030,6 +1030,62 @@ int scsi_get_vpd_page(struct scsi_device *sdev, u8 page, unsigned char *buf,
EXPORT_SYMBOL_GPL(scsi_get_vpd_page);
/**
+ * scsi_lookup_naa - Lookup NAA descriptor in VPD page 0x83
+ * @sdev: The device to ask
+ *
+ * Copy offloading requires us to know the NAA descriptor for both
+ * source and target device. This descriptor is mandatory in the Device
+ * Identification VPD page. Locate this descriptor in the returned VPD
+ * data so we don't have to do lookups for every copy command.
+ */
+static void scsi_lookup_naa(struct scsi_device *sdev)
+{
+ unsigned char *buf = sdev->vpd_pg83;
+ unsigned int len = sdev->vpd_pg83_len;
+
+ if (buf[1] != 0x83 || get_unaligned_be16(&buf[2]) == 0) {
+ sdev_printk(KERN_ERR, sdev,
+ "%s: VPD page 0x83 contains no descriptors\n",
+ __func__);
+ return;
+ }
+
+ buf += 4;
+ len -= 4;
+
+ do {
+ unsigned int desig_len = buf[3] + 4;
+
+ /* Binary code set */
+ if ((buf[0] & 0xf) != 1)
+ goto skip;
+
+ /* Target association */
+ if ((buf[1] >> 4) & 0x3)
+ goto skip;
+
+ /* NAA designator */
+ if ((buf[1] & 0xf) != 0x3)
+ goto skip;
+
+ sdev->naa = buf;
+ sdev->naa_len = desig_len;
+
+ return;
+
+ skip:
+ buf += desig_len;
+ len -= desig_len;
+
+ } while (len > 0);
+
+ sdev_printk(KERN_ERR, sdev,
+ "%s: VPD page 0x83 NAA descriptor not found\n", __func__);
+
+ return;
+}
+
+/**
* scsi_attach_vpd - Attach Vital Product Data to a SCSI device structure
* @sdev: The device to ask
*
@@ -1147,6 +1203,7 @@ retry_pg83:
spin_lock(&sdev->reconfig_lock);
sdev->vpd_pg83 = vpd_buf;
sdev->vpd_pg83_len = result;
+ scsi_lookup_naa(sdev);
spin_unlock(&sdev->reconfig_lock);
}
} else {
diff --git a/drivers/scsi/sd.c b/drivers/scsi/sd.c
index 36d1a23f14be9..129799321eced 100644
--- a/drivers/scsi/sd.c
+++ b/drivers/scsi/sd.c
@@ -100,6 +100,7 @@ MODULE_ALIAS_SCSI_DEVICE(TYPE_RBC);
static void sd_config_discard(struct scsi_disk *, unsigned int);
static void sd_config_write_same(struct scsi_disk *);
+static void sd_config_copy(struct scsi_disk *);
static int sd_revalidate_disk(struct gendisk *);
static void sd_unlock_native_capacity(struct gendisk *disk);
static int sd_probe(struct device *);
@@ -461,6 +462,48 @@ max_write_same_blocks_store(struct device *dev, struct device_attribute *attr,
}
static DEVICE_ATTR_RW(max_write_same_blocks);
+static ssize_t
+max_copy_blocks_show(struct device *dev, struct device_attribute *attr,
+ char *buf)
+{
+ struct scsi_disk *sdkp = to_scsi_disk(dev);
+
+ return snprintf(buf, 20, "%u\n", sdkp->max_copy_blocks);
+}
+
+static ssize_t
+max_copy_blocks_store(struct device *dev, struct device_attribute *attr,
+ const char *buf, size_t count)
+{
+ struct scsi_disk *sdkp = to_scsi_disk(dev);
+ struct scsi_device *sdp = sdkp->device;
+ unsigned long max;
+ int err;
+
+ if (!capable(CAP_SYS_ADMIN))
+ return -EACCES;
+
+ if (sdp->type != TYPE_DISK)
+ return -EINVAL;
+
+ err = kstrtoul(buf, 10, &max);
+
+ if (err)
+ return err;
+
+ if (max == 0)
+ sdp->no_copy = 1;
+ else if (max <= SD_MAX_COPY_BLOCKS) {
+ sdp->no_copy = 0;
+ sdkp->max_copy_blocks = max;
+ }
+
+ sd_config_copy(sdkp);
+
+ return count;
+}
+static DEVICE_ATTR_RW(max_copy_blocks);
+
static struct attribute *sd_disk_attrs[] = {
&dev_attr_cache_type.attr,
&dev_attr_FUA.attr,
@@ -472,6 +515,7 @@ static struct attribute *sd_disk_attrs[] = {
&dev_attr_thin_provisioning.attr,
&dev_attr_provisioning_mode.attr,
&dev_attr_max_write_same_blocks.attr,
+ &dev_attr_max_copy_blocks.attr,
&dev_attr_max_medium_access_timeouts.attr,
NULL,
};
@@ -828,6 +872,103 @@ static int sd_setup_write_same_cmnd(struct scsi_device *sdp, struct request *rq)
return ret;
}
+static void sd_config_copy(struct scsi_disk *sdkp)
+{
+ struct request_queue *q = sdkp->disk->queue;
+ unsigned int logical_block_size = sdkp->device->sector_size;
+
+ if (sdkp->device->no_copy)
+ sdkp->max_copy_blocks = 0;
+
+ /* Segment descriptor 0x02 has a 64k block limit */
+ sdkp->max_copy_blocks = min(sdkp->max_copy_blocks,
+ (u32)SD_MAX_CSD2_BLOCKS);
+
+ blk_queue_max_copy_sectors(q, sdkp->max_copy_blocks *
+ (logical_block_size >> 9));
+}
+
+static int sd_setup_copy_cmnd(struct scsi_device *sdp, struct request *rq)
+{
+ struct scsi_device *src_sdp, *dst_sdp;
+ sector_t src_lba, dst_lba;
+ unsigned int nr_blocks, buf_len, nr_bytes = blk_rq_bytes(rq);
+ int ret;
+ struct bio *bio = rq->bio;
+ struct page *page;
+ unsigned char *buf, *buf2;
+ unsigned int i;
+
+ if (!bio->bi_copy)
+ return BLKPREP_KILL;
+
+ dst_sdp = scsi_disk(rq->rq_disk)->device;
+ src_sdp = scsi_disk(bio->bi_copy->bic_bdev->bd_disk)->device;
+
+ if (src_sdp->vpd_invalid || dst_sdp->vpd_invalid)
+ return BLKPREP_KILL;
+
+ if (src_sdp->no_copy || dst_sdp->no_copy)
+ return BLKPREP_KILL;
+
+ if (src_sdp->sector_size != dst_sdp->sector_size)
+ return BLKPREP_KILL;
+
+ dst_lba = blk_rq_pos(rq) >> (ilog2(dst_sdp->sector_size) - 9);
+ src_lba = bio->bi_copy->bic_sector >> (ilog2(src_sdp->sector_size) - 9);
+ nr_blocks = blk_rq_sectors(rq) >> (ilog2(dst_sdp->sector_size) - 9);
+
+ page = alloc_page(GFP_ATOMIC | __GFP_ZERO);
+ if (!page)
+ return BLKPREP_DEFER;
+
+ rq->buffer = buf = buf2 = page_address(page);
+
+ /* Extended Copy (LID1) Parameter List (16 bytes) */
+ buf[0] = 0; /* LID */
+ buf[1] = 3 << 3; /* LID usage 11b */
+ put_unaligned_be16(32 + 32, &buf[2]); /* 32 bytes per E4 desc. */
+ put_unaligned_be32(28, &buf[8]); /* 28 bytes per B2B desc. */
+ buf += 16;
+
+ /* Source CSCD (32 bytes) */
+ buf[0] = 0xe4; /* Identification desc. */
+ memcpy(&buf[4], src_sdp->naa, src_sdp->naa_len);
+ buf += 32;
+
+ /* Destination CSCD (32 bytes) */
+ buf[0] = 0xe4; /* Identification desc. */
+ memcpy(&buf[4], dst_sdp->naa, dst_sdp->naa_len);
+ buf += 32;
+
+ /* Segment descriptor (28 bytes) */
+ buf[0] = 0x02; /* Block to block desc. */
+ put_unaligned_be16(0x18, &buf[2]); /* Descriptor length */
+ put_unaligned_be16(0, &buf[4]); /* Source is desc. 0 */
+ put_unaligned_be16(1, &buf[6]); /* Dest. is desc. 1 */
+ put_unaligned_be16(nr_blocks, &buf[10]);
+ put_unaligned_be64(src_lba, &buf[12]);
+ put_unaligned_be64(dst_lba, &buf[20]);
+
+ /* CDB */
+ memset(rq->cmd, 0, rq->cmd_len);
+ rq->cmd[0] = EXTENDED_COPY;
+ rq->cmd[1] = 0; /* LID1 */
+ buf_len = 16 + 32 + 32 + 28;
+ put_unaligned_be32(buf_len, &rq->cmd[10]);
+ rq->timeout = SD_COPY_TIMEOUT;
+
+ blk_add_request_payload(rq, page, buf_len);
+ ret = scsi_setup_blk_pc_cmnd(sdp, rq);
+ rq->__data_len = nr_bytes;
+
+ if (ret != BLKPREP_OK) {
+ __free_page(page);
+ rq->buffer = NULL;
+ }
+ return ret;
+}
+
static int scsi_setup_flush_cmnd(struct scsi_device *sdp, struct request *rq)
{
rq->timeout *= SD_FLUSH_TIMEOUT_MULTIPLIER;
@@ -842,7 +983,7 @@ static void sd_unprep_fn(struct request_queue *q, struct request *rq)
{
struct scsi_cmnd *SCpnt = rq->special;
- if (rq->cmd_flags & REQ_DISCARD) {
+ if (rq->cmd_flags & (REQ_DISCARD | REQ_COPY)) {
free_page((unsigned long)rq->buffer);
rq->buffer = NULL;
}
@@ -883,6 +1024,9 @@ static int sd_prep_fn(struct request_queue *q, struct request *rq)
} else if (rq->cmd_flags & REQ_WRITE_SAME) {
ret = sd_setup_write_same_cmnd(sdp, rq);
goto out;
+ } else if (rq->cmd_flags & REQ_COPY) {
+ ret = sd_setup_copy_cmnd(sdp, rq);
+ goto out;
} else if (rq->cmd_flags & REQ_FLUSH) {
ret = scsi_setup_flush_cmnd(sdp, rq);
goto out;
@@ -1663,7 +1807,8 @@ static int sd_done(struct scsi_cmnd *SCpnt)
unsigned char op = SCpnt->cmnd[0];
unsigned char unmap = SCpnt->cmnd[1] & 8;
- if (req->cmd_flags & REQ_DISCARD || req->cmd_flags & REQ_WRITE_SAME) {
+ if (req->cmd_flags & REQ_DISCARD || req->cmd_flags & REQ_WRITE_SAME ||
+ req->cmd_flags & REQ_COPY) {
if (!result) {
good_bytes = blk_rq_bytes(req);
scsi_set_resid(SCpnt, 0);
@@ -1722,6 +1867,14 @@ static int sd_done(struct scsi_cmnd *SCpnt)
/* INVALID COMMAND OPCODE or INVALID FIELD IN CDB */
if (sshdr.asc == 0x20 || sshdr.asc == 0x24) {
switch (op) {
+ case EXTENDED_COPY:
+ sdkp->device->no_copy = 1;
+ sd_config_copy(sdkp);
+
+ good_bytes = 0;
+ req->__data_len = blk_rq_bytes(req);
+ req->cmd_flags |= REQ_QUIET;
+ break;
case UNMAP:
sd_config_discard(sdkp, SD_LBP_DISABLE);
break;
@@ -2684,6 +2837,105 @@ static void sd_read_write_same(struct scsi_disk *sdkp, unsigned char *buffer)
sdkp->ws10 = 1;
}
+static void sd_read_copy_operations(struct scsi_disk *sdkp,
+ unsigned char *buffer)
+{
+ struct scsi_device *sdev = sdkp->device;
+ struct scsi_sense_hdr sshdr;
+ unsigned char cdb[16];
+ unsigned int result, len, i;
+ bool b2b_desc = false, id_desc = false;
+
+ if (sdev->naa_len == 0)
+ return;
+
+ /* Verify that the device has 3PC set in INQUIRY response */
+ if (sdev->inquiry_len < 6 || (sdev->inquiry[5] & (1 << 3)) == 0)
+ return;
+
+ /* Receive Copy Operation Parameters */
+ memset(cdb, 0, 16);
+ cdb[0] = RECEIVE_COPY_RESULTS;
+ cdb[1] = 0x3;
+ put_unaligned_be32(SD_BUF_SIZE, &cdb[10]);
+
+ memset(buffer, 0, SD_BUF_SIZE);
+ result = scsi_execute_req(sdev, cdb, DMA_FROM_DEVICE,
+ buffer, SD_BUF_SIZE, &sshdr,
+ SD_TIMEOUT, SD_MAX_RETRIES, NULL);
+
+ if (!scsi_status_is_good(result)) {
+ sd_printk(KERN_ERR, sdkp,
+ "%s: Receive Copy Operating Parameters failed\n",
+ __func__);
+ return;
+ }
+
+ /* The RCOP response is a minimum of 44 bytes long. First 4
+ * bytes contain the length of the remaining buffer, i.e. 40+
+ * bytes. Trailing the defined fields is a list of supported
+ * descriptors. We need at least 2 descriptors to drive the
+ * target, hence 42.
+ */
+ len = get_unaligned_be32(&buffer[0]);
+ if (len < 42) {
+ sd_printk(KERN_ERR, sdkp, "%s: result too short (%u)\n",
+ __func__, len);
+ return;
+ }
+
+ if ((buffer[4] & 1) == 0) {
+ sd_printk(KERN_ERR, sdkp, "%s: does not support SNLID\n",
+ __func__);
+ return;
+ }
+
+ if (get_unaligned_be16(&buffer[8]) < 2) {
+ sd_printk(KERN_ERR, sdkp,
+ "%s: Need 2 or more CSCD descriptors\n", __func__);
+ return;
+ }
+
+ if (get_unaligned_be16(&buffer[10]) < 1) {
+ sd_printk(KERN_ERR, sdkp,
+ "%s: Need 1 or more segment descriptor\n", __func__);
+ return;
+ }
+
+ if (len - 40 != buffer[43]) {
+ sd_printk(KERN_ERR, sdkp,
+ "%s: Buffer len and descriptor count mismatch " \
+ "(%u vs. %u)\n", __func__, len - 40, buffer[43]);
+ return;
+ }
+
+ for (i = 44 ; i < len + 4 ; i++) {
+ if (buffer[i] == 0x02)
+ b2b_desc = true;
+
+ if (buffer[i] == 0xe4)
+ id_desc = true;
+ }
+
+ if (!b2b_desc) {
+ sd_printk(KERN_ERR, sdkp,
+ "%s: No block 2 block descriptor (0x02)\n",
+ __func__);
+ return;
+ }
+
+ if (!id_desc) {
+ sd_printk(KERN_ERR, sdkp,
+ "%s: No identification descriptor (0xE4)\n",
+ __func__);
+ return;
+ }
+
+ sdkp->max_copy_blocks = get_unaligned_be32(&buffer[16])
+ >> ilog2(sdev->sector_size);
+ sd_config_copy(sdkp);
+}
+
static int sd_try_extended_inquiry(struct scsi_device *sdp)
{
/*
@@ -2744,6 +2996,7 @@ static int sd_revalidate_disk(struct gendisk *disk)
sd_read_cache_type(sdkp, buffer);
sd_read_app_tag_own(sdkp, buffer);
sd_read_write_same(sdkp, buffer);
+ sd_read_copy_operations(sdkp, buffer);
}
sdkp->first_scan = 0;
diff --git a/drivers/scsi/sd.h b/drivers/scsi/sd.h
index 26895ff247c59..9dd63c81420b7 100644
--- a/drivers/scsi/sd.h
+++ b/drivers/scsi/sd.h
@@ -19,6 +19,7 @@
*/
#define SD_FLUSH_TIMEOUT_MULTIPLIER 2
#define SD_WRITE_SAME_TIMEOUT (120 * HZ)
+#define SD_COPY_TIMEOUT (120 * HZ)
/*
* Number of allowed retries
@@ -46,6 +47,8 @@ enum {
enum {
SD_MAX_WS10_BLOCKS = 0xffff,
SD_MAX_WS16_BLOCKS = 0x7fffff,
+ SD_MAX_CSD2_BLOCKS = 0xffff,
+ SD_MAX_COPY_BLOCKS = 0xffffffff,
};
enum {
@@ -66,6 +69,7 @@ struct scsi_disk {
sector_t capacity; /* size in 512-byte sectors */
u32 max_ws_blocks;
u32 max_unmap_blocks;
+ u32 max_copy_blocks;
u32 unmap_granularity;
u32 unmap_alignment;
u32 index;
diff --git a/include/linux/bio.h b/include/linux/bio.h
index 5a4d39b4686be..7766fcf96d622 100644
--- a/include/linux/bio.h
+++ b/include/linux/bio.h
@@ -251,8 +251,8 @@ static inline unsigned bio_segments(struct bio *bio)
struct bvec_iter iter;
/*
- * We special case discard/write same, because they interpret bi_size
- * differently:
+ * We special case discard/write same/copy, because they
+ * interpret bi_size differently:
*/
if (bio->bi_rw & REQ_DISCARD)
@@ -261,6 +261,9 @@ static inline unsigned bio_segments(struct bio *bio)
if (bio->bi_rw & REQ_WRITE_SAME)
return 1;
+ if (bio->bi_rw & REQ_COPY)
+ return 1;
+
bio_for_each_segment(bv, bio, iter)
segs++;
diff --git a/include/linux/blk_types.h b/include/linux/blk_types.h
index bbc3a6c88fce3..f71a26b513a10 100644
--- a/include/linux/blk_types.h
+++ b/include/linux/blk_types.h
@@ -39,6 +39,11 @@ struct bvec_iter {
current bvec */
};
+struct bio_copy {
+ struct block_device *bic_bdev;
+ sector_t bic_sector;
+};
+
/*
* main unit of I/O for the block layer and lower layers (ie drivers and
* stacking drivers)
@@ -81,6 +86,7 @@ struct bio {
#if defined(CONFIG_BLK_DEV_INTEGRITY)
struct bio_integrity_payload *bi_integrity; /* data integrity */
#endif
+ struct bio_copy *bi_copy; /* TODO, use bi_integrity */
unsigned short bi_vcnt; /* how many bio_vec's */
@@ -160,6 +166,7 @@ enum rq_flag_bits {
__REQ_DISCARD, /* request to discard sectors */
__REQ_SECURE, /* secure discard (used with __REQ_DISCARD) */
__REQ_WRITE_SAME, /* write same block many times */
+ __REQ_COPY, /* copy block range */
__REQ_NOIDLE, /* don't anticipate more IO after this one */
__REQ_FUA, /* forced unit access */
@@ -201,6 +208,7 @@ enum rq_flag_bits {
#define REQ_PRIO (1ULL << __REQ_PRIO)
#define REQ_DISCARD (1ULL << __REQ_DISCARD)
#define REQ_WRITE_SAME (1ULL << __REQ_WRITE_SAME)
+#define REQ_COPY (1ULL << __REQ_COPY)
#define REQ_NOIDLE (1ULL << __REQ_NOIDLE)
#define REQ_FAILFAST_MASK \
@@ -208,14 +216,15 @@ enum rq_flag_bits {
#define REQ_COMMON_MASK \
(REQ_WRITE | REQ_FAILFAST_MASK | REQ_SYNC | REQ_META | REQ_PRIO | \
REQ_DISCARD | REQ_WRITE_SAME | REQ_NOIDLE | REQ_FLUSH | REQ_FUA | \
- REQ_SECURE)
+ REQ_SECURE | REQ_COPY)
#define REQ_CLONE_MASK REQ_COMMON_MASK
-#define BIO_NO_ADVANCE_ITER_MASK (REQ_DISCARD|REQ_WRITE_SAME)
+#define BIO_NO_ADVANCE_ITER_MASK (REQ_DISCARD|REQ_WRITE_SAME|REQ_COPY)
/* This mask is used for both bio and request merge checking */
#define REQ_NOMERGE_FLAGS \
- (REQ_NOMERGE | REQ_STARTED | REQ_SOFTBARRIER | REQ_FLUSH | REQ_FUA)
+ (REQ_NOMERGE | REQ_STARTED | REQ_SOFTBARRIER | REQ_FLUSH | REQ_FUA | \
+ REQ_COPY)
#define REQ_RAHEAD (1ULL << __REQ_RAHEAD)
#define REQ_THROTTLED (1ULL << __REQ_THROTTLED)
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h
index 4afa4f8f60909..aa964dabafbb8 100644
--- a/include/linux/blkdev.h
+++ b/include/linux/blkdev.h
@@ -277,6 +277,7 @@ struct queue_limits {
unsigned int io_opt;
unsigned int max_discard_sectors;
unsigned int max_write_same_sectors;
+ unsigned int max_copy_sectors;
unsigned int discard_granularity;
unsigned int discard_alignment;
@@ -960,6 +961,8 @@ extern void blk_queue_max_discard_sectors(struct request_queue *q,
unsigned int max_discard_sectors);
extern void blk_queue_max_write_same_sectors(struct request_queue *q,
unsigned int max_write_same_sectors);
+extern void blk_queue_max_copy_sectors(struct request_queue *q,
+ unsigned int max_copy_sectors);
extern void blk_queue_logical_block_size(struct request_queue *, unsigned short);
extern void blk_queue_physical_block_size(struct request_queue *, unsigned int);
extern void blk_queue_alignment_offset(struct request_queue *q,
@@ -1116,6 +1119,8 @@ extern int blkdev_issue_discard(struct block_device *bdev, sector_t sector,
sector_t nr_sects, gfp_t gfp_mask, unsigned long flags);
extern int blkdev_issue_write_same(struct block_device *bdev, sector_t sector,
sector_t nr_sects, gfp_t gfp_mask, struct page *page);
+extern int blkdev_issue_copy(struct block_device *, sector_t,
+ struct block_device *, sector_t, unsigned int, gfp_t);
extern int blkdev_issue_zeroout(struct block_device *bdev, sector_t sector,
sector_t nr_sects, gfp_t gfp_mask);
static inline int sb_issue_discard(struct super_block *sb, sector_t block,
@@ -1315,6 +1320,16 @@ static inline unsigned int bdev_write_same(struct block_device *bdev)
return 0;
}
+static inline unsigned int bdev_copy_offload(struct block_device *bdev)
+{
+ struct request_queue *q = bdev_get_queue(bdev);
+
+ if (q)
+ return q->limits.max_copy_sectors;
+
+ return 0;
+}
+
static inline int queue_dma_alignment(struct request_queue *q)
{
return q ? q->dma_alignment : 511;
diff --git a/include/scsi/scsi_device.h b/include/scsi/scsi_device.h
index 3ea622075a94a..594f8a96ec531 100644
--- a/include/scsi/scsi_device.h
+++ b/include/scsi/scsi_device.h
@@ -117,6 +117,8 @@ struct scsi_device {
unsigned char *vpd_pg83;
unsigned char vpd_pg80_len;
unsigned char *vpd_pg80;
+ unsigned char naa_len;
+ unsigned char *naa;
spinlock_t reconfig_lock;
unsigned char current_tag; /* current tag */
struct scsi_target *sdev_target; /* used only for single_lun */
@@ -150,6 +152,7 @@ struct scsi_device {
unsigned use_10_for_ms:1; /* first try 10-byte mode sense/select */
unsigned no_report_opcodes:1; /* no REPORT SUPPORTED OPERATION CODES */
unsigned no_write_same:1; /* no WRITE SAME command */
+ unsigned no_copy:1; /* no copy offload */
unsigned use_16_for_rw:1; /* Use read/write(16) over read/write(10) */
unsigned skip_ms_page_8:1; /* do not use MODE SENSE page 0x08 */
unsigned skip_ms_page_3f:1; /* do not use MODE SENSE page 0x3f */
diff --git a/include/uapi/linux/fs.h b/include/uapi/linux/fs.h
index 6c28b61bb6904..66592930e582d 100644
--- a/include/uapi/linux/fs.h
+++ b/include/uapi/linux/fs.h
@@ -146,6 +146,7 @@ struct inodes_stat_t {
#define BLKSECDISCARD _IO(0x12,125)
#define BLKROTATIONAL _IO(0x12,126)
#define BLKZEROOUT _IO(0x12,127)
+#define BLKCOPY _IO(0x12,128)
#define BMAP_IOCTL 1 /* obsolete - kept for compatibility */
#define FIBMAP _IO(0x00,1) /* bmap access */