target_core_cdb.c revision e22a7f075226c51f3f71b922e9eeb4f99fac1475
1/*
2 * CDB emulation for non-READ/WRITE commands.
3 *
4 * Copyright (c) 2002, 2003, 2004, 2005 PyX Technologies, Inc.
5 * Copyright (c) 2005, 2006, 2007 SBE, Inc.
6 * Copyright (c) 2007-2010 Rising Tide Systems
7 * Copyright (c) 2008-2010 Linux-iSCSI.org
8 *
9 * Nicholas A. Bellinger <nab@kernel.org>
10 *
11 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of the GNU General Public License as published by
13 * the Free Software Foundation; either version 2 of the License, or
14 * (at your option) any later version.
15 *
16 * This program is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
19 * GNU General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, write to the Free Software
23 * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
24 */
25
26#include <asm/unaligned.h>
27#include <scsi/scsi.h>
28
29#include <target/target_core_base.h>
30#include <target/target_core_transport.h>
31#include <target/target_core_fabric_ops.h>
32#include "target_core_ua.h"
33
34static void
35target_fill_alua_data(struct se_port *port, unsigned char *buf)
36{
37	struct t10_alua_tg_pt_gp *tg_pt_gp;
38	struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
39
40	/*
41	 * Set SCCS for MAINTENANCE_IN + REPORT_TARGET_PORT_GROUPS.
42	 */
43	buf[5]	= 0x80;
44
45	/*
46	 * Set TPGS field for explict and/or implict ALUA access type
47	 * and opteration.
48	 *
49	 * See spc4r17 section 6.4.2 Table 135
50	 */
51	if (!port)
52		return;
53	tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
54	if (!tg_pt_gp_mem)
55		return;
56
57	spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
58	tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
59	if (tg_pt_gp)
60		buf[5] |= tg_pt_gp->tg_pt_gp_alua_access_type;
61	spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
62}
63
64static int
65target_emulate_inquiry_std(struct se_cmd *cmd)
66{
67	struct se_lun *lun = cmd->se_lun;
68	struct se_device *dev = cmd->se_dev;
69	unsigned char *buf = cmd->t_task_buf;
70
71	/*
72	 * Make sure we at least have 6 bytes of INQUIRY response
73	 * payload going back for EVPD=0
74	 */
75	if (cmd->data_length < 6) {
76		printk(KERN_ERR "SCSI Inquiry payload length: %u"
77			" too small for EVPD=0\n", cmd->data_length);
78		return -EINVAL;
79	}
80
81	buf[0] = dev->transport->get_device_type(dev);
82	if (buf[0] == TYPE_TAPE)
83		buf[1] = 0x80;
84	buf[2] = dev->transport->get_device_rev(dev);
85
86	/*
87	 * Enable SCCS and TPGS fields for Emulated ALUA
88	 */
89	if (dev->se_sub_dev->t10_alua.alua_type == SPC3_ALUA_EMULATED)
90		target_fill_alua_data(lun->lun_sep, buf);
91
92	if (cmd->data_length < 8) {
93		buf[4] = 1; /* Set additional length to 1 */
94		return 0;
95	}
96
97	buf[7] = 0x32; /* Sync=1 and CmdQue=1 */
98
99	/*
100	 * Do not include vendor, product, reversion info in INQUIRY
101	 * response payload for cdbs with a small allocation length.
102	 */
103	if (cmd->data_length < 36) {
104		buf[4] = 3; /* Set additional length to 3 */
105		return 0;
106	}
107
108	snprintf((unsigned char *)&buf[8], 8, "LIO-ORG");
109	snprintf((unsigned char *)&buf[16], 16, "%s",
110		 &dev->se_sub_dev->t10_wwn.model[0]);
111	snprintf((unsigned char *)&buf[32], 4, "%s",
112		 &dev->se_sub_dev->t10_wwn.revision[0]);
113	buf[4] = 31; /* Set additional length to 31 */
114	return 0;
115}
116
117/* unit serial number */
118static int
119target_emulate_evpd_80(struct se_cmd *cmd, unsigned char *buf)
120{
121	struct se_device *dev = cmd->se_dev;
122	u16 len = 0;
123
124	if (dev->se_sub_dev->su_dev_flags &
125			SDF_EMULATED_VPD_UNIT_SERIAL) {
126		u32 unit_serial_len;
127
128		unit_serial_len =
129			strlen(&dev->se_sub_dev->t10_wwn.unit_serial[0]);
130		unit_serial_len++; /* For NULL Terminator */
131
132		if (((len + 4) + unit_serial_len) > cmd->data_length) {
133			len += unit_serial_len;
134			buf[2] = ((len >> 8) & 0xff);
135			buf[3] = (len & 0xff);
136			return 0;
137		}
138		len += sprintf((unsigned char *)&buf[4], "%s",
139			&dev->se_sub_dev->t10_wwn.unit_serial[0]);
140		len++; /* Extra Byte for NULL Terminator */
141		buf[3] = len;
142	}
143	return 0;
144}
145
146/*
147 * Device identification VPD, for a complete list of
148 * DESIGNATOR TYPEs see spc4r17 Table 459.
149 */
150static int
151target_emulate_evpd_83(struct se_cmd *cmd, unsigned char *buf)
152{
153	struct se_device *dev = cmd->se_dev;
154	struct se_lun *lun = cmd->se_lun;
155	struct se_port *port = NULL;
156	struct se_portal_group *tpg = NULL;
157	struct t10_alua_lu_gp_member *lu_gp_mem;
158	struct t10_alua_tg_pt_gp *tg_pt_gp;
159	struct t10_alua_tg_pt_gp_member *tg_pt_gp_mem;
160	unsigned char binary, binary_new;
161	unsigned char *prod = &dev->se_sub_dev->t10_wwn.model[0];
162	u32 prod_len;
163	u32 unit_serial_len, off = 0;
164	int i;
165	u16 len = 0, id_len;
166
167	off = 4;
168
169	/*
170	 * NAA IEEE Registered Extended Assigned designator format, see
171	 * spc4r17 section 7.7.3.6.5
172	 *
173	 * We depend upon a target_core_mod/ConfigFS provided
174	 * /sys/kernel/config/target/core/$HBA/$DEV/wwn/vpd_unit_serial
175	 * value in order to return the NAA id.
176	 */
177	if (!(dev->se_sub_dev->su_dev_flags & SDF_EMULATED_VPD_UNIT_SERIAL))
178		goto check_t10_vend_desc;
179
180	if (off + 20 > cmd->data_length)
181		goto check_t10_vend_desc;
182
183	/* CODE SET == Binary */
184	buf[off++] = 0x1;
185
186	/* Set ASSOICATION == addressed logical unit: 0)b */
187	buf[off] = 0x00;
188
189	/* Identifier/Designator type == NAA identifier */
190	buf[off++] = 0x3;
191	off++;
192
193	/* Identifier/Designator length */
194	buf[off++] = 0x10;
195
196	/*
197	 * Start NAA IEEE Registered Extended Identifier/Designator
198	 */
199	buf[off++] = (0x6 << 4);
200
201	/*
202	 * Use OpenFabrics IEEE Company ID: 00 14 05
203	 */
204	buf[off++] = 0x01;
205	buf[off++] = 0x40;
206	buf[off] = (0x5 << 4);
207
208	/*
209	 * Return ConfigFS Unit Serial Number information for
210	 * VENDOR_SPECIFIC_IDENTIFIER and
211	 * VENDOR_SPECIFIC_IDENTIFIER_EXTENTION
212	 */
213	binary = transport_asciihex_to_binaryhex(
214				&dev->se_sub_dev->t10_wwn.unit_serial[0]);
215	buf[off++] |= (binary & 0xf0) >> 4;
216	for (i = 0; i < 24; i += 2) {
217		binary_new = transport_asciihex_to_binaryhex(
218			&dev->se_sub_dev->t10_wwn.unit_serial[i+2]);
219		buf[off] = (binary & 0x0f) << 4;
220		buf[off++] |= (binary_new & 0xf0) >> 4;
221		binary = binary_new;
222	}
223	len = 20;
224	off = (len + 4);
225
226check_t10_vend_desc:
227	/*
228	 * T10 Vendor Identifier Page, see spc4r17 section 7.7.3.4
229	 */
230	id_len = 8; /* For Vendor field */
231	prod_len = 4; /* For VPD Header */
232	prod_len += 8; /* For Vendor field */
233	prod_len += strlen(prod);
234	prod_len++; /* For : */
235
236	if (dev->se_sub_dev->su_dev_flags &
237			SDF_EMULATED_VPD_UNIT_SERIAL) {
238		unit_serial_len =
239			strlen(&dev->se_sub_dev->t10_wwn.unit_serial[0]);
240		unit_serial_len++; /* For NULL Terminator */
241
242		if ((len + (id_len + 4) +
243		    (prod_len + unit_serial_len)) >
244				cmd->data_length) {
245			len += (prod_len + unit_serial_len);
246			goto check_port;
247		}
248		id_len += sprintf((unsigned char *)&buf[off+12],
249				"%s:%s", prod,
250				&dev->se_sub_dev->t10_wwn.unit_serial[0]);
251	}
252	buf[off] = 0x2; /* ASCII */
253	buf[off+1] = 0x1; /* T10 Vendor ID */
254	buf[off+2] = 0x0;
255	memcpy((unsigned char *)&buf[off+4], "LIO-ORG", 8);
256	/* Extra Byte for NULL Terminator */
257	id_len++;
258	/* Identifier Length */
259	buf[off+3] = id_len;
260	/* Header size for Designation descriptor */
261	len += (id_len + 4);
262	off += (id_len + 4);
263	/*
264	 * struct se_port is only set for INQUIRY VPD=1 through $FABRIC_MOD
265	 */
266check_port:
267	port = lun->lun_sep;
268	if (port) {
269		struct t10_alua_lu_gp *lu_gp;
270		u32 padding, scsi_name_len;
271		u16 lu_gp_id = 0;
272		u16 tg_pt_gp_id = 0;
273		u16 tpgt;
274
275		tpg = port->sep_tpg;
276		/*
277		 * Relative target port identifer, see spc4r17
278		 * section 7.7.3.7
279		 *
280		 * Get the PROTOCOL IDENTIFIER as defined by spc4r17
281		 * section 7.5.1 Table 362
282		 */
283		if (((len + 4) + 8) > cmd->data_length) {
284			len += 8;
285			goto check_tpgi;
286		}
287		buf[off] =
288			(tpg->se_tpg_tfo->get_fabric_proto_ident(tpg) << 4);
289		buf[off++] |= 0x1; /* CODE SET == Binary */
290		buf[off] = 0x80; /* Set PIV=1 */
291		/* Set ASSOICATION == target port: 01b */
292		buf[off] |= 0x10;
293		/* DESIGNATOR TYPE == Relative target port identifer */
294		buf[off++] |= 0x4;
295		off++; /* Skip over Reserved */
296		buf[off++] = 4; /* DESIGNATOR LENGTH */
297		/* Skip over Obsolete field in RTPI payload
298		 * in Table 472 */
299		off += 2;
300		buf[off++] = ((port->sep_rtpi >> 8) & 0xff);
301		buf[off++] = (port->sep_rtpi & 0xff);
302		len += 8; /* Header size + Designation descriptor */
303		/*
304		 * Target port group identifier, see spc4r17
305		 * section 7.7.3.8
306		 *
307		 * Get the PROTOCOL IDENTIFIER as defined by spc4r17
308		 * section 7.5.1 Table 362
309		 */
310check_tpgi:
311		if (dev->se_sub_dev->t10_alua.alua_type !=
312				SPC3_ALUA_EMULATED)
313			goto check_scsi_name;
314
315		if (((len + 4) + 8) > cmd->data_length) {
316			len += 8;
317			goto check_lu_gp;
318		}
319		tg_pt_gp_mem = port->sep_alua_tg_pt_gp_mem;
320		if (!tg_pt_gp_mem)
321			goto check_lu_gp;
322
323		spin_lock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
324		tg_pt_gp = tg_pt_gp_mem->tg_pt_gp;
325		if (!(tg_pt_gp)) {
326			spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
327			goto check_lu_gp;
328		}
329		tg_pt_gp_id = tg_pt_gp->tg_pt_gp_id;
330		spin_unlock(&tg_pt_gp_mem->tg_pt_gp_mem_lock);
331
332		buf[off] =
333			(tpg->se_tpg_tfo->get_fabric_proto_ident(tpg) << 4);
334		buf[off++] |= 0x1; /* CODE SET == Binary */
335		buf[off] = 0x80; /* Set PIV=1 */
336		/* Set ASSOICATION == target port: 01b */
337		buf[off] |= 0x10;
338		/* DESIGNATOR TYPE == Target port group identifier */
339		buf[off++] |= 0x5;
340		off++; /* Skip over Reserved */
341		buf[off++] = 4; /* DESIGNATOR LENGTH */
342		off += 2; /* Skip over Reserved Field */
343		buf[off++] = ((tg_pt_gp_id >> 8) & 0xff);
344		buf[off++] = (tg_pt_gp_id & 0xff);
345		len += 8; /* Header size + Designation descriptor */
346		/*
347		 * Logical Unit Group identifier, see spc4r17
348		 * section 7.7.3.8
349		 */
350check_lu_gp:
351		if (((len + 4) + 8) > cmd->data_length) {
352			len += 8;
353			goto check_scsi_name;
354		}
355		lu_gp_mem = dev->dev_alua_lu_gp_mem;
356		if (!(lu_gp_mem))
357			goto check_scsi_name;
358
359		spin_lock(&lu_gp_mem->lu_gp_mem_lock);
360		lu_gp = lu_gp_mem->lu_gp;
361		if (!(lu_gp)) {
362			spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
363			goto check_scsi_name;
364		}
365		lu_gp_id = lu_gp->lu_gp_id;
366		spin_unlock(&lu_gp_mem->lu_gp_mem_lock);
367
368		buf[off++] |= 0x1; /* CODE SET == Binary */
369		/* DESIGNATOR TYPE == Logical Unit Group identifier */
370		buf[off++] |= 0x6;
371		off++; /* Skip over Reserved */
372		buf[off++] = 4; /* DESIGNATOR LENGTH */
373		off += 2; /* Skip over Reserved Field */
374		buf[off++] = ((lu_gp_id >> 8) & 0xff);
375		buf[off++] = (lu_gp_id & 0xff);
376		len += 8; /* Header size + Designation descriptor */
377		/*
378		 * SCSI name string designator, see spc4r17
379		 * section 7.7.3.11
380		 *
381		 * Get the PROTOCOL IDENTIFIER as defined by spc4r17
382		 * section 7.5.1 Table 362
383		 */
384check_scsi_name:
385		scsi_name_len = strlen(tpg->se_tpg_tfo->tpg_get_wwn(tpg));
386		/* UTF-8 ",t,0x<16-bit TPGT>" + NULL Terminator */
387		scsi_name_len += 10;
388		/* Check for 4-byte padding */
389		padding = ((-scsi_name_len) & 3);
390		if (padding != 0)
391			scsi_name_len += padding;
392		/* Header size + Designation descriptor */
393		scsi_name_len += 4;
394
395		if (((len + 4) + scsi_name_len) > cmd->data_length) {
396			len += scsi_name_len;
397			goto set_len;
398		}
399		buf[off] =
400			(tpg->se_tpg_tfo->get_fabric_proto_ident(tpg) << 4);
401		buf[off++] |= 0x3; /* CODE SET == UTF-8 */
402		buf[off] = 0x80; /* Set PIV=1 */
403		/* Set ASSOICATION == target port: 01b */
404		buf[off] |= 0x10;
405		/* DESIGNATOR TYPE == SCSI name string */
406		buf[off++] |= 0x8;
407		off += 2; /* Skip over Reserved and length */
408		/*
409		 * SCSI name string identifer containing, $FABRIC_MOD
410		 * dependent information.  For LIO-Target and iSCSI
411		 * Target Port, this means "<iSCSI name>,t,0x<TPGT> in
412		 * UTF-8 encoding.
413		 */
414		tpgt = tpg->se_tpg_tfo->tpg_get_tag(tpg);
415		scsi_name_len = sprintf(&buf[off], "%s,t,0x%04x",
416					tpg->se_tpg_tfo->tpg_get_wwn(tpg), tpgt);
417		scsi_name_len += 1 /* Include  NULL terminator */;
418		/*
419		 * The null-terminated, null-padded (see 4.4.2) SCSI
420		 * NAME STRING field contains a UTF-8 format string.
421		 * The number of bytes in the SCSI NAME STRING field
422		 * (i.e., the value in the DESIGNATOR LENGTH field)
423		 * shall be no larger than 256 and shall be a multiple
424		 * of four.
425		 */
426		if (padding)
427			scsi_name_len += padding;
428
429		buf[off-1] = scsi_name_len;
430		off += scsi_name_len;
431		/* Header size + Designation descriptor */
432		len += (scsi_name_len + 4);
433	}
434set_len:
435	buf[2] = ((len >> 8) & 0xff);
436	buf[3] = (len & 0xff); /* Page Length for VPD 0x83 */
437	return 0;
438}
439
440/* Extended INQUIRY Data VPD Page */
441static int
442target_emulate_evpd_86(struct se_cmd *cmd, unsigned char *buf)
443{
444	if (cmd->data_length < 60)
445		return 0;
446
447	buf[2] = 0x3c;
448	/* Set HEADSUP, ORDSUP, SIMPSUP */
449	buf[5] = 0x07;
450
451	/* If WriteCache emulation is enabled, set V_SUP */
452	if (cmd->se_dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0)
453		buf[6] = 0x01;
454	return 0;
455}
456
457/* Block Limits VPD page */
458static int
459target_emulate_evpd_b0(struct se_cmd *cmd, unsigned char *buf)
460{
461	struct se_device *dev = cmd->se_dev;
462	int have_tp = 0;
463
464	/*
465	 * Following sbc3r22 section 6.5.3 Block Limits VPD page, when
466	 * emulate_tpu=1 or emulate_tpws=1 we will be expect a
467	 * different page length for Thin Provisioning.
468	 */
469	if (dev->se_sub_dev->se_dev_attrib.emulate_tpu || dev->se_sub_dev->se_dev_attrib.emulate_tpws)
470		have_tp = 1;
471
472	if (cmd->data_length < (0x10 + 4)) {
473		printk(KERN_INFO "Received data_length: %u"
474			" too small for EVPD 0xb0\n",
475			cmd->data_length);
476		return -EINVAL;
477	}
478
479	if (have_tp && cmd->data_length < (0x3c + 4)) {
480		printk(KERN_INFO "Received data_length: %u"
481			" too small for TPE=1 EVPD 0xb0\n",
482			cmd->data_length);
483		have_tp = 0;
484	}
485
486	buf[0] = dev->transport->get_device_type(dev);
487	buf[3] = have_tp ? 0x3c : 0x10;
488
489	/*
490	 * Set OPTIMAL TRANSFER LENGTH GRANULARITY
491	 */
492	put_unaligned_be16(1, &buf[6]);
493
494	/*
495	 * Set MAXIMUM TRANSFER LENGTH
496	 */
497	put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.max_sectors, &buf[8]);
498
499	/*
500	 * Set OPTIMAL TRANSFER LENGTH
501	 */
502	put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.optimal_sectors, &buf[12]);
503
504	/*
505	 * Exit now if we don't support TP or the initiator sent a too
506	 * short buffer.
507	 */
508	if (!have_tp || cmd->data_length < (0x3c + 4))
509		return 0;
510
511	/*
512	 * Set MAXIMUM UNMAP LBA COUNT
513	 */
514	put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.max_unmap_lba_count, &buf[20]);
515
516	/*
517	 * Set MAXIMUM UNMAP BLOCK DESCRIPTOR COUNT
518	 */
519	put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.max_unmap_block_desc_count,
520			   &buf[24]);
521
522	/*
523	 * Set OPTIMAL UNMAP GRANULARITY
524	 */
525	put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.unmap_granularity, &buf[28]);
526
527	/*
528	 * UNMAP GRANULARITY ALIGNMENT
529	 */
530	put_unaligned_be32(dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment,
531			   &buf[32]);
532	if (dev->se_sub_dev->se_dev_attrib.unmap_granularity_alignment != 0)
533		buf[32] |= 0x80; /* Set the UGAVALID bit */
534
535	return 0;
536}
537
538/* Block Device Characteristics VPD page */
539static int
540target_emulate_evpd_b1(struct se_cmd *cmd, unsigned char *buf)
541{
542	struct se_device *dev = cmd->se_dev;
543
544	buf[0] = dev->transport->get_device_type(dev);
545	buf[3] = 0x3c;
546
547	if (cmd->data_length >= 5 &&
548	    dev->se_sub_dev->se_dev_attrib.is_nonrot)
549		buf[5] = 1;
550
551	return 0;
552}
553
554/* Thin Provisioning VPD */
555static int
556target_emulate_evpd_b2(struct se_cmd *cmd, unsigned char *buf)
557{
558	struct se_device *dev = cmd->se_dev;
559
560	/*
561	 * From sbc3r22 section 6.5.4 Thin Provisioning VPD page:
562	 *
563	 * The PAGE LENGTH field is defined in SPC-4. If the DP bit is set to
564	 * zero, then the page length shall be set to 0004h.  If the DP bit
565	 * is set to one, then the page length shall be set to the value
566	 * defined in table 162.
567	 */
568	buf[0] = dev->transport->get_device_type(dev);
569
570	/*
571	 * Set Hardcoded length mentioned above for DP=0
572	 */
573	put_unaligned_be16(0x0004, &buf[2]);
574
575	/*
576	 * The THRESHOLD EXPONENT field indicates the threshold set size in
577	 * LBAs as a power of 2 (i.e., the threshold set size is equal to
578	 * 2(threshold exponent)).
579	 *
580	 * Note that this is currently set to 0x00 as mkp says it will be
581	 * changing again.  We can enable this once it has settled in T10
582	 * and is actually used by Linux/SCSI ML code.
583	 */
584	buf[4] = 0x00;
585
586	/*
587	 * A TPU bit set to one indicates that the device server supports
588	 * the UNMAP command (see 5.25). A TPU bit set to zero indicates
589	 * that the device server does not support the UNMAP command.
590	 */
591	if (dev->se_sub_dev->se_dev_attrib.emulate_tpu != 0)
592		buf[5] = 0x80;
593
594	/*
595	 * A TPWS bit set to one indicates that the device server supports
596	 * the use of the WRITE SAME (16) command (see 5.42) to unmap LBAs.
597	 * A TPWS bit set to zero indicates that the device server does not
598	 * support the use of the WRITE SAME (16) command to unmap LBAs.
599	 */
600	if (dev->se_sub_dev->se_dev_attrib.emulate_tpws != 0)
601		buf[5] |= 0x40;
602
603	return 0;
604}
605
606static int
607target_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf);
608
609static struct {
610	uint8_t		page;
611	int		(*emulate)(struct se_cmd *, unsigned char *);
612} evpd_handlers[] = {
613	{ .page = 0x00, .emulate = target_emulate_evpd_00 },
614	{ .page = 0x80, .emulate = target_emulate_evpd_80 },
615	{ .page = 0x83, .emulate = target_emulate_evpd_83 },
616	{ .page = 0x86, .emulate = target_emulate_evpd_86 },
617	{ .page = 0xb0, .emulate = target_emulate_evpd_b0 },
618	{ .page = 0xb1, .emulate = target_emulate_evpd_b1 },
619	{ .page = 0xb2, .emulate = target_emulate_evpd_b2 },
620};
621
622/* supported vital product data pages */
623static int
624target_emulate_evpd_00(struct se_cmd *cmd, unsigned char *buf)
625{
626	int p;
627
628	if (cmd->data_length < 8)
629		return 0;
630	/*
631	 * Only report the INQUIRY EVPD=1 pages after a valid NAA
632	 * Registered Extended LUN WWN has been set via ConfigFS
633	 * during device creation/restart.
634	 */
635	if (cmd->se_dev->se_sub_dev->su_dev_flags &
636			SDF_EMULATED_VPD_UNIT_SERIAL) {
637		buf[3] = ARRAY_SIZE(evpd_handlers);
638		for (p = 0; p < min_t(int, ARRAY_SIZE(evpd_handlers),
639				      cmd->data_length - 4); ++p)
640			buf[p + 4] = evpd_handlers[p].page;
641	}
642
643	return 0;
644}
645
646static int
647target_emulate_inquiry(struct se_cmd *cmd)
648{
649	struct se_device *dev = cmd->se_dev;
650	unsigned char *buf = cmd->t_task_buf;
651	unsigned char *cdb = cmd->t_task_cdb;
652	int p;
653
654	if (!(cdb[1] & 0x1))
655		return target_emulate_inquiry_std(cmd);
656
657	/*
658	 * Make sure we at least have 4 bytes of INQUIRY response
659	 * payload for 0x00 going back for EVPD=1.  Note that 0x80
660	 * and 0x83 will check for enough payload data length and
661	 * jump to set_len: label when there is not enough inquiry EVPD
662	 * payload length left for the next outgoing EVPD metadata
663	 */
664	if (cmd->data_length < 4) {
665		printk(KERN_ERR "SCSI Inquiry payload length: %u"
666			" too small for EVPD=1\n", cmd->data_length);
667		return -EINVAL;
668	}
669	buf[0] = dev->transport->get_device_type(dev);
670
671	for (p = 0; p < ARRAY_SIZE(evpd_handlers); ++p)
672		if (cdb[2] == evpd_handlers[p].page) {
673			buf[1] = cdb[2];
674			return evpd_handlers[p].emulate(cmd, buf);
675		}
676
677	printk(KERN_ERR "Unknown VPD Code: 0x%02x\n", cdb[2]);
678	return -EINVAL;
679}
680
681static int
682target_emulate_readcapacity(struct se_cmd *cmd)
683{
684	struct se_device *dev = cmd->se_dev;
685	unsigned char *buf = cmd->t_task_buf;
686	unsigned long long blocks_long = dev->transport->get_blocks(dev);
687	u32 blocks;
688
689	if (blocks_long >= 0x00000000ffffffff)
690		blocks = 0xffffffff;
691	else
692		blocks = (u32)blocks_long;
693
694	buf[0] = (blocks >> 24) & 0xff;
695	buf[1] = (blocks >> 16) & 0xff;
696	buf[2] = (blocks >> 8) & 0xff;
697	buf[3] = blocks & 0xff;
698	buf[4] = (dev->se_sub_dev->se_dev_attrib.block_size >> 24) & 0xff;
699	buf[5] = (dev->se_sub_dev->se_dev_attrib.block_size >> 16) & 0xff;
700	buf[6] = (dev->se_sub_dev->se_dev_attrib.block_size >> 8) & 0xff;
701	buf[7] = dev->se_sub_dev->se_dev_attrib.block_size & 0xff;
702	/*
703	 * Set max 32-bit blocks to signal SERVICE ACTION READ_CAPACITY_16
704	*/
705	if (dev->se_sub_dev->se_dev_attrib.emulate_tpu || dev->se_sub_dev->se_dev_attrib.emulate_tpws)
706		put_unaligned_be32(0xFFFFFFFF, &buf[0]);
707
708	return 0;
709}
710
711static int
712target_emulate_readcapacity_16(struct se_cmd *cmd)
713{
714	struct se_device *dev = cmd->se_dev;
715	unsigned char *buf = cmd->t_task_buf;
716	unsigned long long blocks = dev->transport->get_blocks(dev);
717
718	buf[0] = (blocks >> 56) & 0xff;
719	buf[1] = (blocks >> 48) & 0xff;
720	buf[2] = (blocks >> 40) & 0xff;
721	buf[3] = (blocks >> 32) & 0xff;
722	buf[4] = (blocks >> 24) & 0xff;
723	buf[5] = (blocks >> 16) & 0xff;
724	buf[6] = (blocks >> 8) & 0xff;
725	buf[7] = blocks & 0xff;
726	buf[8] = (dev->se_sub_dev->se_dev_attrib.block_size >> 24) & 0xff;
727	buf[9] = (dev->se_sub_dev->se_dev_attrib.block_size >> 16) & 0xff;
728	buf[10] = (dev->se_sub_dev->se_dev_attrib.block_size >> 8) & 0xff;
729	buf[11] = dev->se_sub_dev->se_dev_attrib.block_size & 0xff;
730	/*
731	 * Set Thin Provisioning Enable bit following sbc3r22 in section
732	 * READ CAPACITY (16) byte 14 if emulate_tpu or emulate_tpws is enabled.
733	 */
734	if (dev->se_sub_dev->se_dev_attrib.emulate_tpu || dev->se_sub_dev->se_dev_attrib.emulate_tpws)
735		buf[14] = 0x80;
736
737	return 0;
738}
739
740static int
741target_modesense_rwrecovery(unsigned char *p)
742{
743	p[0] = 0x01;
744	p[1] = 0x0a;
745
746	return 12;
747}
748
749static int
750target_modesense_control(struct se_device *dev, unsigned char *p)
751{
752	p[0] = 0x0a;
753	p[1] = 0x0a;
754	p[2] = 2;
755	/*
756	 * From spc4r17, section 7.4.6 Control mode Page
757	 *
758	 * Unit Attention interlocks control (UN_INTLCK_CTRL) to code 00b
759	 *
760	 * 00b: The logical unit shall clear any unit attention condition
761	 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
762	 * status and shall not establish a unit attention condition when a com-
763	 * mand is completed with BUSY, TASK SET FULL, or RESERVATION CONFLICT
764	 * status.
765	 *
766	 * 10b: The logical unit shall not clear any unit attention condition
767	 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
768	 * status and shall not establish a unit attention condition when
769	 * a command is completed with BUSY, TASK SET FULL, or RESERVATION
770	 * CONFLICT status.
771	 *
772	 * 11b a The logical unit shall not clear any unit attention condition
773	 * reported in the same I_T_L_Q nexus transaction as a CHECK CONDITION
774	 * status and shall establish a unit attention condition for the
775	 * initiator port associated with the I_T nexus on which the BUSY,
776	 * TASK SET FULL, or RESERVATION CONFLICT status is being returned.
777	 * Depending on the status, the additional sense code shall be set to
778	 * PREVIOUS BUSY STATUS, PREVIOUS TASK SET FULL STATUS, or PREVIOUS
779	 * RESERVATION CONFLICT STATUS. Until it is cleared by a REQUEST SENSE
780	 * command, a unit attention condition shall be established only once
781	 * for a BUSY, TASK SET FULL, or RESERVATION CONFLICT status regardless
782	 * to the number of commands completed with one of those status codes.
783	 */
784	p[4] = (dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl == 2) ? 0x30 :
785	       (dev->se_sub_dev->se_dev_attrib.emulate_ua_intlck_ctrl == 1) ? 0x20 : 0x00;
786	/*
787	 * From spc4r17, section 7.4.6 Control mode Page
788	 *
789	 * Task Aborted Status (TAS) bit set to zero.
790	 *
791	 * A task aborted status (TAS) bit set to zero specifies that aborted
792	 * tasks shall be terminated by the device server without any response
793	 * to the application client. A TAS bit set to one specifies that tasks
794	 * aborted by the actions of an I_T nexus other than the I_T nexus on
795	 * which the command was received shall be completed with TASK ABORTED
796	 * status (see SAM-4).
797	 */
798	p[5] = (dev->se_sub_dev->se_dev_attrib.emulate_tas) ? 0x40 : 0x00;
799	p[8] = 0xff;
800	p[9] = 0xff;
801	p[11] = 30;
802
803	return 12;
804}
805
806static int
807target_modesense_caching(struct se_device *dev, unsigned char *p)
808{
809	p[0] = 0x08;
810	p[1] = 0x12;
811	if (dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0)
812		p[2] = 0x04; /* Write Cache Enable */
813	p[12] = 0x20; /* Disabled Read Ahead */
814
815	return 20;
816}
817
818static void
819target_modesense_write_protect(unsigned char *buf, int type)
820{
821	/*
822	 * I believe that the WP bit (bit 7) in the mode header is the same for
823	 * all device types..
824	 */
825	switch (type) {
826	case TYPE_DISK:
827	case TYPE_TAPE:
828	default:
829		buf[0] |= 0x80; /* WP bit */
830		break;
831	}
832}
833
834static void
835target_modesense_dpofua(unsigned char *buf, int type)
836{
837	switch (type) {
838	case TYPE_DISK:
839		buf[0] |= 0x10; /* DPOFUA bit */
840		break;
841	default:
842		break;
843	}
844}
845
846static int
847target_emulate_modesense(struct se_cmd *cmd, int ten)
848{
849	struct se_device *dev = cmd->se_dev;
850	char *cdb = cmd->t_task_cdb;
851	unsigned char *rbuf = cmd->t_task_buf;
852	int type = dev->transport->get_device_type(dev);
853	int offset = (ten) ? 8 : 4;
854	int length = 0;
855	unsigned char buf[SE_MODE_PAGE_BUF];
856
857	memset(buf, 0, SE_MODE_PAGE_BUF);
858
859	switch (cdb[2] & 0x3f) {
860	case 0x01:
861		length = target_modesense_rwrecovery(&buf[offset]);
862		break;
863	case 0x08:
864		length = target_modesense_caching(dev, &buf[offset]);
865		break;
866	case 0x0a:
867		length = target_modesense_control(dev, &buf[offset]);
868		break;
869	case 0x3f:
870		length = target_modesense_rwrecovery(&buf[offset]);
871		length += target_modesense_caching(dev, &buf[offset+length]);
872		length += target_modesense_control(dev, &buf[offset+length]);
873		break;
874	default:
875		printk(KERN_ERR "Got Unknown Mode Page: 0x%02x\n",
876				cdb[2] & 0x3f);
877		return PYX_TRANSPORT_UNKNOWN_MODE_PAGE;
878	}
879	offset += length;
880
881	if (ten) {
882		offset -= 2;
883		buf[0] = (offset >> 8) & 0xff;
884		buf[1] = offset & 0xff;
885
886		if ((cmd->se_lun->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
887		    (cmd->se_deve &&
888		    (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
889			target_modesense_write_protect(&buf[3], type);
890
891		if ((dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0) &&
892		    (dev->se_sub_dev->se_dev_attrib.emulate_fua_write > 0))
893			target_modesense_dpofua(&buf[3], type);
894
895		if ((offset + 2) > cmd->data_length)
896			offset = cmd->data_length;
897
898	} else {
899		offset -= 1;
900		buf[0] = offset & 0xff;
901
902		if ((cmd->se_lun->lun_access & TRANSPORT_LUNFLAGS_READ_ONLY) ||
903		    (cmd->se_deve &&
904		    (cmd->se_deve->lun_flags & TRANSPORT_LUNFLAGS_READ_ONLY)))
905			target_modesense_write_protect(&buf[2], type);
906
907		if ((dev->se_sub_dev->se_dev_attrib.emulate_write_cache > 0) &&
908		    (dev->se_sub_dev->se_dev_attrib.emulate_fua_write > 0))
909			target_modesense_dpofua(&buf[2], type);
910
911		if ((offset + 1) > cmd->data_length)
912			offset = cmd->data_length;
913	}
914	memcpy(rbuf, buf, offset);
915
916	return 0;
917}
918
919static int
920target_emulate_request_sense(struct se_cmd *cmd)
921{
922	unsigned char *cdb = cmd->t_task_cdb;
923	unsigned char *buf = cmd->t_task_buf;
924	u8 ua_asc = 0, ua_ascq = 0;
925
926	if (cdb[1] & 0x01) {
927		printk(KERN_ERR "REQUEST_SENSE description emulation not"
928			" supported\n");
929		return PYX_TRANSPORT_INVALID_CDB_FIELD;
930	}
931	if (!(core_scsi3_ua_clear_for_request_sense(cmd, &ua_asc, &ua_ascq))) {
932		/*
933		 * CURRENT ERROR, UNIT ATTENTION
934		 */
935		buf[0] = 0x70;
936		buf[SPC_SENSE_KEY_OFFSET] = UNIT_ATTENTION;
937		/*
938		 * Make sure request data length is enough for additional
939		 * sense data.
940		 */
941		if (cmd->data_length <= 18) {
942			buf[7] = 0x00;
943			return 0;
944		}
945		/*
946		 * The Additional Sense Code (ASC) from the UNIT ATTENTION
947		 */
948		buf[SPC_ASC_KEY_OFFSET] = ua_asc;
949		buf[SPC_ASCQ_KEY_OFFSET] = ua_ascq;
950		buf[7] = 0x0A;
951	} else {
952		/*
953		 * CURRENT ERROR, NO SENSE
954		 */
955		buf[0] = 0x70;
956		buf[SPC_SENSE_KEY_OFFSET] = NO_SENSE;
957		/*
958		 * Make sure request data length is enough for additional
959		 * sense data.
960		 */
961		if (cmd->data_length <= 18) {
962			buf[7] = 0x00;
963			return 0;
964		}
965		/*
966		 * NO ADDITIONAL SENSE INFORMATION
967		 */
968		buf[SPC_ASC_KEY_OFFSET] = 0x00;
969		buf[7] = 0x0A;
970	}
971
972	return 0;
973}
974
975/*
976 * Used for TCM/IBLOCK and TCM/FILEIO for block/blk-lib.c level discard support.
977 * Note this is not used for TCM/pSCSI passthrough
978 */
979static int
980target_emulate_unmap(struct se_task *task)
981{
982	struct se_cmd *cmd = task->task_se_cmd;
983	struct se_device *dev = cmd->se_dev;
984	unsigned char *buf = cmd->t_task_buf, *ptr = NULL;
985	unsigned char *cdb = &cmd->t_task_cdb[0];
986	sector_t lba;
987	unsigned int size = cmd->data_length, range;
988	int ret, offset;
989	unsigned short dl, bd_dl;
990
991	/* First UNMAP block descriptor starts at 8 byte offset */
992	offset = 8;
993	size -= 8;
994	dl = get_unaligned_be16(&cdb[0]);
995	bd_dl = get_unaligned_be16(&cdb[2]);
996	ptr = &buf[offset];
997	printk(KERN_INFO "UNMAP: Sub: %s Using dl: %hu bd_dl: %hu size: %hu"
998		" ptr: %p\n", dev->transport->name, dl, bd_dl, size, ptr);
999
1000	while (size) {
1001		lba = get_unaligned_be64(&ptr[0]);
1002		range = get_unaligned_be32(&ptr[8]);
1003		printk(KERN_INFO "UNMAP: Using lba: %llu and range: %u\n",
1004				 (unsigned long long)lba, range);
1005
1006		ret = dev->transport->do_discard(dev, lba, range);
1007		if (ret < 0) {
1008			printk(KERN_ERR "blkdev_issue_discard() failed: %d\n",
1009					ret);
1010			return ret;
1011		}
1012
1013		ptr += 16;
1014		size -= 16;
1015	}
1016
1017	task->task_scsi_status = GOOD;
1018	transport_complete_task(task, 1);
1019	return 0;
1020}
1021
1022/*
1023 * Used for TCM/IBLOCK and TCM/FILEIO for block/blk-lib.c level discard support.
1024 * Note this is not used for TCM/pSCSI passthrough
1025 */
1026static int
1027target_emulate_write_same(struct se_task *task, int write_same32)
1028{
1029	struct se_cmd *cmd = task->task_se_cmd;
1030	struct se_device *dev = cmd->se_dev;
1031	sector_t range;
1032	sector_t lba = cmd->t_task_lba;
1033	unsigned int num_blocks;
1034	int ret;
1035	/*
1036	 * Extract num_blocks from the WRITE_SAME_* CDB.  Then use the explict
1037	 * range when non zero is supplied, otherwise calculate the remaining
1038	 * range based on ->get_blocks() - starting LBA.
1039	 */
1040	if (write_same32)
1041		num_blocks = get_unaligned_be32(&cmd->t_task_cdb[28]);
1042	else
1043		num_blocks = get_unaligned_be32(&cmd->t_task_cdb[10]);
1044
1045	if (num_blocks != 0)
1046		range = num_blocks;
1047	else
1048		range = (dev->transport->get_blocks(dev) - lba);
1049
1050	printk(KERN_INFO "WRITE_SAME UNMAP: LBA: %llu Range: %llu\n",
1051		 (unsigned long long)lba, (unsigned long long)range);
1052
1053	ret = dev->transport->do_discard(dev, lba, range);
1054	if (ret < 0) {
1055		printk(KERN_INFO "blkdev_issue_discard() failed for WRITE_SAME\n");
1056		return ret;
1057	}
1058
1059	task->task_scsi_status = GOOD;
1060	transport_complete_task(task, 1);
1061	return 0;
1062}
1063
1064int
1065transport_emulate_control_cdb(struct se_task *task)
1066{
1067	struct se_cmd *cmd = task->task_se_cmd;
1068	struct se_device *dev = cmd->se_dev;
1069	unsigned short service_action;
1070	int ret = 0;
1071
1072	switch (cmd->t_task_cdb[0]) {
1073	case INQUIRY:
1074		ret = target_emulate_inquiry(cmd);
1075		break;
1076	case READ_CAPACITY:
1077		ret = target_emulate_readcapacity(cmd);
1078		break;
1079	case MODE_SENSE:
1080		ret = target_emulate_modesense(cmd, 0);
1081		break;
1082	case MODE_SENSE_10:
1083		ret = target_emulate_modesense(cmd, 1);
1084		break;
1085	case SERVICE_ACTION_IN:
1086		switch (cmd->t_task_cdb[1] & 0x1f) {
1087		case SAI_READ_CAPACITY_16:
1088			ret = target_emulate_readcapacity_16(cmd);
1089			break;
1090		default:
1091			printk(KERN_ERR "Unsupported SA: 0x%02x\n",
1092				cmd->t_task_cdb[1] & 0x1f);
1093			return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
1094		}
1095		break;
1096	case REQUEST_SENSE:
1097		ret = target_emulate_request_sense(cmd);
1098		break;
1099	case UNMAP:
1100		if (!dev->transport->do_discard) {
1101			printk(KERN_ERR "UNMAP emulation not supported for: %s\n",
1102					dev->transport->name);
1103			return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
1104		}
1105		ret = target_emulate_unmap(task);
1106		break;
1107	case WRITE_SAME_16:
1108		if (!dev->transport->do_discard) {
1109			printk(KERN_ERR "WRITE_SAME_16 emulation not supported"
1110					" for: %s\n", dev->transport->name);
1111			return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
1112		}
1113		ret = target_emulate_write_same(task, 0);
1114		break;
1115	case VARIABLE_LENGTH_CMD:
1116		service_action =
1117			get_unaligned_be16(&cmd->t_task_cdb[8]);
1118		switch (service_action) {
1119		case WRITE_SAME_32:
1120			if (!dev->transport->do_discard) {
1121				printk(KERN_ERR "WRITE_SAME_32 SA emulation not"
1122					" supported for: %s\n",
1123					dev->transport->name);
1124				return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
1125			}
1126			ret = target_emulate_write_same(task, 1);
1127			break;
1128		default:
1129			printk(KERN_ERR "Unsupported VARIABLE_LENGTH_CMD SA:"
1130					" 0x%02x\n", service_action);
1131			break;
1132		}
1133		break;
1134	case SYNCHRONIZE_CACHE:
1135	case 0x91: /* SYNCHRONIZE_CACHE_16: */
1136		if (!dev->transport->do_sync_cache) {
1137			printk(KERN_ERR
1138				"SYNCHRONIZE_CACHE emulation not supported"
1139				" for: %s\n", dev->transport->name);
1140			return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
1141		}
1142		dev->transport->do_sync_cache(task);
1143		break;
1144	case ALLOW_MEDIUM_REMOVAL:
1145	case ERASE:
1146	case REZERO_UNIT:
1147	case SEEK_10:
1148	case SPACE:
1149	case START_STOP:
1150	case TEST_UNIT_READY:
1151	case VERIFY:
1152	case WRITE_FILEMARKS:
1153		break;
1154	default:
1155		printk(KERN_ERR "Unsupported SCSI Opcode: 0x%02x for %s\n",
1156			cmd->t_task_cdb[0], dev->transport->name);
1157		return PYX_TRANSPORT_UNKNOWN_SAM_OPCODE;
1158	}
1159
1160	if (ret < 0)
1161		return ret;
1162	task->task_scsi_status = GOOD;
1163	transport_complete_task(task, 1);
1164
1165	return PYX_TRANSPORT_SENT_TO_TRANSPORT;
1166}
1167