1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Generic SCSI-3 ALUA SCSI Device Handler
5 * Copyright (C) 2007-2010 Hannes Reinecke, SUSE Linux Products GmbH.
8 #include <linux/slab.h>
9 #include <linux/delay.h>
10 #include <linux/module.h>
11 #include <linux/unaligned.h>
12 #include <scsi/scsi.h>
13 #include <scsi/scsi_proto.h>
14 #include <scsi/scsi_dbg.h>
15 #include <scsi/scsi_eh.h>
16 #include <scsi/scsi_dh.h>
18 #define ALUA_DH_NAME "alua"
19 #define ALUA_DH_VER "2.0"
21 #define TPGS_SUPPORT_NONE 0x00
22 #define TPGS_SUPPORT_OPTIMIZED 0x01
23 #define TPGS_SUPPORT_NONOPTIMIZED 0x02
24 #define TPGS_SUPPORT_STANDBY 0x04
25 #define TPGS_SUPPORT_UNAVAILABLE 0x08
26 #define TPGS_SUPPORT_LBA_DEPENDENT 0x10
27 #define TPGS_SUPPORT_OFFLINE 0x40
28 #define TPGS_SUPPORT_TRANSITION 0x80
29 #define TPGS_SUPPORT_ALL 0xdf
31 #define RTPG_FMT_MASK 0x70
32 #define RTPG_FMT_EXT_HDR 0x10
34 #define TPGS_MODE_UNINITIALIZED -1
35 #define TPGS_MODE_NONE 0x0
36 #define TPGS_MODE_IMPLICIT 0x1
37 #define TPGS_MODE_EXPLICIT 0x2
39 #define ALUA_RTPG_SIZE 128
40 #define ALUA_FAILOVER_TIMEOUT 60
41 #define ALUA_FAILOVER_RETRIES 5
42 #define ALUA_RTPG_DELAY_MSECS 5
43 #define ALUA_RTPG_RETRY_DELAY 2
45 /* device handler flags */
46 #define ALUA_OPTIMIZE_STPG 0x01
47 #define ALUA_RTPG_EXT_HDR_UNSUPP 0x02
48 /* State machine flags */
49 #define ALUA_PG_RUN_RTPG 0x10
50 #define ALUA_PG_RUN_STPG 0x20
51 #define ALUA_PG_RUNNING 0x40
53 static uint optimize_stpg
;
54 module_param(optimize_stpg
, uint
, S_IRUGO
|S_IWUSR
);
55 MODULE_PARM_DESC(optimize_stpg
, "Allow use of a non-optimized path, rather than sending a STPG, when implicit TPGS is supported (0=No,1=Yes). Default is 0.");
57 static LIST_HEAD(port_group_list
);
58 static DEFINE_SPINLOCK(port_group_lock
);
59 static struct workqueue_struct
*kaluad_wq
;
61 struct alua_port_group
{
64 struct list_head node
;
65 struct list_head dh_list
;
66 unsigned char device_id_str
[256];
73 unsigned flags
; /* used for optimizing STPG */
74 unsigned char transition_tmo
;
76 unsigned long interval
;
77 struct delayed_work rtpg_work
;
79 struct list_head rtpg_list
;
80 struct scsi_device
*rtpg_sdev
;
84 struct list_head node
;
85 struct alua_port_group __rcu
*pg
;
88 struct scsi_device
*sdev
;
90 struct mutex init_mutex
;
94 struct alua_queue_data
{
95 struct list_head entry
;
96 activate_complete callback_fn
;
100 #define ALUA_POLICY_SWITCH_CURRENT 0
101 #define ALUA_POLICY_SWITCH_ALL 1
103 static void alua_rtpg_work(struct work_struct
*work
);
104 static bool alua_rtpg_queue(struct alua_port_group
*pg
,
105 struct scsi_device
*sdev
,
106 struct alua_queue_data
*qdata
, bool force
);
107 static void alua_check(struct scsi_device
*sdev
, bool force
);
109 static void release_port_group(struct kref
*kref
)
111 struct alua_port_group
*pg
;
113 pg
= container_of(kref
, struct alua_port_group
, kref
);
115 flush_delayed_work(&pg
->rtpg_work
);
116 spin_lock(&port_group_lock
);
118 spin_unlock(&port_group_lock
);
123 * submit_rtpg - Issue a REPORT TARGET GROUP STATES command
124 * @sdev: sdev the command should be sent to
126 static int submit_rtpg(struct scsi_device
*sdev
, unsigned char *buff
,
127 int bufflen
, struct scsi_sense_hdr
*sshdr
, int flags
)
129 u8 cdb
[MAX_COMMAND_SIZE
];
130 blk_opf_t opf
= REQ_OP_DRV_IN
| REQ_FAILFAST_DEV
|
131 REQ_FAILFAST_TRANSPORT
| REQ_FAILFAST_DRIVER
;
132 const struct scsi_exec_args exec_args
= {
136 /* Prepare the command. */
137 memset(cdb
, 0x0, MAX_COMMAND_SIZE
);
138 cdb
[0] = MAINTENANCE_IN
;
139 if (!(flags
& ALUA_RTPG_EXT_HDR_UNSUPP
))
140 cdb
[1] = MI_REPORT_TARGET_PGS
| MI_EXT_HDR_PARAM_FMT
;
142 cdb
[1] = MI_REPORT_TARGET_PGS
;
143 put_unaligned_be32(bufflen
, &cdb
[6]);
145 return scsi_execute_cmd(sdev
, cdb
, opf
, buff
, bufflen
,
146 ALUA_FAILOVER_TIMEOUT
* HZ
,
147 ALUA_FAILOVER_RETRIES
, &exec_args
);
151 * submit_stpg - Issue a SET TARGET PORT GROUP command
153 * Currently we're only setting the current target port group state
154 * to 'active/optimized' and let the array firmware figure out
155 * the states of the remaining groups.
157 static int submit_stpg(struct scsi_device
*sdev
, int group_id
,
158 struct scsi_sense_hdr
*sshdr
)
160 u8 cdb
[MAX_COMMAND_SIZE
];
161 unsigned char stpg_data
[8];
163 blk_opf_t opf
= REQ_OP_DRV_OUT
| REQ_FAILFAST_DEV
|
164 REQ_FAILFAST_TRANSPORT
| REQ_FAILFAST_DRIVER
;
165 const struct scsi_exec_args exec_args
= {
169 /* Prepare the data buffer */
170 memset(stpg_data
, 0, stpg_len
);
171 stpg_data
[4] = SCSI_ACCESS_STATE_OPTIMAL
;
172 put_unaligned_be16(group_id
, &stpg_data
[6]);
174 /* Prepare the command. */
175 memset(cdb
, 0x0, MAX_COMMAND_SIZE
);
176 cdb
[0] = MAINTENANCE_OUT
;
177 cdb
[1] = MO_SET_TARGET_PGS
;
178 put_unaligned_be32(stpg_len
, &cdb
[6]);
180 return scsi_execute_cmd(sdev
, cdb
, opf
, stpg_data
,
181 stpg_len
, ALUA_FAILOVER_TIMEOUT
* HZ
,
182 ALUA_FAILOVER_RETRIES
, &exec_args
);
185 static struct alua_port_group
*alua_find_get_pg(char *id_str
, size_t id_size
,
188 struct alua_port_group
*pg
;
190 if (!id_str
|| !id_size
|| !strlen(id_str
))
193 list_for_each_entry(pg
, &port_group_list
, node
) {
194 if (pg
->group_id
!= group_id
)
196 if (!pg
->device_id_len
|| pg
->device_id_len
!= id_size
)
198 if (strncmp(pg
->device_id_str
, id_str
, id_size
))
200 if (!kref_get_unless_zero(&pg
->kref
))
209 * alua_alloc_pg - Allocate a new port_group structure
211 * @group_id: port group id
212 * @tpgs: target port group settings
214 * Allocate a new port_group structure for a given
217 static struct alua_port_group
*alua_alloc_pg(struct scsi_device
*sdev
,
218 int group_id
, int tpgs
)
220 struct alua_port_group
*pg
, *tmp_pg
;
222 pg
= kzalloc(sizeof(struct alua_port_group
), GFP_KERNEL
);
224 return ERR_PTR(-ENOMEM
);
226 pg
->device_id_len
= scsi_vpd_lun_id(sdev
, pg
->device_id_str
,
227 sizeof(pg
->device_id_str
));
228 if (pg
->device_id_len
<= 0) {
230 * TPGS supported but no device identification found.
231 * Generate private device identification.
233 sdev_printk(KERN_INFO
, sdev
,
234 "%s: No device descriptors found\n",
236 pg
->device_id_str
[0] = '\0';
237 pg
->device_id_len
= 0;
239 pg
->group_id
= group_id
;
241 pg
->state
= SCSI_ACCESS_STATE_OPTIMAL
;
242 pg
->valid_states
= TPGS_SUPPORT_ALL
;
244 pg
->flags
|= ALUA_OPTIMIZE_STPG
;
245 kref_init(&pg
->kref
);
246 INIT_DELAYED_WORK(&pg
->rtpg_work
, alua_rtpg_work
);
247 INIT_LIST_HEAD(&pg
->rtpg_list
);
248 INIT_LIST_HEAD(&pg
->node
);
249 INIT_LIST_HEAD(&pg
->dh_list
);
250 spin_lock_init(&pg
->lock
);
252 spin_lock(&port_group_lock
);
253 tmp_pg
= alua_find_get_pg(pg
->device_id_str
, pg
->device_id_len
,
256 spin_unlock(&port_group_lock
);
261 list_add(&pg
->node
, &port_group_list
);
262 spin_unlock(&port_group_lock
);
268 * alua_check_tpgs - Evaluate TPGS setting
269 * @sdev: device to be checked
271 * Examine the TPGS setting of the sdev to find out if ALUA
274 static int alua_check_tpgs(struct scsi_device
*sdev
)
276 int tpgs
= TPGS_MODE_NONE
;
279 * ALUA support for non-disk devices is fraught with
280 * difficulties, so disable it for now.
282 if (sdev
->type
!= TYPE_DISK
) {
283 sdev_printk(KERN_INFO
, sdev
,
284 "%s: disable for non-disk devices\n",
289 tpgs
= scsi_device_tpgs(sdev
);
291 case TPGS_MODE_EXPLICIT
|TPGS_MODE_IMPLICIT
:
292 sdev_printk(KERN_INFO
, sdev
,
293 "%s: supports implicit and explicit TPGS\n",
296 case TPGS_MODE_EXPLICIT
:
297 sdev_printk(KERN_INFO
, sdev
, "%s: supports explicit TPGS\n",
300 case TPGS_MODE_IMPLICIT
:
301 sdev_printk(KERN_INFO
, sdev
, "%s: supports implicit TPGS\n",
305 sdev_printk(KERN_INFO
, sdev
, "%s: not supported\n",
309 sdev_printk(KERN_INFO
, sdev
,
310 "%s: unsupported TPGS setting %d\n",
312 tpgs
= TPGS_MODE_NONE
;
320 * alua_check_vpd - Evaluate INQUIRY vpd page 0x83
321 * @sdev: device to be checked
323 * Extract the relative target port and the target port group
324 * descriptor from the list of identificators.
326 static int alua_check_vpd(struct scsi_device
*sdev
, struct alua_dh_data
*h
,
329 int rel_port
= -1, group_id
;
330 struct alua_port_group
*pg
, *old_pg
= NULL
;
331 bool pg_updated
= false;
334 group_id
= scsi_vpd_tpg_id(sdev
, &rel_port
);
337 * Internal error; TPGS supported but required
338 * VPD identification descriptors not present.
339 * Disable ALUA support
341 sdev_printk(KERN_INFO
, sdev
,
342 "%s: No target port descriptors found\n",
344 return SCSI_DH_DEV_UNSUPP
;
347 pg
= alua_alloc_pg(sdev
, group_id
, tpgs
);
349 if (PTR_ERR(pg
) == -ENOMEM
)
350 return SCSI_DH_NOMEM
;
351 return SCSI_DH_DEV_UNSUPP
;
353 if (pg
->device_id_len
)
354 sdev_printk(KERN_INFO
, sdev
,
355 "%s: device %s port group %x rel port %x\n",
356 ALUA_DH_NAME
, pg
->device_id_str
,
359 sdev_printk(KERN_INFO
, sdev
,
360 "%s: port group %x rel port %x\n",
361 ALUA_DH_NAME
, group_id
, rel_port
);
365 /* Check for existing port group references */
366 spin_lock(&h
->pg_lock
);
367 old_pg
= rcu_dereference_protected(h
->pg
, lockdep_is_held(&h
->pg_lock
));
369 /* port group has changed. Update to new port group */
371 spin_lock_irqsave(&old_pg
->lock
, flags
);
372 list_del_rcu(&h
->node
);
373 spin_unlock_irqrestore(&old_pg
->lock
, flags
);
375 rcu_assign_pointer(h
->pg
, pg
);
379 spin_lock_irqsave(&pg
->lock
, flags
);
381 list_add_rcu(&h
->node
, &pg
->dh_list
);
382 spin_unlock_irqrestore(&pg
->lock
, flags
);
384 spin_unlock(&h
->pg_lock
);
386 alua_rtpg_queue(pg
, sdev
, NULL
, true);
387 kref_put(&pg
->kref
, release_port_group
);
390 kref_put(&old_pg
->kref
, release_port_group
);
395 static char print_alua_state(unsigned char state
)
398 case SCSI_ACCESS_STATE_OPTIMAL
:
400 case SCSI_ACCESS_STATE_ACTIVE
:
402 case SCSI_ACCESS_STATE_STANDBY
:
404 case SCSI_ACCESS_STATE_UNAVAILABLE
:
406 case SCSI_ACCESS_STATE_LBA
:
408 case SCSI_ACCESS_STATE_OFFLINE
:
410 case SCSI_ACCESS_STATE_TRANSITIONING
:
417 static void alua_handle_state_transition(struct scsi_device
*sdev
)
419 struct alua_dh_data
*h
= sdev
->handler_data
;
420 struct alua_port_group
*pg
;
423 pg
= rcu_dereference(h
->pg
);
425 pg
->state
= SCSI_ACCESS_STATE_TRANSITIONING
;
427 alua_check(sdev
, false);
430 static enum scsi_disposition
alua_check_sense(struct scsi_device
*sdev
,
431 struct scsi_sense_hdr
*sense_hdr
)
433 switch (sense_hdr
->sense_key
) {
435 if (sense_hdr
->asc
== 0x04 && sense_hdr
->ascq
== 0x0a) {
437 * LUN Not Accessible - ALUA state transition
439 alua_handle_state_transition(sdev
);
444 if (sense_hdr
->asc
== 0x04 && sense_hdr
->ascq
== 0x0a) {
446 * LUN Not Accessible - ALUA state transition
448 alua_handle_state_transition(sdev
);
451 if (sense_hdr
->asc
== 0x29 && sense_hdr
->ascq
== 0x00) {
453 * Power On, Reset, or Bus Device Reset.
454 * Might have obscured a state transition,
455 * so schedule a recheck.
457 alua_check(sdev
, true);
458 return ADD_TO_MLQUEUE
;
460 if (sense_hdr
->asc
== 0x29 && sense_hdr
->ascq
== 0x04)
462 * Device internal reset
464 return ADD_TO_MLQUEUE
;
465 if (sense_hdr
->asc
== 0x2a && sense_hdr
->ascq
== 0x01)
467 * Mode Parameters Changed
469 return ADD_TO_MLQUEUE
;
470 if (sense_hdr
->asc
== 0x2a && sense_hdr
->ascq
== 0x06) {
474 alua_check(sdev
, true);
475 return ADD_TO_MLQUEUE
;
477 if (sense_hdr
->asc
== 0x2a && sense_hdr
->ascq
== 0x07) {
479 * Implicit ALUA state transition failed
481 alua_check(sdev
, true);
482 return ADD_TO_MLQUEUE
;
484 if (sense_hdr
->asc
== 0x3f && sense_hdr
->ascq
== 0x03)
486 * Inquiry data has changed
488 return ADD_TO_MLQUEUE
;
489 if (sense_hdr
->asc
== 0x3f && sense_hdr
->ascq
== 0x0e)
491 * REPORTED_LUNS_DATA_HAS_CHANGED is reported
492 * when switching controllers on targets like
493 * Intel Multi-Flex. We can just retry.
495 return ADD_TO_MLQUEUE
;
499 return SCSI_RETURN_NOT_HANDLED
;
503 * alua_tur - Send a TEST UNIT READY
504 * @sdev: device to which the TEST UNIT READY command should be send
506 * Send a TEST UNIT READY to @sdev to figure out the device state
507 * Returns SCSI_DH_RETRY if the sense code is NOT READY/ALUA TRANSITIONING,
508 * SCSI_DH_OK if no error occurred, and SCSI_DH_IO otherwise.
510 static int alua_tur(struct scsi_device
*sdev
)
512 struct scsi_sense_hdr sense_hdr
;
515 retval
= scsi_test_unit_ready(sdev
, ALUA_FAILOVER_TIMEOUT
* HZ
,
516 ALUA_FAILOVER_RETRIES
, &sense_hdr
);
517 if ((sense_hdr
.sense_key
== NOT_READY
||
518 sense_hdr
.sense_key
== UNIT_ATTENTION
) &&
519 sense_hdr
.asc
== 0x04 && sense_hdr
.ascq
== 0x0a)
520 return SCSI_DH_RETRY
;
528 * alua_rtpg - Evaluate REPORT TARGET GROUP STATES
529 * @sdev: the device to be evaluated.
531 * Evaluate the Target Port Group State.
532 * Returns SCSI_DH_DEV_OFFLINED if the path is
533 * found to be unusable.
535 static int alua_rtpg(struct scsi_device
*sdev
, struct alua_port_group
*pg
)
537 struct scsi_sense_hdr sense_hdr
;
538 struct alua_port_group
*tmp_pg
;
539 int len
, k
, off
, bufflen
= ALUA_RTPG_SIZE
;
540 int group_id_old
, state_old
, pref_old
, valid_states_old
;
541 unsigned char *desc
, *buff
;
544 unsigned int tpg_desc_tbl_off
;
545 unsigned char orig_transition_tmo
;
547 bool transitioning_sense
= false;
549 group_id_old
= pg
->group_id
;
550 state_old
= pg
->state
;
552 valid_states_old
= pg
->valid_states
;
555 unsigned long transition_tmo
= ALUA_FAILOVER_TIMEOUT
* HZ
;
557 if (pg
->transition_tmo
)
558 transition_tmo
= pg
->transition_tmo
* HZ
;
560 pg
->expiry
= round_jiffies_up(jiffies
+ transition_tmo
);
563 buff
= kzalloc(bufflen
, GFP_KERNEL
);
565 return SCSI_DH_DEV_TEMP_BUSY
;
569 retval
= submit_rtpg(sdev
, buff
, bufflen
, &sense_hdr
, pg
->flags
);
573 * Some (broken) implementations have a habit of returning
574 * an error during things like firmware update etc.
575 * But if the target only supports active/optimized there's
576 * not much we can do; it's not that we can switch paths
578 * So ignore any errors to avoid spurious failures during
581 if ((pg
->valid_states
& ~TPGS_SUPPORT_OPTIMIZED
) == 0) {
582 sdev_printk(KERN_INFO
, sdev
,
583 "%s: ignoring rtpg result %d\n",
584 ALUA_DH_NAME
, retval
);
588 if (retval
< 0 || !scsi_sense_valid(&sense_hdr
)) {
589 sdev_printk(KERN_INFO
, sdev
,
590 "%s: rtpg failed, result %d\n",
591 ALUA_DH_NAME
, retval
);
594 return SCSI_DH_DEV_TEMP_BUSY
;
595 if (host_byte(retval
) == DID_NO_CONNECT
)
596 return SCSI_DH_RES_TEMP_UNAVAIL
;
601 * submit_rtpg() has failed on existing arrays
602 * when requesting extended header info, and
603 * the array doesn't support extended headers,
604 * even though it shouldn't according to T10.
605 * The retry without rtpg_ext_hdr_req set
607 * Note: some arrays return a sense key of ILLEGAL_REQUEST
608 * with ASC 00h if they don't support the extended header.
610 if (!(pg
->flags
& ALUA_RTPG_EXT_HDR_UNSUPP
) &&
611 sense_hdr
.sense_key
== ILLEGAL_REQUEST
) {
612 pg
->flags
|= ALUA_RTPG_EXT_HDR_UNSUPP
;
616 * If the array returns with 'ALUA state transition'
617 * sense code here it cannot return RTPG data during
618 * transition. So set the state to 'transitioning' directly.
620 if (sense_hdr
.sense_key
== NOT_READY
&&
621 sense_hdr
.asc
== 0x04 && sense_hdr
.ascq
== 0x0a) {
622 transitioning_sense
= true;
626 * Retry on any other UNIT ATTENTION occurred.
628 if (sense_hdr
.sense_key
== UNIT_ATTENTION
)
630 if (err
== SCSI_DH_RETRY
&&
631 pg
->expiry
!= 0 && time_before(jiffies
, pg
->expiry
)) {
632 sdev_printk(KERN_ERR
, sdev
, "%s: rtpg retry\n",
634 scsi_print_sense_hdr(sdev
, ALUA_DH_NAME
, &sense_hdr
);
638 sdev_printk(KERN_ERR
, sdev
, "%s: rtpg failed\n",
640 scsi_print_sense_hdr(sdev
, ALUA_DH_NAME
, &sense_hdr
);
646 len
= get_unaligned_be32(&buff
[0]) + 4;
649 /* Resubmit with the correct length */
652 buff
= kmalloc(bufflen
, GFP_KERNEL
);
654 sdev_printk(KERN_WARNING
, sdev
,
655 "%s: kmalloc buffer failed\n",__func__
);
656 /* Temporary failure, bypass */
658 return SCSI_DH_DEV_TEMP_BUSY
;
663 orig_transition_tmo
= pg
->transition_tmo
;
664 if ((buff
[4] & RTPG_FMT_MASK
) == RTPG_FMT_EXT_HDR
&& buff
[5] != 0)
665 pg
->transition_tmo
= buff
[5];
667 pg
->transition_tmo
= ALUA_FAILOVER_TIMEOUT
;
669 if (orig_transition_tmo
!= pg
->transition_tmo
) {
670 sdev_printk(KERN_INFO
, sdev
,
671 "%s: transition timeout set to %d seconds\n",
672 ALUA_DH_NAME
, pg
->transition_tmo
);
673 pg
->expiry
= jiffies
+ pg
->transition_tmo
* HZ
;
676 if ((buff
[4] & RTPG_FMT_MASK
) == RTPG_FMT_EXT_HDR
)
677 tpg_desc_tbl_off
= 8;
679 tpg_desc_tbl_off
= 4;
681 for (k
= tpg_desc_tbl_off
, desc
= buff
+ tpg_desc_tbl_off
;
683 k
+= off
, desc
+= off
) {
684 u16 group_id
= get_unaligned_be16(&desc
[2]);
686 spin_lock_irqsave(&port_group_lock
, flags
);
687 tmp_pg
= alua_find_get_pg(pg
->device_id_str
, pg
->device_id_len
,
689 spin_unlock_irqrestore(&port_group_lock
, flags
);
691 if (spin_trylock_irqsave(&tmp_pg
->lock
, flags
)) {
692 if ((tmp_pg
== pg
) ||
693 !(tmp_pg
->flags
& ALUA_PG_RUNNING
)) {
694 struct alua_dh_data
*h
;
696 tmp_pg
->state
= desc
[0] & 0x0f;
697 tmp_pg
->pref
= desc
[0] >> 7;
699 list_for_each_entry_rcu(h
,
700 &tmp_pg
->dh_list
, node
) {
703 h
->sdev
->access_state
= desc
[0];
708 tmp_pg
->valid_states
= desc
[1];
709 spin_unlock_irqrestore(&tmp_pg
->lock
, flags
);
711 kref_put(&tmp_pg
->kref
, release_port_group
);
713 off
= 8 + (desc
[7] * 4);
717 spin_lock_irqsave(&pg
->lock
, flags
);
718 if (transitioning_sense
)
719 pg
->state
= SCSI_ACCESS_STATE_TRANSITIONING
;
721 if (group_id_old
!= pg
->group_id
|| state_old
!= pg
->state
||
722 pref_old
!= pg
->pref
|| valid_states_old
!= pg
->valid_states
)
723 sdev_printk(KERN_INFO
, sdev
,
724 "%s: port group %02x state %c %s supports %c%c%c%c%c%c%c\n",
725 ALUA_DH_NAME
, pg
->group_id
, print_alua_state(pg
->state
),
726 pg
->pref
? "preferred" : "non-preferred",
727 pg
->valid_states
&TPGS_SUPPORT_TRANSITION
?'T':'t',
728 pg
->valid_states
&TPGS_SUPPORT_OFFLINE
?'O':'o',
729 pg
->valid_states
&TPGS_SUPPORT_LBA_DEPENDENT
?'L':'l',
730 pg
->valid_states
&TPGS_SUPPORT_UNAVAILABLE
?'U':'u',
731 pg
->valid_states
&TPGS_SUPPORT_STANDBY
?'S':'s',
732 pg
->valid_states
&TPGS_SUPPORT_NONOPTIMIZED
?'N':'n',
733 pg
->valid_states
&TPGS_SUPPORT_OPTIMIZED
?'A':'a');
736 case SCSI_ACCESS_STATE_TRANSITIONING
:
737 if (time_before(jiffies
, pg
->expiry
)) {
738 /* State transition, retry */
739 pg
->interval
= ALUA_RTPG_RETRY_DELAY
;
742 struct alua_dh_data
*h
;
744 /* Transitioning time exceeded, set port to standby */
746 pg
->state
= SCSI_ACCESS_STATE_STANDBY
;
749 list_for_each_entry_rcu(h
, &pg
->dh_list
, node
) {
752 h
->sdev
->access_state
=
753 (pg
->state
& SCSI_ACCESS_STATE_MASK
);
755 h
->sdev
->access_state
|=
756 SCSI_ACCESS_STATE_PREFERRED
;
761 case SCSI_ACCESS_STATE_OFFLINE
:
763 err
= SCSI_DH_DEV_OFFLINED
;
767 /* Useable path if active */
772 spin_unlock_irqrestore(&pg
->lock
, flags
);
778 * alua_stpg - Issue a SET TARGET PORT GROUP command
780 * Issue a SET TARGET PORT GROUP command and evaluate the
781 * response. Returns SCSI_DH_RETRY per default to trigger
782 * a re-evaluation of the target group state or SCSI_DH_OK
783 * if no further action needs to be taken.
785 static unsigned alua_stpg(struct scsi_device
*sdev
, struct alua_port_group
*pg
)
788 struct scsi_sense_hdr sense_hdr
;
790 if (!(pg
->tpgs
& TPGS_MODE_EXPLICIT
)) {
791 /* Only implicit ALUA supported, retry */
792 return SCSI_DH_RETRY
;
795 case SCSI_ACCESS_STATE_OPTIMAL
:
797 case SCSI_ACCESS_STATE_ACTIVE
:
798 if ((pg
->flags
& ALUA_OPTIMIZE_STPG
) &&
800 (pg
->tpgs
& TPGS_MODE_IMPLICIT
))
803 case SCSI_ACCESS_STATE_STANDBY
:
804 case SCSI_ACCESS_STATE_UNAVAILABLE
:
806 case SCSI_ACCESS_STATE_OFFLINE
:
808 case SCSI_ACCESS_STATE_TRANSITIONING
:
811 sdev_printk(KERN_INFO
, sdev
,
812 "%s: stpg failed, unhandled TPGS state %d",
813 ALUA_DH_NAME
, pg
->state
);
814 return SCSI_DH_NOSYS
;
816 retval
= submit_stpg(sdev
, pg
->group_id
, &sense_hdr
);
819 if (retval
< 0 || !scsi_sense_valid(&sense_hdr
)) {
820 sdev_printk(KERN_INFO
, sdev
,
821 "%s: stpg failed, result %d",
822 ALUA_DH_NAME
, retval
);
824 return SCSI_DH_DEV_TEMP_BUSY
;
826 sdev_printk(KERN_INFO
, sdev
, "%s: stpg failed\n",
828 scsi_print_sense_hdr(sdev
, ALUA_DH_NAME
, &sense_hdr
);
832 return SCSI_DH_RETRY
;
836 * The caller must call scsi_device_put() on the returned pointer if it is not
839 static struct scsi_device
* __must_check
840 alua_rtpg_select_sdev(struct alua_port_group
*pg
)
842 struct alua_dh_data
*h
;
843 struct scsi_device
*sdev
= NULL
, *prev_sdev
;
845 lockdep_assert_held(&pg
->lock
);
846 if (WARN_ON(!pg
->rtpg_sdev
))
850 * RCU protection isn't necessary for dh_list here
851 * as we hold pg->lock, but for access to h->pg.
854 list_for_each_entry_rcu(h
, &pg
->dh_list
, node
) {
857 if (h
->sdev
== pg
->rtpg_sdev
) {
861 if (rcu_dereference(h
->pg
) == pg
&&
863 !scsi_device_get(h
->sdev
)) {
871 pr_warn("%s: no device found for rtpg\n",
873 (char *)pg
->device_id_str
: "(nameless PG)"));
877 sdev_printk(KERN_INFO
, sdev
, "rtpg retry on different device\n");
879 prev_sdev
= pg
->rtpg_sdev
;
880 pg
->rtpg_sdev
= sdev
;
885 static void alua_rtpg_work(struct work_struct
*work
)
887 struct alua_port_group
*pg
=
888 container_of(work
, struct alua_port_group
, rtpg_work
.work
);
889 struct scsi_device
*sdev
, *prev_sdev
= NULL
;
890 LIST_HEAD(qdata_list
);
891 int err
= SCSI_DH_OK
;
892 struct alua_queue_data
*qdata
, *tmp
;
893 struct alua_dh_data
*h
;
896 spin_lock_irqsave(&pg
->lock
, flags
);
897 sdev
= pg
->rtpg_sdev
;
899 WARN_ON(pg
->flags
& ALUA_PG_RUN_RTPG
);
900 WARN_ON(pg
->flags
& ALUA_PG_RUN_STPG
);
901 spin_unlock_irqrestore(&pg
->lock
, flags
);
902 kref_put(&pg
->kref
, release_port_group
);
905 pg
->flags
|= ALUA_PG_RUNNING
;
906 if (pg
->flags
& ALUA_PG_RUN_RTPG
) {
907 int state
= pg
->state
;
909 pg
->flags
&= ~ALUA_PG_RUN_RTPG
;
910 spin_unlock_irqrestore(&pg
->lock
, flags
);
911 if (state
== SCSI_ACCESS_STATE_TRANSITIONING
) {
912 if (alua_tur(sdev
) == SCSI_DH_RETRY
) {
913 spin_lock_irqsave(&pg
->lock
, flags
);
914 pg
->flags
&= ~ALUA_PG_RUNNING
;
915 pg
->flags
|= ALUA_PG_RUN_RTPG
;
917 pg
->interval
= ALUA_RTPG_RETRY_DELAY
;
918 spin_unlock_irqrestore(&pg
->lock
, flags
);
919 queue_delayed_work(kaluad_wq
, &pg
->rtpg_work
,
923 /* Send RTPG on failure or if TUR indicates SUCCESS */
925 err
= alua_rtpg(sdev
, pg
);
926 spin_lock_irqsave(&pg
->lock
, flags
);
928 /* If RTPG failed on the current device, try using another */
929 if (err
== SCSI_DH_RES_TEMP_UNAVAIL
&&
930 (prev_sdev
= alua_rtpg_select_sdev(pg
)))
931 err
= SCSI_DH_IMM_RETRY
;
933 if (err
== SCSI_DH_RETRY
|| err
== SCSI_DH_IMM_RETRY
||
934 pg
->flags
& ALUA_PG_RUN_RTPG
) {
935 pg
->flags
&= ~ALUA_PG_RUNNING
;
936 if (err
== SCSI_DH_IMM_RETRY
)
938 else if (!pg
->interval
&& !(pg
->flags
& ALUA_PG_RUN_RTPG
))
939 pg
->interval
= ALUA_RTPG_RETRY_DELAY
;
940 pg
->flags
|= ALUA_PG_RUN_RTPG
;
941 spin_unlock_irqrestore(&pg
->lock
, flags
);
944 if (err
!= SCSI_DH_OK
)
945 pg
->flags
&= ~ALUA_PG_RUN_STPG
;
947 if (pg
->flags
& ALUA_PG_RUN_STPG
) {
948 pg
->flags
&= ~ALUA_PG_RUN_STPG
;
949 spin_unlock_irqrestore(&pg
->lock
, flags
);
950 err
= alua_stpg(sdev
, pg
);
951 spin_lock_irqsave(&pg
->lock
, flags
);
952 if (err
== SCSI_DH_RETRY
|| pg
->flags
& ALUA_PG_RUN_RTPG
) {
953 pg
->flags
|= ALUA_PG_RUN_RTPG
;
955 pg
->flags
&= ~ALUA_PG_RUNNING
;
956 spin_unlock_irqrestore(&pg
->lock
, flags
);
961 list_splice_init(&pg
->rtpg_list
, &qdata_list
);
963 * We went through an RTPG, for good or bad.
964 * Re-enable all devices for the next attempt.
966 list_for_each_entry(h
, &pg
->dh_list
, node
)
968 pg
->rtpg_sdev
= NULL
;
969 spin_unlock_irqrestore(&pg
->lock
, flags
);
972 scsi_device_put(prev_sdev
);
974 list_for_each_entry_safe(qdata
, tmp
, &qdata_list
, entry
) {
975 list_del(&qdata
->entry
);
976 if (qdata
->callback_fn
)
977 qdata
->callback_fn(qdata
->callback_data
, err
);
980 spin_lock_irqsave(&pg
->lock
, flags
);
981 pg
->flags
&= ~ALUA_PG_RUNNING
;
982 spin_unlock_irqrestore(&pg
->lock
, flags
);
983 scsi_device_put(sdev
);
984 kref_put(&pg
->kref
, release_port_group
);
989 scsi_device_put(prev_sdev
);
990 queue_delayed_work(kaluad_wq
, &pg
->rtpg_work
, pg
->interval
* HZ
);
994 * alua_rtpg_queue() - cause RTPG to be submitted asynchronously
995 * @pg: ALUA port group associated with @sdev.
996 * @sdev: SCSI device for which to submit an RTPG.
997 * @qdata: Information about the callback to invoke after the RTPG.
998 * @force: Whether or not to submit an RTPG if a work item that will submit an
999 * RTPG already has been scheduled.
1001 * Returns true if and only if alua_rtpg_work() will be called asynchronously.
1002 * That function is responsible for calling @qdata->fn().
1004 * Context: may be called from atomic context (alua_check()) only if the caller
1005 * holds an sdev reference.
1007 static bool alua_rtpg_queue(struct alua_port_group
*pg
,
1008 struct scsi_device
*sdev
,
1009 struct alua_queue_data
*qdata
, bool force
)
1011 int start_queue
= 0;
1012 unsigned long flags
;
1014 if (WARN_ON_ONCE(!pg
) || scsi_device_get(sdev
))
1017 spin_lock_irqsave(&pg
->lock
, flags
);
1019 list_add_tail(&qdata
->entry
, &pg
->rtpg_list
);
1020 pg
->flags
|= ALUA_PG_RUN_STPG
;
1023 if (pg
->rtpg_sdev
== NULL
) {
1024 struct alua_dh_data
*h
= sdev
->handler_data
;
1027 if (h
&& rcu_dereference(h
->pg
) == pg
) {
1029 pg
->flags
|= ALUA_PG_RUN_RTPG
;
1030 kref_get(&pg
->kref
);
1031 pg
->rtpg_sdev
= sdev
;
1035 } else if (!(pg
->flags
& ALUA_PG_RUN_RTPG
) && force
) {
1036 pg
->flags
|= ALUA_PG_RUN_RTPG
;
1037 /* Do not queue if the worker is already running */
1038 if (!(pg
->flags
& ALUA_PG_RUNNING
)) {
1039 kref_get(&pg
->kref
);
1044 spin_unlock_irqrestore(&pg
->lock
, flags
);
1047 if (queue_delayed_work(kaluad_wq
, &pg
->rtpg_work
,
1048 msecs_to_jiffies(ALUA_RTPG_DELAY_MSECS
)))
1051 kref_put(&pg
->kref
, release_port_group
);
1054 scsi_device_put(sdev
);
1060 * alua_initialize - Initialize ALUA state
1061 * @sdev: the device to be initialized
1063 * For the prep_fn to work correctly we have
1064 * to initialize the ALUA state for the device.
1066 static int alua_initialize(struct scsi_device
*sdev
, struct alua_dh_data
*h
)
1068 int err
= SCSI_DH_DEV_UNSUPP
, tpgs
;
1070 mutex_lock(&h
->init_mutex
);
1071 h
->disabled
= false;
1072 tpgs
= alua_check_tpgs(sdev
);
1073 if (tpgs
!= TPGS_MODE_NONE
)
1074 err
= alua_check_vpd(sdev
, h
, tpgs
);
1075 h
->init_error
= err
;
1076 mutex_unlock(&h
->init_mutex
);
1080 * alua_set_params - set/unset the optimize flag
1081 * @sdev: device on the path to be activated
1082 * params - parameters in the following format
1083 * "no_of_params\0param1\0param2\0param3\0...\0"
1084 * For example, to set the flag pass the following parameters
1085 * from multipath.conf
1086 * hardware_handler "2 alua 1"
1088 static int alua_set_params(struct scsi_device
*sdev
, const char *params
)
1090 struct alua_dh_data
*h
= sdev
->handler_data
;
1091 struct alua_port_group
*pg
= NULL
;
1092 unsigned int optimize
= 0, argc
;
1093 const char *p
= params
;
1094 int result
= SCSI_DH_OK
;
1095 unsigned long flags
;
1097 if ((sscanf(params
, "%u", &argc
) != 1) || (argc
!= 1))
1102 if ((sscanf(p
, "%u", &optimize
) != 1) || (optimize
> 1))
1106 pg
= rcu_dereference(h
->pg
);
1111 spin_lock_irqsave(&pg
->lock
, flags
);
1113 pg
->flags
|= ALUA_OPTIMIZE_STPG
;
1115 pg
->flags
&= ~ALUA_OPTIMIZE_STPG
;
1116 spin_unlock_irqrestore(&pg
->lock
, flags
);
1123 * alua_activate - activate a path
1124 * @sdev: device on the path to be activated
1126 * We're currently switching the port group to be activated only and
1127 * let the array figure out the rest.
1128 * There may be other arrays which require us to switch all port groups
1129 * based on a certain policy. But until we actually encounter them it
1132 static int alua_activate(struct scsi_device
*sdev
,
1133 activate_complete fn
, void *data
)
1135 struct alua_dh_data
*h
= sdev
->handler_data
;
1136 int err
= SCSI_DH_OK
;
1137 struct alua_queue_data
*qdata
;
1138 struct alua_port_group
*pg
;
1140 qdata
= kzalloc(sizeof(*qdata
), GFP_KERNEL
);
1142 err
= SCSI_DH_RES_TEMP_UNAVAIL
;
1145 qdata
->callback_fn
= fn
;
1146 qdata
->callback_data
= data
;
1148 mutex_lock(&h
->init_mutex
);
1150 pg
= rcu_dereference(h
->pg
);
1151 if (!pg
|| !kref_get_unless_zero(&pg
->kref
)) {
1154 err
= h
->init_error
;
1155 mutex_unlock(&h
->init_mutex
);
1159 mutex_unlock(&h
->init_mutex
);
1161 if (alua_rtpg_queue(pg
, sdev
, qdata
, true)) {
1165 err
= SCSI_DH_DEV_OFFLINED
;
1167 kref_put(&pg
->kref
, release_port_group
);
1175 * alua_check - check path status
1176 * @sdev: device on the path to be checked
1178 * Check the device status
1180 static void alua_check(struct scsi_device
*sdev
, bool force
)
1182 struct alua_dh_data
*h
= sdev
->handler_data
;
1183 struct alua_port_group
*pg
;
1186 pg
= rcu_dereference(h
->pg
);
1187 if (!pg
|| !kref_get_unless_zero(&pg
->kref
)) {
1192 alua_rtpg_queue(pg
, sdev
, NULL
, force
);
1193 kref_put(&pg
->kref
, release_port_group
);
1197 * alua_prep_fn - request callback
1199 * Fail I/O to all paths not in state
1200 * active/optimized or active/non-optimized.
1202 static blk_status_t
alua_prep_fn(struct scsi_device
*sdev
, struct request
*req
)
1204 struct alua_dh_data
*h
= sdev
->handler_data
;
1205 struct alua_port_group
*pg
;
1206 unsigned char state
= SCSI_ACCESS_STATE_OPTIMAL
;
1209 pg
= rcu_dereference(h
->pg
);
1215 case SCSI_ACCESS_STATE_OPTIMAL
:
1216 case SCSI_ACCESS_STATE_ACTIVE
:
1217 case SCSI_ACCESS_STATE_LBA
:
1218 case SCSI_ACCESS_STATE_TRANSITIONING
:
1221 req
->rq_flags
|= RQF_QUIET
;
1222 return BLK_STS_IOERR
;
1226 static void alua_rescan(struct scsi_device
*sdev
)
1228 struct alua_dh_data
*h
= sdev
->handler_data
;
1230 alua_initialize(sdev
, h
);
1234 * alua_bus_attach - Attach device handler
1235 * @sdev: device to be attached to
1237 static int alua_bus_attach(struct scsi_device
*sdev
)
1239 struct alua_dh_data
*h
;
1242 h
= kzalloc(sizeof(*h
) , GFP_KERNEL
);
1244 return SCSI_DH_NOMEM
;
1245 spin_lock_init(&h
->pg_lock
);
1246 rcu_assign_pointer(h
->pg
, NULL
);
1247 h
->init_error
= SCSI_DH_OK
;
1249 INIT_LIST_HEAD(&h
->node
);
1251 mutex_init(&h
->init_mutex
);
1252 err
= alua_initialize(sdev
, h
);
1253 if (err
!= SCSI_DH_OK
&& err
!= SCSI_DH_DEV_OFFLINED
)
1256 sdev
->handler_data
= h
;
1264 * alua_bus_detach - Detach device handler
1265 * @sdev: device to be detached from
1267 static void alua_bus_detach(struct scsi_device
*sdev
)
1269 struct alua_dh_data
*h
= sdev
->handler_data
;
1270 struct alua_port_group
*pg
;
1272 spin_lock(&h
->pg_lock
);
1273 pg
= rcu_dereference_protected(h
->pg
, lockdep_is_held(&h
->pg_lock
));
1274 rcu_assign_pointer(h
->pg
, NULL
);
1275 spin_unlock(&h
->pg_lock
);
1277 spin_lock_irq(&pg
->lock
);
1278 list_del_rcu(&h
->node
);
1279 spin_unlock_irq(&pg
->lock
);
1280 kref_put(&pg
->kref
, release_port_group
);
1282 sdev
->handler_data
= NULL
;
1287 static struct scsi_device_handler alua_dh
= {
1288 .name
= ALUA_DH_NAME
,
1289 .module
= THIS_MODULE
,
1290 .attach
= alua_bus_attach
,
1291 .detach
= alua_bus_detach
,
1292 .prep_fn
= alua_prep_fn
,
1293 .check_sense
= alua_check_sense
,
1294 .activate
= alua_activate
,
1295 .rescan
= alua_rescan
,
1296 .set_params
= alua_set_params
,
1299 static int __init
alua_init(void)
1303 kaluad_wq
= alloc_workqueue("kaluad", WQ_MEM_RECLAIM
, 0);
1307 r
= scsi_register_device_handler(&alua_dh
);
1309 printk(KERN_ERR
"%s: Failed to register scsi device handler",
1311 destroy_workqueue(kaluad_wq
);
1316 static void __exit
alua_exit(void)
1318 scsi_unregister_device_handler(&alua_dh
);
1319 destroy_workqueue(kaluad_wq
);
1322 module_init(alua_init
);
1323 module_exit(alua_exit
);
1325 MODULE_DESCRIPTION("DM Multipath ALUA support");
1326 MODULE_AUTHOR("Hannes Reinecke <hare@suse.de>");
1327 MODULE_LICENSE("GPL");
1328 MODULE_VERSION(ALUA_DH_VER
);