1 // SPDX-License-Identifier: GPL-2.0-only
3 * Serial Attached SCSI (SAS) class SCSI Host glue.
5 * Copyright (C) 2005 Adaptec, Inc. All rights reserved.
6 * Copyright (C) 2005 Luben Tuikov <luben_tuikov@adaptec.com>
9 #include <linux/kthread.h>
10 #include <linux/firmware.h>
11 #include <linux/export.h>
12 #include <linux/ctype.h>
13 #include <linux/kernel.h>
15 #include "sas_internal.h"
17 #include <scsi/scsi_host.h>
18 #include <scsi/scsi_device.h>
19 #include <scsi/scsi_tcq.h>
20 #include <scsi/scsi.h>
21 #include <scsi/scsi_eh.h>
22 #include <scsi/scsi_transport.h>
23 #include <scsi/scsi_transport_sas.h>
24 #include <scsi/sas_ata.h>
25 #include "../scsi_sas_internal.h"
26 #include "../scsi_transport_api.h"
27 #include "../scsi_priv.h"
29 #include <linux/err.h>
30 #include <linux/blkdev.h>
31 #include <linux/freezer.h>
32 #include <linux/gfp.h>
33 #include <linux/scatterlist.h>
34 #include <linux/libata.h>
36 /* record final status and free the task */
37 static void sas_end_task(struct scsi_cmnd
*sc
, struct sas_task
*task
)
39 struct task_status_struct
*ts
= &task
->task_status
;
42 if (ts
->resp
== SAS_TASK_UNDELIVERED
) {
45 } else { /* ts->resp == SAS_TASK_COMPLETE */
46 /* task delivered, what happened afterwards? */
48 case SAS_DEV_NO_RESPONSE
:
55 case SAS_DATA_UNDERRUN
:
56 scsi_set_resid(sc
, ts
->residual
);
57 if (scsi_bufflen(sc
) - scsi_get_resid(sc
) < sc
->underflow
)
60 case SAS_DATA_OVERRUN
:
64 hs
= DID_SOFT_ERROR
; /* retry */
66 case SAS_DEVICE_UNKNOWN
:
73 if (ts
->open_rej_reason
== SAS_OREJ_RSVD_RETRY
)
74 hs
= DID_SOFT_ERROR
; /* retry */
78 case SAS_PROTO_RESPONSE
:
79 pr_notice("LLDD:%s sent SAS_PROTO_RESP for an SSP task; please report this\n",
80 task
->dev
->port
->ha
->sas_ha_name
);
82 case SAS_ABORTED_TASK
:
85 case SAM_STAT_CHECK_CONDITION
:
86 memcpy(sc
->sense_buffer
, ts
->buf
,
87 min(SCSI_SENSE_BUFFERSIZE
, ts
->buf_valid_size
));
88 stat
= SAM_STAT_CHECK_CONDITION
;
96 sc
->result
= (hs
<< 16) | stat
;
97 ASSIGN_SAS_TASK(sc
, NULL
);
101 static void sas_scsi_task_done(struct sas_task
*task
)
103 struct scsi_cmnd
*sc
= task
->uldd_task
;
104 struct domain_device
*dev
= task
->dev
;
105 struct sas_ha_struct
*ha
= dev
->port
->ha
;
108 spin_lock_irqsave(&dev
->done_lock
, flags
);
109 if (test_bit(SAS_HA_FROZEN
, &ha
->state
))
112 ASSIGN_SAS_TASK(sc
, NULL
);
113 spin_unlock_irqrestore(&dev
->done_lock
, flags
);
115 if (unlikely(!task
)) {
116 /* task will be completed by the error handler */
117 pr_debug("task done but aborted\n");
122 pr_debug("task_done called with non existing SCSI cmnd!\n");
127 sas_end_task(sc
, task
);
131 static struct sas_task
*sas_create_task(struct scsi_cmnd
*cmd
,
132 struct domain_device
*dev
,
135 struct sas_task
*task
= sas_alloc_task(gfp_flags
);
141 task
->uldd_task
= cmd
;
142 ASSIGN_SAS_TASK(cmd
, task
);
145 task
->task_proto
= task
->dev
->tproto
; /* BUG_ON(!SSP) */
147 task
->ssp_task
.retry_count
= 1;
148 int_to_scsilun(cmd
->device
->lun
, &lun
);
149 memcpy(task
->ssp_task
.LUN
, &lun
.scsi_lun
, 8);
150 task
->ssp_task
.task_attr
= TASK_ATTR_SIMPLE
;
151 task
->ssp_task
.cmd
= cmd
;
153 task
->scatter
= scsi_sglist(cmd
);
154 task
->num_scatter
= scsi_sg_count(cmd
);
155 task
->total_xfer_len
= scsi_bufflen(cmd
);
156 task
->data_dir
= cmd
->sc_data_direction
;
158 task
->task_done
= sas_scsi_task_done
;
163 int sas_queuecommand(struct Scsi_Host
*host
, struct scsi_cmnd
*cmd
)
165 struct sas_internal
*i
= to_sas_internal(host
->transportt
);
166 struct domain_device
*dev
= cmd_to_domain_dev(cmd
);
167 struct sas_task
*task
;
170 /* If the device fell off, no sense in issuing commands */
171 if (test_bit(SAS_DEV_GONE
, &dev
->state
)) {
172 cmd
->result
= DID_BAD_TARGET
<< 16;
176 if (dev_is_sata(dev
)) {
177 spin_lock_irq(dev
->sata_dev
.ap
->lock
);
178 res
= ata_sas_queuecmd(cmd
, dev
->sata_dev
.ap
);
179 spin_unlock_irq(dev
->sata_dev
.ap
->lock
);
183 task
= sas_create_task(cmd
, dev
, GFP_ATOMIC
);
185 return SCSI_MLQUEUE_HOST_BUSY
;
187 res
= i
->dft
->lldd_execute_task(task
, GFP_ATOMIC
);
193 pr_debug("lldd_execute_task returned: %d\n", res
);
194 ASSIGN_SAS_TASK(cmd
, NULL
);
196 if (res
== -SAS_QUEUE_FULL
)
197 cmd
->result
= DID_SOFT_ERROR
<< 16; /* retry */
199 cmd
->result
= DID_ERROR
<< 16;
205 static void sas_eh_finish_cmd(struct scsi_cmnd
*cmd
)
207 struct sas_ha_struct
*sas_ha
= SHOST_TO_SAS_HA(cmd
->device
->host
);
208 struct domain_device
*dev
= cmd_to_domain_dev(cmd
);
209 struct sas_task
*task
= TO_SAS_TASK(cmd
);
211 /* At this point, we only get called following an actual abort
212 * of the task, so we should be guaranteed not to be racing with
213 * any completions from the LLD. Task is freed after this.
215 sas_end_task(cmd
, task
);
217 if (dev_is_sata(dev
)) {
218 /* defer commands to libata so that libata EH can
219 * handle ata qcs correctly
221 list_move_tail(&cmd
->eh_entry
, &sas_ha
->eh_ata_q
);
225 /* now finish the command and move it on to the error
226 * handler done list, this also takes it off the
227 * error handler pending list.
229 scsi_eh_finish_cmd(cmd
, &sas_ha
->eh_done_q
);
232 static void sas_scsi_clear_queue_lu(struct list_head
*error_q
, struct scsi_cmnd
*my_cmd
)
234 struct scsi_cmnd
*cmd
, *n
;
236 list_for_each_entry_safe(cmd
, n
, error_q
, eh_entry
) {
237 if (cmd
->device
->sdev_target
== my_cmd
->device
->sdev_target
&&
238 cmd
->device
->lun
== my_cmd
->device
->lun
)
239 sas_eh_finish_cmd(cmd
);
243 static void sas_scsi_clear_queue_I_T(struct list_head
*error_q
,
244 struct domain_device
*dev
)
246 struct scsi_cmnd
*cmd
, *n
;
248 list_for_each_entry_safe(cmd
, n
, error_q
, eh_entry
) {
249 struct domain_device
*x
= cmd_to_domain_dev(cmd
);
252 sas_eh_finish_cmd(cmd
);
256 static void sas_scsi_clear_queue_port(struct list_head
*error_q
,
257 struct asd_sas_port
*port
)
259 struct scsi_cmnd
*cmd
, *n
;
261 list_for_each_entry_safe(cmd
, n
, error_q
, eh_entry
) {
262 struct domain_device
*dev
= cmd_to_domain_dev(cmd
);
263 struct asd_sas_port
*x
= dev
->port
;
266 sas_eh_finish_cmd(cmd
);
270 enum task_disposition
{
278 static enum task_disposition
sas_scsi_find_task(struct sas_task
*task
)
282 struct sas_internal
*si
=
283 to_sas_internal(task
->dev
->port
->ha
->core
.shost
->transportt
);
285 for (i
= 0; i
< 5; i
++) {
286 pr_notice("%s: aborting task 0x%p\n", __func__
, task
);
287 res
= si
->dft
->lldd_abort_task(task
);
289 spin_lock_irqsave(&task
->task_state_lock
, flags
);
290 if (task
->task_state_flags
& SAS_TASK_STATE_DONE
) {
291 spin_unlock_irqrestore(&task
->task_state_lock
, flags
);
292 pr_debug("%s: task 0x%p is done\n", __func__
, task
);
295 spin_unlock_irqrestore(&task
->task_state_lock
, flags
);
297 if (res
== TMF_RESP_FUNC_COMPLETE
) {
298 pr_notice("%s: task 0x%p is aborted\n",
300 return TASK_IS_ABORTED
;
301 } else if (si
->dft
->lldd_query_task
) {
302 pr_notice("%s: querying task 0x%p\n", __func__
, task
);
303 res
= si
->dft
->lldd_query_task(task
);
305 case TMF_RESP_FUNC_SUCC
:
306 pr_notice("%s: task 0x%p at LU\n", __func__
,
308 return TASK_IS_AT_LU
;
309 case TMF_RESP_FUNC_COMPLETE
:
310 pr_notice("%s: task 0x%p not at LU\n",
312 return TASK_IS_NOT_AT_LU
;
313 case TMF_RESP_FUNC_FAILED
:
314 pr_notice("%s: task 0x%p failed to abort\n",
316 return TASK_ABORT_FAILED
;
324 static int sas_recover_lu(struct domain_device
*dev
, struct scsi_cmnd
*cmd
)
326 int res
= TMF_RESP_FUNC_FAILED
;
328 struct sas_internal
*i
=
329 to_sas_internal(dev
->port
->ha
->core
.shost
->transportt
);
331 int_to_scsilun(cmd
->device
->lun
, &lun
);
333 pr_notice("eh: device %llx LUN %llx has the task\n",
334 SAS_ADDR(dev
->sas_addr
),
337 if (i
->dft
->lldd_abort_task_set
)
338 res
= i
->dft
->lldd_abort_task_set(dev
, lun
.scsi_lun
);
340 if (res
== TMF_RESP_FUNC_FAILED
) {
341 if (i
->dft
->lldd_clear_task_set
)
342 res
= i
->dft
->lldd_clear_task_set(dev
, lun
.scsi_lun
);
345 if (res
== TMF_RESP_FUNC_FAILED
) {
346 if (i
->dft
->lldd_lu_reset
)
347 res
= i
->dft
->lldd_lu_reset(dev
, lun
.scsi_lun
);
353 static int sas_recover_I_T(struct domain_device
*dev
)
355 int res
= TMF_RESP_FUNC_FAILED
;
356 struct sas_internal
*i
=
357 to_sas_internal(dev
->port
->ha
->core
.shost
->transportt
);
359 pr_notice("I_T nexus reset for dev %016llx\n",
360 SAS_ADDR(dev
->sas_addr
));
362 if (i
->dft
->lldd_I_T_nexus_reset
)
363 res
= i
->dft
->lldd_I_T_nexus_reset(dev
);
368 /* take a reference on the last known good phy for this device */
369 struct sas_phy
*sas_get_local_phy(struct domain_device
*dev
)
371 struct sas_ha_struct
*ha
= dev
->port
->ha
;
375 /* a published domain device always has a valid phy, it may be
376 * stale, but it is never NULL
380 spin_lock_irqsave(&ha
->phy_port_lock
, flags
);
382 get_device(&phy
->dev
);
383 spin_unlock_irqrestore(&ha
->phy_port_lock
, flags
);
387 EXPORT_SYMBOL_GPL(sas_get_local_phy
);
389 static void sas_wait_eh(struct domain_device
*dev
)
391 struct sas_ha_struct
*ha
= dev
->port
->ha
;
394 if (dev_is_sata(dev
)) {
395 ata_port_wait_eh(dev
->sata_dev
.ap
);
399 spin_lock_irq(&ha
->lock
);
401 while (test_bit(SAS_DEV_EH_PENDING
, &dev
->state
)) {
402 prepare_to_wait(&ha
->eh_wait_q
, &wait
, TASK_UNINTERRUPTIBLE
);
403 spin_unlock_irq(&ha
->lock
);
405 spin_lock_irq(&ha
->lock
);
407 finish_wait(&ha
->eh_wait_q
, &wait
);
409 spin_unlock_irq(&ha
->lock
);
411 /* make sure SCSI EH is complete */
412 if (scsi_host_in_recovery(ha
->core
.shost
)) {
418 static int sas_queue_reset(struct domain_device
*dev
, int reset_type
,
421 struct sas_ha_struct
*ha
= dev
->port
->ha
;
422 int scheduled
= 0, tries
= 100;
424 /* ata: promote lun reset to bus reset */
425 if (dev_is_sata(dev
)) {
426 sas_ata_schedule_reset(dev
);
428 sas_ata_wait_eh(dev
);
432 while (!scheduled
&& tries
--) {
433 spin_lock_irq(&ha
->lock
);
434 if (!test_bit(SAS_DEV_EH_PENDING
, &dev
->state
) &&
435 !test_bit(reset_type
, &dev
->state
)) {
438 list_add_tail(&dev
->ssp_dev
.eh_list_node
, &ha
->eh_dev_q
);
439 set_bit(SAS_DEV_EH_PENDING
, &dev
->state
);
440 set_bit(reset_type
, &dev
->state
);
441 int_to_scsilun(lun
, &dev
->ssp_dev
.reset_lun
);
442 scsi_schedule_eh(ha
->core
.shost
);
444 spin_unlock_irq(&ha
->lock
);
453 pr_warn("%s reset of %s failed\n",
454 reset_type
== SAS_DEV_LU_RESET
? "LUN" : "Bus",
455 dev_name(&dev
->rphy
->dev
));
460 int sas_eh_abort_handler(struct scsi_cmnd
*cmd
)
462 int res
= TMF_RESP_FUNC_FAILED
;
463 struct sas_task
*task
= TO_SAS_TASK(cmd
);
464 struct Scsi_Host
*host
= cmd
->device
->host
;
465 struct domain_device
*dev
= cmd_to_domain_dev(cmd
);
466 struct sas_internal
*i
= to_sas_internal(host
->transportt
);
469 if (!i
->dft
->lldd_abort_task
)
472 spin_lock_irqsave(host
->host_lock
, flags
);
473 /* We cannot do async aborts for SATA devices */
474 if (dev_is_sata(dev
) && !host
->host_eh_scheduled
) {
475 spin_unlock_irqrestore(host
->host_lock
, flags
);
478 spin_unlock_irqrestore(host
->host_lock
, flags
);
481 res
= i
->dft
->lldd_abort_task(task
);
483 pr_notice("no task to abort\n");
484 if (res
== TMF_RESP_FUNC_SUCC
|| res
== TMF_RESP_FUNC_COMPLETE
)
489 EXPORT_SYMBOL_GPL(sas_eh_abort_handler
);
491 /* Attempt to send a LUN reset message to a device */
492 int sas_eh_device_reset_handler(struct scsi_cmnd
*cmd
)
496 struct Scsi_Host
*host
= cmd
->device
->host
;
497 struct domain_device
*dev
= cmd_to_domain_dev(cmd
);
498 struct sas_internal
*i
= to_sas_internal(host
->transportt
);
500 if (current
!= host
->ehandler
)
501 return sas_queue_reset(dev
, SAS_DEV_LU_RESET
, cmd
->device
->lun
, 0);
503 int_to_scsilun(cmd
->device
->lun
, &lun
);
505 if (!i
->dft
->lldd_lu_reset
)
508 res
= i
->dft
->lldd_lu_reset(dev
, lun
.scsi_lun
);
509 if (res
== TMF_RESP_FUNC_SUCC
|| res
== TMF_RESP_FUNC_COMPLETE
)
515 int sas_eh_target_reset_handler(struct scsi_cmnd
*cmd
)
518 struct Scsi_Host
*host
= cmd
->device
->host
;
519 struct domain_device
*dev
= cmd_to_domain_dev(cmd
);
520 struct sas_internal
*i
= to_sas_internal(host
->transportt
);
522 if (current
!= host
->ehandler
)
523 return sas_queue_reset(dev
, SAS_DEV_RESET
, 0, 0);
525 if (!i
->dft
->lldd_I_T_nexus_reset
)
528 res
= i
->dft
->lldd_I_T_nexus_reset(dev
);
529 if (res
== TMF_RESP_FUNC_SUCC
|| res
== TMF_RESP_FUNC_COMPLETE
||
536 /* Try to reset a device */
537 static int try_to_reset_cmd_device(struct scsi_cmnd
*cmd
)
540 struct Scsi_Host
*shost
= cmd
->device
->host
;
542 if (!shost
->hostt
->eh_device_reset_handler
)
543 goto try_target_reset
;
545 res
= shost
->hostt
->eh_device_reset_handler(cmd
);
550 if (shost
->hostt
->eh_target_reset_handler
)
551 return shost
->hostt
->eh_target_reset_handler(cmd
);
556 static void sas_eh_handle_sas_errors(struct Scsi_Host
*shost
, struct list_head
*work_q
)
558 struct scsi_cmnd
*cmd
, *n
;
559 enum task_disposition res
= TASK_IS_DONE
;
560 int tmf_resp
, need_reset
;
561 struct sas_internal
*i
= to_sas_internal(shost
->transportt
);
563 struct sas_ha_struct
*ha
= SHOST_TO_SAS_HA(shost
);
566 /* clean out any commands that won the completion vs eh race */
567 list_for_each_entry_safe(cmd
, n
, work_q
, eh_entry
) {
568 struct domain_device
*dev
= cmd_to_domain_dev(cmd
);
569 struct sas_task
*task
;
571 spin_lock_irqsave(&dev
->done_lock
, flags
);
572 /* by this point the lldd has either observed
573 * SAS_HA_FROZEN and is leaving the task alone, or has
574 * won the race with eh and decided to complete it
576 task
= TO_SAS_TASK(cmd
);
577 spin_unlock_irqrestore(&dev
->done_lock
, flags
);
580 list_move_tail(&cmd
->eh_entry
, &done
);
584 list_for_each_entry_safe(cmd
, n
, work_q
, eh_entry
) {
585 struct sas_task
*task
= TO_SAS_TASK(cmd
);
587 list_del_init(&cmd
->eh_entry
);
589 spin_lock_irqsave(&task
->task_state_lock
, flags
);
590 need_reset
= task
->task_state_flags
& SAS_TASK_NEED_DEV_RESET
;
591 spin_unlock_irqrestore(&task
->task_state_lock
, flags
);
594 pr_notice("%s: task 0x%p requests reset\n",
599 pr_debug("trying to find task 0x%p\n", task
);
600 res
= sas_scsi_find_task(task
);
604 pr_notice("%s: task 0x%p is done\n", __func__
,
606 sas_eh_finish_cmd(cmd
);
608 case TASK_IS_ABORTED
:
609 pr_notice("%s: task 0x%p is aborted\n",
611 sas_eh_finish_cmd(cmd
);
614 pr_info("task 0x%p is at LU: lu recover\n", task
);
616 tmf_resp
= sas_recover_lu(task
->dev
, cmd
);
617 if (tmf_resp
== TMF_RESP_FUNC_COMPLETE
) {
618 pr_notice("dev %016llx LU %llx is recovered\n",
621 sas_eh_finish_cmd(cmd
);
622 sas_scsi_clear_queue_lu(work_q
, cmd
);
626 case TASK_IS_NOT_AT_LU
:
627 case TASK_ABORT_FAILED
:
628 pr_notice("task 0x%p is not at LU: I_T recover\n",
630 tmf_resp
= sas_recover_I_T(task
->dev
);
631 if (tmf_resp
== TMF_RESP_FUNC_COMPLETE
||
632 tmf_resp
== -ENODEV
) {
633 struct domain_device
*dev
= task
->dev
;
634 pr_notice("I_T %016llx recovered\n",
635 SAS_ADDR(task
->dev
->sas_addr
));
636 sas_eh_finish_cmd(cmd
);
637 sas_scsi_clear_queue_I_T(work_q
, dev
);
640 /* Hammer time :-) */
641 try_to_reset_cmd_device(cmd
);
642 if (i
->dft
->lldd_clear_nexus_port
) {
643 struct asd_sas_port
*port
= task
->dev
->port
;
644 pr_debug("clearing nexus for port:%d\n",
646 res
= i
->dft
->lldd_clear_nexus_port(port
);
647 if (res
== TMF_RESP_FUNC_COMPLETE
) {
648 pr_notice("clear nexus port:%d succeeded\n",
650 sas_eh_finish_cmd(cmd
);
651 sas_scsi_clear_queue_port(work_q
,
656 if (i
->dft
->lldd_clear_nexus_ha
) {
657 pr_debug("clear nexus ha\n");
658 res
= i
->dft
->lldd_clear_nexus_ha(ha
);
659 if (res
== TMF_RESP_FUNC_COMPLETE
) {
660 pr_notice("clear nexus ha succeeded\n");
661 sas_eh_finish_cmd(cmd
);
665 /* If we are here -- this means that no amount
666 * of effort could recover from errors. Quite
667 * possibly the HA just disappeared.
669 pr_err("error from device %llx, LUN %llx couldn't be recovered in any way\n",
670 SAS_ADDR(task
->dev
->sas_addr
),
673 sas_eh_finish_cmd(cmd
);
678 list_splice_tail(&done
, work_q
);
679 list_splice_tail_init(&ha
->eh_ata_q
, work_q
);
683 pr_debug("--- Exit %s -- clear_q\n", __func__
);
684 list_for_each_entry_safe(cmd
, n
, work_q
, eh_entry
)
685 sas_eh_finish_cmd(cmd
);
689 static void sas_eh_handle_resets(struct Scsi_Host
*shost
)
691 struct sas_ha_struct
*ha
= SHOST_TO_SAS_HA(shost
);
692 struct sas_internal
*i
= to_sas_internal(shost
->transportt
);
694 /* handle directed resets to sas devices */
695 spin_lock_irq(&ha
->lock
);
696 while (!list_empty(&ha
->eh_dev_q
)) {
697 struct domain_device
*dev
;
698 struct ssp_device
*ssp
;
700 ssp
= list_entry(ha
->eh_dev_q
.next
, typeof(*ssp
), eh_list_node
);
701 list_del_init(&ssp
->eh_list_node
);
702 dev
= container_of(ssp
, typeof(*dev
), ssp_dev
);
703 kref_get(&dev
->kref
);
704 WARN_ONCE(dev_is_sata(dev
), "ssp reset to ata device?\n");
706 spin_unlock_irq(&ha
->lock
);
708 if (test_and_clear_bit(SAS_DEV_LU_RESET
, &dev
->state
))
709 i
->dft
->lldd_lu_reset(dev
, ssp
->reset_lun
.scsi_lun
);
711 if (test_and_clear_bit(SAS_DEV_RESET
, &dev
->state
))
712 i
->dft
->lldd_I_T_nexus_reset(dev
);
715 spin_lock_irq(&ha
->lock
);
716 clear_bit(SAS_DEV_EH_PENDING
, &dev
->state
);
719 spin_unlock_irq(&ha
->lock
);
723 void sas_scsi_recover_host(struct Scsi_Host
*shost
)
725 struct sas_ha_struct
*ha
= SHOST_TO_SAS_HA(shost
);
726 LIST_HEAD(eh_work_q
);
733 spin_lock_irq(shost
->host_lock
);
734 list_splice_init(&shost
->eh_cmd_q
, &eh_work_q
);
735 spin_unlock_irq(shost
->host_lock
);
737 pr_notice("Enter %s busy: %d failed: %d\n",
738 __func__
, scsi_host_busy(shost
), shost
->host_failed
);
740 * Deal with commands that still have SAS tasks (i.e. they didn't
741 * complete via the normal sas_task completion mechanism),
742 * SAS_HA_FROZEN gives eh dominion over all sas_task completion.
744 set_bit(SAS_HA_FROZEN
, &ha
->state
);
745 sas_eh_handle_sas_errors(shost
, &eh_work_q
);
746 clear_bit(SAS_HA_FROZEN
, &ha
->state
);
747 if (list_empty(&eh_work_q
))
751 * Now deal with SCSI commands that completed ok but have a an error
752 * code (and hopefully sense data) attached. This is roughly what
753 * scsi_unjam_host does, but we skip scsi_eh_abort_cmds because any
754 * command we see here has no sas_task and is thus unknown to the HA.
756 sas_ata_eh(shost
, &eh_work_q
, &ha
->eh_done_q
);
757 if (!scsi_eh_get_sense(&eh_work_q
, &ha
->eh_done_q
))
758 scsi_eh_ready_devs(shost
, &eh_work_q
, &ha
->eh_done_q
);
761 sas_eh_handle_resets(shost
);
763 /* now link into libata eh --- if we have any ata devices */
764 sas_ata_strategy_handler(shost
);
766 scsi_eh_flush_done_q(&ha
->eh_done_q
);
768 /* check if any new eh work was scheduled during the last run */
769 spin_lock_irq(&ha
->lock
);
770 if (ha
->eh_active
== 0) {
771 shost
->host_eh_scheduled
= 0;
774 spin_unlock_irq(&ha
->lock
);
779 pr_notice("--- Exit %s: busy: %d failed: %d tries: %d\n",
780 __func__
, scsi_host_busy(shost
),
781 shost
->host_failed
, tries
);
784 int sas_ioctl(struct scsi_device
*sdev
, unsigned int cmd
, void __user
*arg
)
786 struct domain_device
*dev
= sdev_to_domain_dev(sdev
);
788 if (dev_is_sata(dev
))
789 return ata_sas_scsi_ioctl(dev
->sata_dev
.ap
, sdev
, cmd
, arg
);
794 struct domain_device
*sas_find_dev_by_rphy(struct sas_rphy
*rphy
)
796 struct Scsi_Host
*shost
= dev_to_shost(rphy
->dev
.parent
);
797 struct sas_ha_struct
*ha
= SHOST_TO_SAS_HA(shost
);
798 struct domain_device
*found_dev
= NULL
;
802 spin_lock_irqsave(&ha
->phy_port_lock
, flags
);
803 for (i
= 0; i
< ha
->num_phys
; i
++) {
804 struct asd_sas_port
*port
= ha
->sas_port
[i
];
805 struct domain_device
*dev
;
807 spin_lock(&port
->dev_list_lock
);
808 list_for_each_entry(dev
, &port
->dev_list
, dev_list_node
) {
809 if (rphy
== dev
->rphy
) {
811 spin_unlock(&port
->dev_list_lock
);
815 spin_unlock(&port
->dev_list_lock
);
818 spin_unlock_irqrestore(&ha
->phy_port_lock
, flags
);
823 int sas_target_alloc(struct scsi_target
*starget
)
825 struct sas_rphy
*rphy
= dev_to_rphy(starget
->dev
.parent
);
826 struct domain_device
*found_dev
= sas_find_dev_by_rphy(rphy
);
831 kref_get(&found_dev
->kref
);
832 starget
->hostdata
= found_dev
;
836 #define SAS_DEF_QD 256
838 int sas_slave_configure(struct scsi_device
*scsi_dev
)
840 struct domain_device
*dev
= sdev_to_domain_dev(scsi_dev
);
842 BUG_ON(dev
->rphy
->identify
.device_type
!= SAS_END_DEVICE
);
844 if (dev_is_sata(dev
)) {
845 ata_sas_slave_configure(scsi_dev
, dev
->sata_dev
.ap
);
849 sas_read_port_mode_page(scsi_dev
);
851 if (scsi_dev
->tagged_supported
) {
852 scsi_change_queue_depth(scsi_dev
, SAS_DEF_QD
);
854 pr_notice("device %llx, LUN %llx doesn't support TCQ\n",
855 SAS_ADDR(dev
->sas_addr
), scsi_dev
->lun
);
856 scsi_change_queue_depth(scsi_dev
, 1);
859 scsi_dev
->allow_restart
= 1;
864 int sas_change_queue_depth(struct scsi_device
*sdev
, int depth
)
866 struct domain_device
*dev
= sdev_to_domain_dev(sdev
);
868 if (dev_is_sata(dev
))
869 return __ata_change_queue_depth(dev
->sata_dev
.ap
, sdev
, depth
);
871 if (!sdev
->tagged_supported
)
873 return scsi_change_queue_depth(sdev
, depth
);
876 int sas_bios_param(struct scsi_device
*scsi_dev
,
877 struct block_device
*bdev
,
878 sector_t capacity
, int *hsc
)
882 sector_div(capacity
, 255*63);
889 * Tell an upper layer that it needs to initiate an abort for a given task.
890 * This should only ever be called by an LLDD.
892 void sas_task_abort(struct sas_task
*task
)
894 struct scsi_cmnd
*sc
= task
->uldd_task
;
896 /* Escape for libsas internal commands */
898 struct sas_task_slow
*slow
= task
->slow_task
;
902 if (!del_timer(&slow
->timer
))
904 slow
->timer
.function(&slow
->timer
);
908 if (dev_is_sata(task
->dev
))
909 sas_ata_task_abort(task
);
911 blk_abort_request(sc
->request
);
914 void sas_target_destroy(struct scsi_target
*starget
)
916 struct domain_device
*found_dev
= starget
->hostdata
;
921 starget
->hostdata
= NULL
;
922 sas_put_device(found_dev
);
925 #define SAS_STRING_ADDR_SIZE 16
927 int sas_request_addr(struct Scsi_Host
*shost
, u8
*addr
)
930 const struct firmware
*fw
;
932 res
= request_firmware(&fw
, "sas_addr", &shost
->shost_gendev
);
936 if (fw
->size
< SAS_STRING_ADDR_SIZE
) {
941 res
= hex2bin(addr
, fw
->data
, strnlen(fw
->data
, SAS_ADDR_SIZE
* 2) / 2);
946 release_firmware(fw
);
949 EXPORT_SYMBOL_GPL(sas_request_addr
);
951 EXPORT_SYMBOL_GPL(sas_queuecommand
);
952 EXPORT_SYMBOL_GPL(sas_target_alloc
);
953 EXPORT_SYMBOL_GPL(sas_slave_configure
);
954 EXPORT_SYMBOL_GPL(sas_change_queue_depth
);
955 EXPORT_SYMBOL_GPL(sas_bios_param
);
956 EXPORT_SYMBOL_GPL(sas_task_abort
);
957 EXPORT_SYMBOL_GPL(sas_phy_reset
);
958 EXPORT_SYMBOL_GPL(sas_eh_device_reset_handler
);
959 EXPORT_SYMBOL_GPL(sas_eh_target_reset_handler
);
960 EXPORT_SYMBOL_GPL(sas_target_destroy
);
961 EXPORT_SYMBOL_GPL(sas_ioctl
);