WIP FPC-III support
[linux/fpc-iii.git] / drivers / scsi / isci / remote_node_context.c
blob68333f523b351a8305a829a546b17826d68a0f60
1 /*
2 * This file is provided under a dual BSD/GPLv2 license. When using or
3 * redistributing this file, you may do so under either license.
5 * GPL LICENSE SUMMARY
7 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of version 2 of the GNU General Public License as
11 * published by the Free Software Foundation.
13 * This program is distributed in the hope that it will be useful, but
14 * WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * General Public License for more details.
18 * You should have received a copy of the GNU General Public License
19 * along with this program; if not, write to the Free Software
20 * Foundation, Inc., 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
21 * The full GNU General Public License is included in this distribution
22 * in the file called LICENSE.GPL.
24 * BSD LICENSE
26 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
27 * All rights reserved.
29 * Redistribution and use in source and binary forms, with or without
30 * modification, are permitted provided that the following conditions
31 * are met:
33 * * Redistributions of source code must retain the above copyright
34 * notice, this list of conditions and the following disclaimer.
35 * * Redistributions in binary form must reproduce the above copyright
36 * notice, this list of conditions and the following disclaimer in
37 * the documentation and/or other materials provided with the
38 * distribution.
39 * * Neither the name of Intel Corporation nor the names of its
40 * contributors may be used to endorse or promote products derived
41 * from this software without specific prior written permission.
43 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
44 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
45 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
46 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
47 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
48 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
49 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
50 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
51 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
52 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
53 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
55 #include <scsi/sas_ata.h>
56 #include "host.h"
57 #include "isci.h"
58 #include "remote_device.h"
59 #include "remote_node_context.h"
60 #include "scu_event_codes.h"
61 #include "scu_task_context.h"
63 #undef C
64 #define C(a) (#a)
65 const char *rnc_state_name(enum scis_sds_remote_node_context_states state)
67 static const char * const strings[] = RNC_STATES;
69 if (state >= ARRAY_SIZE(strings))
70 return "UNKNOWN";
72 return strings[state];
74 #undef C
76 /**
78 * @sci_rnc: The state of the remote node context object to check.
80 * This method will return true if the remote node context is in a READY state
81 * otherwise it will return false bool true if the remote node context is in
82 * the ready state. false if the remote node context is not in the ready state.
84 bool sci_remote_node_context_is_ready(
85 struct sci_remote_node_context *sci_rnc)
87 u32 current_state = sci_rnc->sm.current_state_id;
89 if (current_state == SCI_RNC_READY) {
90 return true;
93 return false;
96 bool sci_remote_node_context_is_suspended(struct sci_remote_node_context *sci_rnc)
98 u32 current_state = sci_rnc->sm.current_state_id;
100 if (current_state == SCI_RNC_TX_RX_SUSPENDED)
101 return true;
102 return false;
105 static union scu_remote_node_context *sci_rnc_by_id(struct isci_host *ihost, u16 id)
107 if (id < ihost->remote_node_entries &&
108 ihost->device_table[id])
109 return &ihost->remote_node_context_table[id];
111 return NULL;
114 static void sci_remote_node_context_construct_buffer(struct sci_remote_node_context *sci_rnc)
116 struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
117 struct domain_device *dev = idev->domain_dev;
118 int rni = sci_rnc->remote_node_index;
119 union scu_remote_node_context *rnc;
120 struct isci_host *ihost;
121 __le64 sas_addr;
123 ihost = idev->owning_port->owning_controller;
124 rnc = sci_rnc_by_id(ihost, rni);
126 memset(rnc, 0, sizeof(union scu_remote_node_context)
127 * sci_remote_device_node_count(idev));
129 rnc->ssp.remote_node_index = rni;
130 rnc->ssp.remote_node_port_width = idev->device_port_width;
131 rnc->ssp.logical_port_index = idev->owning_port->physical_port_index;
133 /* sas address is __be64, context ram format is __le64 */
134 sas_addr = cpu_to_le64(SAS_ADDR(dev->sas_addr));
135 rnc->ssp.remote_sas_address_hi = upper_32_bits(sas_addr);
136 rnc->ssp.remote_sas_address_lo = lower_32_bits(sas_addr);
138 rnc->ssp.nexus_loss_timer_enable = true;
139 rnc->ssp.check_bit = false;
140 rnc->ssp.is_valid = false;
141 rnc->ssp.is_remote_node_context = true;
142 rnc->ssp.function_number = 0;
144 rnc->ssp.arbitration_wait_time = 0;
146 if (dev_is_sata(dev)) {
147 rnc->ssp.connection_occupancy_timeout =
148 ihost->user_parameters.stp_max_occupancy_timeout;
149 rnc->ssp.connection_inactivity_timeout =
150 ihost->user_parameters.stp_inactivity_timeout;
151 } else {
152 rnc->ssp.connection_occupancy_timeout =
153 ihost->user_parameters.ssp_max_occupancy_timeout;
154 rnc->ssp.connection_inactivity_timeout =
155 ihost->user_parameters.ssp_inactivity_timeout;
158 rnc->ssp.initial_arbitration_wait_time = 0;
160 /* Open Address Frame Parameters */
161 rnc->ssp.oaf_connection_rate = idev->connection_rate;
162 rnc->ssp.oaf_features = 0;
163 rnc->ssp.oaf_source_zone_group = 0;
164 rnc->ssp.oaf_more_compatibility_features = 0;
168 * @sci_rnc:
169 * @callback:
170 * @callback_parameter:
172 * This method will setup the remote node context object so it will transition
173 * to its ready state. If the remote node context is already setup to
174 * transition to its final state then this function does nothing. none
176 static void sci_remote_node_context_setup_to_resume(
177 struct sci_remote_node_context *sci_rnc,
178 scics_sds_remote_node_context_callback callback,
179 void *callback_parameter,
180 enum sci_remote_node_context_destination_state dest_param)
182 if (sci_rnc->destination_state != RNC_DEST_FINAL) {
183 sci_rnc->destination_state = dest_param;
184 if (callback != NULL) {
185 sci_rnc->user_callback = callback;
186 sci_rnc->user_cookie = callback_parameter;
191 static void sci_remote_node_context_setup_to_destroy(
192 struct sci_remote_node_context *sci_rnc,
193 scics_sds_remote_node_context_callback callback,
194 void *callback_parameter)
196 struct isci_host *ihost = idev_to_ihost(rnc_to_dev(sci_rnc));
198 sci_rnc->destination_state = RNC_DEST_FINAL;
199 sci_rnc->user_callback = callback;
200 sci_rnc->user_cookie = callback_parameter;
202 wake_up(&ihost->eventq);
208 * This method just calls the user callback function and then resets the
209 * callback.
211 static void sci_remote_node_context_notify_user(
212 struct sci_remote_node_context *rnc)
214 if (rnc->user_callback != NULL) {
215 (*rnc->user_callback)(rnc->user_cookie);
217 rnc->user_callback = NULL;
218 rnc->user_cookie = NULL;
222 static void sci_remote_node_context_continue_state_transitions(struct sci_remote_node_context *rnc)
224 switch (rnc->destination_state) {
225 case RNC_DEST_READY:
226 case RNC_DEST_SUSPENDED_RESUME:
227 rnc->destination_state = RNC_DEST_READY;
228 fallthrough;
229 case RNC_DEST_FINAL:
230 sci_remote_node_context_resume(rnc, rnc->user_callback,
231 rnc->user_cookie);
232 break;
233 default:
234 rnc->destination_state = RNC_DEST_UNSPECIFIED;
235 break;
239 static void sci_remote_node_context_validate_context_buffer(struct sci_remote_node_context *sci_rnc)
241 union scu_remote_node_context *rnc_buffer;
242 struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
243 struct domain_device *dev = idev->domain_dev;
244 struct isci_host *ihost = idev->owning_port->owning_controller;
246 rnc_buffer = sci_rnc_by_id(ihost, sci_rnc->remote_node_index);
248 rnc_buffer->ssp.is_valid = true;
250 if (dev_is_sata(dev) && dev->parent) {
251 sci_remote_device_post_request(idev, SCU_CONTEXT_COMMAND_POST_RNC_96);
252 } else {
253 sci_remote_device_post_request(idev, SCU_CONTEXT_COMMAND_POST_RNC_32);
255 if (!dev->parent)
256 sci_port_setup_transports(idev->owning_port,
257 sci_rnc->remote_node_index);
261 static void sci_remote_node_context_invalidate_context_buffer(struct sci_remote_node_context *sci_rnc)
263 union scu_remote_node_context *rnc_buffer;
264 struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
265 struct isci_host *ihost = idev->owning_port->owning_controller;
267 rnc_buffer = sci_rnc_by_id(ihost, sci_rnc->remote_node_index);
269 rnc_buffer->ssp.is_valid = false;
271 sci_remote_device_post_request(rnc_to_dev(sci_rnc),
272 SCU_CONTEXT_COMMAND_POST_RNC_INVALIDATE);
275 static void sci_remote_node_context_initial_state_enter(struct sci_base_state_machine *sm)
277 struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
278 struct isci_remote_device *idev = rnc_to_dev(rnc);
279 struct isci_host *ihost = idev->owning_port->owning_controller;
281 /* Check to see if we have gotten back to the initial state because
282 * someone requested to destroy the remote node context object.
284 if (sm->previous_state_id == SCI_RNC_INVALIDATING) {
285 rnc->destination_state = RNC_DEST_UNSPECIFIED;
286 sci_remote_node_context_notify_user(rnc);
288 smp_wmb();
289 wake_up(&ihost->eventq);
293 static void sci_remote_node_context_posting_state_enter(struct sci_base_state_machine *sm)
295 struct sci_remote_node_context *sci_rnc = container_of(sm, typeof(*sci_rnc), sm);
297 sci_remote_node_context_validate_context_buffer(sci_rnc);
300 static void sci_remote_node_context_invalidating_state_enter(struct sci_base_state_machine *sm)
302 struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
304 /* Terminate all outstanding requests. */
305 sci_remote_device_terminate_requests(rnc_to_dev(rnc));
306 sci_remote_node_context_invalidate_context_buffer(rnc);
309 static void sci_remote_node_context_resuming_state_enter(struct sci_base_state_machine *sm)
311 struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
312 struct isci_remote_device *idev;
313 struct domain_device *dev;
315 idev = rnc_to_dev(rnc);
316 dev = idev->domain_dev;
319 * For direct attached SATA devices we need to clear the TLCR
320 * NCQ to TCi tag mapping on the phy and in cases where we
321 * resume because of a target reset we also need to update
322 * the STPTLDARNI register with the RNi of the device
324 if (dev_is_sata(dev) && !dev->parent)
325 sci_port_setup_transports(idev->owning_port, rnc->remote_node_index);
327 sci_remote_device_post_request(idev, SCU_CONTEXT_COMMAND_POST_RNC_RESUME);
330 static void sci_remote_node_context_ready_state_enter(struct sci_base_state_machine *sm)
332 struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
333 enum sci_remote_node_context_destination_state dest_select;
334 int tell_user = 1;
336 dest_select = rnc->destination_state;
337 rnc->destination_state = RNC_DEST_UNSPECIFIED;
339 if ((dest_select == RNC_DEST_SUSPENDED) ||
340 (dest_select == RNC_DEST_SUSPENDED_RESUME)) {
341 sci_remote_node_context_suspend(
342 rnc, rnc->suspend_reason,
343 SCI_SOFTWARE_SUSPEND_EXPECTED_EVENT);
345 if (dest_select == RNC_DEST_SUSPENDED_RESUME)
346 tell_user = 0; /* Wait until ready again. */
348 if (tell_user)
349 sci_remote_node_context_notify_user(rnc);
352 static void sci_remote_node_context_tx_suspended_state_enter(struct sci_base_state_machine *sm)
354 struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
356 sci_remote_node_context_continue_state_transitions(rnc);
359 static void sci_remote_node_context_tx_rx_suspended_state_enter(struct sci_base_state_machine *sm)
361 struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
362 struct isci_remote_device *idev = rnc_to_dev(rnc);
363 struct isci_host *ihost = idev->owning_port->owning_controller;
364 u32 new_count = rnc->suspend_count + 1;
366 if (new_count == 0)
367 rnc->suspend_count = 1;
368 else
369 rnc->suspend_count = new_count;
370 smp_wmb();
372 /* Terminate outstanding requests pending abort. */
373 sci_remote_device_abort_requests_pending_abort(idev);
375 wake_up(&ihost->eventq);
376 sci_remote_node_context_continue_state_transitions(rnc);
379 static void sci_remote_node_context_await_suspend_state_exit(
380 struct sci_base_state_machine *sm)
382 struct sci_remote_node_context *rnc
383 = container_of(sm, typeof(*rnc), sm);
384 struct isci_remote_device *idev = rnc_to_dev(rnc);
386 if (dev_is_sata(idev->domain_dev))
387 isci_dev_set_hang_detection_timeout(idev, 0);
390 static const struct sci_base_state sci_remote_node_context_state_table[] = {
391 [SCI_RNC_INITIAL] = {
392 .enter_state = sci_remote_node_context_initial_state_enter,
394 [SCI_RNC_POSTING] = {
395 .enter_state = sci_remote_node_context_posting_state_enter,
397 [SCI_RNC_INVALIDATING] = {
398 .enter_state = sci_remote_node_context_invalidating_state_enter,
400 [SCI_RNC_RESUMING] = {
401 .enter_state = sci_remote_node_context_resuming_state_enter,
403 [SCI_RNC_READY] = {
404 .enter_state = sci_remote_node_context_ready_state_enter,
406 [SCI_RNC_TX_SUSPENDED] = {
407 .enter_state = sci_remote_node_context_tx_suspended_state_enter,
409 [SCI_RNC_TX_RX_SUSPENDED] = {
410 .enter_state = sci_remote_node_context_tx_rx_suspended_state_enter,
412 [SCI_RNC_AWAIT_SUSPENSION] = {
413 .exit_state = sci_remote_node_context_await_suspend_state_exit,
417 void sci_remote_node_context_construct(struct sci_remote_node_context *rnc,
418 u16 remote_node_index)
420 memset(rnc, 0, sizeof(struct sci_remote_node_context));
422 rnc->remote_node_index = remote_node_index;
423 rnc->destination_state = RNC_DEST_UNSPECIFIED;
425 sci_init_sm(&rnc->sm, sci_remote_node_context_state_table, SCI_RNC_INITIAL);
428 enum sci_status sci_remote_node_context_event_handler(struct sci_remote_node_context *sci_rnc,
429 u32 event_code)
431 enum scis_sds_remote_node_context_states state;
432 u32 next_state;
434 state = sci_rnc->sm.current_state_id;
435 switch (state) {
436 case SCI_RNC_POSTING:
437 switch (scu_get_event_code(event_code)) {
438 case SCU_EVENT_POST_RNC_COMPLETE:
439 sci_change_state(&sci_rnc->sm, SCI_RNC_READY);
440 break;
441 default:
442 goto out;
444 break;
445 case SCI_RNC_INVALIDATING:
446 if (scu_get_event_code(event_code) == SCU_EVENT_POST_RNC_INVALIDATE_COMPLETE) {
447 if (sci_rnc->destination_state == RNC_DEST_FINAL)
448 next_state = SCI_RNC_INITIAL;
449 else
450 next_state = SCI_RNC_POSTING;
451 sci_change_state(&sci_rnc->sm, next_state);
452 } else {
453 switch (scu_get_event_type(event_code)) {
454 case SCU_EVENT_TYPE_RNC_SUSPEND_TX:
455 case SCU_EVENT_TYPE_RNC_SUSPEND_TX_RX:
456 /* We really dont care if the hardware is going to suspend
457 * the device since it's being invalidated anyway */
458 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
459 "%s: SCIC Remote Node Context 0x%p was "
460 "suspended by hardware while being "
461 "invalidated.\n", __func__, sci_rnc);
462 break;
463 default:
464 goto out;
467 break;
468 case SCI_RNC_RESUMING:
469 if (scu_get_event_code(event_code) == SCU_EVENT_POST_RCN_RELEASE) {
470 sci_change_state(&sci_rnc->sm, SCI_RNC_READY);
471 } else {
472 switch (scu_get_event_type(event_code)) {
473 case SCU_EVENT_TYPE_RNC_SUSPEND_TX:
474 case SCU_EVENT_TYPE_RNC_SUSPEND_TX_RX:
475 /* We really dont care if the hardware is going to suspend
476 * the device since it's being resumed anyway */
477 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
478 "%s: SCIC Remote Node Context 0x%p was "
479 "suspended by hardware while being resumed.\n",
480 __func__, sci_rnc);
481 break;
482 default:
483 goto out;
486 break;
487 case SCI_RNC_READY:
488 switch (scu_get_event_type(event_code)) {
489 case SCU_EVENT_TL_RNC_SUSPEND_TX:
490 sci_change_state(&sci_rnc->sm, SCI_RNC_TX_SUSPENDED);
491 sci_rnc->suspend_type = scu_get_event_type(event_code);
492 break;
493 case SCU_EVENT_TL_RNC_SUSPEND_TX_RX:
494 sci_change_state(&sci_rnc->sm, SCI_RNC_TX_RX_SUSPENDED);
495 sci_rnc->suspend_type = scu_get_event_type(event_code);
496 break;
497 default:
498 goto out;
500 break;
501 case SCI_RNC_AWAIT_SUSPENSION:
502 switch (scu_get_event_type(event_code)) {
503 case SCU_EVENT_TL_RNC_SUSPEND_TX:
504 next_state = SCI_RNC_TX_SUSPENDED;
505 break;
506 case SCU_EVENT_TL_RNC_SUSPEND_TX_RX:
507 next_state = SCI_RNC_TX_RX_SUSPENDED;
508 break;
509 default:
510 goto out;
512 if (sci_rnc->suspend_type == scu_get_event_type(event_code))
513 sci_change_state(&sci_rnc->sm, next_state);
514 break;
515 default:
516 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
517 "%s: invalid state: %s\n", __func__,
518 rnc_state_name(state));
519 return SCI_FAILURE_INVALID_STATE;
521 return SCI_SUCCESS;
523 out:
524 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
525 "%s: code: %#x state: %s\n", __func__, event_code,
526 rnc_state_name(state));
527 return SCI_FAILURE;
531 enum sci_status sci_remote_node_context_destruct(struct sci_remote_node_context *sci_rnc,
532 scics_sds_remote_node_context_callback cb_fn,
533 void *cb_p)
535 enum scis_sds_remote_node_context_states state;
537 state = sci_rnc->sm.current_state_id;
538 switch (state) {
539 case SCI_RNC_INVALIDATING:
540 sci_remote_node_context_setup_to_destroy(sci_rnc, cb_fn, cb_p);
541 return SCI_SUCCESS;
542 case SCI_RNC_POSTING:
543 case SCI_RNC_RESUMING:
544 case SCI_RNC_READY:
545 case SCI_RNC_TX_SUSPENDED:
546 case SCI_RNC_TX_RX_SUSPENDED:
547 sci_remote_node_context_setup_to_destroy(sci_rnc, cb_fn, cb_p);
548 sci_change_state(&sci_rnc->sm, SCI_RNC_INVALIDATING);
549 return SCI_SUCCESS;
550 case SCI_RNC_AWAIT_SUSPENSION:
551 sci_remote_node_context_setup_to_destroy(sci_rnc, cb_fn, cb_p);
552 return SCI_SUCCESS;
553 case SCI_RNC_INITIAL:
554 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
555 "%s: invalid state: %s\n", __func__,
556 rnc_state_name(state));
557 /* We have decided that the destruct request on the remote node context
558 * can not fail since it is either in the initial/destroyed state or is
559 * can be destroyed.
561 return SCI_SUCCESS;
562 default:
563 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
564 "%s: invalid state %s\n", __func__,
565 rnc_state_name(state));
566 return SCI_FAILURE_INVALID_STATE;
570 enum sci_status sci_remote_node_context_suspend(
571 struct sci_remote_node_context *sci_rnc,
572 enum sci_remote_node_suspension_reasons suspend_reason,
573 u32 suspend_type)
575 enum scis_sds_remote_node_context_states state
576 = sci_rnc->sm.current_state_id;
577 struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
578 enum sci_status status = SCI_FAILURE_INVALID_STATE;
579 enum sci_remote_node_context_destination_state dest_param =
580 RNC_DEST_UNSPECIFIED;
582 dev_dbg(scirdev_to_dev(idev),
583 "%s: current state %s, current suspend_type %x dest state %d,"
584 " arg suspend_reason %d, arg suspend_type %x",
585 __func__, rnc_state_name(state), sci_rnc->suspend_type,
586 sci_rnc->destination_state, suspend_reason,
587 suspend_type);
589 /* Disable automatic state continuations if explicitly suspending. */
590 if ((suspend_reason == SCI_HW_SUSPEND) ||
591 (sci_rnc->destination_state == RNC_DEST_FINAL))
592 dest_param = sci_rnc->destination_state;
594 switch (state) {
595 case SCI_RNC_READY:
596 break;
597 case SCI_RNC_INVALIDATING:
598 if (sci_rnc->destination_state == RNC_DEST_FINAL) {
599 dev_warn(scirdev_to_dev(idev),
600 "%s: already destroying %p\n",
601 __func__, sci_rnc);
602 return SCI_FAILURE_INVALID_STATE;
604 fallthrough; /* and handle like SCI_RNC_POSTING */
605 case SCI_RNC_RESUMING:
606 fallthrough; /* and handle like SCI_RNC_POSTING */
607 case SCI_RNC_POSTING:
608 /* Set the destination state to AWAIT - this signals the
609 * entry into the SCI_RNC_READY state that a suspension
610 * needs to be done immediately.
612 if (sci_rnc->destination_state != RNC_DEST_FINAL)
613 sci_rnc->destination_state = RNC_DEST_SUSPENDED;
614 sci_rnc->suspend_type = suspend_type;
615 sci_rnc->suspend_reason = suspend_reason;
616 return SCI_SUCCESS;
618 case SCI_RNC_TX_SUSPENDED:
619 if (suspend_type == SCU_EVENT_TL_RNC_SUSPEND_TX)
620 status = SCI_SUCCESS;
621 break;
622 case SCI_RNC_TX_RX_SUSPENDED:
623 if (suspend_type == SCU_EVENT_TL_RNC_SUSPEND_TX_RX)
624 status = SCI_SUCCESS;
625 break;
626 case SCI_RNC_AWAIT_SUSPENSION:
627 if ((sci_rnc->suspend_type == SCU_EVENT_TL_RNC_SUSPEND_TX_RX)
628 || (suspend_type == sci_rnc->suspend_type))
629 return SCI_SUCCESS;
630 break;
631 default:
632 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
633 "%s: invalid state %s\n", __func__,
634 rnc_state_name(state));
635 return SCI_FAILURE_INVALID_STATE;
637 sci_rnc->destination_state = dest_param;
638 sci_rnc->suspend_type = suspend_type;
639 sci_rnc->suspend_reason = suspend_reason;
641 if (status == SCI_SUCCESS) { /* Already in the destination state? */
642 struct isci_host *ihost = idev->owning_port->owning_controller;
644 wake_up_all(&ihost->eventq); /* Let observers look. */
645 return SCI_SUCCESS;
647 if ((suspend_reason == SCI_SW_SUSPEND_NORMAL) ||
648 (suspend_reason == SCI_SW_SUSPEND_LINKHANG_DETECT)) {
650 if (suspend_reason == SCI_SW_SUSPEND_LINKHANG_DETECT)
651 isci_dev_set_hang_detection_timeout(idev, 0x00000001);
653 sci_remote_device_post_request(
654 idev, SCI_SOFTWARE_SUSPEND_CMD);
656 if (state != SCI_RNC_AWAIT_SUSPENSION)
657 sci_change_state(&sci_rnc->sm, SCI_RNC_AWAIT_SUSPENSION);
659 return SCI_SUCCESS;
662 enum sci_status sci_remote_node_context_resume(struct sci_remote_node_context *sci_rnc,
663 scics_sds_remote_node_context_callback cb_fn,
664 void *cb_p)
666 enum scis_sds_remote_node_context_states state;
667 struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
669 state = sci_rnc->sm.current_state_id;
670 dev_dbg(scirdev_to_dev(idev),
671 "%s: state %s, cb_fn = %p, cb_p = %p; dest_state = %d; "
672 "dev resume path %s\n",
673 __func__, rnc_state_name(state), cb_fn, cb_p,
674 sci_rnc->destination_state,
675 test_bit(IDEV_ABORT_PATH_ACTIVE, &idev->flags)
676 ? "<abort active>" : "<normal>");
678 switch (state) {
679 case SCI_RNC_INITIAL:
680 if (sci_rnc->remote_node_index == SCIC_SDS_REMOTE_NODE_CONTEXT_INVALID_INDEX)
681 return SCI_FAILURE_INVALID_STATE;
683 sci_remote_node_context_setup_to_resume(sci_rnc, cb_fn, cb_p,
684 RNC_DEST_READY);
685 if (!test_bit(IDEV_ABORT_PATH_ACTIVE, &idev->flags)) {
686 sci_remote_node_context_construct_buffer(sci_rnc);
687 sci_change_state(&sci_rnc->sm, SCI_RNC_POSTING);
689 return SCI_SUCCESS;
691 case SCI_RNC_POSTING:
692 case SCI_RNC_INVALIDATING:
693 case SCI_RNC_RESUMING:
694 /* We are still waiting to post when a resume was
695 * requested.
697 switch (sci_rnc->destination_state) {
698 case RNC_DEST_SUSPENDED:
699 case RNC_DEST_SUSPENDED_RESUME:
700 /* Previously waiting to suspend after posting.
701 * Now continue onto resumption.
703 sci_remote_node_context_setup_to_resume(
704 sci_rnc, cb_fn, cb_p,
705 RNC_DEST_SUSPENDED_RESUME);
706 break;
707 default:
708 sci_remote_node_context_setup_to_resume(
709 sci_rnc, cb_fn, cb_p,
710 RNC_DEST_READY);
711 break;
713 return SCI_SUCCESS;
715 case SCI_RNC_TX_SUSPENDED:
716 case SCI_RNC_TX_RX_SUSPENDED:
718 struct domain_device *dev = idev->domain_dev;
719 /* If this is an expander attached SATA device we must
720 * invalidate and repost the RNC since this is the only
721 * way to clear the TCi to NCQ tag mapping table for
722 * the RNi. All other device types we can just resume.
724 sci_remote_node_context_setup_to_resume(
725 sci_rnc, cb_fn, cb_p, RNC_DEST_READY);
727 if (!test_bit(IDEV_ABORT_PATH_ACTIVE, &idev->flags)) {
728 if ((dev_is_sata(dev) && dev->parent) ||
729 (sci_rnc->destination_state == RNC_DEST_FINAL))
730 sci_change_state(&sci_rnc->sm,
731 SCI_RNC_INVALIDATING);
732 else
733 sci_change_state(&sci_rnc->sm,
734 SCI_RNC_RESUMING);
737 return SCI_SUCCESS;
739 case SCI_RNC_AWAIT_SUSPENSION:
740 sci_remote_node_context_setup_to_resume(
741 sci_rnc, cb_fn, cb_p, RNC_DEST_SUSPENDED_RESUME);
742 return SCI_SUCCESS;
743 default:
744 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
745 "%s: invalid state %s\n", __func__,
746 rnc_state_name(state));
747 return SCI_FAILURE_INVALID_STATE;
751 enum sci_status sci_remote_node_context_start_io(struct sci_remote_node_context *sci_rnc,
752 struct isci_request *ireq)
754 enum scis_sds_remote_node_context_states state;
756 state = sci_rnc->sm.current_state_id;
758 switch (state) {
759 case SCI_RNC_READY:
760 return SCI_SUCCESS;
761 case SCI_RNC_TX_SUSPENDED:
762 case SCI_RNC_TX_RX_SUSPENDED:
763 case SCI_RNC_AWAIT_SUSPENSION:
764 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
765 "%s: invalid state %s\n", __func__,
766 rnc_state_name(state));
767 return SCI_FAILURE_REMOTE_DEVICE_RESET_REQUIRED;
768 default:
769 dev_dbg(scirdev_to_dev(rnc_to_dev(sci_rnc)),
770 "%s: invalid state %s\n", __func__,
771 rnc_state_name(state));
772 return SCI_FAILURE_INVALID_STATE;
776 enum sci_status sci_remote_node_context_start_task(
777 struct sci_remote_node_context *sci_rnc,
778 struct isci_request *ireq,
779 scics_sds_remote_node_context_callback cb_fn,
780 void *cb_p)
782 enum sci_status status = sci_remote_node_context_resume(sci_rnc,
783 cb_fn, cb_p);
784 if (status != SCI_SUCCESS)
785 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
786 "%s: resume failed: %d\n", __func__, status);
787 return status;
790 int sci_remote_node_context_is_safe_to_abort(
791 struct sci_remote_node_context *sci_rnc)
793 enum scis_sds_remote_node_context_states state;
795 state = sci_rnc->sm.current_state_id;
796 switch (state) {
797 case SCI_RNC_INVALIDATING:
798 case SCI_RNC_TX_RX_SUSPENDED:
799 return 1;
800 case SCI_RNC_POSTING:
801 case SCI_RNC_RESUMING:
802 case SCI_RNC_READY:
803 case SCI_RNC_TX_SUSPENDED:
804 case SCI_RNC_AWAIT_SUSPENSION:
805 case SCI_RNC_INITIAL:
806 return 0;
807 default:
808 dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
809 "%s: invalid state %d\n", __func__, state);
810 return 0;