Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
remote_node_context.c
Go to the documentation of this file.
1 /*
2  * This file is provided under a dual BSD/GPLv2 license. When using or
3  * redistributing this file, you may do so under either license.
4  *
5  * GPL LICENSE SUMMARY
6  *
7  * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
8  *
9  * This program is free software; you can redistribute it and/or modify
10  * it under the terms of version 2 of the GNU General Public License as
11  * published by the Free Software Foundation.
12  *
13  * This program is distributed in the hope that it will be useful, but
14  * WITHOUT ANY WARRANTY; without even the implied warranty of
15  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16  * General Public License for more details.
17  *
18  * You should have received a copy of the GNU General Public License
19  * along with this program; if not, write to the Free Software
20  * Foundation, Inc., 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
21  * The full GNU General Public License is included in this distribution
22  * in the file called LICENSE.GPL.
23  *
24  * BSD LICENSE
25  *
26  * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved.
27  * All rights reserved.
28  *
29  * Redistribution and use in source and binary forms, with or without
30  * modification, are permitted provided that the following conditions
31  * are met:
32  *
33  * * Redistributions of source code must retain the above copyright
34  * notice, this list of conditions and the following disclaimer.
35  * * Redistributions in binary form must reproduce the above copyright
36  * notice, this list of conditions and the following disclaimer in
37  * the documentation and/or other materials provided with the
38  * distribution.
39  * * Neither the name of Intel Corporation nor the names of its
40  * contributors may be used to endorse or promote products derived
41  * from this software without specific prior written permission.
42  *
43  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
44  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
45  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
46  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
47  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
48  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
49  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
50  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
51  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
52  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
53  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
54  */
55 #include <scsi/sas_ata.h>
56 #include "host.h"
57 #include "isci.h"
58 #include "remote_device.h"
59 #include "remote_node_context.h"
60 #include "scu_event_codes.h"
61 #include "scu_task_context.h"
62 
63 #undef C
64 #define C(a) (#a)
65 const char *rnc_state_name(enum scis_sds_remote_node_context_states state)
66 {
67  static const char * const strings[] = RNC_STATES;
68 
69  return strings[state];
70 }
71 #undef C
72 
82  struct sci_remote_node_context *sci_rnc)
83 {
84  u32 current_state = sci_rnc->sm.current_state_id;
85 
86  if (current_state == SCI_RNC_READY) {
87  return true;
88  }
89 
90  return false;
91 }
92 
94 {
95  u32 current_state = sci_rnc->sm.current_state_id;
96 
97  if (current_state == SCI_RNC_TX_RX_SUSPENDED)
98  return true;
99  return false;
100 }
101 
102 static union scu_remote_node_context *sci_rnc_by_id(struct isci_host *ihost, u16 id)
103 {
104  if (id < ihost->remote_node_entries &&
105  ihost->device_table[id])
106  return &ihost->remote_node_context_table[id];
107 
108  return NULL;
109 }
110 
111 static void sci_remote_node_context_construct_buffer(struct sci_remote_node_context *sci_rnc)
112 {
113  struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
114  struct domain_device *dev = idev->domain_dev;
115  int rni = sci_rnc->remote_node_index;
116  union scu_remote_node_context *rnc;
117  struct isci_host *ihost;
119 
120  ihost = idev->owning_port->owning_controller;
121  rnc = sci_rnc_by_id(ihost, rni);
122 
123  memset(rnc, 0, sizeof(union scu_remote_node_context)
124  * sci_remote_device_node_count(idev));
125 
126  rnc->ssp.remote_node_index = rni;
127  rnc->ssp.remote_node_port_width = idev->device_port_width;
128  rnc->ssp.logical_port_index = idev->owning_port->physical_port_index;
129 
130  /* sas address is __be64, context ram format is __le64 */
131  sas_addr = cpu_to_le64(SAS_ADDR(dev->sas_addr));
132  rnc->ssp.remote_sas_address_hi = upper_32_bits(sas_addr);
133  rnc->ssp.remote_sas_address_lo = lower_32_bits(sas_addr);
134 
135  rnc->ssp.nexus_loss_timer_enable = true;
136  rnc->ssp.check_bit = false;
137  rnc->ssp.is_valid = false;
138  rnc->ssp.is_remote_node_context = true;
139  rnc->ssp.function_number = 0;
140 
141  rnc->ssp.arbitration_wait_time = 0;
142 
143  if (dev_is_sata(dev)) {
144  rnc->ssp.connection_occupancy_timeout =
145  ihost->user_parameters.stp_max_occupancy_timeout;
146  rnc->ssp.connection_inactivity_timeout =
147  ihost->user_parameters.stp_inactivity_timeout;
148  } else {
149  rnc->ssp.connection_occupancy_timeout =
150  ihost->user_parameters.ssp_max_occupancy_timeout;
151  rnc->ssp.connection_inactivity_timeout =
152  ihost->user_parameters.ssp_inactivity_timeout;
153  }
154 
155  rnc->ssp.initial_arbitration_wait_time = 0;
156 
157  /* Open Address Frame Parameters */
158  rnc->ssp.oaf_connection_rate = idev->connection_rate;
159  rnc->ssp.oaf_features = 0;
160  rnc->ssp.oaf_source_zone_group = 0;
161  rnc->ssp.oaf_more_compatibility_features = 0;
162 }
173 static void sci_remote_node_context_setup_to_resume(
174  struct sci_remote_node_context *sci_rnc,
176  void *callback_parameter,
178 {
179  if (sci_rnc->destination_state != RNC_DEST_FINAL) {
180  sci_rnc->destination_state = dest_param;
181  if (callback != NULL) {
182  sci_rnc->user_callback = callback;
183  sci_rnc->user_cookie = callback_parameter;
184  }
185  }
186 }
187 
188 static void sci_remote_node_context_setup_to_destroy(
189  struct sci_remote_node_context *sci_rnc,
191  void *callback_parameter)
192 {
193  struct isci_host *ihost = idev_to_ihost(rnc_to_dev(sci_rnc));
194 
196  sci_rnc->user_callback = callback;
197  sci_rnc->user_cookie = callback_parameter;
198 
199  wake_up(&ihost->eventq);
200 }
201 
208 static void sci_remote_node_context_notify_user(
209  struct sci_remote_node_context *rnc)
210 {
211  if (rnc->user_callback != NULL) {
212  (*rnc->user_callback)(rnc->user_cookie);
213 
214  rnc->user_callback = NULL;
215  rnc->user_cookie = NULL;
216  }
217 }
218 
219 static void sci_remote_node_context_continue_state_transitions(struct sci_remote_node_context *rnc)
220 {
221  switch (rnc->destination_state) {
222  case RNC_DEST_READY:
225  /* Fall through... */
226  case RNC_DEST_FINAL:
228  rnc->user_cookie);
229  break;
230  default:
232  break;
233  }
234 }
235 
236 static void sci_remote_node_context_validate_context_buffer(struct sci_remote_node_context *sci_rnc)
237 {
238  union scu_remote_node_context *rnc_buffer;
239  struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
240  struct domain_device *dev = idev->domain_dev;
241  struct isci_host *ihost = idev->owning_port->owning_controller;
242 
243  rnc_buffer = sci_rnc_by_id(ihost, sci_rnc->remote_node_index);
244 
245  rnc_buffer->ssp.is_valid = true;
246 
247  if (dev_is_sata(dev) && dev->parent) {
249  } else {
251 
252  if (!dev->parent)
254  sci_rnc->remote_node_index);
255  }
256 }
257 
258 static void sci_remote_node_context_invalidate_context_buffer(struct sci_remote_node_context *sci_rnc)
259 {
260  union scu_remote_node_context *rnc_buffer;
261  struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
262  struct isci_host *ihost = idev->owning_port->owning_controller;
263 
264  rnc_buffer = sci_rnc_by_id(ihost, sci_rnc->remote_node_index);
265 
266  rnc_buffer->ssp.is_valid = false;
267 
268  sci_remote_device_post_request(rnc_to_dev(sci_rnc),
270 }
271 
272 static void sci_remote_node_context_initial_state_enter(struct sci_base_state_machine *sm)
273 {
274  struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
275  struct isci_remote_device *idev = rnc_to_dev(rnc);
276  struct isci_host *ihost = idev->owning_port->owning_controller;
277 
278  /* Check to see if we have gotten back to the initial state because
279  * someone requested to destroy the remote node context object.
280  */
281  if (sm->previous_state_id == SCI_RNC_INVALIDATING) {
283  sci_remote_node_context_notify_user(rnc);
284 
285  smp_wmb();
286  wake_up(&ihost->eventq);
287  }
288 }
289 
290 static void sci_remote_node_context_posting_state_enter(struct sci_base_state_machine *sm)
291 {
292  struct sci_remote_node_context *sci_rnc = container_of(sm, typeof(*sci_rnc), sm);
293 
294  sci_remote_node_context_validate_context_buffer(sci_rnc);
295 }
296 
297 static void sci_remote_node_context_invalidating_state_enter(struct sci_base_state_machine *sm)
298 {
299  struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
300 
301  /* Terminate all outstanding requests. */
302  sci_remote_device_terminate_requests(rnc_to_dev(rnc));
303  sci_remote_node_context_invalidate_context_buffer(rnc);
304 }
305 
306 static void sci_remote_node_context_resuming_state_enter(struct sci_base_state_machine *sm)
307 {
308  struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
309  struct isci_remote_device *idev;
310  struct domain_device *dev;
311 
312  idev = rnc_to_dev(rnc);
313  dev = idev->domain_dev;
314 
315  /*
316  * For direct attached SATA devices we need to clear the TLCR
317  * NCQ to TCi tag mapping on the phy and in cases where we
318  * resume because of a target reset we also need to update
319  * the STPTLDARNI register with the RNi of the device
320  */
321  if (dev_is_sata(dev) && !dev->parent)
323 
325 }
326 
327 static void sci_remote_node_context_ready_state_enter(struct sci_base_state_machine *sm)
328 {
329  struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
331  int tell_user = 1;
332 
333  dest_select = rnc->destination_state;
335 
336  if ((dest_select == RNC_DEST_SUSPENDED) ||
337  (dest_select == RNC_DEST_SUSPENDED_RESUME)) {
339  rnc, rnc->suspend_reason,
341 
342  if (dest_select == RNC_DEST_SUSPENDED_RESUME)
343  tell_user = 0; /* Wait until ready again. */
344  }
345  if (tell_user)
346  sci_remote_node_context_notify_user(rnc);
347 }
348 
349 static void sci_remote_node_context_tx_suspended_state_enter(struct sci_base_state_machine *sm)
350 {
351  struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
352 
353  sci_remote_node_context_continue_state_transitions(rnc);
354 }
355 
356 static void sci_remote_node_context_tx_rx_suspended_state_enter(struct sci_base_state_machine *sm)
357 {
358  struct sci_remote_node_context *rnc = container_of(sm, typeof(*rnc), sm);
359  struct isci_remote_device *idev = rnc_to_dev(rnc);
360  struct isci_host *ihost = idev->owning_port->owning_controller;
361  u32 new_count = rnc->suspend_count + 1;
362 
363  if (new_count == 0)
364  rnc->suspend_count = 1;
365  else
366  rnc->suspend_count = new_count;
367  smp_wmb();
368 
369  /* Terminate outstanding requests pending abort. */
371 
372  wake_up(&ihost->eventq);
373  sci_remote_node_context_continue_state_transitions(rnc);
374 }
375 
376 static void sci_remote_node_context_await_suspend_state_exit(
377  struct sci_base_state_machine *sm)
378 {
379  struct sci_remote_node_context *rnc
380  = container_of(sm, typeof(*rnc), sm);
381  struct isci_remote_device *idev = rnc_to_dev(rnc);
382 
383  if (dev_is_sata(idev->domain_dev))
385 }
386 
387 static const struct sci_base_state sci_remote_node_context_state_table[] = {
388  [SCI_RNC_INITIAL] = {
389  .enter_state = sci_remote_node_context_initial_state_enter,
390  },
391  [SCI_RNC_POSTING] = {
392  .enter_state = sci_remote_node_context_posting_state_enter,
393  },
394  [SCI_RNC_INVALIDATING] = {
395  .enter_state = sci_remote_node_context_invalidating_state_enter,
396  },
397  [SCI_RNC_RESUMING] = {
398  .enter_state = sci_remote_node_context_resuming_state_enter,
399  },
400  [SCI_RNC_READY] = {
401  .enter_state = sci_remote_node_context_ready_state_enter,
402  },
403  [SCI_RNC_TX_SUSPENDED] = {
404  .enter_state = sci_remote_node_context_tx_suspended_state_enter,
405  },
406  [SCI_RNC_TX_RX_SUSPENDED] = {
407  .enter_state = sci_remote_node_context_tx_rx_suspended_state_enter,
408  },
409  [SCI_RNC_AWAIT_SUSPENSION] = {
410  .exit_state = sci_remote_node_context_await_suspend_state_exit,
411  },
412 };
413 
415  u16 remote_node_index)
416 {
417  memset(rnc, 0, sizeof(struct sci_remote_node_context));
418 
419  rnc->remote_node_index = remote_node_index;
421 
422  sci_init_sm(&rnc->sm, sci_remote_node_context_state_table, SCI_RNC_INITIAL);
423 }
424 
426  u32 event_code)
427 {
428  enum scis_sds_remote_node_context_states state;
429  u32 next_state;
430 
431  state = sci_rnc->sm.current_state_id;
432  switch (state) {
433  case SCI_RNC_POSTING:
434  switch (scu_get_event_code(event_code)) {
436  sci_change_state(&sci_rnc->sm, SCI_RNC_READY);
437  break;
438  default:
439  goto out;
440  }
441  break;
442  case SCI_RNC_INVALIDATING:
444  if (sci_rnc->destination_state == RNC_DEST_FINAL)
445  next_state = SCI_RNC_INITIAL;
446  else
447  next_state = SCI_RNC_POSTING;
448  sci_change_state(&sci_rnc->sm, next_state);
449  } else {
450  switch (scu_get_event_type(event_code)) {
453  /* We really dont care if the hardware is going to suspend
454  * the device since it's being invalidated anyway */
455  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
456  "%s: SCIC Remote Node Context 0x%p was "
457  "suspeneded by hardware while being "
458  "invalidated.\n", __func__, sci_rnc);
459  break;
460  default:
461  goto out;
462  }
463  }
464  break;
465  case SCI_RNC_RESUMING:
466  if (scu_get_event_code(event_code) == SCU_EVENT_POST_RCN_RELEASE) {
467  sci_change_state(&sci_rnc->sm, SCI_RNC_READY);
468  } else {
469  switch (scu_get_event_type(event_code)) {
472  /* We really dont care if the hardware is going to suspend
473  * the device since it's being resumed anyway */
474  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
475  "%s: SCIC Remote Node Context 0x%p was "
476  "suspeneded by hardware while being resumed.\n",
477  __func__, sci_rnc);
478  break;
479  default:
480  goto out;
481  }
482  }
483  break;
484  case SCI_RNC_READY:
485  switch (scu_get_event_type(event_code)) {
487  sci_change_state(&sci_rnc->sm, SCI_RNC_TX_SUSPENDED);
488  sci_rnc->suspend_type = scu_get_event_type(event_code);
489  break;
491  sci_change_state(&sci_rnc->sm, SCI_RNC_TX_RX_SUSPENDED);
492  sci_rnc->suspend_type = scu_get_event_type(event_code);
493  break;
494  default:
495  goto out;
496  }
497  break;
498  case SCI_RNC_AWAIT_SUSPENSION:
499  switch (scu_get_event_type(event_code)) {
501  next_state = SCI_RNC_TX_SUSPENDED;
502  break;
504  next_state = SCI_RNC_TX_RX_SUSPENDED;
505  break;
506  default:
507  goto out;
508  }
509  if (sci_rnc->suspend_type == scu_get_event_type(event_code))
510  sci_change_state(&sci_rnc->sm, next_state);
511  break;
512  default:
513  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
514  "%s: invalid state: %s\n", __func__,
515  rnc_state_name(state));
517  }
518  return SCI_SUCCESS;
519 
520  out:
521  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
522  "%s: code: %#x state: %s\n", __func__, event_code,
523  rnc_state_name(state));
524  return SCI_FAILURE;
525 
526 }
527 
530  void *cb_p)
531 {
532  enum scis_sds_remote_node_context_states state;
533 
534  state = sci_rnc->sm.current_state_id;
535  switch (state) {
536  case SCI_RNC_INVALIDATING:
537  sci_remote_node_context_setup_to_destroy(sci_rnc, cb_fn, cb_p);
538  return SCI_SUCCESS;
539  case SCI_RNC_POSTING:
540  case SCI_RNC_RESUMING:
541  case SCI_RNC_READY:
542  case SCI_RNC_TX_SUSPENDED:
543  case SCI_RNC_TX_RX_SUSPENDED:
544  sci_remote_node_context_setup_to_destroy(sci_rnc, cb_fn, cb_p);
545  sci_change_state(&sci_rnc->sm, SCI_RNC_INVALIDATING);
546  return SCI_SUCCESS;
547  case SCI_RNC_AWAIT_SUSPENSION:
548  sci_remote_node_context_setup_to_destroy(sci_rnc, cb_fn, cb_p);
549  return SCI_SUCCESS;
550  case SCI_RNC_INITIAL:
551  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
552  "%s: invalid state: %s\n", __func__,
553  rnc_state_name(state));
554  /* We have decided that the destruct request on the remote node context
555  * can not fail since it is either in the initial/destroyed state or is
556  * can be destroyed.
557  */
558  return SCI_SUCCESS;
559  default:
560  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
561  "%s: invalid state %s\n", __func__,
562  rnc_state_name(state));
564  }
565 }
566 
568  struct sci_remote_node_context *sci_rnc,
569  enum sci_remote_node_suspension_reasons suspend_reason,
570  u32 suspend_type)
571 {
572  enum scis_sds_remote_node_context_states state
573  = sci_rnc->sm.current_state_id;
574  struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
578 
579  dev_dbg(scirdev_to_dev(idev),
580  "%s: current state %s, current suspend_type %x dest state %d,"
581  " arg suspend_reason %d, arg suspend_type %x",
582  __func__, rnc_state_name(state), sci_rnc->suspend_type,
583  sci_rnc->destination_state, suspend_reason,
584  suspend_type);
585 
586  /* Disable automatic state continuations if explicitly suspending. */
587  if ((suspend_reason == SCI_HW_SUSPEND) ||
588  (sci_rnc->destination_state == RNC_DEST_FINAL))
589  dest_param = sci_rnc->destination_state;
590 
591  switch (state) {
592  case SCI_RNC_READY:
593  break;
594  case SCI_RNC_INVALIDATING:
595  if (sci_rnc->destination_state == RNC_DEST_FINAL) {
596  dev_warn(scirdev_to_dev(idev),
597  "%s: already destroying %p\n",
598  __func__, sci_rnc);
600  }
601  /* Fall through and handle like SCI_RNC_POSTING */
602  case SCI_RNC_RESUMING:
603  /* Fall through and handle like SCI_RNC_POSTING */
604  case SCI_RNC_POSTING:
605  /* Set the destination state to AWAIT - this signals the
606  * entry into the SCI_RNC_READY state that a suspension
607  * needs to be done immediately.
608  */
609  if (sci_rnc->destination_state != RNC_DEST_FINAL)
611  sci_rnc->suspend_type = suspend_type;
612  sci_rnc->suspend_reason = suspend_reason;
613  return SCI_SUCCESS;
614 
615  case SCI_RNC_TX_SUSPENDED:
616  if (suspend_type == SCU_EVENT_TL_RNC_SUSPEND_TX)
617  status = SCI_SUCCESS;
618  break;
619  case SCI_RNC_TX_RX_SUSPENDED:
620  if (suspend_type == SCU_EVENT_TL_RNC_SUSPEND_TX_RX)
621  status = SCI_SUCCESS;
622  break;
623  case SCI_RNC_AWAIT_SUSPENSION:
625  || (suspend_type == sci_rnc->suspend_type))
626  return SCI_SUCCESS;
627  break;
628  default:
629  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
630  "%s: invalid state %s\n", __func__,
631  rnc_state_name(state));
633  }
634  sci_rnc->destination_state = dest_param;
635  sci_rnc->suspend_type = suspend_type;
636  sci_rnc->suspend_reason = suspend_reason;
637 
638  if (status == SCI_SUCCESS) { /* Already in the destination state? */
639  struct isci_host *ihost = idev->owning_port->owning_controller;
640 
641  wake_up_all(&ihost->eventq); /* Let observers look. */
642  return SCI_SUCCESS;
643  }
644  if ((suspend_reason == SCI_SW_SUSPEND_NORMAL) ||
645  (suspend_reason == SCI_SW_SUSPEND_LINKHANG_DETECT)) {
646 
647  if (suspend_reason == SCI_SW_SUSPEND_LINKHANG_DETECT)
648  isci_dev_set_hang_detection_timeout(idev, 0x00000001);
649 
652  }
653  if (state != SCI_RNC_AWAIT_SUSPENSION)
654  sci_change_state(&sci_rnc->sm, SCI_RNC_AWAIT_SUSPENSION);
655 
656  return SCI_SUCCESS;
657 }
658 
661  void *cb_p)
662 {
663  enum scis_sds_remote_node_context_states state;
664  struct isci_remote_device *idev = rnc_to_dev(sci_rnc);
665 
666  state = sci_rnc->sm.current_state_id;
667  dev_dbg(scirdev_to_dev(idev),
668  "%s: state %s, cb_fn = %p, cb_p = %p; dest_state = %d; "
669  "dev resume path %s\n",
670  __func__, rnc_state_name(state), cb_fn, cb_p,
671  sci_rnc->destination_state,
673  ? "<abort active>" : "<normal>");
674 
675  switch (state) {
676  case SCI_RNC_INITIAL:
679 
680  sci_remote_node_context_setup_to_resume(sci_rnc, cb_fn, cb_p,
682  if (!test_bit(IDEV_ABORT_PATH_ACTIVE, &idev->flags)) {
683  sci_remote_node_context_construct_buffer(sci_rnc);
684  sci_change_state(&sci_rnc->sm, SCI_RNC_POSTING);
685  }
686  return SCI_SUCCESS;
687 
688  case SCI_RNC_POSTING:
689  case SCI_RNC_INVALIDATING:
690  case SCI_RNC_RESUMING:
691  /* We are still waiting to post when a resume was
692  * requested.
693  */
694  switch (sci_rnc->destination_state) {
695  case RNC_DEST_SUSPENDED:
697  /* Previously waiting to suspend after posting.
698  * Now continue onto resumption.
699  */
700  sci_remote_node_context_setup_to_resume(
701  sci_rnc, cb_fn, cb_p,
703  break;
704  default:
705  sci_remote_node_context_setup_to_resume(
706  sci_rnc, cb_fn, cb_p,
708  break;
709  }
710  return SCI_SUCCESS;
711 
712  case SCI_RNC_TX_SUSPENDED:
713  case SCI_RNC_TX_RX_SUSPENDED:
714  {
715  struct domain_device *dev = idev->domain_dev;
716  /* If this is an expander attached SATA device we must
717  * invalidate and repost the RNC since this is the only
718  * way to clear the TCi to NCQ tag mapping table for
719  * the RNi. All other device types we can just resume.
720  */
721  sci_remote_node_context_setup_to_resume(
722  sci_rnc, cb_fn, cb_p, RNC_DEST_READY);
723 
724  if (!test_bit(IDEV_ABORT_PATH_ACTIVE, &idev->flags)) {
725  if ((dev_is_sata(dev) && dev->parent) ||
726  (sci_rnc->destination_state == RNC_DEST_FINAL))
727  sci_change_state(&sci_rnc->sm,
728  SCI_RNC_INVALIDATING);
729  else
730  sci_change_state(&sci_rnc->sm,
731  SCI_RNC_RESUMING);
732  }
733  }
734  return SCI_SUCCESS;
735 
736  case SCI_RNC_AWAIT_SUSPENSION:
737  sci_remote_node_context_setup_to_resume(
738  sci_rnc, cb_fn, cb_p, RNC_DEST_SUSPENDED_RESUME);
739  return SCI_SUCCESS;
740  default:
741  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
742  "%s: invalid state %s\n", __func__,
743  rnc_state_name(state));
745  }
746 }
747 
749  struct isci_request *ireq)
750 {
751  enum scis_sds_remote_node_context_states state;
752 
753  state = sci_rnc->sm.current_state_id;
754 
755  switch (state) {
756  case SCI_RNC_READY:
757  return SCI_SUCCESS;
758  case SCI_RNC_TX_SUSPENDED:
759  case SCI_RNC_TX_RX_SUSPENDED:
760  case SCI_RNC_AWAIT_SUSPENSION:
761  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
762  "%s: invalid state %s\n", __func__,
763  rnc_state_name(state));
765  default:
766  dev_dbg(scirdev_to_dev(rnc_to_dev(sci_rnc)),
767  "%s: invalid state %s\n", __func__,
768  rnc_state_name(state));
770  }
771 }
772 
774  struct sci_remote_node_context *sci_rnc,
775  struct isci_request *ireq,
777  void *cb_p)
778 {
780  cb_fn, cb_p);
781  if (status != SCI_SUCCESS)
782  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
783  "%s: resume failed: %d\n", __func__, status);
784  return status;
785 }
786 
788  struct sci_remote_node_context *sci_rnc)
789 {
790  enum scis_sds_remote_node_context_states state;
791 
792  state = sci_rnc->sm.current_state_id;
793  switch (state) {
794  case SCI_RNC_INVALIDATING:
795  case SCI_RNC_TX_RX_SUSPENDED:
796  return 1;
797  case SCI_RNC_POSTING:
798  case SCI_RNC_RESUMING:
799  case SCI_RNC_READY:
800  case SCI_RNC_TX_SUSPENDED:
801  case SCI_RNC_AWAIT_SUSPENSION:
802  case SCI_RNC_INITIAL:
803  return 0;
804  default:
805  dev_warn(scirdev_to_dev(rnc_to_dev(sci_rnc)),
806  "%s: invalid state %d\n", __func__, state);
807  return 0;
808  }
809 }