Linux Kernel  3.7.1
 All Data Structures Namespaces Files Functions Variables Typedefs Enumerations Enumerator Macros Groups Pages
lib.c
Go to the documentation of this file.
1 /******************************************************************************
2  *
3  * GPL LICENSE SUMMARY
4  *
5  * Copyright(c) 2008 - 2012 Intel Corporation. All rights reserved.
6  *
7  * This program is free software; you can redistribute it and/or modify
8  * it under the terms of version 2 of the GNU General Public License as
9  * published by the Free Software Foundation.
10  *
11  * This program is distributed in the hope that it will be useful, but
12  * WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14  * General Public License for more details.
15  *
16  * You should have received a copy of the GNU General Public License
17  * along with this program; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
19  * USA
20  *
21  * The full GNU General Public License is included in this distribution
22  * in the file called LICENSE.GPL.
23  *
24  * Contact Information:
25  * Intel Linux Wireless <[email protected]>
26  * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
27  *
28  *****************************************************************************/
29 #include <linux/etherdevice.h>
30 #include <linux/kernel.h>
31 #include <linux/module.h>
32 #include <linux/init.h>
33 #include <linux/sched.h>
34 #include <net/mac80211.h>
35 
36 #include "iwl-io.h"
37 #include "iwl-agn-hw.h"
38 #include "iwl-trans.h"
39 #include "iwl-modparams.h"
40 
41 #include "dev.h"
42 #include "agn.h"
43 
45 {
46  return (addr >= IWLAGN_RTC_DATA_LOWER_BOUND) &&
48 }
49 
51 {
52  struct iwlagn_tx_power_dbm_cmd tx_power_cmd;
53  u8 tx_ant_cfg_cmd;
54 
56  "TX Power requested while scanning!\n"))
57  return -EAGAIN;
58 
59  /* half dBm need to multiply */
60  tx_power_cmd.global_lmt = (s8)(2 * priv->tx_power_user_lmt);
61 
62  if (tx_power_cmd.global_lmt > priv->eeprom_data->max_tx_pwr_half_dbm) {
63  /*
64  * For the newer devices which using enhanced/extend tx power
65  * table in EEPROM, the format is in half dBm. driver need to
66  * convert to dBm format before report to mac80211.
67  * By doing so, there is a possibility of 1/2 dBm resolution
68  * lost. driver will perform "round-up" operation before
69  * reporting, but it will cause 1/2 dBm tx power over the
70  * regulatory limit. Perform the checking here, if the
71  * "tx_power_user_lmt" is higher than EEPROM value (in
72  * half-dBm format), lower the tx power based on EEPROM
73  */
74  tx_power_cmd.global_lmt =
75  priv->eeprom_data->max_tx_pwr_half_dbm;
76  }
77  tx_power_cmd.flags = IWLAGN_TX_POWER_NO_CLOSED;
78  tx_power_cmd.srv_chan_lmt = IWLAGN_TX_POWER_AUTO;
79 
80  if (IWL_UCODE_API(priv->fw->ucode_ver) == 1)
81  tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD_V1;
82  else
83  tx_ant_cfg_cmd = REPLY_TX_POWER_DBM_CMD;
84 
85  return iwl_dvm_send_cmd_pdu(priv, tx_ant_cfg_cmd, CMD_SYNC,
86  sizeof(tx_power_cmd), &tx_power_cmd);
87 }
88 
90 {
91  lockdep_assert_held(&priv->statistics.lock);
92 
93  /* store temperature from correct statistics (in Celsius) */
94  priv->temperature = le32_to_cpu(priv->statistics.common.temperature);
95  iwl_tt_handler(priv);
96 }
97 
98 int iwlagn_hwrate_to_mac80211_idx(u32 rate_n_flags, enum ieee80211_band band)
99 {
100  int idx = 0;
101  int band_offset = 0;
102 
103  /* HT rate format: mac80211 wants an MCS number, which is just LSB */
104  if (rate_n_flags & RATE_MCS_HT_MSK) {
105  idx = (rate_n_flags & 0xff);
106  return idx;
107  /* Legacy rate format, search for match in table */
108  } else {
109  if (band == IEEE80211_BAND_5GHZ)
110  band_offset = IWL_FIRST_OFDM_RATE;
111  for (idx = band_offset; idx < IWL_RATE_COUNT_LEGACY; idx++)
112  if (iwl_rates[idx].plcp == (rate_n_flags & 0xFF))
113  return idx - band_offset;
114  }
115 
116  return -1;
117 }
118 
120  struct ieee80211_vif *vif, bool add)
121 {
122  struct iwl_vif_priv *vif_priv = (void *)vif->drv_priv;
123 
124  if (add)
125  return iwlagn_add_bssid_station(priv, vif_priv->ctx,
126  vif->bss_conf.bssid,
127  &vif_priv->ibss_bssid_sta_id);
128  return iwl_remove_station(priv, vif_priv->ibss_bssid_sta_id,
129  vif->bss_conf.bssid);
130 }
131 
139 int iwlagn_txfifo_flush(struct iwl_priv *priv, u16 flush_control)
140 {
141  struct iwl_txfifo_flush_cmd flush_cmd;
142  struct iwl_host_cmd cmd = {
144  .len = { sizeof(struct iwl_txfifo_flush_cmd), },
145  .flags = CMD_SYNC,
146  .data = { &flush_cmd, },
147  };
148 
149  might_sleep();
150 
151  memset(&flush_cmd, 0, sizeof(flush_cmd));
152  if (flush_control & BIT(IWL_RXON_CTX_BSS))
156  if ((flush_control & BIT(IWL_RXON_CTX_PAN)) &&
157  (priv->valid_contexts != BIT(IWL_RXON_CTX_BSS)))
158  flush_cmd.fifo_control |= IWL_PAN_SCD_VO_MSK |
162 
163  if (priv->eeprom_data->sku & EEPROM_SKU_CAP_11N_ENABLE)
164  flush_cmd.fifo_control |= IWL_AGG_TX_QUEUE_MSK;
165 
166  IWL_DEBUG_INFO(priv, "fifo queue control: 0X%x\n",
167  flush_cmd.fifo_control);
168  flush_cmd.flush_control = cpu_to_le16(flush_control);
169 
170  return iwl_dvm_send_cmd(priv, &cmd);
171 }
172 
173 void iwlagn_dev_txfifo_flush(struct iwl_priv *priv, u16 flush_control)
174 {
175  mutex_lock(&priv->mutex);
176  ieee80211_stop_queues(priv->hw);
177  if (iwlagn_txfifo_flush(priv, IWL_DROP_ALL)) {
178  IWL_ERR(priv, "flush request fail\n");
179  goto done;
180  }
181  IWL_DEBUG_INFO(priv, "wait transmit/flush all frames\n");
182  iwl_trans_wait_tx_queue_empty(priv->trans);
183 done:
184  ieee80211_wake_queues(priv->hw);
185  mutex_unlock(&priv->mutex);
186 }
187 
188 /*
189  * BT coex
190  */
191 /* Notmal TDM */
192 static const __le32 iwlagn_def_3w_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
193  cpu_to_le32(0xaaaaaaaa),
194  cpu_to_le32(0xaaaaaaaa),
195  cpu_to_le32(0xaeaaaaaa),
196  cpu_to_le32(0xaaaaaaaa),
197  cpu_to_le32(0xcc00ff28),
198  cpu_to_le32(0x0000aaaa),
199  cpu_to_le32(0xcc00aaaa),
200  cpu_to_le32(0x0000aaaa),
201  cpu_to_le32(0xc0004000),
202  cpu_to_le32(0x00004000),
203  cpu_to_le32(0xf0005000),
204  cpu_to_le32(0xf0005000),
205 };
206 
207 
208 /* Loose Coex */
209 static const __le32 iwlagn_loose_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
210  cpu_to_le32(0xaaaaaaaa),
211  cpu_to_le32(0xaaaaaaaa),
212  cpu_to_le32(0xaeaaaaaa),
213  cpu_to_le32(0xaaaaaaaa),
214  cpu_to_le32(0xcc00ff28),
215  cpu_to_le32(0x0000aaaa),
216  cpu_to_le32(0xcc00aaaa),
217  cpu_to_le32(0x0000aaaa),
218  cpu_to_le32(0x00000000),
219  cpu_to_le32(0x00000000),
220  cpu_to_le32(0xf0005000),
221  cpu_to_le32(0xf0005000),
222 };
223 
224 /* Full concurrency */
225 static const __le32 iwlagn_concurrent_lookup[IWLAGN_BT_DECISION_LUT_SIZE] = {
226  cpu_to_le32(0xaaaaaaaa),
227  cpu_to_le32(0xaaaaaaaa),
228  cpu_to_le32(0xaaaaaaaa),
229  cpu_to_le32(0xaaaaaaaa),
230  cpu_to_le32(0xaaaaaaaa),
231  cpu_to_le32(0xaaaaaaaa),
232  cpu_to_le32(0xaaaaaaaa),
233  cpu_to_le32(0xaaaaaaaa),
234  cpu_to_le32(0x00000000),
235  cpu_to_le32(0x00000000),
236  cpu_to_le32(0x00000000),
237  cpu_to_le32(0x00000000),
238 };
239 
241 {
242  struct iwl_basic_bt_cmd basic = {
244  .bt3_timer_t7_value = IWLAGN_BT3_T7_DEFAULT,
245  .bt3_prio_sample_time = IWLAGN_BT3_PRIO_SAMPLE_DEFAULT,
246  .bt3_timer_t2_value = IWLAGN_BT3_T2_DEFAULT,
247  };
248  struct iwl_bt_cmd_v1 bt_cmd_v1;
249  struct iwl_bt_cmd_v2 bt_cmd_v2;
250  int ret;
251 
252  BUILD_BUG_ON(sizeof(iwlagn_def_3w_lookup) !=
253  sizeof(basic.bt3_lookup_table));
254 
255  if (priv->cfg->bt_params) {
256  /*
257  * newer generation of devices (2000 series and newer)
258  * use the version 2 of the bt command
259  * we need to make sure sending the host command
260  * with correct data structure to avoid uCode assert
261  */
262  if (priv->cfg->bt_params->bt_session_2) {
263  bt_cmd_v2.prio_boost = cpu_to_le32(
264  priv->cfg->bt_params->bt_prio_boost);
265  bt_cmd_v2.tx_prio_boost = 0;
266  bt_cmd_v2.rx_prio_boost = 0;
267  } else {
268  /* older version only has 8 bits */
269  WARN_ON(priv->cfg->bt_params->bt_prio_boost & ~0xFF);
270  bt_cmd_v1.prio_boost =
271  priv->cfg->bt_params->bt_prio_boost;
272  bt_cmd_v1.tx_prio_boost = 0;
273  bt_cmd_v1.rx_prio_boost = 0;
274  }
275  } else {
276  IWL_ERR(priv, "failed to construct BT Coex Config\n");
277  return;
278  }
279 
280  /*
281  * Possible situations when BT needs to take over for receive,
282  * at the same time where STA needs to response to AP's frame(s),
283  * reduce the tx power of the required response frames, by that,
284  * allow the concurrent BT receive & WiFi transmit
285  * (BT - ANT A, WiFi -ANT B), without interference to one another
286  *
287  * Reduced tx power apply to control frames only (ACK/Back/CTS)
288  * when indicated by the BT config command
289  */
290  basic.kill_ack_mask = priv->kill_ack_mask;
291  basic.kill_cts_mask = priv->kill_cts_mask;
292  if (priv->reduced_txpower)
294  basic.valid = priv->bt_valid;
295 
296  /*
297  * Configure BT coex mode to "no coexistence" when the
298  * user disabled BT coexistence, we have no interface
299  * (might be in monitor mode), or the interface is in
300  * IBSS mode (no proper uCode support for coex then).
301  */
302  if (!iwlwifi_mod_params.bt_coex_active ||
303  priv->iw_mode == NL80211_IFTYPE_ADHOC) {
305  } else {
308 
309  if (!priv->bt_enable_pspoll)
311  else
313 
314  if (priv->bt_ch_announce)
316  IWL_DEBUG_COEX(priv, "BT coex flag: 0X%x\n", basic.flags);
317  }
318  priv->bt_enable_flag = basic.flags;
319  if (priv->bt_full_concurrent)
320  memcpy(basic.bt3_lookup_table, iwlagn_concurrent_lookup,
321  sizeof(iwlagn_concurrent_lookup));
322  else
323  memcpy(basic.bt3_lookup_table, iwlagn_def_3w_lookup,
324  sizeof(iwlagn_def_3w_lookup));
325 
326  IWL_DEBUG_COEX(priv, "BT coex %s in %s mode\n",
327  basic.flags ? "active" : "disabled",
328  priv->bt_full_concurrent ?
329  "full concurrency" : "3-wire");
330 
331  if (priv->cfg->bt_params->bt_session_2) {
332  memcpy(&bt_cmd_v2.basic, &basic,
333  sizeof(basic));
335  CMD_SYNC, sizeof(bt_cmd_v2), &bt_cmd_v2);
336  } else {
337  memcpy(&bt_cmd_v1.basic, &basic,
338  sizeof(basic));
340  CMD_SYNC, sizeof(bt_cmd_v1), &bt_cmd_v1);
341  }
342  if (ret)
343  IWL_ERR(priv, "failed to send BT Coex Config\n");
344 
345 }
346 
347 void iwlagn_bt_adjust_rssi_monitor(struct iwl_priv *priv, bool rssi_ena)
348 {
349  struct iwl_rxon_context *ctx, *found_ctx = NULL;
350  bool found_ap = false;
351 
352  lockdep_assert_held(&priv->mutex);
353 
354  /* Check whether AP or GO mode is active. */
355  if (rssi_ena) {
356  for_each_context(priv, ctx) {
357  if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_AP &&
358  iwl_is_associated_ctx(ctx)) {
359  found_ap = true;
360  break;
361  }
362  }
363  }
364 
365  /*
366  * If disable was received or If GO/AP mode, disable RSSI
367  * measurements.
368  */
369  if (!rssi_ena || found_ap) {
370  if (priv->cur_rssi_ctx) {
371  ctx = priv->cur_rssi_ctx;
373  priv->cur_rssi_ctx = NULL;
374  }
375  return;
376  }
377 
378  /*
379  * If rssi measurements need to be enabled, consider all cases now.
380  * Figure out how many contexts are active.
381  */
382  for_each_context(priv, ctx) {
383  if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION &&
384  iwl_is_associated_ctx(ctx)) {
385  found_ctx = ctx;
386  break;
387  }
388  }
389 
390  /*
391  * rssi monitor already enabled for the correct interface...nothing
392  * to do.
393  */
394  if (found_ctx == priv->cur_rssi_ctx)
395  return;
396 
397  /*
398  * Figure out if rssi monitor is currently enabled, and needs
399  * to be changed. If rssi monitor is already enabled, disable
400  * it first else just enable rssi measurements on the
401  * interface found above.
402  */
403  if (priv->cur_rssi_ctx) {
404  ctx = priv->cur_rssi_ctx;
405  if (ctx->vif)
407  }
408 
409  priv->cur_rssi_ctx = found_ctx;
410 
411  if (!found_ctx)
412  return;
413 
417 }
418 
419 static bool iwlagn_bt_traffic_is_sco(struct iwl_bt_uart_msg *uart_msg)
420 {
421  return BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3 >>
423 }
424 
425 static void iwlagn_bt_traffic_change_work(struct work_struct *work)
426 {
427  struct iwl_priv *priv =
429  struct iwl_rxon_context *ctx;
430  int smps_request = -1;
431 
433  /* bt coex disabled */
434  return;
435  }
436 
437  /*
438  * Note: bt_traffic_load can be overridden by scan complete and
439  * coex profile notifications. Ignore that since only bad consequence
440  * can be not matching debug print with actual state.
441  */
442  IWL_DEBUG_COEX(priv, "BT traffic load changes: %d\n",
443  priv->bt_traffic_load);
444 
445  switch (priv->bt_traffic_load) {
447  if (priv->bt_status)
448  smps_request = IEEE80211_SMPS_DYNAMIC;
449  else
450  smps_request = IEEE80211_SMPS_AUTOMATIC;
451  break;
453  smps_request = IEEE80211_SMPS_DYNAMIC;
454  break;
457  smps_request = IEEE80211_SMPS_STATIC;
458  break;
459  default:
460  IWL_ERR(priv, "Invalid BT traffic load: %d\n",
461  priv->bt_traffic_load);
462  break;
463  }
464 
465  mutex_lock(&priv->mutex);
466 
467  /*
468  * We can not send command to firmware while scanning. When the scan
469  * complete we will schedule this work again. We do check with mutex
470  * locked to prevent new scan request to arrive. We do not check
471  * STATUS_SCANNING to avoid race when queue_work two times from
472  * different notifications, but quit and not perform any work at all.
473  */
474  if (test_bit(STATUS_SCAN_HW, &priv->status))
475  goto out;
476 
478 
479  if (smps_request != -1) {
480  priv->current_ht_config.smps = smps_request;
481  for_each_context(priv, ctx) {
482  if (ctx->vif && ctx->vif->type == NL80211_IFTYPE_STATION)
483  ieee80211_request_smps(ctx->vif, smps_request);
484  }
485  }
486 
487  /*
488  * Dynamic PS poll related functionality. Adjust RSSI measurements if
489  * necessary.
490  */
492 out:
493  mutex_unlock(&priv->mutex);
494 }
495 
496 /*
497  * If BT sco traffic, and RSSI monitor is enabled, move measurements to the
498  * correct interface or disable it if this is the last interface to be
499  * removed.
500  */
502 {
503  if (priv->bt_is_sco &&
505  iwlagn_bt_adjust_rssi_monitor(priv, true);
506  else
507  iwlagn_bt_adjust_rssi_monitor(priv, false);
508 }
509 
510 static void iwlagn_print_uartmsg(struct iwl_priv *priv,
511  struct iwl_bt_uart_msg *uart_msg)
512 {
513  IWL_DEBUG_COEX(priv, "Message Type = 0x%X, SSN = 0x%X, "
514  "Update Req = 0x%X\n",
515  (BT_UART_MSG_FRAME1MSGTYPE_MSK & uart_msg->frame1) >>
517  (BT_UART_MSG_FRAME1SSN_MSK & uart_msg->frame1) >>
519  (BT_UART_MSG_FRAME1UPDATEREQ_MSK & uart_msg->frame1) >>
521 
522  IWL_DEBUG_COEX(priv, "Open connections = 0x%X, Traffic load = 0x%X, "
523  "Chl_SeqN = 0x%X, In band = 0x%X\n",
528  (BT_UART_MSG_FRAME2CHLSEQN_MSK & uart_msg->frame2) >>
530  (BT_UART_MSG_FRAME2INBAND_MSK & uart_msg->frame2) >>
532 
533  IWL_DEBUG_COEX(priv, "SCO/eSCO = 0x%X, Sniff = 0x%X, A2DP = 0x%X, "
534  "ACL = 0x%X, Master = 0x%X, OBEX = 0x%X\n",
535  (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3) >>
537  (BT_UART_MSG_FRAME3SNIFF_MSK & uart_msg->frame3) >>
539  (BT_UART_MSG_FRAME3A2DP_MSK & uart_msg->frame3) >>
541  (BT_UART_MSG_FRAME3ACL_MSK & uart_msg->frame3) >>
543  (BT_UART_MSG_FRAME3MASTER_MSK & uart_msg->frame3) >>
545  (BT_UART_MSG_FRAME3OBEX_MSK & uart_msg->frame3) >>
547 
548  IWL_DEBUG_COEX(priv, "Idle duration = 0x%X\n",
551 
552  IWL_DEBUG_COEX(priv, "Tx Activity = 0x%X, Rx Activity = 0x%X, "
553  "eSCO Retransmissions = 0x%X\n",
560 
561  IWL_DEBUG_COEX(priv, "Sniff Interval = 0x%X, Discoverable = 0x%X\n",
566 
567  IWL_DEBUG_COEX(priv, "Sniff Activity = 0x%X, Page = "
568  "0x%X, Inquiry = 0x%X, Connectable = 0x%X\n",
571  (BT_UART_MSG_FRAME7PAGE_MSK & uart_msg->frame7) >>
573  (BT_UART_MSG_FRAME7INQUIRY_MSK & uart_msg->frame7) >>
577 }
578 
579 static bool iwlagn_set_kill_msk(struct iwl_priv *priv,
580  struct iwl_bt_uart_msg *uart_msg)
581 {
582  bool need_update = false;
583  u8 kill_msk = IWL_BT_KILL_REDUCE;
584  static const __le32 bt_kill_ack_msg[3] = {
588  static const __le32 bt_kill_cts_msg[3] = {
592 
593  if (!priv->reduced_txpower)
594  kill_msk = (BT_UART_MSG_FRAME3SCOESCO_MSK & uart_msg->frame3)
596  if (priv->kill_ack_mask != bt_kill_ack_msg[kill_msk] ||
597  priv->kill_cts_mask != bt_kill_cts_msg[kill_msk]) {
599  priv->kill_ack_mask = bt_kill_ack_msg[kill_msk];
601  priv->kill_cts_mask = bt_kill_cts_msg[kill_msk];
602  need_update = true;
603  }
604  return need_update;
605 }
606 
607 /*
608  * Upon RSSI changes, sends a bt config command with following changes
609  * 1. enable/disable "reduced control frames tx power
610  * 2. update the "kill)ack_mask" and "kill_cts_mask"
611  *
612  * If "reduced tx power" is enabled, uCode shall
613  * 1. ACK/Back/CTS rate shall reduced to 6Mbps
614  * 2. not use duplciate 20/40MHz mode
615  */
616 static bool iwlagn_fill_txpower_mode(struct iwl_priv *priv,
617  struct iwl_bt_uart_msg *uart_msg)
618 {
619  bool need_update = false;
620  struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
621  int ave_rssi;
622 
623  if (!ctx->vif || (ctx->vif->type != NL80211_IFTYPE_STATION)) {
624  IWL_DEBUG_INFO(priv, "BSS ctx not active or not in sta mode\n");
625  return false;
626  }
627 
628  ave_rssi = ieee80211_ave_rssi(ctx->vif);
629  if (!ave_rssi) {
630  /* no rssi data, no changes to reduce tx power */
631  IWL_DEBUG_COEX(priv, "no rssi data available\n");
632  return need_update;
633  }
634  if (!priv->reduced_txpower &&
635  !iwl_is_associated(priv, IWL_RXON_CTX_PAN) &&
637  (uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
639  !(uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
641  /* enabling reduced tx power */
642  priv->reduced_txpower = true;
644  need_update = true;
645  } else if (priv->reduced_txpower &&
646  (iwl_is_associated(priv, IWL_RXON_CTX_PAN) ||
648  (uart_msg->frame3 & (BT_UART_MSG_FRAME3SCOESCO_MSK |
650  !(uart_msg->frame3 & (BT_UART_MSG_FRAME3ACL_MSK |
652  /* disable reduced tx power */
653  priv->reduced_txpower = false;
655  need_update = true;
656  }
657 
658  return need_update;
659 }
660 
662  struct iwl_rx_cmd_buffer *rxb,
663  struct iwl_device_cmd *cmd)
664 {
665  struct iwl_rx_packet *pkt = rxb_addr(rxb);
666  struct iwl_bt_coex_profile_notif *coex = (void *)pkt->data;
667  struct iwl_bt_uart_msg *uart_msg = &coex->last_bt_uart_msg;
668 
670  /* bt coex disabled */
671  return 0;
672  }
673 
674  IWL_DEBUG_COEX(priv, "BT Coex notification:\n");
675  IWL_DEBUG_COEX(priv, " status: %d\n", coex->bt_status);
676  IWL_DEBUG_COEX(priv, " traffic load: %d\n", coex->bt_traffic_load);
677  IWL_DEBUG_COEX(priv, " CI compliance: %d\n",
678  coex->bt_ci_compliance);
679  iwlagn_print_uartmsg(priv, uart_msg);
680 
682  priv->bt_is_sco = iwlagn_bt_traffic_is_sco(uart_msg);
683 
684  if (priv->iw_mode != NL80211_IFTYPE_ADHOC) {
685  if (priv->bt_status != coex->bt_status ||
686  priv->last_bt_traffic_load != coex->bt_traffic_load) {
687  if (coex->bt_status) {
688  /* BT on */
689  if (!priv->bt_ch_announce)
690  priv->bt_traffic_load =
692  else
693  priv->bt_traffic_load =
694  coex->bt_traffic_load;
695  } else {
696  /* BT off */
697  priv->bt_traffic_load =
699  }
700  priv->bt_status = coex->bt_status;
701  queue_work(priv->workqueue,
702  &priv->bt_traffic_change_work);
703  }
704  }
705 
706  /* schedule to send runtime bt_config */
707  /* check reduce power before change ack/cts kill mask */
708  if (iwlagn_fill_txpower_mode(priv, uart_msg) ||
709  iwlagn_set_kill_msk(priv, uart_msg))
710  queue_work(priv->workqueue, &priv->bt_runtime_config);
711 
712 
713  /* FIXME: based on notification, adjust the prio_boost */
714 
715  priv->bt_ci_compliance = coex->bt_ci_compliance;
716  return 0;
717 }
718 
720 {
723 }
724 
726 {
728  iwlagn_bt_traffic_change_work);
729 }
730 
732 {
734 }
735 
736 static bool is_single_rx_stream(struct iwl_priv *priv)
737 {
738  return priv->current_ht_config.smps == IEEE80211_SMPS_STATIC ||
739  priv->current_ht_config.single_chain_sufficient;
740 }
741 
742 #define IWL_NUM_RX_CHAINS_MULTIPLE 3
743 #define IWL_NUM_RX_CHAINS_SINGLE 2
744 #define IWL_NUM_IDLE_CHAINS_DUAL 2
745 #define IWL_NUM_IDLE_CHAINS_SINGLE 1
746 
747 /*
748  * Determine how many receiver/antenna chains to use.
749  *
750  * More provides better reception via diversity. Fewer saves power
751  * at the expense of throughput, but only when not in powersave to
752  * start with.
753  *
754  * MIMO (dual stream) requires at least 2, but works better with 3.
755  * This does not determine *which* chains to use, just how many.
756  */
757 static int iwl_get_active_rx_chain_count(struct iwl_priv *priv)
758 {
759  if (priv->cfg->bt_params &&
760  priv->cfg->bt_params->advanced_bt_coexist &&
761  (priv->bt_full_concurrent ||
763  /*
764  * only use chain 'A' in bt high traffic load or
765  * full concurrency mode
766  */
768  }
769  /* # of Rx chains to use when expecting MIMO. */
770  if (is_single_rx_stream(priv))
772  else
774 }
775 
776 /*
777  * When we are in power saving mode, unless device support spatial
778  * multiplexing power save, use the active count for rx chain count.
779  */
780 static int iwl_get_idle_rx_chain_count(struct iwl_priv *priv, int active_cnt)
781 {
782  /* # Rx chains when idling, depending on SMPS mode */
783  switch (priv->current_ht_config.smps) {
788  case IEEE80211_SMPS_OFF:
789  return active_cnt;
790  default:
791  WARN(1, "invalid SMPS mode %d",
792  priv->current_ht_config.smps);
793  return active_cnt;
794  }
795 }
796 
797 /* up to 4 chains */
798 static u8 iwl_count_chain_bitmap(u32 chain_bitmap)
799 {
800  u8 res;
801  res = (chain_bitmap & BIT(0)) >> 0;
802  res += (chain_bitmap & BIT(1)) >> 1;
803  res += (chain_bitmap & BIT(2)) >> 2;
804  res += (chain_bitmap & BIT(3)) >> 3;
805  return res;
806 }
807 
814 void iwlagn_set_rxon_chain(struct iwl_priv *priv, struct iwl_rxon_context *ctx)
815 {
816  bool is_single = is_single_rx_stream(priv);
817  bool is_cam = !test_bit(STATUS_POWER_PMI, &priv->status);
818  u8 idle_rx_cnt, active_rx_cnt, valid_rx_cnt;
819  u32 active_chains;
820  u16 rx_chain;
821 
822  /* Tell uCode which antennas are actually connected.
823  * Before first association, we assume all antennas are connected.
824  * Just after first association, iwl_chain_noise_calibration()
825  * checks which antennas actually *are* connected. */
826  if (priv->chain_noise_data.active_chains)
827  active_chains = priv->chain_noise_data.active_chains;
828  else
829  active_chains = priv->eeprom_data->valid_rx_ant;
830 
831  if (priv->cfg->bt_params &&
832  priv->cfg->bt_params->advanced_bt_coexist &&
833  (priv->bt_full_concurrent ||
835  /*
836  * only use chain 'A' in bt high traffic load or
837  * full concurrency mode
838  */
839  active_chains = first_antenna(active_chains);
840  }
841 
842  rx_chain = active_chains << RXON_RX_CHAIN_VALID_POS;
843 
844  /* How many receivers should we use? */
845  active_rx_cnt = iwl_get_active_rx_chain_count(priv);
846  idle_rx_cnt = iwl_get_idle_rx_chain_count(priv, active_rx_cnt);
847 
848 
849  /* correct rx chain count according hw settings
850  * and chain noise calibration
851  */
852  valid_rx_cnt = iwl_count_chain_bitmap(active_chains);
853  if (valid_rx_cnt < active_rx_cnt)
854  active_rx_cnt = valid_rx_cnt;
855 
856  if (valid_rx_cnt < idle_rx_cnt)
857  idle_rx_cnt = valid_rx_cnt;
858 
859  rx_chain |= active_rx_cnt << RXON_RX_CHAIN_MIMO_CNT_POS;
860  rx_chain |= idle_rx_cnt << RXON_RX_CHAIN_CNT_POS;
861 
862  ctx->staging.rx_chain = cpu_to_le16(rx_chain);
863 
864  if (!is_single && (active_rx_cnt >= IWL_NUM_RX_CHAINS_SINGLE) && is_cam)
865  ctx->staging.rx_chain |= RXON_RX_CHAIN_MIMO_FORCE_MSK;
866  else
867  ctx->staging.rx_chain &= ~RXON_RX_CHAIN_MIMO_FORCE_MSK;
868 
869  IWL_DEBUG_ASSOC(priv, "rx_chain=0x%X active=%d idle=%d\n",
870  ctx->staging.rx_chain,
871  active_rx_cnt, idle_rx_cnt);
872 
873  WARN_ON(active_rx_cnt == 0 || idle_rx_cnt == 0 ||
874  active_rx_cnt < idle_rx_cnt);
875 }
876 
877 u8 iwl_toggle_tx_ant(struct iwl_priv *priv, u8 ant, u8 valid)
878 {
879  int i;
880  u8 ind = ant;
881 
882  if (priv->band == IEEE80211_BAND_2GHZ &&
884  return 0;
885 
886  for (i = 0; i < RATE_ANT_NUM - 1; i++) {
887  ind = (ind + 1) < RATE_ANT_NUM ? ind + 1 : 0;
888  if (valid & BIT(ind))
889  return ind;
890  }
891  return ant;
892 }
893 
894 #ifdef CONFIG_PM_SLEEP
895 static void iwlagn_convert_p1k(u16 *p1k, __le16 *out)
896 {
897  int i;
898 
899  for (i = 0; i < IWLAGN_P1K_SIZE; i++)
900  out[i] = cpu_to_le16(p1k[i]);
901 }
902 
903 struct wowlan_key_data {
904  struct iwl_rxon_context *ctx;
905  struct iwlagn_wowlan_rsc_tsc_params_cmd *rsc_tsc;
906  struct iwlagn_wowlan_tkip_params_cmd *tkip;
907  const u8 *bssid;
908  bool error, use_rsc_tsc, use_tkip;
909 };
910 
911 
912 static void iwlagn_wowlan_program_keys(struct ieee80211_hw *hw,
913  struct ieee80211_vif *vif,
914  struct ieee80211_sta *sta,
915  struct ieee80211_key_conf *key,
916  void *_data)
917 {
918  struct iwl_priv *priv = IWL_MAC80211_GET_DVM(hw);
919  struct wowlan_key_data *data = _data;
920  struct iwl_rxon_context *ctx = data->ctx;
921  struct aes_sc *aes_sc, *aes_tx_sc = NULL;
922  struct tkip_sc *tkip_sc, *tkip_tx_sc = NULL;
923  struct iwlagn_p1k_cache *rx_p1ks;
924  u8 *rx_mic_key;
925  struct ieee80211_key_seq seq;
926  u32 cur_rx_iv32 = 0;
927  u16 p1k[IWLAGN_P1K_SIZE];
928  int ret, i;
929 
930  mutex_lock(&priv->mutex);
931 
932  if ((key->cipher == WLAN_CIPHER_SUITE_WEP40 ||
933  key->cipher == WLAN_CIPHER_SUITE_WEP104) &&
934  !sta && !ctx->key_mapping_keys)
935  ret = iwl_set_default_wep_key(priv, ctx, key);
936  else
937  ret = iwl_set_dynamic_key(priv, ctx, key, sta);
938 
939  if (ret) {
940  IWL_ERR(priv, "Error setting key during suspend!\n");
941  data->error = true;
942  }
943 
944  switch (key->cipher) {
946  if (sta) {
947  tkip_sc = data->rsc_tsc->all_tsc_rsc.tkip.unicast_rsc;
948  tkip_tx_sc = &data->rsc_tsc->all_tsc_rsc.tkip.tsc;
949 
950  rx_p1ks = data->tkip->rx_uni;
951 
952  ieee80211_get_key_tx_seq(key, &seq);
953  tkip_tx_sc->iv16 = cpu_to_le16(seq.tkip.iv16);
954  tkip_tx_sc->iv32 = cpu_to_le32(seq.tkip.iv32);
955 
956  ieee80211_get_tkip_p1k_iv(key, seq.tkip.iv32, p1k);
957  iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
958 
959  memcpy(data->tkip->mic_keys.tx,
962 
963  rx_mic_key = data->tkip->mic_keys.rx_unicast;
964  } else {
965  tkip_sc =
966  data->rsc_tsc->all_tsc_rsc.tkip.multicast_rsc;
967  rx_p1ks = data->tkip->rx_multi;
968  rx_mic_key = data->tkip->mic_keys.rx_mcast;
969  }
970 
971  /*
972  * For non-QoS this relies on the fact that both the uCode and
973  * mac80211 use TID 0 (as they need to to avoid replay attacks)
974  * for checking the IV in the frames.
975  */
976  for (i = 0; i < IWLAGN_NUM_RSC; i++) {
977  ieee80211_get_key_rx_seq(key, i, &seq);
978  tkip_sc[i].iv16 = cpu_to_le16(seq.tkip.iv16);
979  tkip_sc[i].iv32 = cpu_to_le32(seq.tkip.iv32);
980  /* wrapping isn't allowed, AP must rekey */
981  if (seq.tkip.iv32 > cur_rx_iv32)
982  cur_rx_iv32 = seq.tkip.iv32;
983  }
984 
985  ieee80211_get_tkip_rx_p1k(key, data->bssid, cur_rx_iv32, p1k);
986  iwlagn_convert_p1k(p1k, rx_p1ks[0].p1k);
987  ieee80211_get_tkip_rx_p1k(key, data->bssid,
988  cur_rx_iv32 + 1, p1k);
989  iwlagn_convert_p1k(p1k, rx_p1ks[1].p1k);
990 
991  memcpy(rx_mic_key,
994 
995  data->use_tkip = true;
996  data->use_rsc_tsc = true;
997  break;
999  if (sta) {
1000  u8 *pn = seq.ccmp.pn;
1001 
1002  aes_sc = data->rsc_tsc->all_tsc_rsc.aes.unicast_rsc;
1003  aes_tx_sc = &data->rsc_tsc->all_tsc_rsc.aes.tsc;
1004 
1005  ieee80211_get_key_tx_seq(key, &seq);
1006  aes_tx_sc->pn = cpu_to_le64(
1007  (u64)pn[5] |
1008  ((u64)pn[4] << 8) |
1009  ((u64)pn[3] << 16) |
1010  ((u64)pn[2] << 24) |
1011  ((u64)pn[1] << 32) |
1012  ((u64)pn[0] << 40));
1013  } else
1014  aes_sc = data->rsc_tsc->all_tsc_rsc.aes.multicast_rsc;
1015 
1016  /*
1017  * For non-QoS this relies on the fact that both the uCode and
1018  * mac80211 use TID 0 for checking the IV in the frames.
1019  */
1020  for (i = 0; i < IWLAGN_NUM_RSC; i++) {
1021  u8 *pn = seq.ccmp.pn;
1022 
1023  ieee80211_get_key_rx_seq(key, i, &seq);
1024  aes_sc->pn = cpu_to_le64(
1025  (u64)pn[5] |
1026  ((u64)pn[4] << 8) |
1027  ((u64)pn[3] << 16) |
1028  ((u64)pn[2] << 24) |
1029  ((u64)pn[1] << 32) |
1030  ((u64)pn[0] << 40));
1031  }
1032  data->use_rsc_tsc = true;
1033  break;
1034  }
1035 
1036  mutex_unlock(&priv->mutex);
1037 }
1038 
1039 int iwlagn_send_patterns(struct iwl_priv *priv,
1040  struct cfg80211_wowlan *wowlan)
1041 {
1042  struct iwlagn_wowlan_patterns_cmd *pattern_cmd;
1043  struct iwl_host_cmd cmd = {
1045  .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1046  .flags = CMD_SYNC,
1047  };
1048  int i, err;
1049 
1050  if (!wowlan->n_patterns)
1051  return 0;
1052 
1053  cmd.len[0] = sizeof(*pattern_cmd) +
1054  wowlan->n_patterns * sizeof(struct iwlagn_wowlan_pattern);
1055 
1056  pattern_cmd = kmalloc(cmd.len[0], GFP_KERNEL);
1057  if (!pattern_cmd)
1058  return -ENOMEM;
1059 
1060  pattern_cmd->n_patterns = cpu_to_le32(wowlan->n_patterns);
1061 
1062  for (i = 0; i < wowlan->n_patterns; i++) {
1063  int mask_len = DIV_ROUND_UP(wowlan->patterns[i].pattern_len, 8);
1064 
1065  memcpy(&pattern_cmd->patterns[i].mask,
1066  wowlan->patterns[i].mask, mask_len);
1067  memcpy(&pattern_cmd->patterns[i].pattern,
1068  wowlan->patterns[i].pattern,
1069  wowlan->patterns[i].pattern_len);
1070  pattern_cmd->patterns[i].mask_size = mask_len;
1071  pattern_cmd->patterns[i].pattern_size =
1072  wowlan->patterns[i].pattern_len;
1073  }
1074 
1075  cmd.data[0] = pattern_cmd;
1076  err = iwl_dvm_send_cmd(priv, &cmd);
1077  kfree(pattern_cmd);
1078  return err;
1079 }
1080 
1081 int iwlagn_suspend(struct iwl_priv *priv, struct cfg80211_wowlan *wowlan)
1082 {
1083  struct iwlagn_wowlan_wakeup_filter_cmd wakeup_filter_cmd;
1084  struct iwl_rxon_cmd rxon;
1085  struct iwl_rxon_context *ctx = &priv->contexts[IWL_RXON_CTX_BSS];
1086  struct iwlagn_wowlan_kek_kck_material_cmd kek_kck_cmd;
1087  struct iwlagn_wowlan_tkip_params_cmd tkip_cmd = {};
1088  struct iwlagn_d3_config_cmd d3_cfg_cmd = {};
1089  struct wowlan_key_data key_data = {
1090  .ctx = ctx,
1091  .bssid = ctx->active.bssid_addr,
1092  .use_rsc_tsc = false,
1093  .tkip = &tkip_cmd,
1094  .use_tkip = false,
1095  };
1096  int ret, i;
1097  u16 seq;
1098 
1099  key_data.rsc_tsc = kzalloc(sizeof(*key_data.rsc_tsc), GFP_KERNEL);
1100  if (!key_data.rsc_tsc)
1101  return -ENOMEM;
1102 
1103  memset(&wakeup_filter_cmd, 0, sizeof(wakeup_filter_cmd));
1104 
1105  /*
1106  * We know the last used seqno, and the uCode expects to know that
1107  * one, it will increment before TX.
1108  */
1110  wakeup_filter_cmd.non_qos_seq = cpu_to_le16(seq);
1111 
1112  /*
1113  * For QoS counters, we store the one to use next, so subtract 0x10
1114  * since the uCode will add 0x10 before using the value.
1115  */
1116  for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1117  seq = priv->tid_data[IWL_AP_ID][i].seq_number;
1118  seq -= 0x10;
1119  wakeup_filter_cmd.qos_seq[i] = cpu_to_le16(seq);
1120  }
1121 
1122  if (wowlan->disconnect)
1123  wakeup_filter_cmd.enabled |=
1126  if (wowlan->magic_pkt)
1127  wakeup_filter_cmd.enabled |=
1129  if (wowlan->gtk_rekey_failure)
1130  wakeup_filter_cmd.enabled |=
1132  if (wowlan->eap_identity_req)
1133  wakeup_filter_cmd.enabled |=
1135  if (wowlan->four_way_handshake)
1136  wakeup_filter_cmd.enabled |=
1138  if (wowlan->n_patterns)
1139  wakeup_filter_cmd.enabled |=
1141 
1142  if (wowlan->rfkill_release)
1143  d3_cfg_cmd.wakeup_flags |=
1145 
1146  iwl_scan_cancel_timeout(priv, 200);
1147 
1148  memcpy(&rxon, &ctx->active, sizeof(rxon));
1149 
1150  priv->ucode_loaded = false;
1151  iwl_trans_stop_device(priv->trans);
1152 
1153  priv->wowlan = true;
1154 
1156  if (ret)
1157  goto out;
1158 
1159  /* now configure WoWLAN ucode */
1160  ret = iwl_alive_start(priv);
1161  if (ret)
1162  goto out;
1163 
1164  memcpy(&ctx->staging, &rxon, sizeof(rxon));
1165  ret = iwlagn_commit_rxon(priv, ctx);
1166  if (ret)
1167  goto out;
1168 
1169  ret = iwl_power_update_mode(priv, true);
1170  if (ret)
1171  goto out;
1172 
1173  if (!iwlwifi_mod_params.sw_crypto) {
1174  /* mark all keys clear */
1175  priv->ucode_key_table = 0;
1176  ctx->key_mapping_keys = 0;
1177 
1178  /*
1179  * This needs to be unlocked due to lock ordering
1180  * constraints. Since we're in the suspend path
1181  * that isn't really a problem though.
1182  */
1183  mutex_unlock(&priv->mutex);
1184  ieee80211_iter_keys(priv->hw, ctx->vif,
1185  iwlagn_wowlan_program_keys,
1186  &key_data);
1187  mutex_lock(&priv->mutex);
1188  if (key_data.error) {
1189  ret = -EIO;
1190  goto out;
1191  }
1192 
1193  if (key_data.use_rsc_tsc) {
1194  struct iwl_host_cmd rsc_tsc_cmd = {
1196  .flags = CMD_SYNC,
1197  .data[0] = key_data.rsc_tsc,
1198  .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
1199  .len[0] = sizeof(*key_data.rsc_tsc),
1200  };
1201 
1202  ret = iwl_dvm_send_cmd(priv, &rsc_tsc_cmd);
1203  if (ret)
1204  goto out;
1205  }
1206 
1207  if (key_data.use_tkip) {
1208  ret = iwl_dvm_send_cmd_pdu(priv,
1210  CMD_SYNC, sizeof(tkip_cmd),
1211  &tkip_cmd);
1212  if (ret)
1213  goto out;
1214  }
1215 
1216  if (priv->have_rekey_data) {
1217  memset(&kek_kck_cmd, 0, sizeof(kek_kck_cmd));
1218  memcpy(kek_kck_cmd.kck, priv->kck, NL80211_KCK_LEN);
1219  kek_kck_cmd.kck_len = cpu_to_le16(NL80211_KCK_LEN);
1220  memcpy(kek_kck_cmd.kek, priv->kek, NL80211_KEK_LEN);
1221  kek_kck_cmd.kek_len = cpu_to_le16(NL80211_KEK_LEN);
1222  kek_kck_cmd.replay_ctr = priv->replay_ctr;
1223 
1224  ret = iwl_dvm_send_cmd_pdu(priv,
1226  CMD_SYNC, sizeof(kek_kck_cmd),
1227  &kek_kck_cmd);
1228  if (ret)
1229  goto out;
1230  }
1231  }
1232 
1234  sizeof(d3_cfg_cmd), &d3_cfg_cmd);
1235  if (ret)
1236  goto out;
1237 
1239  CMD_SYNC, sizeof(wakeup_filter_cmd),
1240  &wakeup_filter_cmd);
1241  if (ret)
1242  goto out;
1243 
1244  ret = iwlagn_send_patterns(priv, wowlan);
1245  out:
1246  kfree(key_data.rsc_tsc);
1247  return ret;
1248 }
1249 #endif
1250 
1251 int iwl_dvm_send_cmd(struct iwl_priv *priv, struct iwl_host_cmd *cmd)
1252 {
1253  if (iwl_is_rfkill(priv) || iwl_is_ctkill(priv)) {
1254  IWL_WARN(priv, "Not sending command - %s KILL\n",
1255  iwl_is_rfkill(priv) ? "RF" : "CT");
1256  return -EIO;
1257  }
1258 
1259  if (test_bit(STATUS_FW_ERROR, &priv->status)) {
1260  IWL_ERR(priv, "Command %s failed: FW Error\n",
1261  iwl_dvm_get_cmd_string(cmd->id));
1262  return -EIO;
1263  }
1264 
1265  /*
1266  * Synchronous commands from this op-mode must hold
1267  * the mutex, this ensures we don't try to send two
1268  * (or more) synchronous commands at a time.
1269  */
1270  if (!(cmd->flags & CMD_ASYNC))
1271  lockdep_assert_held(&priv->mutex);
1272 
1273  if (priv->ucode_owner == IWL_OWNERSHIP_TM &&
1274  !(cmd->flags & CMD_ON_DEMAND)) {
1275  IWL_DEBUG_HC(priv, "tm own the uCode, no regular hcmd send\n");
1276  return -EIO;
1277  }
1278 
1279  return iwl_trans_send_cmd(priv->trans, cmd);
1280 }
1281 
1282 int iwl_dvm_send_cmd_pdu(struct iwl_priv *priv, u8 id,
1283  u32 flags, u16 len, const void *data)
1284 {
1285  struct iwl_host_cmd cmd = {
1286  .id = id,
1287  .len = { len, },
1288  .data = { data, },
1289  .flags = flags,
1290  };
1291 
1292  return iwl_dvm_send_cmd(priv, &cmd);
1293 }