1 /******************************************************************************
3 * Copyright(c) 2003 - 2007 Intel Corporation. All rights reserved.
5 * Portions of this file are derived from the ipw3945 project, as well
6 * as portions of the ieee80211 subsystem header files.
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms of version 2 of the GNU General Public License as
10 * published by the Free Software Foundation.
12 * This program is distributed in the hope that it will be useful, but WITHOUT
13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
17 * You should have received a copy of the GNU General Public License along with
18 * this program; if not, write to the Free Software Foundation, Inc.,
19 * 51 Franklin Street, Fifth Floor, Boston, MA 02110, USA
21 * The full GNU General Public License is included in this distribution in the
22 * file called LICENSE.
24 * Contact Information:
25 * James P. Ketrenos <ipw2100-admin@linux.intel.com>
26 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
28 *****************************************************************************/
31 * NOTE: This file (iwl-base.c) is used to build to multiple hardware targets
32 * by defining IWL to either 3945 or 4965. The Makefile used when building
33 * the base targets will create base-3945.o and base-4965.o
35 * The eventual goal is to move as many of the #if IWL / #endif blocks out of
36 * this file and into the hardware specific implementation files (iwl-XXXX.c)
37 * and leave only the common (non #ifdef sprinkled) code in this file
40 #include <linux/kernel.h>
41 #include <linux/module.h>
42 #include <linux/version.h>
43 #include <linux/init.h>
44 #include <linux/pci.h>
45 #include <linux/dma-mapping.h>
46 #include <linux/delay.h>
47 #include <linux/skbuff.h>
48 #include <linux/netdevice.h>
49 #include <linux/wireless.h>
50 #include <linux/firmware.h>
51 #include <linux/etherdevice.h>
52 #include <linux/if_arp.h>
54 #include <net/ieee80211_radiotap.h>
55 #include <net/mac80211.h>
57 #include <asm/div64.h>
60 #include "iwl-helpers.h"
62 #ifdef CONFIG_IWL4965_DEBUG
66 static int iwl_tx_queue_update_write_ptr(struct iwl_priv
*priv
,
67 struct iwl_tx_queue
*txq
);
69 /******************************************************************************
73 ******************************************************************************/
75 /* module parameters */
76 static int iwl_param_disable_hw_scan
;
77 static int iwl_param_debug
;
78 static int iwl_param_disable
; /* def: enable radio */
79 static int iwl_param_antenna
; /* def: 0 = both antennas (use diversity) */
80 int iwl_param_hwcrypto
; /* def: using software encryption */
81 static int iwl_param_qos_enable
= 1;
82 int iwl_param_queues_num
= IWL_MAX_NUM_QUEUES
;
85 * module name, copyright, version, etc.
86 * NOTE: DRV_NAME is defined in iwlwifi.h for use by iwl-debug.h and printk
89 #define DRV_DESCRIPTION "Intel(R) Wireless WiFi Link 4965AGN driver for Linux"
91 #ifdef CONFIG_IWL4965_DEBUG
97 #ifdef CONFIG_IWL4965_SPECTRUM_MEASUREMENT
103 #define IWLWIFI_VERSION "1.1.19k" VD VS
104 #define DRV_COPYRIGHT "Copyright(c) 2003-2007 Intel Corporation"
105 #define DRV_VERSION IWLWIFI_VERSION
107 /* Change firmware file name, using "-" and incrementing number,
108 * *only* when uCode interface or architecture changes so that it
109 * is not compatible with earlier drivers.
110 * This number will also appear in << 8 position of 1st dword of uCode file */
111 #define IWL4965_UCODE_API "-1"
113 MODULE_DESCRIPTION(DRV_DESCRIPTION
);
114 MODULE_VERSION(DRV_VERSION
);
115 MODULE_AUTHOR(DRV_COPYRIGHT
);
116 MODULE_LICENSE("GPL");
118 __le16
*ieee80211_get_qos_ctrl(struct ieee80211_hdr
*hdr
)
120 u16 fc
= le16_to_cpu(hdr
->frame_control
);
121 int hdr_len
= ieee80211_get_hdrlen(fc
);
123 if ((fc
& 0x00cc) == (IEEE80211_STYPE_QOS_DATA
| IEEE80211_FTYPE_DATA
))
124 return (__le16
*) ((u8
*) hdr
+ hdr_len
- QOS_CONTROL_LEN
);
128 static const struct ieee80211_hw_mode
*iwl_get_hw_mode(
129 struct iwl_priv
*priv
, int mode
)
133 for (i
= 0; i
< 3; i
++)
134 if (priv
->modes
[i
].mode
== mode
)
135 return &priv
->modes
[i
];
140 static int iwl_is_empty_essid(const char *essid
, int essid_len
)
142 /* Single white space is for Linksys APs */
143 if (essid_len
== 1 && essid
[0] == ' ')
146 /* Otherwise, if the entire essid is 0, we assume it is hidden */
149 if (essid
[essid_len
] != '\0')
156 static const char *iwl_escape_essid(const char *essid
, u8 essid_len
)
158 static char escaped
[IW_ESSID_MAX_SIZE
* 2 + 1];
159 const char *s
= essid
;
162 if (iwl_is_empty_essid(essid
, essid_len
)) {
163 memcpy(escaped
, "<hidden>", sizeof("<hidden>"));
167 essid_len
= min(essid_len
, (u8
) IW_ESSID_MAX_SIZE
);
168 while (essid_len
--) {
180 static void iwl_print_hex_dump(int level
, void *p
, u32 len
)
182 #ifdef CONFIG_IWL4965_DEBUG
183 if (!(iwl_debug_level
& level
))
186 print_hex_dump(KERN_DEBUG
, "iwl data: ", DUMP_PREFIX_OFFSET
, 16, 1,
191 /*************** DMA-QUEUE-GENERAL-FUNCTIONS *****
194 * Theory of operation
196 * A queue is a circular buffers with 'Read' and 'Write' pointers.
197 * 2 empty entries always kept in the buffer to protect from overflow.
199 * For Tx queue, there are low mark and high mark limits. If, after queuing
200 * the packet for Tx, free space become < low mark, Tx queue stopped. When
201 * reclaiming packets (on 'tx done IRQ), if free space become > high mark,
204 * The IWL operates with six queues, one receive queue in the device's
205 * sram, one transmit queue for sending commands to the device firmware,
206 * and four transmit queues for data.
207 ***************************************************/
209 static int iwl_queue_space(const struct iwl_queue
*q
)
211 int s
= q
->read_ptr
- q
->write_ptr
;
213 if (q
->read_ptr
> q
->write_ptr
)
218 /* keep some reserve to not confuse empty and full situations */
225 /* XXX: n_bd must be power-of-two size */
226 static inline int iwl_queue_inc_wrap(int index
, int n_bd
)
228 return ++index
& (n_bd
- 1);
231 /* XXX: n_bd must be power-of-two size */
232 static inline int iwl_queue_dec_wrap(int index
, int n_bd
)
234 return --index
& (n_bd
- 1);
237 static inline int x2_queue_used(const struct iwl_queue
*q
, int i
)
239 return q
->write_ptr
> q
->read_ptr
?
240 (i
>= q
->read_ptr
&& i
< q
->write_ptr
) :
241 !(i
< q
->read_ptr
&& i
>= q
->write_ptr
);
244 static inline u8
get_cmd_index(struct iwl_queue
*q
, u32 index
, int is_huge
)
249 return index
& (q
->n_window
- 1);
252 static int iwl_queue_init(struct iwl_priv
*priv
, struct iwl_queue
*q
,
253 int count
, int slots_num
, u32 id
)
256 q
->n_window
= slots_num
;
259 /* count must be power-of-two size, otherwise iwl_queue_inc_wrap
260 * and iwl_queue_dec_wrap are broken. */
261 BUG_ON(!is_power_of_2(count
));
263 /* slots_num must be power-of-two size, otherwise
264 * get_cmd_index is broken. */
265 BUG_ON(!is_power_of_2(slots_num
));
267 q
->low_mark
= q
->n_window
/ 4;
271 q
->high_mark
= q
->n_window
/ 8;
272 if (q
->high_mark
< 2)
275 q
->write_ptr
= q
->read_ptr
= 0;
280 static int iwl_tx_queue_alloc(struct iwl_priv
*priv
,
281 struct iwl_tx_queue
*txq
, u32 id
)
283 struct pci_dev
*dev
= priv
->pci_dev
;
285 if (id
!= IWL_CMD_QUEUE_NUM
) {
286 txq
->txb
= kmalloc(sizeof(txq
->txb
[0]) *
287 TFD_QUEUE_SIZE_MAX
, GFP_KERNEL
);
289 IWL_ERROR("kmalloc for auxiliary BD "
290 "structures failed\n");
296 txq
->bd
= pci_alloc_consistent(dev
,
297 sizeof(txq
->bd
[0]) * TFD_QUEUE_SIZE_MAX
,
301 IWL_ERROR("pci_alloc_consistent(%zd) failed\n",
302 sizeof(txq
->bd
[0]) * TFD_QUEUE_SIZE_MAX
);
318 int iwl_tx_queue_init(struct iwl_priv
*priv
,
319 struct iwl_tx_queue
*txq
, int slots_num
, u32 txq_id
)
321 struct pci_dev
*dev
= priv
->pci_dev
;
325 /* allocate command space + one big command for scan since scan
326 * command is very huge the system will not have two scan at the
328 len
= sizeof(struct iwl_cmd
) * slots_num
;
329 if (txq_id
== IWL_CMD_QUEUE_NUM
)
330 len
+= IWL_MAX_SCAN_SIZE
;
331 txq
->cmd
= pci_alloc_consistent(dev
, len
, &txq
->dma_addr_cmd
);
335 rc
= iwl_tx_queue_alloc(priv
, txq
, txq_id
);
337 pci_free_consistent(dev
, len
, txq
->cmd
, txq
->dma_addr_cmd
);
341 txq
->need_update
= 0;
343 /* TFD_QUEUE_SIZE_MAX must be power-of-two size, otherwise
344 * iwl_queue_inc_wrap and iwl_queue_dec_wrap are broken. */
345 BUILD_BUG_ON(TFD_QUEUE_SIZE_MAX
& (TFD_QUEUE_SIZE_MAX
- 1));
346 iwl_queue_init(priv
, &txq
->q
, TFD_QUEUE_SIZE_MAX
, slots_num
, txq_id
);
348 iwl_hw_tx_queue_init(priv
, txq
);
354 * iwl_tx_queue_free - Deallocate DMA queue.
355 * @txq: Transmit queue to deallocate.
357 * Empty queue by removing and destroying all BD's.
358 * Free all buffers. txq itself is not freed.
361 void iwl_tx_queue_free(struct iwl_priv
*priv
, struct iwl_tx_queue
*txq
)
363 struct iwl_queue
*q
= &txq
->q
;
364 struct pci_dev
*dev
= priv
->pci_dev
;
370 /* first, empty all BD's */
371 for (; q
->write_ptr
!= q
->read_ptr
;
372 q
->read_ptr
= iwl_queue_inc_wrap(q
->read_ptr
, q
->n_bd
))
373 iwl_hw_txq_free_tfd(priv
, txq
);
375 len
= sizeof(struct iwl_cmd
) * q
->n_window
;
376 if (q
->id
== IWL_CMD_QUEUE_NUM
)
377 len
+= IWL_MAX_SCAN_SIZE
;
379 pci_free_consistent(dev
, len
, txq
->cmd
, txq
->dma_addr_cmd
);
381 /* free buffers belonging to queue itself */
383 pci_free_consistent(dev
, sizeof(struct iwl_tfd_frame
) *
384 txq
->q
.n_bd
, txq
->bd
, txq
->q
.dma_addr
);
391 /* 0 fill whole structure */
392 memset(txq
, 0, sizeof(*txq
));
395 const u8 BROADCAST_ADDR
[ETH_ALEN
] = { 0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF };
397 /*************** STATION TABLE MANAGEMENT ****
399 * NOTE: This needs to be overhauled to better synchronize between
400 * how the iwl-4965.c is using iwl_hw_find_station vs. iwl-3945.c
402 * mac80211 should also be examined to determine if sta_info is duplicating
403 * the functionality provided here
406 /**************************************************************/
408 #if 0 /* temporary disable till we add real remove station */
409 static u8
iwl_remove_station(struct iwl_priv
*priv
, const u8
*addr
, int is_ap
)
411 int index
= IWL_INVALID_STATION
;
415 spin_lock_irqsave(&priv
->sta_lock
, flags
);
419 else if (is_broadcast_ether_addr(addr
))
420 index
= priv
->hw_setting
.bcast_sta_id
;
422 for (i
= IWL_STA_ID
; i
< priv
->hw_setting
.max_stations
; i
++)
423 if (priv
->stations
[i
].used
&&
424 !compare_ether_addr(priv
->stations
[i
].sta
.sta
.addr
,
430 if (unlikely(index
== IWL_INVALID_STATION
))
433 if (priv
->stations
[index
].used
) {
434 priv
->stations
[index
].used
= 0;
435 priv
->num_stations
--;
438 BUG_ON(priv
->num_stations
< 0);
441 spin_unlock_irqrestore(&priv
->sta_lock
, flags
);
446 static void iwl_clear_stations_table(struct iwl_priv
*priv
)
450 spin_lock_irqsave(&priv
->sta_lock
, flags
);
452 priv
->num_stations
= 0;
453 memset(priv
->stations
, 0, sizeof(priv
->stations
));
455 spin_unlock_irqrestore(&priv
->sta_lock
, flags
);
458 u8
iwl_add_station(struct iwl_priv
*priv
, const u8
*addr
, int is_ap
, u8 flags
)
461 int index
= IWL_INVALID_STATION
;
462 struct iwl_station_entry
*station
;
463 unsigned long flags_spin
;
464 DECLARE_MAC_BUF(mac
);
466 spin_lock_irqsave(&priv
->sta_lock
, flags_spin
);
469 else if (is_broadcast_ether_addr(addr
))
470 index
= priv
->hw_setting
.bcast_sta_id
;
472 for (i
= IWL_STA_ID
; i
< priv
->hw_setting
.max_stations
; i
++) {
473 if (!compare_ether_addr(priv
->stations
[i
].sta
.sta
.addr
,
479 if (!priv
->stations
[i
].used
&&
480 index
== IWL_INVALID_STATION
)
485 /* These two conditions has the same outcome but keep them separate
486 since they have different meaning */
487 if (unlikely(index
== IWL_INVALID_STATION
)) {
488 spin_unlock_irqrestore(&priv
->sta_lock
, flags_spin
);
492 if (priv
->stations
[index
].used
&&
493 !compare_ether_addr(priv
->stations
[index
].sta
.sta
.addr
, addr
)) {
494 spin_unlock_irqrestore(&priv
->sta_lock
, flags_spin
);
499 IWL_DEBUG_ASSOC("Add STA ID %d: %s\n", index
, print_mac(mac
, addr
));
500 station
= &priv
->stations
[index
];
502 priv
->num_stations
++;
504 memset(&station
->sta
, 0, sizeof(struct iwl_addsta_cmd
));
505 memcpy(station
->sta
.sta
.addr
, addr
, ETH_ALEN
);
506 station
->sta
.mode
= 0;
507 station
->sta
.sta
.sta_id
= index
;
508 station
->sta
.station_flags
= 0;
510 #ifdef CONFIG_IWL4965_HT
511 /* BCAST station and IBSS stations do not work in HT mode */
512 if (index
!= priv
->hw_setting
.bcast_sta_id
&&
513 priv
->iw_mode
!= IEEE80211_IF_TYPE_IBSS
)
514 iwl4965_set_ht_add_station(priv
, index
);
515 #endif /*CONFIG_IWL4965_HT*/
517 spin_unlock_irqrestore(&priv
->sta_lock
, flags_spin
);
518 iwl_send_add_station(priv
, &station
->sta
, flags
);
523 /*************** DRIVER STATUS FUNCTIONS *****/
525 static inline int iwl_is_ready(struct iwl_priv
*priv
)
527 /* The adapter is 'ready' if READY and GEO_CONFIGURED bits are
528 * set but EXIT_PENDING is not */
529 return test_bit(STATUS_READY
, &priv
->status
) &&
530 test_bit(STATUS_GEO_CONFIGURED
, &priv
->status
) &&
531 !test_bit(STATUS_EXIT_PENDING
, &priv
->status
);
534 static inline int iwl_is_alive(struct iwl_priv
*priv
)
536 return test_bit(STATUS_ALIVE
, &priv
->status
);
539 static inline int iwl_is_init(struct iwl_priv
*priv
)
541 return test_bit(STATUS_INIT
, &priv
->status
);
544 static inline int iwl_is_rfkill(struct iwl_priv
*priv
)
546 return test_bit(STATUS_RF_KILL_HW
, &priv
->status
) ||
547 test_bit(STATUS_RF_KILL_SW
, &priv
->status
);
550 static inline int iwl_is_ready_rf(struct iwl_priv
*priv
)
553 if (iwl_is_rfkill(priv
))
556 return iwl_is_ready(priv
);
559 /*************** HOST COMMAND QUEUE FUNCTIONS *****/
561 #define IWL_CMD(x) case x : return #x
563 static const char *get_cmd_string(u8 cmd
)
566 IWL_CMD(REPLY_ALIVE
);
567 IWL_CMD(REPLY_ERROR
);
569 IWL_CMD(REPLY_RXON_ASSOC
);
570 IWL_CMD(REPLY_QOS_PARAM
);
571 IWL_CMD(REPLY_RXON_TIMING
);
572 IWL_CMD(REPLY_ADD_STA
);
573 IWL_CMD(REPLY_REMOVE_STA
);
574 IWL_CMD(REPLY_REMOVE_ALL_STA
);
576 IWL_CMD(REPLY_RATE_SCALE
);
577 IWL_CMD(REPLY_LEDS_CMD
);
578 IWL_CMD(REPLY_TX_LINK_QUALITY_CMD
);
579 IWL_CMD(RADAR_NOTIFICATION
);
580 IWL_CMD(REPLY_QUIET_CMD
);
581 IWL_CMD(REPLY_CHANNEL_SWITCH
);
582 IWL_CMD(CHANNEL_SWITCH_NOTIFICATION
);
583 IWL_CMD(REPLY_SPECTRUM_MEASUREMENT_CMD
);
584 IWL_CMD(SPECTRUM_MEASURE_NOTIFICATION
);
585 IWL_CMD(POWER_TABLE_CMD
);
586 IWL_CMD(PM_SLEEP_NOTIFICATION
);
587 IWL_CMD(PM_DEBUG_STATISTIC_NOTIFIC
);
588 IWL_CMD(REPLY_SCAN_CMD
);
589 IWL_CMD(REPLY_SCAN_ABORT_CMD
);
590 IWL_CMD(SCAN_START_NOTIFICATION
);
591 IWL_CMD(SCAN_RESULTS_NOTIFICATION
);
592 IWL_CMD(SCAN_COMPLETE_NOTIFICATION
);
593 IWL_CMD(BEACON_NOTIFICATION
);
594 IWL_CMD(REPLY_TX_BEACON
);
595 IWL_CMD(WHO_IS_AWAKE_NOTIFICATION
);
596 IWL_CMD(QUIET_NOTIFICATION
);
597 IWL_CMD(REPLY_TX_PWR_TABLE_CMD
);
598 IWL_CMD(MEASURE_ABORT_NOTIFICATION
);
599 IWL_CMD(REPLY_BT_CONFIG
);
600 IWL_CMD(REPLY_STATISTICS_CMD
);
601 IWL_CMD(STATISTICS_NOTIFICATION
);
602 IWL_CMD(REPLY_CARD_STATE_CMD
);
603 IWL_CMD(CARD_STATE_NOTIFICATION
);
604 IWL_CMD(MISSED_BEACONS_NOTIFICATION
);
605 IWL_CMD(REPLY_CT_KILL_CONFIG_CMD
);
606 IWL_CMD(SENSITIVITY_CMD
);
607 IWL_CMD(REPLY_PHY_CALIBRATION_CMD
);
608 IWL_CMD(REPLY_RX_PHY_CMD
);
609 IWL_CMD(REPLY_RX_MPDU_CMD
);
610 IWL_CMD(REPLY_4965_RX
);
611 IWL_CMD(REPLY_COMPRESSED_BA
);
618 #define HOST_COMPLETE_TIMEOUT (HZ / 2)
621 * iwl_enqueue_hcmd - enqueue a uCode command
622 * @priv: device private data point
623 * @cmd: a point to the ucode command structure
625 * The function returns < 0 values to indicate the operation is
626 * failed. On success, it turns the index (> 0) of command in the
629 static int iwl_enqueue_hcmd(struct iwl_priv
*priv
, struct iwl_host_cmd
*cmd
)
631 struct iwl_tx_queue
*txq
= &priv
->txq
[IWL_CMD_QUEUE_NUM
];
632 struct iwl_queue
*q
= &txq
->q
;
633 struct iwl_tfd_frame
*tfd
;
635 struct iwl_cmd
*out_cmd
;
637 u16 fix_size
= (u16
)(cmd
->len
+ sizeof(out_cmd
->hdr
));
638 dma_addr_t phys_addr
;
642 /* If any of the command structures end up being larger than
643 * the TFD_MAX_PAYLOAD_SIZE, and it sent as a 'small' command then
644 * we will need to increase the size of the TFD entries */
645 BUG_ON((fix_size
> TFD_MAX_PAYLOAD_SIZE
) &&
646 !(cmd
->meta
.flags
& CMD_SIZE_HUGE
));
648 if (iwl_queue_space(q
) < ((cmd
->meta
.flags
& CMD_ASYNC
) ? 2 : 1)) {
649 IWL_ERROR("No space for Tx\n");
653 spin_lock_irqsave(&priv
->hcmd_lock
, flags
);
655 tfd
= &txq
->bd
[q
->write_ptr
];
656 memset(tfd
, 0, sizeof(*tfd
));
658 control_flags
= (u32
*) tfd
;
660 idx
= get_cmd_index(q
, q
->write_ptr
, cmd
->meta
.flags
& CMD_SIZE_HUGE
);
661 out_cmd
= &txq
->cmd
[idx
];
663 out_cmd
->hdr
.cmd
= cmd
->id
;
664 memcpy(&out_cmd
->meta
, &cmd
->meta
, sizeof(cmd
->meta
));
665 memcpy(&out_cmd
->cmd
.payload
, cmd
->data
, cmd
->len
);
667 /* At this point, the out_cmd now has all of the incoming cmd
670 out_cmd
->hdr
.flags
= 0;
671 out_cmd
->hdr
.sequence
= cpu_to_le16(QUEUE_TO_SEQ(IWL_CMD_QUEUE_NUM
) |
672 INDEX_TO_SEQ(q
->write_ptr
));
673 if (out_cmd
->meta
.flags
& CMD_SIZE_HUGE
)
674 out_cmd
->hdr
.sequence
|= cpu_to_le16(SEQ_HUGE_FRAME
);
676 phys_addr
= txq
->dma_addr_cmd
+ sizeof(txq
->cmd
[0]) * idx
+
677 offsetof(struct iwl_cmd
, hdr
);
678 iwl_hw_txq_attach_buf_to_tfd(priv
, tfd
, phys_addr
, fix_size
);
680 IWL_DEBUG_HC("Sending command %s (#%x), seq: 0x%04X, "
681 "%d bytes at %d[%d]:%d\n",
682 get_cmd_string(out_cmd
->hdr
.cmd
),
683 out_cmd
->hdr
.cmd
, le16_to_cpu(out_cmd
->hdr
.sequence
),
684 fix_size
, q
->write_ptr
, idx
, IWL_CMD_QUEUE_NUM
);
686 txq
->need_update
= 1;
687 ret
= iwl4965_tx_queue_update_wr_ptr(priv
, txq
, 0);
688 q
->write_ptr
= iwl_queue_inc_wrap(q
->write_ptr
, q
->n_bd
);
689 iwl_tx_queue_update_write_ptr(priv
, txq
);
691 spin_unlock_irqrestore(&priv
->hcmd_lock
, flags
);
692 return ret
? ret
: idx
;
695 static int iwl_send_cmd_async(struct iwl_priv
*priv
, struct iwl_host_cmd
*cmd
)
699 BUG_ON(!(cmd
->meta
.flags
& CMD_ASYNC
));
701 /* An asynchronous command can not expect an SKB to be set. */
702 BUG_ON(cmd
->meta
.flags
& CMD_WANT_SKB
);
704 /* An asynchronous command MUST have a callback. */
705 BUG_ON(!cmd
->meta
.u
.callback
);
707 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
710 ret
= iwl_enqueue_hcmd(priv
, cmd
);
712 IWL_ERROR("Error sending %s: iwl_enqueue_hcmd failed: %d\n",
713 get_cmd_string(cmd
->id
), ret
);
719 static int iwl_send_cmd_sync(struct iwl_priv
*priv
, struct iwl_host_cmd
*cmd
)
723 static atomic_t entry
= ATOMIC_INIT(0); /* reentrance protection */
725 BUG_ON(cmd
->meta
.flags
& CMD_ASYNC
);
727 /* A synchronous command can not have a callback set. */
728 BUG_ON(cmd
->meta
.u
.callback
!= NULL
);
730 if (atomic_xchg(&entry
, 1)) {
731 IWL_ERROR("Error sending %s: Already sending a host command\n",
732 get_cmd_string(cmd
->id
));
736 set_bit(STATUS_HCMD_ACTIVE
, &priv
->status
);
738 if (cmd
->meta
.flags
& CMD_WANT_SKB
)
739 cmd
->meta
.source
= &cmd
->meta
;
741 cmd_idx
= iwl_enqueue_hcmd(priv
, cmd
);
744 IWL_ERROR("Error sending %s: iwl_enqueue_hcmd failed: %d\n",
745 get_cmd_string(cmd
->id
), ret
);
749 ret
= wait_event_interruptible_timeout(priv
->wait_command_queue
,
750 !test_bit(STATUS_HCMD_ACTIVE
, &priv
->status
),
751 HOST_COMPLETE_TIMEOUT
);
753 if (test_bit(STATUS_HCMD_ACTIVE
, &priv
->status
)) {
754 IWL_ERROR("Error sending %s: time out after %dms.\n",
755 get_cmd_string(cmd
->id
),
756 jiffies_to_msecs(HOST_COMPLETE_TIMEOUT
));
758 clear_bit(STATUS_HCMD_ACTIVE
, &priv
->status
);
764 if (test_bit(STATUS_RF_KILL_HW
, &priv
->status
)) {
765 IWL_DEBUG_INFO("Command %s aborted: RF KILL Switch\n",
766 get_cmd_string(cmd
->id
));
770 if (test_bit(STATUS_FW_ERROR
, &priv
->status
)) {
771 IWL_DEBUG_INFO("Command %s failed: FW Error\n",
772 get_cmd_string(cmd
->id
));
776 if ((cmd
->meta
.flags
& CMD_WANT_SKB
) && !cmd
->meta
.u
.skb
) {
777 IWL_ERROR("Error: Response NULL in '%s'\n",
778 get_cmd_string(cmd
->id
));
787 if (cmd
->meta
.flags
& CMD_WANT_SKB
) {
788 struct iwl_cmd
*qcmd
;
790 /* Cancel the CMD_WANT_SKB flag for the cmd in the
791 * TX cmd queue. Otherwise in case the cmd comes
792 * in later, it will possibly set an invalid
793 * address (cmd->meta.source). */
794 qcmd
= &priv
->txq
[IWL_CMD_QUEUE_NUM
].cmd
[cmd_idx
];
795 qcmd
->meta
.flags
&= ~CMD_WANT_SKB
;
798 if (cmd
->meta
.u
.skb
) {
799 dev_kfree_skb_any(cmd
->meta
.u
.skb
);
800 cmd
->meta
.u
.skb
= NULL
;
803 atomic_set(&entry
, 0);
807 int iwl_send_cmd(struct iwl_priv
*priv
, struct iwl_host_cmd
*cmd
)
809 if (cmd
->meta
.flags
& CMD_ASYNC
)
810 return iwl_send_cmd_async(priv
, cmd
);
812 return iwl_send_cmd_sync(priv
, cmd
);
815 int iwl_send_cmd_pdu(struct iwl_priv
*priv
, u8 id
, u16 len
, const void *data
)
817 struct iwl_host_cmd cmd
= {
823 return iwl_send_cmd_sync(priv
, &cmd
);
826 static int __must_check
iwl_send_cmd_u32(struct iwl_priv
*priv
, u8 id
, u32 val
)
828 struct iwl_host_cmd cmd
= {
834 return iwl_send_cmd_sync(priv
, &cmd
);
837 int iwl_send_statistics_request(struct iwl_priv
*priv
)
839 return iwl_send_cmd_u32(priv
, REPLY_STATISTICS_CMD
, 0);
843 * iwl_rxon_add_station - add station into station table.
845 * there is only one AP station with id= IWL_AP_ID
846 * NOTE: mutex must be held before calling the this fnction
848 static int iwl_rxon_add_station(struct iwl_priv
*priv
,
849 const u8
*addr
, int is_ap
)
853 sta_id
= iwl_add_station(priv
, addr
, is_ap
, 0);
854 iwl4965_add_station(priv
, addr
, is_ap
);
860 * iwl_set_rxon_channel - Set the phymode and channel values in staging RXON
861 * @phymode: MODE_IEEE80211A sets to 5.2GHz; all else set to 2.4GHz
862 * @channel: Any channel valid for the requested phymode
864 * In addition to setting the staging RXON, priv->phymode is also set.
866 * NOTE: Does not commit to the hardware; it sets appropriate bit fields
867 * in the staging RXON flag structure based on the phymode
869 static int iwl_set_rxon_channel(struct iwl_priv
*priv
, u8 phymode
, u16 channel
)
871 if (!iwl_get_channel_info(priv
, phymode
, channel
)) {
872 IWL_DEBUG_INFO("Could not set channel to %d [%d]\n",
877 if ((le16_to_cpu(priv
->staging_rxon
.channel
) == channel
) &&
878 (priv
->phymode
== phymode
))
881 priv
->staging_rxon
.channel
= cpu_to_le16(channel
);
882 if (phymode
== MODE_IEEE80211A
)
883 priv
->staging_rxon
.flags
&= ~RXON_FLG_BAND_24G_MSK
;
885 priv
->staging_rxon
.flags
|= RXON_FLG_BAND_24G_MSK
;
887 priv
->phymode
= phymode
;
889 IWL_DEBUG_INFO("Staging channel set to %d [%d]\n", channel
, phymode
);
895 * iwl_check_rxon_cmd - validate RXON structure is valid
897 * NOTE: This is really only useful during development and can eventually
898 * be #ifdef'd out once the driver is stable and folks aren't actively
901 static int iwl_check_rxon_cmd(struct iwl_rxon_cmd
*rxon
)
906 if (rxon
->flags
& RXON_FLG_BAND_24G_MSK
) {
907 error
|= le32_to_cpu(rxon
->flags
&
908 (RXON_FLG_TGJ_NARROW_BAND_MSK
|
909 RXON_FLG_RADAR_DETECT_MSK
));
911 IWL_WARNING("check 24G fields %d | %d\n",
914 error
|= (rxon
->flags
& RXON_FLG_SHORT_SLOT_MSK
) ?
915 0 : le32_to_cpu(RXON_FLG_SHORT_SLOT_MSK
);
917 IWL_WARNING("check 52 fields %d | %d\n",
919 error
|= le32_to_cpu(rxon
->flags
& RXON_FLG_CCK_MSK
);
921 IWL_WARNING("check 52 CCK %d | %d\n",
924 error
|= (rxon
->node_addr
[0] | rxon
->bssid_addr
[0]) & 0x1;
926 IWL_WARNING("check mac addr %d | %d\n", counter
++, error
);
928 /* make sure basic rates 6Mbps and 1Mbps are supported */
929 error
|= (((rxon
->ofdm_basic_rates
& IWL_RATE_6M_MASK
) == 0) &&
930 ((rxon
->cck_basic_rates
& IWL_RATE_1M_MASK
) == 0));
932 IWL_WARNING("check basic rate %d | %d\n", counter
++, error
);
934 error
|= (le16_to_cpu(rxon
->assoc_id
) > 2007);
936 IWL_WARNING("check assoc id %d | %d\n", counter
++, error
);
938 error
|= ((rxon
->flags
& (RXON_FLG_CCK_MSK
| RXON_FLG_SHORT_SLOT_MSK
))
939 == (RXON_FLG_CCK_MSK
| RXON_FLG_SHORT_SLOT_MSK
));
941 IWL_WARNING("check CCK and short slot %d | %d\n",
944 error
|= ((rxon
->flags
& (RXON_FLG_CCK_MSK
| RXON_FLG_AUTO_DETECT_MSK
))
945 == (RXON_FLG_CCK_MSK
| RXON_FLG_AUTO_DETECT_MSK
));
947 IWL_WARNING("check CCK & auto detect %d | %d\n",
950 error
|= ((rxon
->flags
& (RXON_FLG_AUTO_DETECT_MSK
|
951 RXON_FLG_TGG_PROTECT_MSK
)) == RXON_FLG_TGG_PROTECT_MSK
);
953 IWL_WARNING("check TGG and auto detect %d | %d\n",
957 IWL_WARNING("Tuning to channel %d\n",
958 le16_to_cpu(rxon
->channel
));
961 IWL_ERROR("Not a valid iwl_rxon_assoc_cmd field values\n");
968 * iwl_full_rxon_required - determine if RXON_ASSOC can be used in RXON commit
969 * @priv: staging_rxon is compared to active_rxon
971 * If the RXON structure is changing sufficient to require a new
972 * tune or to clear and reset the RXON_FILTER_ASSOC_MSK then return 1
973 * to indicate a new tune is required.
975 static int iwl_full_rxon_required(struct iwl_priv
*priv
)
978 /* These items are only settable from the full RXON command */
979 if (!(priv
->active_rxon
.filter_flags
& RXON_FILTER_ASSOC_MSK
) ||
980 compare_ether_addr(priv
->staging_rxon
.bssid_addr
,
981 priv
->active_rxon
.bssid_addr
) ||
982 compare_ether_addr(priv
->staging_rxon
.node_addr
,
983 priv
->active_rxon
.node_addr
) ||
984 compare_ether_addr(priv
->staging_rxon
.wlap_bssid_addr
,
985 priv
->active_rxon
.wlap_bssid_addr
) ||
986 (priv
->staging_rxon
.dev_type
!= priv
->active_rxon
.dev_type
) ||
987 (priv
->staging_rxon
.channel
!= priv
->active_rxon
.channel
) ||
988 (priv
->staging_rxon
.air_propagation
!=
989 priv
->active_rxon
.air_propagation
) ||
990 (priv
->staging_rxon
.ofdm_ht_single_stream_basic_rates
!=
991 priv
->active_rxon
.ofdm_ht_single_stream_basic_rates
) ||
992 (priv
->staging_rxon
.ofdm_ht_dual_stream_basic_rates
!=
993 priv
->active_rxon
.ofdm_ht_dual_stream_basic_rates
) ||
994 (priv
->staging_rxon
.rx_chain
!= priv
->active_rxon
.rx_chain
) ||
995 (priv
->staging_rxon
.assoc_id
!= priv
->active_rxon
.assoc_id
))
998 /* flags, filter_flags, ofdm_basic_rates, and cck_basic_rates can
999 * be updated with the RXON_ASSOC command -- however only some
1000 * flag transitions are allowed using RXON_ASSOC */
1002 /* Check if we are not switching bands */
1003 if ((priv
->staging_rxon
.flags
& RXON_FLG_BAND_24G_MSK
) !=
1004 (priv
->active_rxon
.flags
& RXON_FLG_BAND_24G_MSK
))
1007 /* Check if we are switching association toggle */
1008 if ((priv
->staging_rxon
.filter_flags
& RXON_FILTER_ASSOC_MSK
) !=
1009 (priv
->active_rxon
.filter_flags
& RXON_FILTER_ASSOC_MSK
))
1015 static int iwl_send_rxon_assoc(struct iwl_priv
*priv
)
1018 struct iwl_rx_packet
*res
= NULL
;
1019 struct iwl_rxon_assoc_cmd rxon_assoc
;
1020 struct iwl_host_cmd cmd
= {
1021 .id
= REPLY_RXON_ASSOC
,
1022 .len
= sizeof(rxon_assoc
),
1023 .meta
.flags
= CMD_WANT_SKB
,
1024 .data
= &rxon_assoc
,
1026 const struct iwl_rxon_cmd
*rxon1
= &priv
->staging_rxon
;
1027 const struct iwl_rxon_cmd
*rxon2
= &priv
->active_rxon
;
1029 if ((rxon1
->flags
== rxon2
->flags
) &&
1030 (rxon1
->filter_flags
== rxon2
->filter_flags
) &&
1031 (rxon1
->cck_basic_rates
== rxon2
->cck_basic_rates
) &&
1032 (rxon1
->ofdm_ht_single_stream_basic_rates
==
1033 rxon2
->ofdm_ht_single_stream_basic_rates
) &&
1034 (rxon1
->ofdm_ht_dual_stream_basic_rates
==
1035 rxon2
->ofdm_ht_dual_stream_basic_rates
) &&
1036 (rxon1
->rx_chain
== rxon2
->rx_chain
) &&
1037 (rxon1
->ofdm_basic_rates
== rxon2
->ofdm_basic_rates
)) {
1038 IWL_DEBUG_INFO("Using current RXON_ASSOC. Not resending.\n");
1042 rxon_assoc
.flags
= priv
->staging_rxon
.flags
;
1043 rxon_assoc
.filter_flags
= priv
->staging_rxon
.filter_flags
;
1044 rxon_assoc
.ofdm_basic_rates
= priv
->staging_rxon
.ofdm_basic_rates
;
1045 rxon_assoc
.cck_basic_rates
= priv
->staging_rxon
.cck_basic_rates
;
1046 rxon_assoc
.reserved
= 0;
1047 rxon_assoc
.ofdm_ht_single_stream_basic_rates
=
1048 priv
->staging_rxon
.ofdm_ht_single_stream_basic_rates
;
1049 rxon_assoc
.ofdm_ht_dual_stream_basic_rates
=
1050 priv
->staging_rxon
.ofdm_ht_dual_stream_basic_rates
;
1051 rxon_assoc
.rx_chain_select_flags
= priv
->staging_rxon
.rx_chain
;
1053 rc
= iwl_send_cmd_sync(priv
, &cmd
);
1057 res
= (struct iwl_rx_packet
*)cmd
.meta
.u
.skb
->data
;
1058 if (res
->hdr
.flags
& IWL_CMD_FAILED_MSK
) {
1059 IWL_ERROR("Bad return from REPLY_RXON_ASSOC command\n");
1063 priv
->alloc_rxb_skb
--;
1064 dev_kfree_skb_any(cmd
.meta
.u
.skb
);
1070 * iwl_commit_rxon - commit staging_rxon to hardware
1072 * The RXON command in staging_rxon is committed to the hardware and
1073 * the active_rxon structure is updated with the new data. This
1074 * function correctly transitions out of the RXON_ASSOC_MSK state if
1075 * a HW tune is required based on the RXON structure changes.
1077 static int iwl_commit_rxon(struct iwl_priv
*priv
)
1079 /* cast away the const for active_rxon in this function */
1080 struct iwl_rxon_cmd
*active_rxon
= (void *)&priv
->active_rxon
;
1081 DECLARE_MAC_BUF(mac
);
1084 if (!iwl_is_alive(priv
))
1087 /* always get timestamp with Rx frame */
1088 priv
->staging_rxon
.flags
|= RXON_FLG_TSF2HOST_MSK
;
1090 rc
= iwl_check_rxon_cmd(&priv
->staging_rxon
);
1092 IWL_ERROR("Invalid RXON configuration. Not committing.\n");
1096 /* If we don't need to send a full RXON, we can use
1097 * iwl_rxon_assoc_cmd which is used to reconfigure filter
1098 * and other flags for the current radio configuration. */
1099 if (!iwl_full_rxon_required(priv
)) {
1100 rc
= iwl_send_rxon_assoc(priv
);
1102 IWL_ERROR("Error setting RXON_ASSOC "
1103 "configuration (%d).\n", rc
);
1107 memcpy(active_rxon
, &priv
->staging_rxon
, sizeof(*active_rxon
));
1112 /* station table will be cleared */
1113 priv
->assoc_station_added
= 0;
1115 #ifdef CONFIG_IWL4965_SENSITIVITY
1116 priv
->sensitivity_data
.state
= IWL_SENS_CALIB_NEED_REINIT
;
1117 if (!priv
->error_recovering
)
1118 priv
->start_calib
= 0;
1120 iwl4965_init_sensitivity(priv
, CMD_ASYNC
, 1);
1121 #endif /* CONFIG_IWL4965_SENSITIVITY */
1123 /* If we are currently associated and the new config requires
1124 * an RXON_ASSOC and the new config wants the associated mask enabled,
1125 * we must clear the associated from the active configuration
1126 * before we apply the new config */
1127 if (iwl_is_associated(priv
) &&
1128 (priv
->staging_rxon
.filter_flags
& RXON_FILTER_ASSOC_MSK
)) {
1129 IWL_DEBUG_INFO("Toggling associated bit on current RXON\n");
1130 active_rxon
->filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
1132 rc
= iwl_send_cmd_pdu(priv
, REPLY_RXON
,
1133 sizeof(struct iwl_rxon_cmd
),
1134 &priv
->active_rxon
);
1136 /* If the mask clearing failed then we set
1137 * active_rxon back to what it was previously */
1139 active_rxon
->filter_flags
|= RXON_FILTER_ASSOC_MSK
;
1140 IWL_ERROR("Error clearing ASSOC_MSK on current "
1141 "configuration (%d).\n", rc
);
1146 IWL_DEBUG_INFO("Sending RXON\n"
1147 "* with%s RXON_FILTER_ASSOC_MSK\n"
1150 ((priv
->staging_rxon
.filter_flags
&
1151 RXON_FILTER_ASSOC_MSK
) ? "" : "out"),
1152 le16_to_cpu(priv
->staging_rxon
.channel
),
1153 print_mac(mac
, priv
->staging_rxon
.bssid_addr
));
1155 /* Apply the new configuration */
1156 rc
= iwl_send_cmd_pdu(priv
, REPLY_RXON
,
1157 sizeof(struct iwl_rxon_cmd
), &priv
->staging_rxon
);
1159 IWL_ERROR("Error setting new configuration (%d).\n", rc
);
1163 iwl_clear_stations_table(priv
);
1165 #ifdef CONFIG_IWL4965_SENSITIVITY
1166 if (!priv
->error_recovering
)
1167 priv
->start_calib
= 0;
1169 priv
->sensitivity_data
.state
= IWL_SENS_CALIB_NEED_REINIT
;
1170 iwl4965_init_sensitivity(priv
, CMD_ASYNC
, 1);
1171 #endif /* CONFIG_IWL4965_SENSITIVITY */
1173 memcpy(active_rxon
, &priv
->staging_rxon
, sizeof(*active_rxon
));
1175 /* If we issue a new RXON command which required a tune then we must
1176 * send a new TXPOWER command or we won't be able to Tx any frames */
1177 rc
= iwl_hw_reg_send_txpower(priv
);
1179 IWL_ERROR("Error setting Tx power (%d).\n", rc
);
1183 /* Add the broadcast address so we can send broadcast frames */
1184 if (iwl_rxon_add_station(priv
, BROADCAST_ADDR
, 0) ==
1185 IWL_INVALID_STATION
) {
1186 IWL_ERROR("Error adding BROADCAST address for transmit.\n");
1190 /* If we have set the ASSOC_MSK and we are in BSS mode then
1191 * add the IWL_AP_ID to the station rate table */
1192 if (iwl_is_associated(priv
) &&
1193 (priv
->iw_mode
== IEEE80211_IF_TYPE_STA
)) {
1194 if (iwl_rxon_add_station(priv
, priv
->active_rxon
.bssid_addr
, 1)
1195 == IWL_INVALID_STATION
) {
1196 IWL_ERROR("Error adding AP address for transmit.\n");
1199 priv
->assoc_station_added
= 1;
1205 static int iwl_send_bt_config(struct iwl_priv
*priv
)
1207 struct iwl_bt_cmd bt_cmd
= {
1215 return iwl_send_cmd_pdu(priv
, REPLY_BT_CONFIG
,
1216 sizeof(struct iwl_bt_cmd
), &bt_cmd
);
1219 static int iwl_send_scan_abort(struct iwl_priv
*priv
)
1222 struct iwl_rx_packet
*res
;
1223 struct iwl_host_cmd cmd
= {
1224 .id
= REPLY_SCAN_ABORT_CMD
,
1225 .meta
.flags
= CMD_WANT_SKB
,
1228 /* If there isn't a scan actively going on in the hardware
1229 * then we are in between scan bands and not actually
1230 * actively scanning, so don't send the abort command */
1231 if (!test_bit(STATUS_SCAN_HW
, &priv
->status
)) {
1232 clear_bit(STATUS_SCAN_ABORTING
, &priv
->status
);
1236 rc
= iwl_send_cmd_sync(priv
, &cmd
);
1238 clear_bit(STATUS_SCAN_ABORTING
, &priv
->status
);
1242 res
= (struct iwl_rx_packet
*)cmd
.meta
.u
.skb
->data
;
1243 if (res
->u
.status
!= CAN_ABORT_STATUS
) {
1244 /* The scan abort will return 1 for success or
1245 * 2 for "failure". A failure condition can be
1246 * due to simply not being in an active scan which
1247 * can occur if we send the scan abort before we
1248 * the microcode has notified us that a scan is
1250 IWL_DEBUG_INFO("SCAN_ABORT returned %d.\n", res
->u
.status
);
1251 clear_bit(STATUS_SCAN_ABORTING
, &priv
->status
);
1252 clear_bit(STATUS_SCAN_HW
, &priv
->status
);
1255 dev_kfree_skb_any(cmd
.meta
.u
.skb
);
1260 static int iwl_card_state_sync_callback(struct iwl_priv
*priv
,
1261 struct iwl_cmd
*cmd
,
1262 struct sk_buff
*skb
)
1270 * Use: Sets the internal card state to enable, disable, or halt
1272 * When in the 'enable' state the card operates as normal.
1273 * When in the 'disable' state, the card enters into a low power mode.
1274 * When in the 'halt' state, the card is shut down and must be fully
1275 * restarted to come back on.
1277 static int iwl_send_card_state(struct iwl_priv
*priv
, u32 flags
, u8 meta_flag
)
1279 struct iwl_host_cmd cmd
= {
1280 .id
= REPLY_CARD_STATE_CMD
,
1283 .meta
.flags
= meta_flag
,
1286 if (meta_flag
& CMD_ASYNC
)
1287 cmd
.meta
.u
.callback
= iwl_card_state_sync_callback
;
1289 return iwl_send_cmd(priv
, &cmd
);
1292 static int iwl_add_sta_sync_callback(struct iwl_priv
*priv
,
1293 struct iwl_cmd
*cmd
, struct sk_buff
*skb
)
1295 struct iwl_rx_packet
*res
= NULL
;
1298 IWL_ERROR("Error: Response NULL in REPLY_ADD_STA.\n");
1302 res
= (struct iwl_rx_packet
*)skb
->data
;
1303 if (res
->hdr
.flags
& IWL_CMD_FAILED_MSK
) {
1304 IWL_ERROR("Bad return from REPLY_ADD_STA (0x%08X)\n",
1309 switch (res
->u
.add_sta
.status
) {
1310 case ADD_STA_SUCCESS_MSK
:
1316 /* We didn't cache the SKB; let the caller free it */
1320 int iwl_send_add_station(struct iwl_priv
*priv
,
1321 struct iwl_addsta_cmd
*sta
, u8 flags
)
1323 struct iwl_rx_packet
*res
= NULL
;
1325 struct iwl_host_cmd cmd
= {
1326 .id
= REPLY_ADD_STA
,
1327 .len
= sizeof(struct iwl_addsta_cmd
),
1328 .meta
.flags
= flags
,
1332 if (flags
& CMD_ASYNC
)
1333 cmd
.meta
.u
.callback
= iwl_add_sta_sync_callback
;
1335 cmd
.meta
.flags
|= CMD_WANT_SKB
;
1337 rc
= iwl_send_cmd(priv
, &cmd
);
1339 if (rc
|| (flags
& CMD_ASYNC
))
1342 res
= (struct iwl_rx_packet
*)cmd
.meta
.u
.skb
->data
;
1343 if (res
->hdr
.flags
& IWL_CMD_FAILED_MSK
) {
1344 IWL_ERROR("Bad return from REPLY_ADD_STA (0x%08X)\n",
1350 switch (res
->u
.add_sta
.status
) {
1351 case ADD_STA_SUCCESS_MSK
:
1352 IWL_DEBUG_INFO("REPLY_ADD_STA PASSED\n");
1356 IWL_WARNING("REPLY_ADD_STA failed\n");
1361 priv
->alloc_rxb_skb
--;
1362 dev_kfree_skb_any(cmd
.meta
.u
.skb
);
1367 static int iwl_update_sta_key_info(struct iwl_priv
*priv
,
1368 struct ieee80211_key_conf
*keyconf
,
1371 unsigned long flags
;
1372 __le16 key_flags
= 0;
1374 switch (keyconf
->alg
) {
1376 key_flags
|= STA_KEY_FLG_CCMP
;
1377 key_flags
|= cpu_to_le16(
1378 keyconf
->keyidx
<< STA_KEY_FLG_KEYID_POS
);
1379 key_flags
&= ~STA_KEY_FLG_INVALID
;
1386 spin_lock_irqsave(&priv
->sta_lock
, flags
);
1387 priv
->stations
[sta_id
].keyinfo
.alg
= keyconf
->alg
;
1388 priv
->stations
[sta_id
].keyinfo
.keylen
= keyconf
->keylen
;
1389 memcpy(priv
->stations
[sta_id
].keyinfo
.key
, keyconf
->key
,
1392 memcpy(priv
->stations
[sta_id
].sta
.key
.key
, keyconf
->key
,
1394 priv
->stations
[sta_id
].sta
.key
.key_flags
= key_flags
;
1395 priv
->stations
[sta_id
].sta
.sta
.modify_mask
= STA_MODIFY_KEY_MASK
;
1396 priv
->stations
[sta_id
].sta
.mode
= STA_CONTROL_MODIFY_MSK
;
1398 spin_unlock_irqrestore(&priv
->sta_lock
, flags
);
1400 IWL_DEBUG_INFO("hwcrypto: modify ucode station key info\n");
1401 iwl_send_add_station(priv
, &priv
->stations
[sta_id
].sta
, 0);
1405 static int iwl_clear_sta_key_info(struct iwl_priv
*priv
, u8 sta_id
)
1407 unsigned long flags
;
1409 spin_lock_irqsave(&priv
->sta_lock
, flags
);
1410 memset(&priv
->stations
[sta_id
].keyinfo
, 0, sizeof(struct iwl_hw_key
));
1411 memset(&priv
->stations
[sta_id
].sta
.key
, 0, sizeof(struct iwl_keyinfo
));
1412 priv
->stations
[sta_id
].sta
.key
.key_flags
= STA_KEY_FLG_NO_ENC
;
1413 priv
->stations
[sta_id
].sta
.sta
.modify_mask
= STA_MODIFY_KEY_MASK
;
1414 priv
->stations
[sta_id
].sta
.mode
= STA_CONTROL_MODIFY_MSK
;
1415 spin_unlock_irqrestore(&priv
->sta_lock
, flags
);
1417 IWL_DEBUG_INFO("hwcrypto: clear ucode station key info\n");
1418 iwl_send_add_station(priv
, &priv
->stations
[sta_id
].sta
, 0);
1422 static void iwl_clear_free_frames(struct iwl_priv
*priv
)
1424 struct list_head
*element
;
1426 IWL_DEBUG_INFO("%d frames on pre-allocated heap on clear.\n",
1427 priv
->frames_count
);
1429 while (!list_empty(&priv
->free_frames
)) {
1430 element
= priv
->free_frames
.next
;
1432 kfree(list_entry(element
, struct iwl_frame
, list
));
1433 priv
->frames_count
--;
1436 if (priv
->frames_count
) {
1437 IWL_WARNING("%d frames still in use. Did we lose one?\n",
1438 priv
->frames_count
);
1439 priv
->frames_count
= 0;
1443 static struct iwl_frame
*iwl_get_free_frame(struct iwl_priv
*priv
)
1445 struct iwl_frame
*frame
;
1446 struct list_head
*element
;
1447 if (list_empty(&priv
->free_frames
)) {
1448 frame
= kzalloc(sizeof(*frame
), GFP_KERNEL
);
1450 IWL_ERROR("Could not allocate frame!\n");
1454 priv
->frames_count
++;
1458 element
= priv
->free_frames
.next
;
1460 return list_entry(element
, struct iwl_frame
, list
);
1463 static void iwl_free_frame(struct iwl_priv
*priv
, struct iwl_frame
*frame
)
1465 memset(frame
, 0, sizeof(*frame
));
1466 list_add(&frame
->list
, &priv
->free_frames
);
1469 unsigned int iwl_fill_beacon_frame(struct iwl_priv
*priv
,
1470 struct ieee80211_hdr
*hdr
,
1471 const u8
*dest
, int left
)
1474 if (!iwl_is_associated(priv
) || !priv
->ibss_beacon
||
1475 ((priv
->iw_mode
!= IEEE80211_IF_TYPE_IBSS
) &&
1476 (priv
->iw_mode
!= IEEE80211_IF_TYPE_AP
)))
1479 if (priv
->ibss_beacon
->len
> left
)
1482 memcpy(hdr
, priv
->ibss_beacon
->data
, priv
->ibss_beacon
->len
);
1484 return priv
->ibss_beacon
->len
;
1487 int iwl_rate_index_from_plcp(int plcp
)
1491 if (plcp
& RATE_MCS_HT_MSK
) {
1494 if (i
>= IWL_RATE_MIMO_6M_PLCP
)
1495 i
= i
- IWL_RATE_MIMO_6M_PLCP
;
1497 i
+= IWL_FIRST_OFDM_RATE
;
1498 /* skip 9M not supported in ht*/
1499 if (i
>= IWL_RATE_9M_INDEX
)
1501 if ((i
>= IWL_FIRST_OFDM_RATE
) &&
1502 (i
<= IWL_LAST_OFDM_RATE
))
1505 for (i
= 0; i
< ARRAY_SIZE(iwl_rates
); i
++)
1506 if (iwl_rates
[i
].plcp
== (plcp
&0xFF))
1512 static u8
iwl_rate_get_lowest_plcp(int rate_mask
)
1516 for (i
= IWL_RATE_1M_INDEX
; i
!= IWL_RATE_INVALID
;
1517 i
= iwl_rates
[i
].next_ieee
) {
1518 if (rate_mask
& (1 << i
))
1519 return iwl_rates
[i
].plcp
;
1522 return IWL_RATE_INVALID
;
1525 static int iwl_send_beacon_cmd(struct iwl_priv
*priv
)
1527 struct iwl_frame
*frame
;
1528 unsigned int frame_size
;
1532 frame
= iwl_get_free_frame(priv
);
1535 IWL_ERROR("Could not obtain free frame buffer for beacon "
1540 if (!(priv
->staging_rxon
.flags
& RXON_FLG_BAND_24G_MSK
)) {
1541 rate
= iwl_rate_get_lowest_plcp(priv
->active_rate_basic
&
1543 if (rate
== IWL_INVALID_RATE
)
1544 rate
= IWL_RATE_6M_PLCP
;
1546 rate
= iwl_rate_get_lowest_plcp(priv
->active_rate_basic
& 0xF);
1547 if (rate
== IWL_INVALID_RATE
)
1548 rate
= IWL_RATE_1M_PLCP
;
1551 frame_size
= iwl_hw_get_beacon_cmd(priv
, frame
, rate
);
1553 rc
= iwl_send_cmd_pdu(priv
, REPLY_TX_BEACON
, frame_size
,
1556 iwl_free_frame(priv
, frame
);
1561 /******************************************************************************
1563 * EEPROM related functions
1565 ******************************************************************************/
1567 static void get_eeprom_mac(struct iwl_priv
*priv
, u8
*mac
)
1569 memcpy(mac
, priv
->eeprom
.mac_address
, 6);
1573 * iwl_eeprom_init - read EEPROM contents
1575 * Load the EEPROM from adapter into priv->eeprom
1577 * NOTE: This routine uses the non-debug IO access functions.
1579 int iwl_eeprom_init(struct iwl_priv
*priv
)
1581 u16
*e
= (u16
*)&priv
->eeprom
;
1582 u32 gp
= iwl_read32(priv
, CSR_EEPROM_GP
);
1584 int sz
= sizeof(priv
->eeprom
);
1589 /* The EEPROM structure has several padding buffers within it
1590 * and when adding new EEPROM maps is subject to programmer errors
1591 * which may be very difficult to identify without explicitly
1592 * checking the resulting size of the eeprom map. */
1593 BUILD_BUG_ON(sizeof(priv
->eeprom
) != IWL_EEPROM_IMAGE_SIZE
);
1595 if ((gp
& CSR_EEPROM_GP_VALID_MSK
) == CSR_EEPROM_GP_BAD_SIGNATURE
) {
1596 IWL_ERROR("EEPROM not found, EEPROM_GP=0x%08x", gp
);
1600 rc
= iwl_eeprom_acquire_semaphore(priv
);
1602 IWL_ERROR("Failed to acquire EEPROM semaphore.\n");
1606 /* eeprom is an array of 16bit values */
1607 for (addr
= 0; addr
< sz
; addr
+= sizeof(u16
)) {
1608 _iwl_write32(priv
, CSR_EEPROM_REG
, addr
<< 1);
1609 _iwl_clear_bit(priv
, CSR_EEPROM_REG
, CSR_EEPROM_REG_BIT_CMD
);
1611 for (i
= 0; i
< IWL_EEPROM_ACCESS_TIMEOUT
;
1612 i
+= IWL_EEPROM_ACCESS_DELAY
) {
1613 r
= _iwl_read_direct32(priv
, CSR_EEPROM_REG
);
1614 if (r
& CSR_EEPROM_REG_READ_VALID_MSK
)
1616 udelay(IWL_EEPROM_ACCESS_DELAY
);
1619 if (!(r
& CSR_EEPROM_REG_READ_VALID_MSK
)) {
1620 IWL_ERROR("Time out reading EEPROM[%d]", addr
);
1624 e
[addr
/ 2] = le16_to_cpu(r
>> 16);
1629 iwl_eeprom_release_semaphore(priv
);
1633 /******************************************************************************
1635 * Misc. internal state and helper functions
1637 ******************************************************************************/
1638 #ifdef CONFIG_IWL4965_DEBUG
1641 * iwl_report_frame - dump frame to syslog during debug sessions
1643 * hack this function to show different aspects of received frames,
1644 * including selective frame dumps.
1645 * group100 parameter selects whether to show 1 out of 100 good frames.
1647 * TODO: ieee80211_hdr stuff is common to 3945 and 4965, so frame type
1648 * info output is okay, but some of this stuff (e.g. iwl_rx_frame_stats)
1649 * is 3945-specific and gives bad output for 4965. Need to split the
1650 * functionality, keep common stuff here.
1652 void iwl_report_frame(struct iwl_priv
*priv
,
1653 struct iwl_rx_packet
*pkt
,
1654 struct ieee80211_hdr
*header
, int group100
)
1657 u32 print_summary
= 0;
1658 u32 print_dump
= 0; /* set to 1 to dump all frames' contents */
1675 struct iwl_rx_frame_stats
*rx_stats
= IWL_RX_STATS(pkt
);
1676 struct iwl_rx_frame_hdr
*rx_hdr
= IWL_RX_HDR(pkt
);
1677 struct iwl_rx_frame_end
*rx_end
= IWL_RX_END(pkt
);
1678 u8
*data
= IWL_RX_DATA(pkt
);
1681 fc
= le16_to_cpu(header
->frame_control
);
1682 seq_ctl
= le16_to_cpu(header
->seq_ctrl
);
1685 channel
= le16_to_cpu(rx_hdr
->channel
);
1686 phy_flags
= le16_to_cpu(rx_hdr
->phy_flags
);
1687 rate_sym
= rx_hdr
->rate
;
1688 length
= le16_to_cpu(rx_hdr
->len
);
1690 /* end-of-frame status and timestamp */
1691 status
= le32_to_cpu(rx_end
->status
);
1692 bcn_tmr
= le32_to_cpu(rx_end
->beacon_timestamp
);
1693 tsf_low
= le64_to_cpu(rx_end
->timestamp
) & 0x0ffffffff;
1694 tsf
= le64_to_cpu(rx_end
->timestamp
);
1696 /* signal statistics */
1697 rssi
= rx_stats
->rssi
;
1698 agc
= rx_stats
->agc
;
1699 sig_avg
= le16_to_cpu(rx_stats
->sig_avg
);
1700 noise_diff
= le16_to_cpu(rx_stats
->noise_diff
);
1702 to_us
= !compare_ether_addr(header
->addr1
, priv
->mac_addr
);
1704 /* if data frame is to us and all is good,
1705 * (optionally) print summary for only 1 out of every 100 */
1706 if (to_us
&& (fc
& ~IEEE80211_FCTL_PROTECTED
) ==
1707 (IEEE80211_FCTL_FROMDS
| IEEE80211_FTYPE_DATA
)) {
1710 print_summary
= 1; /* print each frame */
1711 else if (priv
->framecnt_to_us
< 100) {
1712 priv
->framecnt_to_us
++;
1715 priv
->framecnt_to_us
= 0;
1720 /* print summary for all other frames */
1724 if (print_summary
) {
1729 title
= "100Frames";
1730 else if (fc
& IEEE80211_FCTL_RETRY
)
1732 else if (ieee80211_is_assoc_response(fc
))
1734 else if (ieee80211_is_reassoc_response(fc
))
1736 else if (ieee80211_is_probe_response(fc
)) {
1738 print_dump
= 1; /* dump frame contents */
1739 } else if (ieee80211_is_beacon(fc
)) {
1741 print_dump
= 1; /* dump frame contents */
1742 } else if (ieee80211_is_atim(fc
))
1744 else if (ieee80211_is_auth(fc
))
1746 else if (ieee80211_is_deauth(fc
))
1748 else if (ieee80211_is_disassoc(fc
))
1753 rate
= iwl_rate_index_from_plcp(rate_sym
);
1757 rate
= iwl_rates
[rate
].ieee
/ 2;
1759 /* print frame summary.
1760 * MAC addresses show just the last byte (for brevity),
1761 * but you can hack it to show more, if you'd like to. */
1763 IWL_DEBUG_RX("%s: mhd=0x%04x, dst=0x%02x, "
1764 "len=%u, rssi=%d, chnl=%d, rate=%u, \n",
1765 title
, fc
, header
->addr1
[5],
1766 length
, rssi
, channel
, rate
);
1768 /* src/dst addresses assume managed mode */
1769 IWL_DEBUG_RX("%s: 0x%04x, dst=0x%02x, "
1770 "src=0x%02x, rssi=%u, tim=%lu usec, "
1771 "phy=0x%02x, chnl=%d\n",
1772 title
, fc
, header
->addr1
[5],
1773 header
->addr3
[5], rssi
,
1774 tsf_low
- priv
->scan_start_tsf
,
1775 phy_flags
, channel
);
1779 iwl_print_hex_dump(IWL_DL_RX
, data
, length
);
1783 static void iwl_unset_hw_setting(struct iwl_priv
*priv
)
1785 if (priv
->hw_setting
.shared_virt
)
1786 pci_free_consistent(priv
->pci_dev
,
1787 sizeof(struct iwl_shared
),
1788 priv
->hw_setting
.shared_virt
,
1789 priv
->hw_setting
.shared_phys
);
1793 * iwl_supported_rate_to_ie - fill in the supported rate in IE field
1795 * return : set the bit for each supported rate insert in ie
1797 static u16
iwl_supported_rate_to_ie(u8
*ie
, u16 supported_rate
,
1798 u16 basic_rate
, int *left
)
1800 u16 ret_rates
= 0, bit
;
1805 for (bit
= 1, i
= 0; i
< IWL_RATE_COUNT
; i
++, bit
<<= 1) {
1806 if (bit
& supported_rate
) {
1808 rates
[*cnt
] = iwl_rates
[i
].ieee
|
1809 ((bit
& basic_rate
) ? 0x80 : 0x00);
1813 (*cnt
>= IWL_SUPPORTED_RATES_IE_LEN
))
1821 #ifdef CONFIG_IWL4965_HT
1822 void static iwl_set_ht_capab(struct ieee80211_hw
*hw
,
1823 struct ieee80211_ht_capability
*ht_cap
,
1828 * iwl_fill_probe_req - fill in all required fields and IE for probe request
1830 static u16
iwl_fill_probe_req(struct iwl_priv
*priv
,
1831 struct ieee80211_mgmt
*frame
,
1832 int left
, int is_direct
)
1836 u16 active_rates
, ret_rates
, cck_rates
, active_rate_basic
;
1838 /* Make sure there is enough space for the probe request,
1839 * two mandatory IEs and the data */
1845 frame
->frame_control
= cpu_to_le16(IEEE80211_STYPE_PROBE_REQ
);
1846 memcpy(frame
->da
, BROADCAST_ADDR
, ETH_ALEN
);
1847 memcpy(frame
->sa
, priv
->mac_addr
, ETH_ALEN
);
1848 memcpy(frame
->bssid
, BROADCAST_ADDR
, ETH_ALEN
);
1849 frame
->seq_ctrl
= 0;
1851 /* fill in our indirect SSID IE */
1858 pos
= &(frame
->u
.probe_req
.variable
[0]);
1859 *pos
++ = WLAN_EID_SSID
;
1862 /* fill in our direct SSID IE... */
1865 left
-= 2 + priv
->essid_len
;
1868 /* ... fill it in... */
1869 *pos
++ = WLAN_EID_SSID
;
1870 *pos
++ = priv
->essid_len
;
1871 memcpy(pos
, priv
->essid
, priv
->essid_len
);
1872 pos
+= priv
->essid_len
;
1873 len
+= 2 + priv
->essid_len
;
1876 /* fill in supported rate */
1882 /* ... fill it in... */
1883 *pos
++ = WLAN_EID_SUPP_RATES
;
1886 /* exclude 60M rate */
1887 active_rates
= priv
->rates_mask
;
1888 active_rates
&= ~IWL_RATE_60M_MASK
;
1890 active_rate_basic
= active_rates
& IWL_BASIC_RATES_MASK
;
1892 cck_rates
= IWL_CCK_RATES_MASK
& active_rates
;
1893 ret_rates
= iwl_supported_rate_to_ie(pos
, cck_rates
,
1894 active_rate_basic
, &left
);
1895 active_rates
&= ~ret_rates
;
1897 ret_rates
= iwl_supported_rate_to_ie(pos
, active_rates
,
1898 active_rate_basic
, &left
);
1899 active_rates
&= ~ret_rates
;
1903 if (active_rates
== 0)
1906 /* fill in supported extended rate */
1911 /* ... fill it in... */
1912 *pos
++ = WLAN_EID_EXT_SUPP_RATES
;
1914 iwl_supported_rate_to_ie(pos
, active_rates
,
1915 active_rate_basic
, &left
);
1919 #ifdef CONFIG_IWL4965_HT
1920 if (is_direct
&& priv
->is_ht_enabled
) {
1921 u8 use_wide_chan
= 1;
1923 if (priv
->channel_width
!= IWL_CHANNEL_WIDTH_40MHZ
)
1926 *pos
++ = WLAN_EID_HT_CAPABILITY
;
1927 *pos
++ = sizeof(struct ieee80211_ht_capability
);
1928 iwl_set_ht_capab(NULL
, (struct ieee80211_ht_capability
*)pos
,
1930 len
+= 2 + sizeof(struct ieee80211_ht_capability
);
1932 #endif /*CONFIG_IWL4965_HT */
1941 #ifdef CONFIG_IWL4965_QOS
1942 static int iwl_send_qos_params_command(struct iwl_priv
*priv
,
1943 struct iwl_qosparam_cmd
*qos
)
1946 return iwl_send_cmd_pdu(priv
, REPLY_QOS_PARAM
,
1947 sizeof(struct iwl_qosparam_cmd
), qos
);
1950 static void iwl_reset_qos(struct iwl_priv
*priv
)
1956 unsigned long flags
;
1959 spin_lock_irqsave(&priv
->lock
, flags
);
1960 priv
->qos_data
.qos_active
= 0;
1962 if (priv
->iw_mode
== IEEE80211_IF_TYPE_IBSS
) {
1963 if (priv
->qos_data
.qos_enable
)
1964 priv
->qos_data
.qos_active
= 1;
1965 if (!(priv
->active_rate
& 0xfff0)) {
1969 } else if (priv
->iw_mode
== IEEE80211_IF_TYPE_AP
) {
1970 if (priv
->qos_data
.qos_enable
)
1971 priv
->qos_data
.qos_active
= 1;
1972 } else if (!(priv
->staging_rxon
.flags
& RXON_FLG_SHORT_SLOT_MSK
)) {
1977 if (priv
->qos_data
.qos_active
)
1980 priv
->qos_data
.def_qos_parm
.ac
[0].cw_min
= cpu_to_le16(cw_min
);
1981 priv
->qos_data
.def_qos_parm
.ac
[0].cw_max
= cpu_to_le16(cw_max
);
1982 priv
->qos_data
.def_qos_parm
.ac
[0].aifsn
= aifs
;
1983 priv
->qos_data
.def_qos_parm
.ac
[0].edca_txop
= 0;
1984 priv
->qos_data
.def_qos_parm
.ac
[0].reserved1
= 0;
1986 if (priv
->qos_data
.qos_active
) {
1988 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_min
= cpu_to_le16(cw_min
);
1989 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_max
= cpu_to_le16(cw_max
);
1990 priv
->qos_data
.def_qos_parm
.ac
[i
].aifsn
= 7;
1991 priv
->qos_data
.def_qos_parm
.ac
[i
].edca_txop
= 0;
1992 priv
->qos_data
.def_qos_parm
.ac
[i
].reserved1
= 0;
1995 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_min
=
1996 cpu_to_le16((cw_min
+ 1) / 2 - 1);
1997 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_max
=
1998 cpu_to_le16(cw_max
);
1999 priv
->qos_data
.def_qos_parm
.ac
[i
].aifsn
= 2;
2001 priv
->qos_data
.def_qos_parm
.ac
[i
].edca_txop
=
2004 priv
->qos_data
.def_qos_parm
.ac
[i
].edca_txop
=
2006 priv
->qos_data
.def_qos_parm
.ac
[i
].reserved1
= 0;
2009 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_min
=
2010 cpu_to_le16((cw_min
+ 1) / 4 - 1);
2011 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_max
=
2012 cpu_to_le16((cw_max
+ 1) / 2 - 1);
2013 priv
->qos_data
.def_qos_parm
.ac
[i
].aifsn
= 2;
2014 priv
->qos_data
.def_qos_parm
.ac
[i
].reserved1
= 0;
2016 priv
->qos_data
.def_qos_parm
.ac
[i
].edca_txop
=
2019 priv
->qos_data
.def_qos_parm
.ac
[i
].edca_txop
=
2022 for (i
= 1; i
< 4; i
++) {
2023 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_min
=
2024 cpu_to_le16(cw_min
);
2025 priv
->qos_data
.def_qos_parm
.ac
[i
].cw_max
=
2026 cpu_to_le16(cw_max
);
2027 priv
->qos_data
.def_qos_parm
.ac
[i
].aifsn
= aifs
;
2028 priv
->qos_data
.def_qos_parm
.ac
[i
].edca_txop
= 0;
2029 priv
->qos_data
.def_qos_parm
.ac
[i
].reserved1
= 0;
2032 IWL_DEBUG_QOS("set QoS to default \n");
2034 spin_unlock_irqrestore(&priv
->lock
, flags
);
2037 static void iwl_activate_qos(struct iwl_priv
*priv
, u8 force
)
2039 unsigned long flags
;
2041 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
2044 if (!priv
->qos_data
.qos_enable
)
2047 spin_lock_irqsave(&priv
->lock
, flags
);
2048 priv
->qos_data
.def_qos_parm
.qos_flags
= 0;
2050 if (priv
->qos_data
.qos_cap
.q_AP
.queue_request
&&
2051 !priv
->qos_data
.qos_cap
.q_AP
.txop_request
)
2052 priv
->qos_data
.def_qos_parm
.qos_flags
|=
2053 QOS_PARAM_FLG_TXOP_TYPE_MSK
;
2054 if (priv
->qos_data
.qos_active
)
2055 priv
->qos_data
.def_qos_parm
.qos_flags
|=
2056 QOS_PARAM_FLG_UPDATE_EDCA_MSK
;
2058 #ifdef CONFIG_IWL4965_HT
2059 if (priv
->is_ht_enabled
&& priv
->current_assoc_ht
.is_ht
)
2060 priv
->qos_data
.def_qos_parm
.qos_flags
|= QOS_PARAM_FLG_TGN_MSK
;
2061 #endif /* CONFIG_IWL4965_HT */
2063 spin_unlock_irqrestore(&priv
->lock
, flags
);
2065 if (force
|| iwl_is_associated(priv
)) {
2066 IWL_DEBUG_QOS("send QoS cmd with Qos active=%d FLAGS=0x%X\n",
2067 priv
->qos_data
.qos_active
,
2068 priv
->qos_data
.def_qos_parm
.qos_flags
);
2070 iwl_send_qos_params_command(priv
,
2071 &(priv
->qos_data
.def_qos_parm
));
2075 #endif /* CONFIG_IWL4965_QOS */
2077 * Power management (not Tx power!) functions
2079 #define MSEC_TO_USEC 1024
2081 #define NOSLP __constant_cpu_to_le16(0), 0, 0
2082 #define SLP IWL_POWER_DRIVER_ALLOW_SLEEP_MSK, 0, 0
2083 #define SLP_TIMEOUT(T) __constant_cpu_to_le32((T) * MSEC_TO_USEC)
2084 #define SLP_VEC(X0, X1, X2, X3, X4) {__constant_cpu_to_le32(X0), \
2085 __constant_cpu_to_le32(X1), \
2086 __constant_cpu_to_le32(X2), \
2087 __constant_cpu_to_le32(X3), \
2088 __constant_cpu_to_le32(X4)}
2091 /* default power management (not Tx power) table values */
2093 static struct iwl_power_vec_entry range_0
[IWL_POWER_AC
] = {
2094 {{NOSLP
, SLP_TIMEOUT(0), SLP_TIMEOUT(0), SLP_VEC(0, 0, 0, 0, 0)}, 0},
2095 {{SLP
, SLP_TIMEOUT(200), SLP_TIMEOUT(500), SLP_VEC(1, 2, 3, 4, 4)}, 0},
2096 {{SLP
, SLP_TIMEOUT(200), SLP_TIMEOUT(300), SLP_VEC(2, 4, 6, 7, 7)}, 0},
2097 {{SLP
, SLP_TIMEOUT(50), SLP_TIMEOUT(100), SLP_VEC(2, 6, 9, 9, 10)}, 0},
2098 {{SLP
, SLP_TIMEOUT(50), SLP_TIMEOUT(25), SLP_VEC(2, 7, 9, 9, 10)}, 1},
2099 {{SLP
, SLP_TIMEOUT(25), SLP_TIMEOUT(25), SLP_VEC(4, 7, 10, 10, 10)}, 1}
2103 static struct iwl_power_vec_entry range_1
[IWL_POWER_AC
] = {
2104 {{NOSLP
, SLP_TIMEOUT(0), SLP_TIMEOUT(0), SLP_VEC(0, 0, 0, 0, 0)}, 0},
2105 {{SLP
, SLP_TIMEOUT(200), SLP_TIMEOUT(500),
2106 SLP_VEC(1, 2, 3, 4, 0xFF)}, 0},
2107 {{SLP
, SLP_TIMEOUT(200), SLP_TIMEOUT(300),
2108 SLP_VEC(2, 4, 6, 7, 0xFF)}, 0},
2109 {{SLP
, SLP_TIMEOUT(50), SLP_TIMEOUT(100),
2110 SLP_VEC(2, 6, 9, 9, 0xFF)}, 0},
2111 {{SLP
, SLP_TIMEOUT(50), SLP_TIMEOUT(25), SLP_VEC(2, 7, 9, 9, 0xFF)}, 0},
2112 {{SLP
, SLP_TIMEOUT(25), SLP_TIMEOUT(25),
2113 SLP_VEC(4, 7, 10, 10, 0xFF)}, 0}
2116 int iwl_power_init_handle(struct iwl_priv
*priv
)
2119 struct iwl_power_mgr
*pow_data
;
2120 int size
= sizeof(struct iwl_power_vec_entry
) * IWL_POWER_AC
;
2123 IWL_DEBUG_POWER("Initialize power \n");
2125 pow_data
= &(priv
->power_data
);
2127 memset(pow_data
, 0, sizeof(*pow_data
));
2129 pow_data
->active_index
= IWL_POWER_RANGE_0
;
2130 pow_data
->dtim_val
= 0xffff;
2132 memcpy(&pow_data
->pwr_range_0
[0], &range_0
[0], size
);
2133 memcpy(&pow_data
->pwr_range_1
[0], &range_1
[0], size
);
2135 rc
= pci_read_config_word(priv
->pci_dev
, PCI_LINK_CTRL
, &pci_pm
);
2139 struct iwl_powertable_cmd
*cmd
;
2141 IWL_DEBUG_POWER("adjust power command flags\n");
2143 for (i
= 0; i
< IWL_POWER_AC
; i
++) {
2144 cmd
= &pow_data
->pwr_range_0
[i
].cmd
;
2147 cmd
->flags
&= ~IWL_POWER_PCI_PM_MSK
;
2149 cmd
->flags
|= IWL_POWER_PCI_PM_MSK
;
2155 static int iwl_update_power_cmd(struct iwl_priv
*priv
,
2156 struct iwl_powertable_cmd
*cmd
, u32 mode
)
2161 struct iwl_power_vec_entry
*range
;
2163 struct iwl_power_mgr
*pow_data
;
2165 if (mode
> IWL_POWER_INDEX_5
) {
2166 IWL_DEBUG_POWER("Error invalid power mode \n");
2169 pow_data
= &(priv
->power_data
);
2171 if (pow_data
->active_index
== IWL_POWER_RANGE_0
)
2172 range
= &pow_data
->pwr_range_0
[0];
2174 range
= &pow_data
->pwr_range_1
[1];
2176 memcpy(cmd
, &range
[mode
].cmd
, sizeof(struct iwl_powertable_cmd
));
2178 #ifdef IWL_MAC80211_DISABLE
2179 if (priv
->assoc_network
!= NULL
) {
2180 unsigned long flags
;
2182 period
= priv
->assoc_network
->tim
.tim_period
;
2184 #endif /*IWL_MAC80211_DISABLE */
2185 skip
= range
[mode
].no_dtim
;
2194 cmd
->flags
&= ~IWL_POWER_SLEEP_OVER_DTIM_MSK
;
2196 __le32 slp_itrvl
= cmd
->sleep_interval
[IWL_POWER_VEC_SIZE
- 1];
2197 max_sleep
= (le32_to_cpu(slp_itrvl
) / period
) * period
;
2198 cmd
->flags
|= IWL_POWER_SLEEP_OVER_DTIM_MSK
;
2201 for (i
= 0; i
< IWL_POWER_VEC_SIZE
; i
++) {
2202 if (le32_to_cpu(cmd
->sleep_interval
[i
]) > max_sleep
)
2203 cmd
->sleep_interval
[i
] = cpu_to_le32(max_sleep
);
2206 IWL_DEBUG_POWER("Flags value = 0x%08X\n", cmd
->flags
);
2207 IWL_DEBUG_POWER("Tx timeout = %u\n", le32_to_cpu(cmd
->tx_data_timeout
));
2208 IWL_DEBUG_POWER("Rx timeout = %u\n", le32_to_cpu(cmd
->rx_data_timeout
));
2209 IWL_DEBUG_POWER("Sleep interval vector = { %d , %d , %d , %d , %d }\n",
2210 le32_to_cpu(cmd
->sleep_interval
[0]),
2211 le32_to_cpu(cmd
->sleep_interval
[1]),
2212 le32_to_cpu(cmd
->sleep_interval
[2]),
2213 le32_to_cpu(cmd
->sleep_interval
[3]),
2214 le32_to_cpu(cmd
->sleep_interval
[4]));
2219 static int iwl_send_power_mode(struct iwl_priv
*priv
, u32 mode
)
2221 u32 final_mode
= mode
;
2223 struct iwl_powertable_cmd cmd
;
2225 /* If on battery, set to 3,
2226 * if plugged into AC power, set to CAM ("continuously aware mode"),
2227 * else user level */
2229 case IWL_POWER_BATTERY
:
2230 final_mode
= IWL_POWER_INDEX_3
;
2233 final_mode
= IWL_POWER_MODE_CAM
;
2240 cmd
.keep_alive_beacons
= 0;
2242 iwl_update_power_cmd(priv
, &cmd
, final_mode
);
2244 rc
= iwl_send_cmd_pdu(priv
, POWER_TABLE_CMD
, sizeof(cmd
), &cmd
);
2246 if (final_mode
== IWL_POWER_MODE_CAM
)
2247 clear_bit(STATUS_POWER_PMI
, &priv
->status
);
2249 set_bit(STATUS_POWER_PMI
, &priv
->status
);
2254 int iwl_is_network_packet(struct iwl_priv
*priv
, struct ieee80211_hdr
*header
)
2256 /* Filter incoming packets to determine if they are targeted toward
2257 * this network, discarding packets coming from ourselves */
2258 switch (priv
->iw_mode
) {
2259 case IEEE80211_IF_TYPE_IBSS
: /* Header: Dest. | Source | BSSID */
2260 /* packets from our adapter are dropped (echo) */
2261 if (!compare_ether_addr(header
->addr2
, priv
->mac_addr
))
2263 /* {broad,multi}cast packets to our IBSS go through */
2264 if (is_multicast_ether_addr(header
->addr1
))
2265 return !compare_ether_addr(header
->addr3
, priv
->bssid
);
2266 /* packets to our adapter go through */
2267 return !compare_ether_addr(header
->addr1
, priv
->mac_addr
);
2268 case IEEE80211_IF_TYPE_STA
: /* Header: Dest. | AP{BSSID} | Source */
2269 /* packets from our adapter are dropped (echo) */
2270 if (!compare_ether_addr(header
->addr3
, priv
->mac_addr
))
2272 /* {broad,multi}cast packets to our BSS go through */
2273 if (is_multicast_ether_addr(header
->addr1
))
2274 return !compare_ether_addr(header
->addr2
, priv
->bssid
);
2275 /* packets to our adapter go through */
2276 return !compare_ether_addr(header
->addr1
, priv
->mac_addr
);
2282 #define TX_STATUS_ENTRY(x) case TX_STATUS_FAIL_ ## x: return #x
2284 static const char *iwl_get_tx_fail_reason(u32 status
)
2286 switch (status
& TX_STATUS_MSK
) {
2287 case TX_STATUS_SUCCESS
:
2289 TX_STATUS_ENTRY(SHORT_LIMIT
);
2290 TX_STATUS_ENTRY(LONG_LIMIT
);
2291 TX_STATUS_ENTRY(FIFO_UNDERRUN
);
2292 TX_STATUS_ENTRY(MGMNT_ABORT
);
2293 TX_STATUS_ENTRY(NEXT_FRAG
);
2294 TX_STATUS_ENTRY(LIFE_EXPIRE
);
2295 TX_STATUS_ENTRY(DEST_PS
);
2296 TX_STATUS_ENTRY(ABORTED
);
2297 TX_STATUS_ENTRY(BT_RETRY
);
2298 TX_STATUS_ENTRY(STA_INVALID
);
2299 TX_STATUS_ENTRY(FRAG_DROPPED
);
2300 TX_STATUS_ENTRY(TID_DISABLE
);
2301 TX_STATUS_ENTRY(FRAME_FLUSHED
);
2302 TX_STATUS_ENTRY(INSUFFICIENT_CF_POLL
);
2303 TX_STATUS_ENTRY(TX_LOCKED
);
2304 TX_STATUS_ENTRY(NO_BEACON_ON_RADAR
);
2311 * iwl_scan_cancel - Cancel any currently executing HW scan
2313 * NOTE: priv->mutex is not required before calling this function
2315 static int iwl_scan_cancel(struct iwl_priv
*priv
)
2317 if (!test_bit(STATUS_SCAN_HW
, &priv
->status
)) {
2318 clear_bit(STATUS_SCANNING
, &priv
->status
);
2322 if (test_bit(STATUS_SCANNING
, &priv
->status
)) {
2323 if (!test_bit(STATUS_SCAN_ABORTING
, &priv
->status
)) {
2324 IWL_DEBUG_SCAN("Queuing scan abort.\n");
2325 set_bit(STATUS_SCAN_ABORTING
, &priv
->status
);
2326 queue_work(priv
->workqueue
, &priv
->abort_scan
);
2329 IWL_DEBUG_SCAN("Scan abort already in progress.\n");
2331 return test_bit(STATUS_SCANNING
, &priv
->status
);
2338 * iwl_scan_cancel_timeout - Cancel any currently executing HW scan
2339 * @ms: amount of time to wait (in milliseconds) for scan to abort
2341 * NOTE: priv->mutex must be held before calling this function
2343 static int iwl_scan_cancel_timeout(struct iwl_priv
*priv
, unsigned long ms
)
2345 unsigned long now
= jiffies
;
2348 ret
= iwl_scan_cancel(priv
);
2350 mutex_unlock(&priv
->mutex
);
2351 while (!time_after(jiffies
, now
+ msecs_to_jiffies(ms
)) &&
2352 test_bit(STATUS_SCANNING
, &priv
->status
))
2354 mutex_lock(&priv
->mutex
);
2356 return test_bit(STATUS_SCANNING
, &priv
->status
);
2362 static void iwl_sequence_reset(struct iwl_priv
*priv
)
2364 /* Reset ieee stats */
2366 /* We don't reset the net_device_stats (ieee->stats) on
2369 priv
->last_seq_num
= -1;
2370 priv
->last_frag_num
= -1;
2371 priv
->last_packet_time
= 0;
2373 iwl_scan_cancel(priv
);
2376 #define MAX_UCODE_BEACON_INTERVAL 4096
2377 #define INTEL_CONN_LISTEN_INTERVAL __constant_cpu_to_le16(0xA)
2379 static __le16
iwl_adjust_beacon_interval(u16 beacon_val
)
2382 u16 beacon_factor
= 0;
2385 (beacon_val
+ MAX_UCODE_BEACON_INTERVAL
)
2386 / MAX_UCODE_BEACON_INTERVAL
;
2387 new_val
= beacon_val
/ beacon_factor
;
2389 return cpu_to_le16(new_val
);
2392 static void iwl_setup_rxon_timing(struct iwl_priv
*priv
)
2394 u64 interval_tm_unit
;
2396 unsigned long flags
;
2397 struct ieee80211_conf
*conf
= NULL
;
2400 conf
= ieee80211_get_hw_conf(priv
->hw
);
2402 spin_lock_irqsave(&priv
->lock
, flags
);
2403 priv
->rxon_timing
.timestamp
.dw
[1] = cpu_to_le32(priv
->timestamp1
);
2404 priv
->rxon_timing
.timestamp
.dw
[0] = cpu_to_le32(priv
->timestamp0
);
2406 priv
->rxon_timing
.listen_interval
= INTEL_CONN_LISTEN_INTERVAL
;
2408 tsf
= priv
->timestamp1
;
2409 tsf
= ((tsf
<< 32) | priv
->timestamp0
);
2411 beacon_int
= priv
->beacon_int
;
2412 spin_unlock_irqrestore(&priv
->lock
, flags
);
2414 if (priv
->iw_mode
== IEEE80211_IF_TYPE_STA
) {
2415 if (beacon_int
== 0) {
2416 priv
->rxon_timing
.beacon_interval
= cpu_to_le16(100);
2417 priv
->rxon_timing
.beacon_init_val
= cpu_to_le32(102400);
2419 priv
->rxon_timing
.beacon_interval
=
2420 cpu_to_le16(beacon_int
);
2421 priv
->rxon_timing
.beacon_interval
=
2422 iwl_adjust_beacon_interval(
2423 le16_to_cpu(priv
->rxon_timing
.beacon_interval
));
2426 priv
->rxon_timing
.atim_window
= 0;
2428 priv
->rxon_timing
.beacon_interval
=
2429 iwl_adjust_beacon_interval(conf
->beacon_int
);
2430 /* TODO: we need to get atim_window from upper stack
2431 * for now we set to 0 */
2432 priv
->rxon_timing
.atim_window
= 0;
2436 (le16_to_cpu(priv
->rxon_timing
.beacon_interval
) * 1024);
2437 result
= do_div(tsf
, interval_tm_unit
);
2438 priv
->rxon_timing
.beacon_init_val
=
2439 cpu_to_le32((u32
) ((u64
) interval_tm_unit
- result
));
2442 ("beacon interval %d beacon timer %d beacon tim %d\n",
2443 le16_to_cpu(priv
->rxon_timing
.beacon_interval
),
2444 le32_to_cpu(priv
->rxon_timing
.beacon_init_val
),
2445 le16_to_cpu(priv
->rxon_timing
.atim_window
));
2448 static int iwl_scan_initiate(struct iwl_priv
*priv
)
2450 if (priv
->iw_mode
== IEEE80211_IF_TYPE_AP
) {
2451 IWL_ERROR("APs don't scan.\n");
2455 if (!iwl_is_ready_rf(priv
)) {
2456 IWL_DEBUG_SCAN("Aborting scan due to not ready.\n");
2460 if (test_bit(STATUS_SCANNING
, &priv
->status
)) {
2461 IWL_DEBUG_SCAN("Scan already in progress.\n");
2465 if (test_bit(STATUS_SCAN_ABORTING
, &priv
->status
)) {
2466 IWL_DEBUG_SCAN("Scan request while abort pending. "
2471 IWL_DEBUG_INFO("Starting scan...\n");
2472 priv
->scan_bands
= 2;
2473 set_bit(STATUS_SCANNING
, &priv
->status
);
2474 priv
->scan_start
= jiffies
;
2475 priv
->scan_pass_start
= priv
->scan_start
;
2477 queue_work(priv
->workqueue
, &priv
->request_scan
);
2482 static int iwl_set_rxon_hwcrypto(struct iwl_priv
*priv
, int hw_decrypt
)
2484 struct iwl_rxon_cmd
*rxon
= &priv
->staging_rxon
;
2487 rxon
->filter_flags
&= ~RXON_FILTER_DIS_DECRYPT_MSK
;
2489 rxon
->filter_flags
|= RXON_FILTER_DIS_DECRYPT_MSK
;
2494 static void iwl_set_flags_for_phymode(struct iwl_priv
*priv
, u8 phymode
)
2496 if (phymode
== MODE_IEEE80211A
) {
2497 priv
->staging_rxon
.flags
&=
2498 ~(RXON_FLG_BAND_24G_MSK
| RXON_FLG_AUTO_DETECT_MSK
2499 | RXON_FLG_CCK_MSK
);
2500 priv
->staging_rxon
.flags
|= RXON_FLG_SHORT_SLOT_MSK
;
2502 /* Copied from iwl_bg_post_associate() */
2503 if (priv
->assoc_capability
& WLAN_CAPABILITY_SHORT_SLOT_TIME
)
2504 priv
->staging_rxon
.flags
|= RXON_FLG_SHORT_SLOT_MSK
;
2506 priv
->staging_rxon
.flags
&= ~RXON_FLG_SHORT_SLOT_MSK
;
2508 if (priv
->iw_mode
== IEEE80211_IF_TYPE_IBSS
)
2509 priv
->staging_rxon
.flags
&= ~RXON_FLG_SHORT_SLOT_MSK
;
2511 priv
->staging_rxon
.flags
|= RXON_FLG_BAND_24G_MSK
;
2512 priv
->staging_rxon
.flags
|= RXON_FLG_AUTO_DETECT_MSK
;
2513 priv
->staging_rxon
.flags
&= ~RXON_FLG_CCK_MSK
;
2518 * initialize rxon structure with default values from eeprom
2520 static void iwl_connection_init_rx_config(struct iwl_priv
*priv
)
2522 const struct iwl_channel_info
*ch_info
;
2524 memset(&priv
->staging_rxon
, 0, sizeof(priv
->staging_rxon
));
2526 switch (priv
->iw_mode
) {
2527 case IEEE80211_IF_TYPE_AP
:
2528 priv
->staging_rxon
.dev_type
= RXON_DEV_TYPE_AP
;
2531 case IEEE80211_IF_TYPE_STA
:
2532 priv
->staging_rxon
.dev_type
= RXON_DEV_TYPE_ESS
;
2533 priv
->staging_rxon
.filter_flags
= RXON_FILTER_ACCEPT_GRP_MSK
;
2536 case IEEE80211_IF_TYPE_IBSS
:
2537 priv
->staging_rxon
.dev_type
= RXON_DEV_TYPE_IBSS
;
2538 priv
->staging_rxon
.flags
= RXON_FLG_SHORT_PREAMBLE_MSK
;
2539 priv
->staging_rxon
.filter_flags
= RXON_FILTER_BCON_AWARE_MSK
|
2540 RXON_FILTER_ACCEPT_GRP_MSK
;
2543 case IEEE80211_IF_TYPE_MNTR
:
2544 priv
->staging_rxon
.dev_type
= RXON_DEV_TYPE_SNIFFER
;
2545 priv
->staging_rxon
.filter_flags
= RXON_FILTER_PROMISC_MSK
|
2546 RXON_FILTER_CTL2HOST_MSK
| RXON_FILTER_ACCEPT_GRP_MSK
;
2551 /* TODO: Figure out when short_preamble would be set and cache from
2553 if (!hw_to_local(priv
->hw
)->short_preamble
)
2554 priv
->staging_rxon
.flags
&= ~RXON_FLG_SHORT_PREAMBLE_MSK
;
2556 priv
->staging_rxon
.flags
|= RXON_FLG_SHORT_PREAMBLE_MSK
;
2559 ch_info
= iwl_get_channel_info(priv
, priv
->phymode
,
2560 le16_to_cpu(priv
->staging_rxon
.channel
));
2563 ch_info
= &priv
->channel_info
[0];
2566 * in some case A channels are all non IBSS
2567 * in this case force B/G channel
2569 if ((priv
->iw_mode
== IEEE80211_IF_TYPE_IBSS
) &&
2570 !(is_channel_ibss(ch_info
)))
2571 ch_info
= &priv
->channel_info
[0];
2573 priv
->staging_rxon
.channel
= cpu_to_le16(ch_info
->channel
);
2574 if (is_channel_a_band(ch_info
))
2575 priv
->phymode
= MODE_IEEE80211A
;
2577 priv
->phymode
= MODE_IEEE80211G
;
2579 iwl_set_flags_for_phymode(priv
, priv
->phymode
);
2581 priv
->staging_rxon
.ofdm_basic_rates
=
2582 (IWL_OFDM_RATES_MASK
>> IWL_FIRST_OFDM_RATE
) & 0xFF;
2583 priv
->staging_rxon
.cck_basic_rates
=
2584 (IWL_CCK_RATES_MASK
>> IWL_FIRST_CCK_RATE
) & 0xF;
2586 priv
->staging_rxon
.flags
&= ~(RXON_FLG_CHANNEL_MODE_MIXED_MSK
|
2587 RXON_FLG_CHANNEL_MODE_PURE_40_MSK
);
2588 memcpy(priv
->staging_rxon
.node_addr
, priv
->mac_addr
, ETH_ALEN
);
2589 memcpy(priv
->staging_rxon
.wlap_bssid_addr
, priv
->mac_addr
, ETH_ALEN
);
2590 priv
->staging_rxon
.ofdm_ht_single_stream_basic_rates
= 0xff;
2591 priv
->staging_rxon
.ofdm_ht_dual_stream_basic_rates
= 0xff;
2592 iwl4965_set_rxon_chain(priv
);
2595 static int iwl_set_mode(struct iwl_priv
*priv
, int mode
)
2597 if (!iwl_is_ready_rf(priv
))
2600 if (mode
== IEEE80211_IF_TYPE_IBSS
) {
2601 const struct iwl_channel_info
*ch_info
;
2603 ch_info
= iwl_get_channel_info(priv
,
2605 le16_to_cpu(priv
->staging_rxon
.channel
));
2607 if (!ch_info
|| !is_channel_ibss(ch_info
)) {
2608 IWL_ERROR("channel %d not IBSS channel\n",
2609 le16_to_cpu(priv
->staging_rxon
.channel
));
2614 cancel_delayed_work(&priv
->scan_check
);
2615 if (iwl_scan_cancel_timeout(priv
, 100)) {
2616 IWL_WARNING("Aborted scan still in progress after 100ms\n");
2617 IWL_DEBUG_MAC80211("leaving - scan abort failed.\n");
2621 priv
->iw_mode
= mode
;
2623 iwl_connection_init_rx_config(priv
);
2624 memcpy(priv
->staging_rxon
.node_addr
, priv
->mac_addr
, ETH_ALEN
);
2626 iwl_clear_stations_table(priv
);
2628 iwl_commit_rxon(priv
);
2633 static void iwl_build_tx_cmd_hwcrypto(struct iwl_priv
*priv
,
2634 struct ieee80211_tx_control
*ctl
,
2635 struct iwl_cmd
*cmd
,
2636 struct sk_buff
*skb_frag
,
2639 struct iwl_hw_key
*keyinfo
= &priv
->stations
[ctl
->key_idx
].keyinfo
;
2641 switch (keyinfo
->alg
) {
2643 cmd
->cmd
.tx
.sec_ctl
= TX_CMD_SEC_CCM
;
2644 memcpy(cmd
->cmd
.tx
.key
, keyinfo
->key
, keyinfo
->keylen
);
2645 IWL_DEBUG_TX("tx_cmd with aes hwcrypto\n");
2650 cmd
->cmd
.tx
.sec_ctl
= TX_CMD_SEC_TKIP
;
2653 memcpy(cmd
->cmd
.tx
.tkip_mic
.byte
, skb_frag
->tail
- 8,
2656 memset(cmd
->cmd
.tx
.tkip_mic
.byte
, 0, 8);
2661 cmd
->cmd
.tx
.sec_ctl
= TX_CMD_SEC_WEP
|
2662 (ctl
->key_idx
& TX_CMD_SEC_MSK
) << TX_CMD_SEC_SHIFT
;
2664 if (keyinfo
->keylen
== 13)
2665 cmd
->cmd
.tx
.sec_ctl
|= TX_CMD_SEC_KEY128
;
2667 memcpy(&cmd
->cmd
.tx
.key
[3], keyinfo
->key
, keyinfo
->keylen
);
2669 IWL_DEBUG_TX("Configuring packet for WEP encryption "
2670 "with key %d\n", ctl
->key_idx
);
2674 printk(KERN_ERR
"Unknown encode alg %d\n", keyinfo
->alg
);
2680 * handle build REPLY_TX command notification.
2682 static void iwl_build_tx_cmd_basic(struct iwl_priv
*priv
,
2683 struct iwl_cmd
*cmd
,
2684 struct ieee80211_tx_control
*ctrl
,
2685 struct ieee80211_hdr
*hdr
,
2686 int is_unicast
, u8 std_id
)
2689 u16 fc
= le16_to_cpu(hdr
->frame_control
);
2690 __le32 tx_flags
= cmd
->cmd
.tx
.tx_flags
;
2692 cmd
->cmd
.tx
.stop_time
.life_time
= TX_CMD_LIFE_TIME_INFINITE
;
2693 if (!(ctrl
->flags
& IEEE80211_TXCTL_NO_ACK
)) {
2694 tx_flags
|= TX_CMD_FLG_ACK_MSK
;
2695 if ((fc
& IEEE80211_FCTL_FTYPE
) == IEEE80211_FTYPE_MGMT
)
2696 tx_flags
|= TX_CMD_FLG_SEQ_CTL_MSK
;
2697 if (ieee80211_is_probe_response(fc
) &&
2698 !(le16_to_cpu(hdr
->seq_ctrl
) & 0xf))
2699 tx_flags
|= TX_CMD_FLG_TSF_MSK
;
2701 tx_flags
&= (~TX_CMD_FLG_ACK_MSK
);
2702 tx_flags
|= TX_CMD_FLG_SEQ_CTL_MSK
;
2705 cmd
->cmd
.tx
.sta_id
= std_id
;
2706 if (ieee80211_get_morefrag(hdr
))
2707 tx_flags
|= TX_CMD_FLG_MORE_FRAG_MSK
;
2709 qc
= ieee80211_get_qos_ctrl(hdr
);
2711 cmd
->cmd
.tx
.tid_tspec
= (u8
) (le16_to_cpu(*qc
) & 0xf);
2712 tx_flags
&= ~TX_CMD_FLG_SEQ_CTL_MSK
;
2714 tx_flags
|= TX_CMD_FLG_SEQ_CTL_MSK
;
2716 if (ctrl
->flags
& IEEE80211_TXCTL_USE_RTS_CTS
) {
2717 tx_flags
|= TX_CMD_FLG_RTS_MSK
;
2718 tx_flags
&= ~TX_CMD_FLG_CTS_MSK
;
2719 } else if (ctrl
->flags
& IEEE80211_TXCTL_USE_CTS_PROTECT
) {
2720 tx_flags
&= ~TX_CMD_FLG_RTS_MSK
;
2721 tx_flags
|= TX_CMD_FLG_CTS_MSK
;
2724 if ((tx_flags
& TX_CMD_FLG_RTS_MSK
) || (tx_flags
& TX_CMD_FLG_CTS_MSK
))
2725 tx_flags
|= TX_CMD_FLG_FULL_TXOP_PROT_MSK
;
2727 tx_flags
&= ~(TX_CMD_FLG_ANT_SEL_MSK
);
2728 if ((fc
& IEEE80211_FCTL_FTYPE
) == IEEE80211_FTYPE_MGMT
) {
2729 if ((fc
& IEEE80211_FCTL_STYPE
) == IEEE80211_STYPE_ASSOC_REQ
||
2730 (fc
& IEEE80211_FCTL_STYPE
) == IEEE80211_STYPE_REASSOC_REQ
)
2731 cmd
->cmd
.tx
.timeout
.pm_frame_timeout
= cpu_to_le16(3);
2733 cmd
->cmd
.tx
.timeout
.pm_frame_timeout
= cpu_to_le16(2);
2735 cmd
->cmd
.tx
.timeout
.pm_frame_timeout
= 0;
2737 cmd
->cmd
.tx
.driver_txop
= 0;
2738 cmd
->cmd
.tx
.tx_flags
= tx_flags
;
2739 cmd
->cmd
.tx
.next_frame_len
= 0;
2742 static int iwl_get_sta_id(struct iwl_priv
*priv
, struct ieee80211_hdr
*hdr
)
2745 u16 fc
= le16_to_cpu(hdr
->frame_control
);
2746 DECLARE_MAC_BUF(mac
);
2748 /* If this frame is broadcast or not data then use the broadcast
2750 if (((fc
& IEEE80211_FCTL_FTYPE
) != IEEE80211_FTYPE_DATA
) ||
2751 is_multicast_ether_addr(hdr
->addr1
))
2752 return priv
->hw_setting
.bcast_sta_id
;
2754 switch (priv
->iw_mode
) {
2756 /* If this frame is part of a BSS network (we're a station), then
2757 * we use the AP's station id */
2758 case IEEE80211_IF_TYPE_STA
:
2761 /* If we are an AP, then find the station, or use BCAST */
2762 case IEEE80211_IF_TYPE_AP
:
2763 sta_id
= iwl_hw_find_station(priv
, hdr
->addr1
);
2764 if (sta_id
!= IWL_INVALID_STATION
)
2766 return priv
->hw_setting
.bcast_sta_id
;
2768 /* If this frame is part of a IBSS network, then we use the
2769 * target specific station id */
2770 case IEEE80211_IF_TYPE_IBSS
:
2771 sta_id
= iwl_hw_find_station(priv
, hdr
->addr1
);
2772 if (sta_id
!= IWL_INVALID_STATION
)
2775 sta_id
= iwl_add_station(priv
, hdr
->addr1
, 0, CMD_ASYNC
);
2777 if (sta_id
!= IWL_INVALID_STATION
)
2780 IWL_DEBUG_DROP("Station %s not in station map. "
2781 "Defaulting to broadcast...\n",
2782 print_mac(mac
, hdr
->addr1
));
2783 iwl_print_hex_dump(IWL_DL_DROP
, (u8
*) hdr
, sizeof(*hdr
));
2784 return priv
->hw_setting
.bcast_sta_id
;
2787 IWL_WARNING("Unknown mode of operation: %d", priv
->iw_mode
);
2788 return priv
->hw_setting
.bcast_sta_id
;
2793 * start REPLY_TX command process
2795 static int iwl_tx_skb(struct iwl_priv
*priv
,
2796 struct sk_buff
*skb
, struct ieee80211_tx_control
*ctl
)
2798 struct ieee80211_hdr
*hdr
= (struct ieee80211_hdr
*)skb
->data
;
2799 struct iwl_tfd_frame
*tfd
;
2801 int txq_id
= ctl
->queue
;
2802 struct iwl_tx_queue
*txq
= NULL
;
2803 struct iwl_queue
*q
= NULL
;
2804 dma_addr_t phys_addr
;
2805 dma_addr_t txcmd_phys
;
2806 struct iwl_cmd
*out_cmd
= NULL
;
2807 u16 len
, idx
, len_org
;
2808 u8 id
, hdr_len
, unicast
;
2813 u8 wait_write_ptr
= 0;
2814 unsigned long flags
;
2817 spin_lock_irqsave(&priv
->lock
, flags
);
2818 if (iwl_is_rfkill(priv
)) {
2819 IWL_DEBUG_DROP("Dropping - RF KILL\n");
2823 if (!priv
->interface_id
) {
2824 IWL_DEBUG_DROP("Dropping - !priv->interface_id\n");
2828 if ((ctl
->tx_rate
& 0xFF) == IWL_INVALID_RATE
) {
2829 IWL_ERROR("ERROR: No TX rate available.\n");
2833 unicast
= !is_multicast_ether_addr(hdr
->addr1
);
2836 fc
= le16_to_cpu(hdr
->frame_control
);
2838 #ifdef CONFIG_IWL4965_DEBUG
2839 if (ieee80211_is_auth(fc
))
2840 IWL_DEBUG_TX("Sending AUTH frame\n");
2841 else if (ieee80211_is_assoc_request(fc
))
2842 IWL_DEBUG_TX("Sending ASSOC frame\n");
2843 else if (ieee80211_is_reassoc_request(fc
))
2844 IWL_DEBUG_TX("Sending REASSOC frame\n");
2847 if (!iwl_is_associated(priv
) &&
2848 ((fc
& IEEE80211_FCTL_FTYPE
) == IEEE80211_FTYPE_DATA
)) {
2849 IWL_DEBUG_DROP("Dropping - !iwl_is_associated\n");
2853 spin_unlock_irqrestore(&priv
->lock
, flags
);
2855 hdr_len
= ieee80211_get_hdrlen(fc
);
2856 sta_id
= iwl_get_sta_id(priv
, hdr
);
2857 if (sta_id
== IWL_INVALID_STATION
) {
2858 DECLARE_MAC_BUF(mac
);
2860 IWL_DEBUG_DROP("Dropping - INVALID STATION: %s\n",
2861 print_mac(mac
, hdr
->addr1
));
2865 IWL_DEBUG_RATE("station Id %d\n", sta_id
);
2867 qc
= ieee80211_get_qos_ctrl(hdr
);
2869 u8 tid
= (u8
)(le16_to_cpu(*qc
) & 0xf);
2870 seq_number
= priv
->stations
[sta_id
].tid
[tid
].seq_number
&
2872 hdr
->seq_ctrl
= cpu_to_le16(seq_number
) |
2874 __constant_cpu_to_le16(IEEE80211_SCTL_FRAG
));
2876 #ifdef CONFIG_IWL4965_HT
2877 #ifdef CONFIG_IWL4965_HT_AGG
2878 /* aggregation is on for this <sta,tid> */
2879 if (ctl
->flags
& IEEE80211_TXCTL_HT_MPDU_AGG
)
2880 txq_id
= priv
->stations
[sta_id
].tid
[tid
].agg
.txq_id
;
2881 #endif /* CONFIG_IWL4965_HT_AGG */
2882 #endif /* CONFIG_IWL4965_HT */
2884 txq
= &priv
->txq
[txq_id
];
2887 spin_lock_irqsave(&priv
->lock
, flags
);
2889 tfd
= &txq
->bd
[q
->write_ptr
];
2890 memset(tfd
, 0, sizeof(*tfd
));
2891 control_flags
= (u32
*) tfd
;
2892 idx
= get_cmd_index(q
, q
->write_ptr
, 0);
2894 memset(&(txq
->txb
[q
->write_ptr
]), 0, sizeof(struct iwl_tx_info
));
2895 txq
->txb
[q
->write_ptr
].skb
[0] = skb
;
2896 memcpy(&(txq
->txb
[q
->write_ptr
].status
.control
),
2897 ctl
, sizeof(struct ieee80211_tx_control
));
2898 out_cmd
= &txq
->cmd
[idx
];
2899 memset(&out_cmd
->hdr
, 0, sizeof(out_cmd
->hdr
));
2900 memset(&out_cmd
->cmd
.tx
, 0, sizeof(out_cmd
->cmd
.tx
));
2901 out_cmd
->hdr
.cmd
= REPLY_TX
;
2902 out_cmd
->hdr
.sequence
= cpu_to_le16((u16
)(QUEUE_TO_SEQ(txq_id
) |
2903 INDEX_TO_SEQ(q
->write_ptr
)));
2904 /* copy frags header */
2905 memcpy(out_cmd
->cmd
.tx
.hdr
, hdr
, hdr_len
);
2907 /* hdr = (struct ieee80211_hdr *)out_cmd->cmd.tx.hdr; */
2908 len
= priv
->hw_setting
.tx_cmd_len
+
2909 sizeof(struct iwl_cmd_header
) + hdr_len
;
2912 len
= (len
+ 3) & ~3;
2919 txcmd_phys
= txq
->dma_addr_cmd
+ sizeof(struct iwl_cmd
) * idx
+
2920 offsetof(struct iwl_cmd
, hdr
);
2922 iwl_hw_txq_attach_buf_to_tfd(priv
, tfd
, txcmd_phys
, len
);
2924 if (!(ctl
->flags
& IEEE80211_TXCTL_DO_NOT_ENCRYPT
))
2925 iwl_build_tx_cmd_hwcrypto(priv
, ctl
, out_cmd
, skb
, 0);
2927 /* 802.11 null functions have no payload... */
2928 len
= skb
->len
- hdr_len
;
2930 phys_addr
= pci_map_single(priv
->pci_dev
, skb
->data
+ hdr_len
,
2931 len
, PCI_DMA_TODEVICE
);
2932 iwl_hw_txq_attach_buf_to_tfd(priv
, tfd
, phys_addr
, len
);
2936 out_cmd
->cmd
.tx
.tx_flags
|= TX_CMD_FLG_MH_PAD_MSK
;
2938 len
= (u16
)skb
->len
;
2939 out_cmd
->cmd
.tx
.len
= cpu_to_le16(len
);
2941 /* TODO need this for burst mode later on */
2942 iwl_build_tx_cmd_basic(priv
, out_cmd
, ctl
, hdr
, unicast
, sta_id
);
2944 /* set is_hcca to 0; it probably will never be implemented */
2945 iwl_hw_build_tx_cmd_rate(priv
, out_cmd
, ctl
, hdr
, sta_id
, 0);
2947 iwl4965_tx_cmd(priv
, out_cmd
, sta_id
, txcmd_phys
,
2948 hdr
, hdr_len
, ctl
, NULL
);
2950 if (!ieee80211_get_morefrag(hdr
)) {
2951 txq
->need_update
= 1;
2953 u8 tid
= (u8
)(le16_to_cpu(*qc
) & 0xf);
2954 priv
->stations
[sta_id
].tid
[tid
].seq_number
= seq_number
;
2958 txq
->need_update
= 0;
2961 iwl_print_hex_dump(IWL_DL_TX
, out_cmd
->cmd
.payload
,
2962 sizeof(out_cmd
->cmd
.tx
));
2964 iwl_print_hex_dump(IWL_DL_TX
, (u8
*)out_cmd
->cmd
.tx
.hdr
,
2965 ieee80211_get_hdrlen(fc
));
2967 iwl4965_tx_queue_update_wr_ptr(priv
, txq
, len
);
2969 q
->write_ptr
= iwl_queue_inc_wrap(q
->write_ptr
, q
->n_bd
);
2970 rc
= iwl_tx_queue_update_write_ptr(priv
, txq
);
2971 spin_unlock_irqrestore(&priv
->lock
, flags
);
2976 if ((iwl_queue_space(q
) < q
->high_mark
)
2977 && priv
->mac80211_registered
) {
2978 if (wait_write_ptr
) {
2979 spin_lock_irqsave(&priv
->lock
, flags
);
2980 txq
->need_update
= 1;
2981 iwl_tx_queue_update_write_ptr(priv
, txq
);
2982 spin_unlock_irqrestore(&priv
->lock
, flags
);
2985 ieee80211_stop_queue(priv
->hw
, ctl
->queue
);
2991 spin_unlock_irqrestore(&priv
->lock
, flags
);
2996 static void iwl_set_rate(struct iwl_priv
*priv
)
2998 const struct ieee80211_hw_mode
*hw
= NULL
;
2999 struct ieee80211_rate
*rate
;
3002 hw
= iwl_get_hw_mode(priv
, priv
->phymode
);
3004 IWL_ERROR("Failed to set rate: unable to get hw mode\n");
3008 priv
->active_rate
= 0;
3009 priv
->active_rate_basic
= 0;
3011 IWL_DEBUG_RATE("Setting rates for 802.11%c\n",
3012 hw
->mode
== MODE_IEEE80211A
?
3013 'a' : ((hw
->mode
== MODE_IEEE80211B
) ? 'b' : 'g'));
3015 for (i
= 0; i
< hw
->num_rates
; i
++) {
3016 rate
= &(hw
->rates
[i
]);
3017 if ((rate
->val
< IWL_RATE_COUNT
) &&
3018 (rate
->flags
& IEEE80211_RATE_SUPPORTED
)) {
3019 IWL_DEBUG_RATE("Adding rate index %d (plcp %d)%s\n",
3020 rate
->val
, iwl_rates
[rate
->val
].plcp
,
3021 (rate
->flags
& IEEE80211_RATE_BASIC
) ?
3023 priv
->active_rate
|= (1 << rate
->val
);
3024 if (rate
->flags
& IEEE80211_RATE_BASIC
)
3025 priv
->active_rate_basic
|= (1 << rate
->val
);
3027 IWL_DEBUG_RATE("Not adding rate %d (plcp %d)\n",
3028 rate
->val
, iwl_rates
[rate
->val
].plcp
);
3031 IWL_DEBUG_RATE("Set active_rate = %0x, active_rate_basic = %0x\n",
3032 priv
->active_rate
, priv
->active_rate_basic
);
3035 * If a basic rate is configured, then use it (adding IWL_RATE_1M_MASK)
3036 * otherwise set it to the default of all CCK rates and 6, 12, 24 for
3039 if (priv
->active_rate_basic
& IWL_CCK_BASIC_RATES_MASK
)
3040 priv
->staging_rxon
.cck_basic_rates
=
3041 ((priv
->active_rate_basic
&
3042 IWL_CCK_RATES_MASK
) >> IWL_FIRST_CCK_RATE
) & 0xF;
3044 priv
->staging_rxon
.cck_basic_rates
=
3045 (IWL_CCK_BASIC_RATES_MASK
>> IWL_FIRST_CCK_RATE
) & 0xF;
3047 if (priv
->active_rate_basic
& IWL_OFDM_BASIC_RATES_MASK
)
3048 priv
->staging_rxon
.ofdm_basic_rates
=
3049 ((priv
->active_rate_basic
&
3050 (IWL_OFDM_BASIC_RATES_MASK
| IWL_RATE_6M_MASK
)) >>
3051 IWL_FIRST_OFDM_RATE
) & 0xFF;
3053 priv
->staging_rxon
.ofdm_basic_rates
=
3054 (IWL_OFDM_BASIC_RATES_MASK
>> IWL_FIRST_OFDM_RATE
) & 0xFF;
3057 static void iwl_radio_kill_sw(struct iwl_priv
*priv
, int disable_radio
)
3059 unsigned long flags
;
3061 if (!!disable_radio
== test_bit(STATUS_RF_KILL_SW
, &priv
->status
))
3064 IWL_DEBUG_RF_KILL("Manual SW RF KILL set to: RADIO %s\n",
3065 disable_radio
? "OFF" : "ON");
3067 if (disable_radio
) {
3068 iwl_scan_cancel(priv
);
3069 /* FIXME: This is a workaround for AP */
3070 if (priv
->iw_mode
!= IEEE80211_IF_TYPE_AP
) {
3071 spin_lock_irqsave(&priv
->lock
, flags
);
3072 iwl_write32(priv
, CSR_UCODE_DRV_GP1_SET
,
3073 CSR_UCODE_SW_BIT_RFKILL
);
3074 spin_unlock_irqrestore(&priv
->lock
, flags
);
3075 iwl_send_card_state(priv
, CARD_STATE_CMD_DISABLE
, 0);
3076 set_bit(STATUS_RF_KILL_SW
, &priv
->status
);
3081 spin_lock_irqsave(&priv
->lock
, flags
);
3082 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
, CSR_UCODE_SW_BIT_RFKILL
);
3084 clear_bit(STATUS_RF_KILL_SW
, &priv
->status
);
3085 spin_unlock_irqrestore(&priv
->lock
, flags
);
3090 spin_lock_irqsave(&priv
->lock
, flags
);
3091 iwl_read32(priv
, CSR_UCODE_DRV_GP1
);
3092 if (!iwl_grab_nic_access(priv
))
3093 iwl_release_nic_access(priv
);
3094 spin_unlock_irqrestore(&priv
->lock
, flags
);
3096 if (test_bit(STATUS_RF_KILL_HW
, &priv
->status
)) {
3097 IWL_DEBUG_RF_KILL("Can not turn radio back on - "
3098 "disabled by HW switch\n");
3102 queue_work(priv
->workqueue
, &priv
->restart
);
3106 void iwl_set_decrypted_flag(struct iwl_priv
*priv
, struct sk_buff
*skb
,
3107 u32 decrypt_res
, struct ieee80211_rx_status
*stats
)
3110 le16_to_cpu(((struct ieee80211_hdr
*)skb
->data
)->frame_control
);
3112 if (priv
->active_rxon
.filter_flags
& RXON_FILTER_DIS_DECRYPT_MSK
)
3115 if (!(fc
& IEEE80211_FCTL_PROTECTED
))
3118 IWL_DEBUG_RX("decrypt_res:0x%x\n", decrypt_res
);
3119 switch (decrypt_res
& RX_RES_STATUS_SEC_TYPE_MSK
) {
3120 case RX_RES_STATUS_SEC_TYPE_TKIP
:
3121 if ((decrypt_res
& RX_RES_STATUS_DECRYPT_TYPE_MSK
) ==
3122 RX_RES_STATUS_BAD_ICV_MIC
)
3123 stats
->flag
|= RX_FLAG_MMIC_ERROR
;
3124 case RX_RES_STATUS_SEC_TYPE_WEP
:
3125 case RX_RES_STATUS_SEC_TYPE_CCMP
:
3126 if ((decrypt_res
& RX_RES_STATUS_DECRYPT_TYPE_MSK
) ==
3127 RX_RES_STATUS_DECRYPT_OK
) {
3128 IWL_DEBUG_RX("hw decrypt successfully!!!\n");
3129 stats
->flag
|= RX_FLAG_DECRYPTED
;
3138 void iwl_handle_data_packet_monitor(struct iwl_priv
*priv
,
3139 struct iwl_rx_mem_buffer
*rxb
,
3140 void *data
, short len
,
3141 struct ieee80211_rx_status
*stats
,
3144 struct iwl_rt_rx_hdr
*iwl_rt
;
3146 /* First cache any information we need before we overwrite
3147 * the information provided in the skb from the hardware */
3148 s8 signal
= stats
->ssi
;
3150 int rate
= stats
->rate
;
3151 u64 tsf
= stats
->mactime
;
3152 __le16 phy_flags_hw
= cpu_to_le16(phy_flags
);
3154 /* We received data from the HW, so stop the watchdog */
3155 if (len
> IWL_RX_BUF_SIZE
- sizeof(*iwl_rt
)) {
3156 IWL_DEBUG_DROP("Dropping too large packet in monitor\n");
3160 /* copy the frame data to write after where the radiotap header goes */
3161 iwl_rt
= (void *)rxb
->skb
->data
;
3162 memmove(iwl_rt
->payload
, data
, len
);
3164 iwl_rt
->rt_hdr
.it_version
= PKTHDR_RADIOTAP_VERSION
;
3165 iwl_rt
->rt_hdr
.it_pad
= 0; /* always good to zero */
3167 /* total header + data */
3168 iwl_rt
->rt_hdr
.it_len
= cpu_to_le16(sizeof(*iwl_rt
));
3170 /* Set the size of the skb to the size of the frame */
3171 skb_put(rxb
->skb
, sizeof(*iwl_rt
) + len
);
3173 /* Big bitfield of all the fields we provide in radiotap */
3174 iwl_rt
->rt_hdr
.it_present
=
3175 cpu_to_le32((1 << IEEE80211_RADIOTAP_TSFT
) |
3176 (1 << IEEE80211_RADIOTAP_FLAGS
) |
3177 (1 << IEEE80211_RADIOTAP_RATE
) |
3178 (1 << IEEE80211_RADIOTAP_CHANNEL
) |
3179 (1 << IEEE80211_RADIOTAP_DBM_ANTSIGNAL
) |
3180 (1 << IEEE80211_RADIOTAP_DBM_ANTNOISE
) |
3181 (1 << IEEE80211_RADIOTAP_ANTENNA
));
3183 /* Zero the flags, we'll add to them as we go */
3184 iwl_rt
->rt_flags
= 0;
3186 iwl_rt
->rt_tsf
= cpu_to_le64(tsf
);
3188 /* Convert to dBm */
3189 iwl_rt
->rt_dbmsignal
= signal
;
3190 iwl_rt
->rt_dbmnoise
= noise
;
3192 /* Convert the channel frequency and set the flags */
3193 iwl_rt
->rt_channelMHz
= cpu_to_le16(stats
->freq
);
3194 if (!(phy_flags_hw
& RX_RES_PHY_FLAGS_BAND_24_MSK
))
3195 iwl_rt
->rt_chbitmask
=
3196 cpu_to_le16((IEEE80211_CHAN_OFDM
| IEEE80211_CHAN_5GHZ
));
3197 else if (phy_flags_hw
& RX_RES_PHY_FLAGS_MOD_CCK_MSK
)
3198 iwl_rt
->rt_chbitmask
=
3199 cpu_to_le16((IEEE80211_CHAN_CCK
| IEEE80211_CHAN_2GHZ
));
3201 iwl_rt
->rt_chbitmask
=
3202 cpu_to_le16((IEEE80211_CHAN_OFDM
| IEEE80211_CHAN_2GHZ
));
3204 rate
= iwl_rate_index_from_plcp(rate
);
3206 iwl_rt
->rt_rate
= 0;
3208 iwl_rt
->rt_rate
= iwl_rates
[rate
].ieee
;
3210 /* antenna number */
3211 iwl_rt
->rt_antenna
=
3212 le16_to_cpu(phy_flags_hw
& RX_RES_PHY_FLAGS_ANTENNA_MSK
) >> 4;
3214 /* set the preamble flag if we have it */
3215 if (phy_flags_hw
& RX_RES_PHY_FLAGS_SHORT_PREAMBLE_MSK
)
3216 iwl_rt
->rt_flags
|= IEEE80211_RADIOTAP_F_SHORTPRE
;
3218 IWL_DEBUG_RX("Rx packet of %d bytes.\n", rxb
->skb
->len
);
3220 stats
->flag
|= RX_FLAG_RADIOTAP
;
3221 ieee80211_rx_irqsafe(priv
->hw
, rxb
->skb
, stats
);
3226 #define IWL_PACKET_RETRY_TIME HZ
3228 int is_duplicate_packet(struct iwl_priv
*priv
, struct ieee80211_hdr
*header
)
3230 u16 sc
= le16_to_cpu(header
->seq_ctrl
);
3231 u16 seq
= (sc
& IEEE80211_SCTL_SEQ
) >> 4;
3232 u16 frag
= sc
& IEEE80211_SCTL_FRAG
;
3233 u16
*last_seq
, *last_frag
;
3234 unsigned long *last_time
;
3236 switch (priv
->iw_mode
) {
3237 case IEEE80211_IF_TYPE_IBSS
:{
3238 struct list_head
*p
;
3239 struct iwl_ibss_seq
*entry
= NULL
;
3240 u8
*mac
= header
->addr2
;
3241 int index
= mac
[5] & (IWL_IBSS_MAC_HASH_SIZE
- 1);
3243 __list_for_each(p
, &priv
->ibss_mac_hash
[index
]) {
3244 entry
= list_entry(p
, struct iwl_ibss_seq
, list
);
3245 if (!compare_ether_addr(entry
->mac
, mac
))
3248 if (p
== &priv
->ibss_mac_hash
[index
]) {
3249 entry
= kzalloc(sizeof(*entry
), GFP_ATOMIC
);
3251 IWL_ERROR("Cannot malloc new mac entry\n");
3254 memcpy(entry
->mac
, mac
, ETH_ALEN
);
3255 entry
->seq_num
= seq
;
3256 entry
->frag_num
= frag
;
3257 entry
->packet_time
= jiffies
;
3258 list_add(&entry
->list
, &priv
->ibss_mac_hash
[index
]);
3261 last_seq
= &entry
->seq_num
;
3262 last_frag
= &entry
->frag_num
;
3263 last_time
= &entry
->packet_time
;
3266 case IEEE80211_IF_TYPE_STA
:
3267 last_seq
= &priv
->last_seq_num
;
3268 last_frag
= &priv
->last_frag_num
;
3269 last_time
= &priv
->last_packet_time
;
3274 if ((*last_seq
== seq
) &&
3275 time_after(*last_time
+ IWL_PACKET_RETRY_TIME
, jiffies
)) {
3276 if (*last_frag
== frag
)
3278 if (*last_frag
+ 1 != frag
)
3279 /* out-of-order fragment */
3285 *last_time
= jiffies
;
3292 #ifdef CONFIG_IWL4965_SPECTRUM_MEASUREMENT
3294 #include "iwl-spectrum.h"
3296 #define BEACON_TIME_MASK_LOW 0x00FFFFFF
3297 #define BEACON_TIME_MASK_HIGH 0xFF000000
3298 #define TIME_UNIT 1024
3301 * extended beacon time format
3302 * time in usec will be changed into a 32-bit value in 8:24 format
3303 * the high 1 byte is the beacon counts
3304 * the lower 3 bytes is the time in usec within one beacon interval
3307 static u32
iwl_usecs_to_beacons(u32 usec
, u32 beacon_interval
)
3311 u32 interval
= beacon_interval
* 1024;
3313 if (!interval
|| !usec
)
3316 quot
= (usec
/ interval
) & (BEACON_TIME_MASK_HIGH
>> 24);
3317 rem
= (usec
% interval
) & BEACON_TIME_MASK_LOW
;
3319 return (quot
<< 24) + rem
;
3322 /* base is usually what we get from ucode with each received frame,
3323 * the same as HW timer counter counting down
3326 static __le32
iwl_add_beacon_time(u32 base
, u32 addon
, u32 beacon_interval
)
3328 u32 base_low
= base
& BEACON_TIME_MASK_LOW
;
3329 u32 addon_low
= addon
& BEACON_TIME_MASK_LOW
;
3330 u32 interval
= beacon_interval
* TIME_UNIT
;
3331 u32 res
= (base
& BEACON_TIME_MASK_HIGH
) +
3332 (addon
& BEACON_TIME_MASK_HIGH
);
3334 if (base_low
> addon_low
)
3335 res
+= base_low
- addon_low
;
3336 else if (base_low
< addon_low
) {
3337 res
+= interval
+ base_low
- addon_low
;
3342 return cpu_to_le32(res
);
3345 static int iwl_get_measurement(struct iwl_priv
*priv
,
3346 struct ieee80211_measurement_params
*params
,
3349 struct iwl_spectrum_cmd spectrum
;
3350 struct iwl_rx_packet
*res
;
3351 struct iwl_host_cmd cmd
= {
3352 .id
= REPLY_SPECTRUM_MEASUREMENT_CMD
,
3353 .data
= (void *)&spectrum
,
3354 .meta
.flags
= CMD_WANT_SKB
,
3356 u32 add_time
= le64_to_cpu(params
->start_time
);
3358 int spectrum_resp_status
;
3359 int duration
= le16_to_cpu(params
->duration
);
3361 if (iwl_is_associated(priv
))
3363 iwl_usecs_to_beacons(
3364 le64_to_cpu(params
->start_time
) - priv
->last_tsf
,
3365 le16_to_cpu(priv
->rxon_timing
.beacon_interval
));
3367 memset(&spectrum
, 0, sizeof(spectrum
));
3369 spectrum
.channel_count
= cpu_to_le16(1);
3371 RXON_FLG_TSF2HOST_MSK
| RXON_FLG_ANT_A_MSK
| RXON_FLG_DIS_DIV_MSK
;
3372 spectrum
.filter_flags
= MEASUREMENT_FILTER_FLAG
;
3373 cmd
.len
= sizeof(spectrum
);
3374 spectrum
.len
= cpu_to_le16(cmd
.len
- sizeof(spectrum
.len
));
3376 if (iwl_is_associated(priv
))
3377 spectrum
.start_time
=
3378 iwl_add_beacon_time(priv
->last_beacon_time
,
3380 le16_to_cpu(priv
->rxon_timing
.beacon_interval
));
3382 spectrum
.start_time
= 0;
3384 spectrum
.channels
[0].duration
= cpu_to_le32(duration
* TIME_UNIT
);
3385 spectrum
.channels
[0].channel
= params
->channel
;
3386 spectrum
.channels
[0].type
= type
;
3387 if (priv
->active_rxon
.flags
& RXON_FLG_BAND_24G_MSK
)
3388 spectrum
.flags
|= RXON_FLG_BAND_24G_MSK
|
3389 RXON_FLG_AUTO_DETECT_MSK
| RXON_FLG_TGG_PROTECT_MSK
;
3391 rc
= iwl_send_cmd_sync(priv
, &cmd
);
3395 res
= (struct iwl_rx_packet
*)cmd
.meta
.u
.skb
->data
;
3396 if (res
->hdr
.flags
& IWL_CMD_FAILED_MSK
) {
3397 IWL_ERROR("Bad return from REPLY_RX_ON_ASSOC command\n");
3401 spectrum_resp_status
= le16_to_cpu(res
->u
.spectrum
.status
);
3402 switch (spectrum_resp_status
) {
3403 case 0: /* Command will be handled */
3404 if (res
->u
.spectrum
.id
!= 0xff) {
3406 ("Replaced existing measurement: %d\n",
3407 res
->u
.spectrum
.id
);
3408 priv
->measurement_status
&= ~MEASUREMENT_READY
;
3410 priv
->measurement_status
|= MEASUREMENT_ACTIVE
;
3414 case 1: /* Command will not be handled */
3419 dev_kfree_skb_any(cmd
.meta
.u
.skb
);
3425 static void iwl_txstatus_to_ieee(struct iwl_priv
*priv
,
3426 struct iwl_tx_info
*tx_sta
)
3429 tx_sta
->status
.ack_signal
= 0;
3430 tx_sta
->status
.excessive_retries
= 0;
3431 tx_sta
->status
.queue_length
= 0;
3432 tx_sta
->status
.queue_number
= 0;
3435 ieee80211_tx_status_irqsafe(priv
->hw
,
3436 tx_sta
->skb
[0], &(tx_sta
->status
));
3438 ieee80211_tx_status(priv
->hw
,
3439 tx_sta
->skb
[0], &(tx_sta
->status
));
3441 tx_sta
->skb
[0] = NULL
;
3445 * iwl_tx_queue_reclaim - Reclaim Tx queue entries no more used by NIC.
3447 * When FW advances 'R' index, all entries between old and
3448 * new 'R' index need to be reclaimed. As result, some free space
3449 * forms. If there is enough free space (> low mark), wake Tx queue.
3451 int iwl_tx_queue_reclaim(struct iwl_priv
*priv
, int txq_id
, int index
)
3453 struct iwl_tx_queue
*txq
= &priv
->txq
[txq_id
];
3454 struct iwl_queue
*q
= &txq
->q
;
3457 if ((index
>= q
->n_bd
) || (x2_queue_used(q
, index
) == 0)) {
3458 IWL_ERROR("Read index for DMA queue txq id (%d), index %d, "
3459 "is out of range [0-%d] %d %d.\n", txq_id
,
3460 index
, q
->n_bd
, q
->write_ptr
, q
->read_ptr
);
3464 for (index
= iwl_queue_inc_wrap(index
, q
->n_bd
);
3465 q
->read_ptr
!= index
;
3466 q
->read_ptr
= iwl_queue_inc_wrap(q
->read_ptr
, q
->n_bd
)) {
3467 if (txq_id
!= IWL_CMD_QUEUE_NUM
) {
3468 iwl_txstatus_to_ieee(priv
,
3469 &(txq
->txb
[txq
->q
.read_ptr
]));
3470 iwl_hw_txq_free_tfd(priv
, txq
);
3471 } else if (nfreed
> 1) {
3472 IWL_ERROR("HCMD skipped: index (%d) %d %d\n", index
,
3473 q
->write_ptr
, q
->read_ptr
);
3474 queue_work(priv
->workqueue
, &priv
->restart
);
3479 if (iwl_queue_space(q
) > q
->low_mark
&& (txq_id
>= 0) &&
3480 (txq_id
!= IWL_CMD_QUEUE_NUM
) &&
3481 priv
->mac80211_registered
)
3482 ieee80211_wake_queue(priv
->hw
, txq_id
);
3488 static int iwl_is_tx_success(u32 status
)
3490 status
&= TX_STATUS_MSK
;
3491 return (status
== TX_STATUS_SUCCESS
)
3492 || (status
== TX_STATUS_DIRECT_DONE
);
3495 /******************************************************************************
3497 * Generic RX handler implementations
3499 ******************************************************************************/
3500 #ifdef CONFIG_IWL4965_HT
3501 #ifdef CONFIG_IWL4965_HT_AGG
3503 static inline int iwl_get_ra_sta_id(struct iwl_priv
*priv
,
3504 struct ieee80211_hdr
*hdr
)
3506 if (priv
->iw_mode
== IEEE80211_IF_TYPE_STA
)
3509 u8
*da
= ieee80211_get_DA(hdr
);
3510 return iwl_hw_find_station(priv
, da
);
3514 static struct ieee80211_hdr
*iwl_tx_queue_get_hdr(
3515 struct iwl_priv
*priv
, int txq_id
, int idx
)
3517 if (priv
->txq
[txq_id
].txb
[idx
].skb
[0])
3518 return (struct ieee80211_hdr
*)priv
->txq
[txq_id
].
3519 txb
[idx
].skb
[0]->data
;
3523 static inline u32
iwl_get_scd_ssn(struct iwl_tx_resp
*tx_resp
)
3525 __le32
*scd_ssn
= (__le32
*)((u32
*)&tx_resp
->status
+
3526 tx_resp
->frame_count
);
3527 return le32_to_cpu(*scd_ssn
) & MAX_SN
;
3530 static int iwl4965_tx_status_reply_tx(struct iwl_priv
*priv
,
3531 struct iwl_ht_agg
*agg
,
3532 struct iwl_tx_resp
*tx_resp
,
3536 __le32
*frame_status
= &tx_resp
->status
;
3537 struct ieee80211_tx_status
*tx_status
= NULL
;
3538 struct ieee80211_hdr
*hdr
= NULL
;
3543 if (agg
->wait_for_ba
)
3544 IWL_DEBUG_TX_REPLY("got tx repsons w/o back\n");
3546 agg
->frame_count
= tx_resp
->frame_count
;
3547 agg
->start_idx
= start_idx
;
3548 agg
->rate_n_flags
= le32_to_cpu(tx_resp
->rate_n_flags
);
3549 agg
->bitmap0
= agg
->bitmap1
= 0;
3551 if (agg
->frame_count
== 1) {
3552 struct iwl_tx_queue
*txq
;
3553 status
= le32_to_cpu(frame_status
[0]);
3555 txq_id
= agg
->txq_id
;
3556 txq
= &priv
->txq
[txq_id
];
3557 /* FIXME: code repetition */
3558 IWL_DEBUG_TX_REPLY("FrameCnt = %d, StartIdx=%d \n",
3559 agg
->frame_count
, agg
->start_idx
);
3561 tx_status
= &(priv
->txq
[txq_id
].txb
[txq
->q
.read_ptr
].status
);
3562 tx_status
->retry_count
= tx_resp
->failure_frame
;
3563 tx_status
->queue_number
= status
& 0xff;
3564 tx_status
->queue_length
= tx_resp
->bt_kill_count
;
3565 tx_status
->queue_length
|= tx_resp
->failure_rts
;
3567 tx_status
->flags
= iwl_is_tx_success(status
)?
3568 IEEE80211_TX_STATUS_ACK
: 0;
3569 tx_status
->control
.tx_rate
=
3570 iwl_hw_get_rate_n_flags(tx_resp
->rate_n_flags
);
3571 /* FIXME: code repetition end */
3573 IWL_DEBUG_TX_REPLY("1 Frame 0x%x failure :%d\n",
3574 status
& 0xff, tx_resp
->failure_frame
);
3575 IWL_DEBUG_TX_REPLY("Rate Info rate_n_flags=%x\n",
3576 iwl_hw_get_rate_n_flags(tx_resp
->rate_n_flags
));
3578 agg
->wait_for_ba
= 0;
3581 int start
= agg
->start_idx
;
3583 for (i
= 0; i
< agg
->frame_count
; i
++) {
3585 status
= le32_to_cpu(frame_status
[i
]);
3587 idx
= SEQ_TO_INDEX(seq
);
3588 txq_id
= SEQ_TO_QUEUE(seq
);
3590 if (status
& (AGG_TX_STATE_FEW_BYTES_MSK
|
3591 AGG_TX_STATE_ABORT_MSK
))
3594 IWL_DEBUG_TX_REPLY("FrameCnt = %d, txq_id=%d idx=%d\n",
3595 agg
->frame_count
, txq_id
, idx
);
3597 hdr
= iwl_tx_queue_get_hdr(priv
, txq_id
, idx
);
3599 sc
= le16_to_cpu(hdr
->seq_ctrl
);
3600 if (idx
!= (SEQ_TO_SN(sc
) & 0xff)) {
3601 IWL_ERROR("BUG_ON idx doesn't match seq control"
3602 " idx=%d, seq_idx=%d, seq=%d\n",
3608 IWL_DEBUG_TX_REPLY("AGG Frame i=%d idx %d seq=%d\n",
3609 i
, idx
, SEQ_TO_SN(sc
));
3613 sh
= (start
- idx
) + 0xff;
3614 bitmap
= bitmap
<< sh
;
3617 } else if (sh
< -64)
3618 sh
= 0xff - (start
- idx
);
3622 bitmap
= bitmap
<< sh
;
3625 bitmap
|= (1 << sh
);
3626 IWL_DEBUG_TX_REPLY("start=%d bitmap=0x%x\n",
3627 start
, (u32
)(bitmap
& 0xFFFFFFFF));
3630 agg
->bitmap0
= bitmap
& 0xFFFFFFFF;
3631 agg
->bitmap1
= bitmap
>> 32;
3632 agg
->start_idx
= start
;
3633 agg
->rate_n_flags
= le32_to_cpu(tx_resp
->rate_n_flags
);
3634 IWL_DEBUG_TX_REPLY("Frames %d start_idx=%d bitmap=0x%x\n",
3635 agg
->frame_count
, agg
->start_idx
,
3639 agg
->wait_for_ba
= 1;
3646 static void iwl_rx_reply_tx(struct iwl_priv
*priv
,
3647 struct iwl_rx_mem_buffer
*rxb
)
3649 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3650 u16 sequence
= le16_to_cpu(pkt
->hdr
.sequence
);
3651 int txq_id
= SEQ_TO_QUEUE(sequence
);
3652 int index
= SEQ_TO_INDEX(sequence
);
3653 struct iwl_tx_queue
*txq
= &priv
->txq
[txq_id
];
3654 struct ieee80211_tx_status
*tx_status
;
3655 struct iwl_tx_resp
*tx_resp
= (void *)&pkt
->u
.raw
[0];
3656 u32 status
= le32_to_cpu(tx_resp
->status
);
3657 #ifdef CONFIG_IWL4965_HT
3658 #ifdef CONFIG_IWL4965_HT_AGG
3663 if ((index
>= txq
->q
.n_bd
) || (x2_queue_used(&txq
->q
, index
) == 0)) {
3664 IWL_ERROR("Read index for DMA queue txq_id (%d) index %d "
3665 "is out of range [0-%d] %d %d\n", txq_id
,
3666 index
, txq
->q
.n_bd
, txq
->q
.write_ptr
,
3671 #ifdef CONFIG_IWL4965_HT
3672 #ifdef CONFIG_IWL4965_HT_AGG
3673 if (txq
->sched_retry
) {
3674 const u32 scd_ssn
= iwl_get_scd_ssn(tx_resp
);
3675 struct ieee80211_hdr
*hdr
=
3676 iwl_tx_queue_get_hdr(priv
, txq_id
, index
);
3677 struct iwl_ht_agg
*agg
= NULL
;
3678 __le16
*qc
= ieee80211_get_qos_ctrl(hdr
);
3681 IWL_ERROR("BUG_ON qc is null!!!!\n");
3685 tid
= le16_to_cpu(*qc
) & 0xf;
3687 sta_id
= iwl_get_ra_sta_id(priv
, hdr
);
3688 if (unlikely(sta_id
== IWL_INVALID_STATION
)) {
3689 IWL_ERROR("Station not known for\n");
3693 agg
= &priv
->stations
[sta_id
].tid
[tid
].agg
;
3695 iwl4965_tx_status_reply_tx(priv
, agg
, tx_resp
, index
);
3697 if ((tx_resp
->frame_count
== 1) &&
3698 !iwl_is_tx_success(status
)) {
3699 /* TODO: send BAR */
3702 if ((txq
->q
.read_ptr
!= (scd_ssn
& 0xff))) {
3703 index
= iwl_queue_dec_wrap(scd_ssn
& 0xff, txq
->q
.n_bd
);
3704 IWL_DEBUG_TX_REPLY("Retry scheduler reclaim scd_ssn "
3705 "%d index %d\n", scd_ssn
, index
);
3706 iwl_tx_queue_reclaim(priv
, txq_id
, index
);
3709 #endif /* CONFIG_IWL4965_HT_AGG */
3710 #endif /* CONFIG_IWL4965_HT */
3711 tx_status
= &(txq
->txb
[txq
->q
.read_ptr
].status
);
3713 tx_status
->retry_count
= tx_resp
->failure_frame
;
3714 tx_status
->queue_number
= status
;
3715 tx_status
->queue_length
= tx_resp
->bt_kill_count
;
3716 tx_status
->queue_length
|= tx_resp
->failure_rts
;
3719 iwl_is_tx_success(status
) ? IEEE80211_TX_STATUS_ACK
: 0;
3721 tx_status
->control
.tx_rate
=
3722 iwl_hw_get_rate_n_flags(tx_resp
->rate_n_flags
);
3724 IWL_DEBUG_TX("Tx queue %d Status %s (0x%08x) rate_n_flags 0x%x "
3725 "retries %d\n", txq_id
, iwl_get_tx_fail_reason(status
),
3726 status
, le32_to_cpu(tx_resp
->rate_n_flags
),
3727 tx_resp
->failure_frame
);
3729 IWL_DEBUG_TX_REPLY("Tx queue reclaim %d\n", index
);
3731 iwl_tx_queue_reclaim(priv
, txq_id
, index
);
3732 #ifdef CONFIG_IWL4965_HT
3733 #ifdef CONFIG_IWL4965_HT_AGG
3735 #endif /* CONFIG_IWL4965_HT_AGG */
3736 #endif /* CONFIG_IWL4965_HT */
3738 if (iwl_check_bits(status
, TX_ABORT_REQUIRED_MSK
))
3739 IWL_ERROR("TODO: Implement Tx ABORT REQUIRED!!!\n");
3743 static void iwl_rx_reply_alive(struct iwl_priv
*priv
,
3744 struct iwl_rx_mem_buffer
*rxb
)
3746 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3747 struct iwl_alive_resp
*palive
;
3748 struct delayed_work
*pwork
;
3750 palive
= &pkt
->u
.alive_frame
;
3752 IWL_DEBUG_INFO("Alive ucode status 0x%08X revision "
3754 palive
->is_valid
, palive
->ver_type
,
3755 palive
->ver_subtype
);
3757 if (palive
->ver_subtype
== INITIALIZE_SUBTYPE
) {
3758 IWL_DEBUG_INFO("Initialization Alive received.\n");
3759 memcpy(&priv
->card_alive_init
,
3760 &pkt
->u
.alive_frame
,
3761 sizeof(struct iwl_init_alive_resp
));
3762 pwork
= &priv
->init_alive_start
;
3764 IWL_DEBUG_INFO("Runtime Alive received.\n");
3765 memcpy(&priv
->card_alive
, &pkt
->u
.alive_frame
,
3766 sizeof(struct iwl_alive_resp
));
3767 pwork
= &priv
->alive_start
;
3770 /* We delay the ALIVE response by 5ms to
3771 * give the HW RF Kill time to activate... */
3772 if (palive
->is_valid
== UCODE_VALID_OK
)
3773 queue_delayed_work(priv
->workqueue
, pwork
,
3774 msecs_to_jiffies(5));
3776 IWL_WARNING("uCode did not respond OK.\n");
3779 static void iwl_rx_reply_add_sta(struct iwl_priv
*priv
,
3780 struct iwl_rx_mem_buffer
*rxb
)
3782 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3784 IWL_DEBUG_RX("Received REPLY_ADD_STA: 0x%02X\n", pkt
->u
.status
);
3788 static void iwl_rx_reply_error(struct iwl_priv
*priv
,
3789 struct iwl_rx_mem_buffer
*rxb
)
3791 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3793 IWL_ERROR("Error Reply type 0x%08X cmd %s (0x%02X) "
3794 "seq 0x%04X ser 0x%08X\n",
3795 le32_to_cpu(pkt
->u
.err_resp
.error_type
),
3796 get_cmd_string(pkt
->u
.err_resp
.cmd_id
),
3797 pkt
->u
.err_resp
.cmd_id
,
3798 le16_to_cpu(pkt
->u
.err_resp
.bad_cmd_seq_num
),
3799 le32_to_cpu(pkt
->u
.err_resp
.error_info
));
3802 #define TX_STATUS_ENTRY(x) case TX_STATUS_FAIL_ ## x: return #x
3804 static void iwl_rx_csa(struct iwl_priv
*priv
, struct iwl_rx_mem_buffer
*rxb
)
3806 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3807 struct iwl_rxon_cmd
*rxon
= (void *)&priv
->active_rxon
;
3808 struct iwl_csa_notification
*csa
= &(pkt
->u
.csa_notif
);
3809 IWL_DEBUG_11H("CSA notif: channel %d, status %d\n",
3810 le16_to_cpu(csa
->channel
), le32_to_cpu(csa
->status
));
3811 rxon
->channel
= csa
->channel
;
3812 priv
->staging_rxon
.channel
= csa
->channel
;
3815 static void iwl_rx_spectrum_measure_notif(struct iwl_priv
*priv
,
3816 struct iwl_rx_mem_buffer
*rxb
)
3818 #ifdef CONFIG_IWL4965_SPECTRUM_MEASUREMENT
3819 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3820 struct iwl_spectrum_notification
*report
= &(pkt
->u
.spectrum_notif
);
3822 if (!report
->state
) {
3823 IWL_DEBUG(IWL_DL_11H
| IWL_DL_INFO
,
3824 "Spectrum Measure Notification: Start\n");
3828 memcpy(&priv
->measure_report
, report
, sizeof(*report
));
3829 priv
->measurement_status
|= MEASUREMENT_READY
;
3833 static void iwl_rx_pm_sleep_notif(struct iwl_priv
*priv
,
3834 struct iwl_rx_mem_buffer
*rxb
)
3836 #ifdef CONFIG_IWL4965_DEBUG
3837 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3838 struct iwl_sleep_notification
*sleep
= &(pkt
->u
.sleep_notif
);
3839 IWL_DEBUG_RX("sleep mode: %d, src: %d\n",
3840 sleep
->pm_sleep_mode
, sleep
->pm_wakeup_src
);
3844 static void iwl_rx_pm_debug_statistics_notif(struct iwl_priv
*priv
,
3845 struct iwl_rx_mem_buffer
*rxb
)
3847 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3848 IWL_DEBUG_RADIO("Dumping %d bytes of unhandled "
3849 "notification for %s:\n",
3850 le32_to_cpu(pkt
->len
), get_cmd_string(pkt
->hdr
.cmd
));
3851 iwl_print_hex_dump(IWL_DL_RADIO
, pkt
->u
.raw
, le32_to_cpu(pkt
->len
));
3854 static void iwl_bg_beacon_update(struct work_struct
*work
)
3856 struct iwl_priv
*priv
=
3857 container_of(work
, struct iwl_priv
, beacon_update
);
3858 struct sk_buff
*beacon
;
3860 /* Pull updated AP beacon from mac80211. will fail if not in AP mode */
3861 beacon
= ieee80211_beacon_get(priv
->hw
, priv
->interface_id
, NULL
);
3864 IWL_ERROR("update beacon failed\n");
3868 mutex_lock(&priv
->mutex
);
3869 /* new beacon skb is allocated every time; dispose previous.*/
3870 if (priv
->ibss_beacon
)
3871 dev_kfree_skb(priv
->ibss_beacon
);
3873 priv
->ibss_beacon
= beacon
;
3874 mutex_unlock(&priv
->mutex
);
3876 iwl_send_beacon_cmd(priv
);
3879 static void iwl_rx_beacon_notif(struct iwl_priv
*priv
,
3880 struct iwl_rx_mem_buffer
*rxb
)
3882 #ifdef CONFIG_IWL4965_DEBUG
3883 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3884 struct iwl_beacon_notif
*beacon
= &(pkt
->u
.beacon_status
);
3885 u8 rate
= iwl_hw_get_rate(beacon
->beacon_notify_hdr
.rate_n_flags
);
3887 IWL_DEBUG_RX("beacon status %x retries %d iss %d "
3888 "tsf %d %d rate %d\n",
3889 le32_to_cpu(beacon
->beacon_notify_hdr
.status
) & TX_STATUS_MSK
,
3890 beacon
->beacon_notify_hdr
.failure_frame
,
3891 le32_to_cpu(beacon
->ibss_mgr_status
),
3892 le32_to_cpu(beacon
->high_tsf
),
3893 le32_to_cpu(beacon
->low_tsf
), rate
);
3896 if ((priv
->iw_mode
== IEEE80211_IF_TYPE_AP
) &&
3897 (!test_bit(STATUS_EXIT_PENDING
, &priv
->status
)))
3898 queue_work(priv
->workqueue
, &priv
->beacon_update
);
3901 /* Service response to REPLY_SCAN_CMD (0x80) */
3902 static void iwl_rx_reply_scan(struct iwl_priv
*priv
,
3903 struct iwl_rx_mem_buffer
*rxb
)
3905 #ifdef CONFIG_IWL4965_DEBUG
3906 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3907 struct iwl_scanreq_notification
*notif
=
3908 (struct iwl_scanreq_notification
*)pkt
->u
.raw
;
3910 IWL_DEBUG_RX("Scan request status = 0x%x\n", notif
->status
);
3914 /* Service SCAN_START_NOTIFICATION (0x82) */
3915 static void iwl_rx_scan_start_notif(struct iwl_priv
*priv
,
3916 struct iwl_rx_mem_buffer
*rxb
)
3918 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3919 struct iwl_scanstart_notification
*notif
=
3920 (struct iwl_scanstart_notification
*)pkt
->u
.raw
;
3921 priv
->scan_start_tsf
= le32_to_cpu(notif
->tsf_low
);
3922 IWL_DEBUG_SCAN("Scan start: "
3924 "(TSF: 0x%08X:%08X) - %d (beacon timer %u)\n",
3926 notif
->band
? "bg" : "a",
3928 notif
->tsf_low
, notif
->status
, notif
->beacon_timer
);
3931 /* Service SCAN_RESULTS_NOTIFICATION (0x83) */
3932 static void iwl_rx_scan_results_notif(struct iwl_priv
*priv
,
3933 struct iwl_rx_mem_buffer
*rxb
)
3935 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3936 struct iwl_scanresults_notification
*notif
=
3937 (struct iwl_scanresults_notification
*)pkt
->u
.raw
;
3939 IWL_DEBUG_SCAN("Scan ch.res: "
3941 "(TSF: 0x%08X:%08X) - %d "
3942 "elapsed=%lu usec (%dms since last)\n",
3944 notif
->band
? "bg" : "a",
3945 le32_to_cpu(notif
->tsf_high
),
3946 le32_to_cpu(notif
->tsf_low
),
3947 le32_to_cpu(notif
->statistics
[0]),
3948 le32_to_cpu(notif
->tsf_low
) - priv
->scan_start_tsf
,
3949 jiffies_to_msecs(elapsed_jiffies
3950 (priv
->last_scan_jiffies
, jiffies
)));
3952 priv
->last_scan_jiffies
= jiffies
;
3955 /* Service SCAN_COMPLETE_NOTIFICATION (0x84) */
3956 static void iwl_rx_scan_complete_notif(struct iwl_priv
*priv
,
3957 struct iwl_rx_mem_buffer
*rxb
)
3959 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
3960 struct iwl_scancomplete_notification
*scan_notif
= (void *)pkt
->u
.raw
;
3962 IWL_DEBUG_SCAN("Scan complete: %d channels (TSF 0x%08X:%08X) - %d\n",
3963 scan_notif
->scanned_channels
,
3964 scan_notif
->tsf_low
,
3965 scan_notif
->tsf_high
, scan_notif
->status
);
3967 /* The HW is no longer scanning */
3968 clear_bit(STATUS_SCAN_HW
, &priv
->status
);
3970 /* The scan completion notification came in, so kill that timer... */
3971 cancel_delayed_work(&priv
->scan_check
);
3973 IWL_DEBUG_INFO("Scan pass on %sGHz took %dms\n",
3974 (priv
->scan_bands
== 2) ? "2.4" : "5.2",
3975 jiffies_to_msecs(elapsed_jiffies
3976 (priv
->scan_pass_start
, jiffies
)));
3978 /* Remove this scanned band from the list
3979 * of pending bands to scan */
3982 /* If a request to abort was given, or the scan did not succeed
3983 * then we reset the scan state machine and terminate,
3984 * re-queuing another scan if one has been requested */
3985 if (test_bit(STATUS_SCAN_ABORTING
, &priv
->status
)) {
3986 IWL_DEBUG_INFO("Aborted scan completed.\n");
3987 clear_bit(STATUS_SCAN_ABORTING
, &priv
->status
);
3989 /* If there are more bands on this scan pass reschedule */
3990 if (priv
->scan_bands
> 0)
3994 priv
->last_scan_jiffies
= jiffies
;
3995 IWL_DEBUG_INFO("Setting scan to off\n");
3997 clear_bit(STATUS_SCANNING
, &priv
->status
);
3999 IWL_DEBUG_INFO("Scan took %dms\n",
4000 jiffies_to_msecs(elapsed_jiffies(priv
->scan_start
, jiffies
)));
4002 queue_work(priv
->workqueue
, &priv
->scan_completed
);
4007 priv
->scan_pass_start
= jiffies
;
4008 queue_work(priv
->workqueue
, &priv
->request_scan
);
4011 /* Handle notification from uCode that card's power state is changing
4012 * due to software, hardware, or critical temperature RFKILL */
4013 static void iwl_rx_card_state_notif(struct iwl_priv
*priv
,
4014 struct iwl_rx_mem_buffer
*rxb
)
4016 struct iwl_rx_packet
*pkt
= (void *)rxb
->skb
->data
;
4017 u32 flags
= le32_to_cpu(pkt
->u
.card_state_notif
.flags
);
4018 unsigned long status
= priv
->status
;
4020 IWL_DEBUG_RF_KILL("Card state received: HW:%s SW:%s\n",
4021 (flags
& HW_CARD_DISABLED
) ? "Kill" : "On",
4022 (flags
& SW_CARD_DISABLED
) ? "Kill" : "On");
4024 if (flags
& (SW_CARD_DISABLED
| HW_CARD_DISABLED
|
4025 RF_CARD_DISABLED
)) {
4027 iwl_write32(priv
, CSR_UCODE_DRV_GP1_SET
,
4028 CSR_UCODE_DRV_GP1_BIT_CMD_BLOCKED
);
4030 if (!iwl_grab_nic_access(priv
)) {
4032 priv
, HBUS_TARG_MBX_C
,
4033 HBUS_TARG_MBX_C_REG_BIT_CMD_BLOCKED
);
4035 iwl_release_nic_access(priv
);
4038 if (!(flags
& RXON_CARD_DISABLED
)) {
4039 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
,
4040 CSR_UCODE_DRV_GP1_BIT_CMD_BLOCKED
);
4041 if (!iwl_grab_nic_access(priv
)) {
4043 priv
, HBUS_TARG_MBX_C
,
4044 HBUS_TARG_MBX_C_REG_BIT_CMD_BLOCKED
);
4046 iwl_release_nic_access(priv
);
4050 if (flags
& RF_CARD_DISABLED
) {
4051 iwl_write32(priv
, CSR_UCODE_DRV_GP1_SET
,
4052 CSR_UCODE_DRV_GP1_REG_BIT_CT_KILL_EXIT
);
4053 iwl_read32(priv
, CSR_UCODE_DRV_GP1
);
4054 if (!iwl_grab_nic_access(priv
))
4055 iwl_release_nic_access(priv
);
4059 if (flags
& HW_CARD_DISABLED
)
4060 set_bit(STATUS_RF_KILL_HW
, &priv
->status
);
4062 clear_bit(STATUS_RF_KILL_HW
, &priv
->status
);
4065 if (flags
& SW_CARD_DISABLED
)
4066 set_bit(STATUS_RF_KILL_SW
, &priv
->status
);
4068 clear_bit(STATUS_RF_KILL_SW
, &priv
->status
);
4070 if (!(flags
& RXON_CARD_DISABLED
))
4071 iwl_scan_cancel(priv
);
4073 if ((test_bit(STATUS_RF_KILL_HW
, &status
) !=
4074 test_bit(STATUS_RF_KILL_HW
, &priv
->status
)) ||
4075 (test_bit(STATUS_RF_KILL_SW
, &status
) !=
4076 test_bit(STATUS_RF_KILL_SW
, &priv
->status
)))
4077 queue_work(priv
->workqueue
, &priv
->rf_kill
);
4079 wake_up_interruptible(&priv
->wait_command_queue
);
4083 * iwl_setup_rx_handlers - Initialize Rx handler callbacks
4085 * Setup the RX handlers for each of the reply types sent from the uCode
4088 * This function chains into the hardware specific files for them to setup
4089 * any hardware specific handlers as well.
4091 static void iwl_setup_rx_handlers(struct iwl_priv
*priv
)
4093 priv
->rx_handlers
[REPLY_ALIVE
] = iwl_rx_reply_alive
;
4094 priv
->rx_handlers
[REPLY_ADD_STA
] = iwl_rx_reply_add_sta
;
4095 priv
->rx_handlers
[REPLY_ERROR
] = iwl_rx_reply_error
;
4096 priv
->rx_handlers
[CHANNEL_SWITCH_NOTIFICATION
] = iwl_rx_csa
;
4097 priv
->rx_handlers
[SPECTRUM_MEASURE_NOTIFICATION
] =
4098 iwl_rx_spectrum_measure_notif
;
4099 priv
->rx_handlers
[PM_SLEEP_NOTIFICATION
] = iwl_rx_pm_sleep_notif
;
4100 priv
->rx_handlers
[PM_DEBUG_STATISTIC_NOTIFIC
] =
4101 iwl_rx_pm_debug_statistics_notif
;
4102 priv
->rx_handlers
[BEACON_NOTIFICATION
] = iwl_rx_beacon_notif
;
4104 /* NOTE: iwl_rx_statistics is different based on whether
4105 * the build is for the 3945 or the 4965. See the
4106 * corresponding implementation in iwl-XXXX.c
4108 * The same handler is used for both the REPLY to a
4109 * discrete statistics request from the host as well as
4110 * for the periodic statistics notification from the uCode
4112 priv
->rx_handlers
[REPLY_STATISTICS_CMD
] = iwl_hw_rx_statistics
;
4113 priv
->rx_handlers
[STATISTICS_NOTIFICATION
] = iwl_hw_rx_statistics
;
4115 priv
->rx_handlers
[REPLY_SCAN_CMD
] = iwl_rx_reply_scan
;
4116 priv
->rx_handlers
[SCAN_START_NOTIFICATION
] = iwl_rx_scan_start_notif
;
4117 priv
->rx_handlers
[SCAN_RESULTS_NOTIFICATION
] =
4118 iwl_rx_scan_results_notif
;
4119 priv
->rx_handlers
[SCAN_COMPLETE_NOTIFICATION
] =
4120 iwl_rx_scan_complete_notif
;
4121 priv
->rx_handlers
[CARD_STATE_NOTIFICATION
] = iwl_rx_card_state_notif
;
4122 priv
->rx_handlers
[REPLY_TX
] = iwl_rx_reply_tx
;
4124 /* Setup hardware specific Rx handlers */
4125 iwl_hw_rx_handler_setup(priv
);
4129 * iwl_tx_cmd_complete - Pull unused buffers off the queue and reclaim them
4130 * @rxb: Rx buffer to reclaim
4132 * If an Rx buffer has an async callback associated with it the callback
4133 * will be executed. The attached skb (if present) will only be freed
4134 * if the callback returns 1
4136 static void iwl_tx_cmd_complete(struct iwl_priv
*priv
,
4137 struct iwl_rx_mem_buffer
*rxb
)
4139 struct iwl_rx_packet
*pkt
= (struct iwl_rx_packet
*)rxb
->skb
->data
;
4140 u16 sequence
= le16_to_cpu(pkt
->hdr
.sequence
);
4141 int txq_id
= SEQ_TO_QUEUE(sequence
);
4142 int index
= SEQ_TO_INDEX(sequence
);
4143 int huge
= sequence
& SEQ_HUGE_FRAME
;
4145 struct iwl_cmd
*cmd
;
4147 /* If a Tx command is being handled and it isn't in the actual
4148 * command queue then there a command routing bug has been introduced
4149 * in the queue management code. */
4150 if (txq_id
!= IWL_CMD_QUEUE_NUM
)
4151 IWL_ERROR("Error wrong command queue %d command id 0x%X\n",
4152 txq_id
, pkt
->hdr
.cmd
);
4153 BUG_ON(txq_id
!= IWL_CMD_QUEUE_NUM
);
4155 cmd_index
= get_cmd_index(&priv
->txq
[IWL_CMD_QUEUE_NUM
].q
, index
, huge
);
4156 cmd
= &priv
->txq
[IWL_CMD_QUEUE_NUM
].cmd
[cmd_index
];
4158 /* Input error checking is done when commands are added to queue. */
4159 if (cmd
->meta
.flags
& CMD_WANT_SKB
) {
4160 cmd
->meta
.source
->u
.skb
= rxb
->skb
;
4162 } else if (cmd
->meta
.u
.callback
&&
4163 !cmd
->meta
.u
.callback(priv
, cmd
, rxb
->skb
))
4166 iwl_tx_queue_reclaim(priv
, txq_id
, index
);
4168 if (!(cmd
->meta
.flags
& CMD_ASYNC
)) {
4169 clear_bit(STATUS_HCMD_ACTIVE
, &priv
->status
);
4170 wake_up_interruptible(&priv
->wait_command_queue
);
4174 /************************** RX-FUNCTIONS ****************************/
4176 * Rx theory of operation
4178 * The host allocates 32 DMA target addresses and passes the host address
4179 * to the firmware at register IWL_RFDS_TABLE_LOWER + N * RFD_SIZE where N is
4183 * The host/firmware share two index registers for managing the Rx buffers.
4185 * The READ index maps to the first position that the firmware may be writing
4186 * to -- the driver can read up to (but not including) this position and get
4188 * The READ index is managed by the firmware once the card is enabled.
4190 * The WRITE index maps to the last position the driver has read from -- the
4191 * position preceding WRITE is the last slot the firmware can place a packet.
4193 * The queue is empty (no good data) if WRITE = READ - 1, and is full if
4196 * During initialization the host sets up the READ queue position to the first
4197 * INDEX position, and WRITE to the last (READ - 1 wrapped)
4199 * When the firmware places a packet in a buffer it will advance the READ index
4200 * and fire the RX interrupt. The driver can then query the READ index and
4201 * process as many packets as possible, moving the WRITE index forward as it
4202 * resets the Rx queue buffers with new memory.
4204 * The management in the driver is as follows:
4205 * + A list of pre-allocated SKBs is stored in iwl->rxq->rx_free. When
4206 * iwl->rxq->free_count drops to or below RX_LOW_WATERMARK, work is scheduled
4207 * to replenish the iwl->rxq->rx_free.
4208 * + In iwl_rx_replenish (scheduled) if 'processed' != 'read' then the
4209 * iwl->rxq is replenished and the READ INDEX is updated (updating the
4210 * 'processed' and 'read' driver indexes as well)
4211 * + A received packet is processed and handed to the kernel network stack,
4212 * detached from the iwl->rxq. The driver 'processed' index is updated.
4213 * + The Host/Firmware iwl->rxq is replenished at tasklet time from the rx_free
4214 * list. If there are no allocated buffers in iwl->rxq->rx_free, the READ
4215 * INDEX is not incremented and iwl->status(RX_STALLED) is set. If there
4216 * were enough free buffers and RX_STALLED is set it is cleared.
4221 * iwl_rx_queue_alloc() Allocates rx_free
4222 * iwl_rx_replenish() Replenishes rx_free list from rx_used, and calls
4223 * iwl_rx_queue_restock
4224 * iwl_rx_queue_restock() Moves available buffers from rx_free into Rx
4225 * queue, updates firmware pointers, and updates
4226 * the WRITE index. If insufficient rx_free buffers
4227 * are available, schedules iwl_rx_replenish
4229 * -- enable interrupts --
4230 * ISR - iwl_rx() Detach iwl_rx_mem_buffers from pool up to the
4231 * READ INDEX, detaching the SKB from the pool.
4232 * Moves the packet buffer from queue to rx_used.
4233 * Calls iwl_rx_queue_restock to refill any empty
4240 * iwl_rx_queue_space - Return number of free slots available in queue.
4242 static int iwl_rx_queue_space(const struct iwl_rx_queue
*q
)
4244 int s
= q
->read
- q
->write
;
4247 /* keep some buffer to not confuse full and empty queue */
4255 * iwl_rx_queue_update_write_ptr - Update the write pointer for the RX queue
4257 * NOTE: This function has 3945 and 4965 specific code sections
4258 * but is declared in base due to the majority of the
4259 * implementation being the same (only a numeric constant is
4263 int iwl_rx_queue_update_write_ptr(struct iwl_priv
*priv
, struct iwl_rx_queue
*q
)
4267 unsigned long flags
;
4269 spin_lock_irqsave(&q
->lock
, flags
);
4271 if (q
->need_update
== 0)
4274 if (test_bit(STATUS_POWER_PMI
, &priv
->status
)) {
4275 reg
= iwl_read32(priv
, CSR_UCODE_DRV_GP1
);
4277 if (reg
& CSR_UCODE_DRV_GP1_BIT_MAC_SLEEP
) {
4278 iwl_set_bit(priv
, CSR_GP_CNTRL
,
4279 CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ
);
4283 rc
= iwl_grab_nic_access(priv
);
4287 iwl_write_direct32(priv
, FH_RSCSR_CHNL0_WPTR
,
4289 iwl_release_nic_access(priv
);
4291 iwl_write32(priv
, FH_RSCSR_CHNL0_WPTR
, q
->write
& ~0x7);
4297 spin_unlock_irqrestore(&q
->lock
, flags
);
4302 * iwl_dma_addr2rbd_ptr - convert a DMA address to a uCode read buffer pointer.
4304 * NOTE: This function has 3945 and 4965 specific code paths in it.
4306 static inline __le32
iwl_dma_addr2rbd_ptr(struct iwl_priv
*priv
,
4307 dma_addr_t dma_addr
)
4309 return cpu_to_le32((u32
)(dma_addr
>> 8));
4314 * iwl_rx_queue_restock - refill RX queue from pre-allocated pool
4316 * If there are slots in the RX queue that need to be restocked,
4317 * and we have free pre-allocated buffers, fill the ranks as much
4318 * as we can pulling from rx_free.
4320 * This moves the 'write' index forward to catch up with 'processed', and
4321 * also updates the memory address in the firmware to reference the new
4324 static int iwl_rx_queue_restock(struct iwl_priv
*priv
)
4326 struct iwl_rx_queue
*rxq
= &priv
->rxq
;
4327 struct list_head
*element
;
4328 struct iwl_rx_mem_buffer
*rxb
;
4329 unsigned long flags
;
4332 spin_lock_irqsave(&rxq
->lock
, flags
);
4333 write
= rxq
->write
& ~0x7;
4334 while ((iwl_rx_queue_space(rxq
) > 0) && (rxq
->free_count
)) {
4335 element
= rxq
->rx_free
.next
;
4336 rxb
= list_entry(element
, struct iwl_rx_mem_buffer
, list
);
4338 rxq
->bd
[rxq
->write
] = iwl_dma_addr2rbd_ptr(priv
, rxb
->dma_addr
);
4339 rxq
->queue
[rxq
->write
] = rxb
;
4340 rxq
->write
= (rxq
->write
+ 1) & RX_QUEUE_MASK
;
4343 spin_unlock_irqrestore(&rxq
->lock
, flags
);
4344 /* If the pre-allocated buffer pool is dropping low, schedule to
4346 if (rxq
->free_count
<= RX_LOW_WATERMARK
)
4347 queue_work(priv
->workqueue
, &priv
->rx_replenish
);
4350 /* If we've added more space for the firmware to place data, tell it */
4351 if ((write
!= (rxq
->write
& ~0x7))
4352 || (abs(rxq
->write
- rxq
->read
) > 7)) {
4353 spin_lock_irqsave(&rxq
->lock
, flags
);
4354 rxq
->need_update
= 1;
4355 spin_unlock_irqrestore(&rxq
->lock
, flags
);
4356 rc
= iwl_rx_queue_update_write_ptr(priv
, rxq
);
4365 * iwl_rx_replenish - Move all used packet from rx_used to rx_free
4367 * When moving to rx_free an SKB is allocated for the slot.
4369 * Also restock the Rx queue via iwl_rx_queue_restock.
4370 * This is called as a scheduled work item (except for during initialization)
4372 void iwl_rx_replenish(void *data
)
4374 struct iwl_priv
*priv
= data
;
4375 struct iwl_rx_queue
*rxq
= &priv
->rxq
;
4376 struct list_head
*element
;
4377 struct iwl_rx_mem_buffer
*rxb
;
4378 unsigned long flags
;
4379 spin_lock_irqsave(&rxq
->lock
, flags
);
4380 while (!list_empty(&rxq
->rx_used
)) {
4381 element
= rxq
->rx_used
.next
;
4382 rxb
= list_entry(element
, struct iwl_rx_mem_buffer
, list
);
4384 alloc_skb(IWL_RX_BUF_SIZE
, __GFP_NOWARN
| GFP_ATOMIC
);
4386 if (net_ratelimit())
4387 printk(KERN_CRIT DRV_NAME
4388 ": Can not allocate SKB buffers\n");
4389 /* We don't reschedule replenish work here -- we will
4390 * call the restock method and if it still needs
4391 * more buffers it will schedule replenish */
4394 priv
->alloc_rxb_skb
++;
4397 pci_map_single(priv
->pci_dev
, rxb
->skb
->data
,
4398 IWL_RX_BUF_SIZE
, PCI_DMA_FROMDEVICE
);
4399 list_add_tail(&rxb
->list
, &rxq
->rx_free
);
4402 spin_unlock_irqrestore(&rxq
->lock
, flags
);
4404 spin_lock_irqsave(&priv
->lock
, flags
);
4405 iwl_rx_queue_restock(priv
);
4406 spin_unlock_irqrestore(&priv
->lock
, flags
);
4409 /* Assumes that the skb field of the buffers in 'pool' is kept accurate.
4410 * If an SKB has been detached, the POOL needs to have it's SKB set to NULL
4411 * This free routine walks the list of POOL entries and if SKB is set to
4412 * non NULL it is unmapped and freed
4414 static void iwl_rx_queue_free(struct iwl_priv
*priv
, struct iwl_rx_queue
*rxq
)
4417 for (i
= 0; i
< RX_QUEUE_SIZE
+ RX_FREE_BUFFERS
; i
++) {
4418 if (rxq
->pool
[i
].skb
!= NULL
) {
4419 pci_unmap_single(priv
->pci_dev
,
4420 rxq
->pool
[i
].dma_addr
,
4421 IWL_RX_BUF_SIZE
, PCI_DMA_FROMDEVICE
);
4422 dev_kfree_skb(rxq
->pool
[i
].skb
);
4426 pci_free_consistent(priv
->pci_dev
, 4 * RX_QUEUE_SIZE
, rxq
->bd
,
4431 int iwl_rx_queue_alloc(struct iwl_priv
*priv
)
4433 struct iwl_rx_queue
*rxq
= &priv
->rxq
;
4434 struct pci_dev
*dev
= priv
->pci_dev
;
4437 spin_lock_init(&rxq
->lock
);
4438 INIT_LIST_HEAD(&rxq
->rx_free
);
4439 INIT_LIST_HEAD(&rxq
->rx_used
);
4440 rxq
->bd
= pci_alloc_consistent(dev
, 4 * RX_QUEUE_SIZE
, &rxq
->dma_addr
);
4443 /* Fill the rx_used queue with _all_ of the Rx buffers */
4444 for (i
= 0; i
< RX_FREE_BUFFERS
+ RX_QUEUE_SIZE
; i
++)
4445 list_add_tail(&rxq
->pool
[i
].list
, &rxq
->rx_used
);
4446 /* Set us so that we have processed and used all buffers, but have
4447 * not restocked the Rx queue with fresh buffers */
4448 rxq
->read
= rxq
->write
= 0;
4449 rxq
->free_count
= 0;
4450 rxq
->need_update
= 0;
4454 void iwl_rx_queue_reset(struct iwl_priv
*priv
, struct iwl_rx_queue
*rxq
)
4456 unsigned long flags
;
4458 spin_lock_irqsave(&rxq
->lock
, flags
);
4459 INIT_LIST_HEAD(&rxq
->rx_free
);
4460 INIT_LIST_HEAD(&rxq
->rx_used
);
4461 /* Fill the rx_used queue with _all_ of the Rx buffers */
4462 for (i
= 0; i
< RX_FREE_BUFFERS
+ RX_QUEUE_SIZE
; i
++) {
4463 /* In the reset function, these buffers may have been allocated
4464 * to an SKB, so we need to unmap and free potential storage */
4465 if (rxq
->pool
[i
].skb
!= NULL
) {
4466 pci_unmap_single(priv
->pci_dev
,
4467 rxq
->pool
[i
].dma_addr
,
4468 IWL_RX_BUF_SIZE
, PCI_DMA_FROMDEVICE
);
4469 priv
->alloc_rxb_skb
--;
4470 dev_kfree_skb(rxq
->pool
[i
].skb
);
4471 rxq
->pool
[i
].skb
= NULL
;
4473 list_add_tail(&rxq
->pool
[i
].list
, &rxq
->rx_used
);
4476 /* Set us so that we have processed and used all buffers, but have
4477 * not restocked the Rx queue with fresh buffers */
4478 rxq
->read
= rxq
->write
= 0;
4479 rxq
->free_count
= 0;
4480 spin_unlock_irqrestore(&rxq
->lock
, flags
);
4483 /* Convert linear signal-to-noise ratio into dB */
4484 static u8 ratio2dB
[100] = {
4485 /* 0 1 2 3 4 5 6 7 8 9 */
4486 0, 0, 6, 10, 12, 14, 16, 17, 18, 19, /* 00 - 09 */
4487 20, 21, 22, 22, 23, 23, 24, 25, 26, 26, /* 10 - 19 */
4488 26, 26, 26, 27, 27, 28, 28, 28, 29, 29, /* 20 - 29 */
4489 29, 30, 30, 30, 31, 31, 31, 31, 32, 32, /* 30 - 39 */
4490 32, 32, 32, 33, 33, 33, 33, 33, 34, 34, /* 40 - 49 */
4491 34, 34, 34, 34, 35, 35, 35, 35, 35, 35, /* 50 - 59 */
4492 36, 36, 36, 36, 36, 36, 36, 37, 37, 37, /* 60 - 69 */
4493 37, 37, 37, 37, 37, 38, 38, 38, 38, 38, /* 70 - 79 */
4494 38, 38, 38, 38, 38, 39, 39, 39, 39, 39, /* 80 - 89 */
4495 39, 39, 39, 39, 39, 40, 40, 40, 40, 40 /* 90 - 99 */
4498 /* Calculates a relative dB value from a ratio of linear
4499 * (i.e. not dB) signal levels.
4500 * Conversion assumes that levels are voltages (20*log), not powers (10*log). */
4501 int iwl_calc_db_from_ratio(int sig_ratio
)
4503 /* 1000:1 or higher just report as 60 dB */
4504 if (sig_ratio
>= 1000)
4507 /* 100:1 or higher, divide by 10 and use table,
4508 * add 20 dB to make up for divide by 10 */
4509 if (sig_ratio
>= 100)
4510 return (20 + (int)ratio2dB
[sig_ratio
/10]);
4512 /* We shouldn't see this */
4516 /* Use table for ratios 1:1 - 99:1 */
4517 return (int)ratio2dB
[sig_ratio
];
4520 #define PERFECT_RSSI (-20) /* dBm */
4521 #define WORST_RSSI (-95) /* dBm */
4522 #define RSSI_RANGE (PERFECT_RSSI - WORST_RSSI)
4524 /* Calculate an indication of rx signal quality (a percentage, not dBm!).
4525 * See http://www.ces.clemson.edu/linux/signal_quality.shtml for info
4526 * about formulas used below. */
4527 int iwl_calc_sig_qual(int rssi_dbm
, int noise_dbm
)
4530 int degradation
= PERFECT_RSSI
- rssi_dbm
;
4532 /* If we get a noise measurement, use signal-to-noise ratio (SNR)
4533 * as indicator; formula is (signal dbm - noise dbm).
4534 * SNR at or above 40 is a great signal (100%).
4535 * Below that, scale to fit SNR of 0 - 40 dB within 0 - 100% indicator.
4536 * Weakest usable signal is usually 10 - 15 dB SNR. */
4538 if (rssi_dbm
- noise_dbm
>= 40)
4540 else if (rssi_dbm
< noise_dbm
)
4542 sig_qual
= ((rssi_dbm
- noise_dbm
) * 5) / 2;
4544 /* Else use just the signal level.
4545 * This formula is a least squares fit of data points collected and
4546 * compared with a reference system that had a percentage (%) display
4547 * for signal quality. */
4549 sig_qual
= (100 * (RSSI_RANGE
* RSSI_RANGE
) - degradation
*
4550 (15 * RSSI_RANGE
+ 62 * degradation
)) /
4551 (RSSI_RANGE
* RSSI_RANGE
);
4555 else if (sig_qual
< 1)
4562 * iwl_rx_handle - Main entry function for receiving responses from the uCode
4564 * Uses the priv->rx_handlers callback function array to invoke
4565 * the appropriate handlers, including command responses,
4566 * frame-received notifications, and other notifications.
4568 static void iwl_rx_handle(struct iwl_priv
*priv
)
4570 struct iwl_rx_mem_buffer
*rxb
;
4571 struct iwl_rx_packet
*pkt
;
4572 struct iwl_rx_queue
*rxq
= &priv
->rxq
;
4575 unsigned long flags
;
4577 r
= iwl_hw_get_rx_read(priv
);
4580 /* Rx interrupt, but nothing sent from uCode */
4582 IWL_DEBUG(IWL_DL_RX
| IWL_DL_ISR
, "r = %d, i = %d\n", r
, i
);
4585 rxb
= rxq
->queue
[i
];
4587 /* If an RXB doesn't have a queue slot associated with it
4588 * then a bug has been introduced in the queue refilling
4589 * routines -- catch it here */
4590 BUG_ON(rxb
== NULL
);
4592 rxq
->queue
[i
] = NULL
;
4594 pci_dma_sync_single_for_cpu(priv
->pci_dev
, rxb
->dma_addr
,
4596 PCI_DMA_FROMDEVICE
);
4597 pkt
= (struct iwl_rx_packet
*)rxb
->skb
->data
;
4599 /* Reclaim a command buffer only if this packet is a response
4600 * to a (driver-originated) command.
4601 * If the packet (e.g. Rx frame) originated from uCode,
4602 * there is no command buffer to reclaim.
4603 * Ucode should set SEQ_RX_FRAME bit if ucode-originated,
4604 * but apparently a few don't get set; catch them here. */
4605 reclaim
= !(pkt
->hdr
.sequence
& SEQ_RX_FRAME
) &&
4606 (pkt
->hdr
.cmd
!= REPLY_RX_PHY_CMD
) &&
4607 (pkt
->hdr
.cmd
!= REPLY_4965_RX
) &&
4608 (pkt
->hdr
.cmd
!= REPLY_COMPRESSED_BA
) &&
4609 (pkt
->hdr
.cmd
!= STATISTICS_NOTIFICATION
) &&
4610 (pkt
->hdr
.cmd
!= REPLY_TX
);
4612 /* Based on type of command response or notification,
4613 * handle those that need handling via function in
4614 * rx_handlers table. See iwl_setup_rx_handlers() */
4615 if (priv
->rx_handlers
[pkt
->hdr
.cmd
]) {
4616 IWL_DEBUG(IWL_DL_HOST_COMMAND
| IWL_DL_RX
| IWL_DL_ISR
,
4617 "r = %d, i = %d, %s, 0x%02x\n", r
, i
,
4618 get_cmd_string(pkt
->hdr
.cmd
), pkt
->hdr
.cmd
);
4619 priv
->rx_handlers
[pkt
->hdr
.cmd
] (priv
, rxb
);
4621 /* No handling needed */
4622 IWL_DEBUG(IWL_DL_HOST_COMMAND
| IWL_DL_RX
| IWL_DL_ISR
,
4623 "r %d i %d No handler needed for %s, 0x%02x\n",
4624 r
, i
, get_cmd_string(pkt
->hdr
.cmd
),
4629 /* Invoke any callbacks, transfer the skb to caller,
4630 * and fire off the (possibly) blocking iwl_send_cmd()
4631 * as we reclaim the driver command queue */
4632 if (rxb
&& rxb
->skb
)
4633 iwl_tx_cmd_complete(priv
, rxb
);
4635 IWL_WARNING("Claim null rxb?\n");
4638 /* For now we just don't re-use anything. We can tweak this
4639 * later to try and re-use notification packets and SKBs that
4640 * fail to Rx correctly */
4641 if (rxb
->skb
!= NULL
) {
4642 priv
->alloc_rxb_skb
--;
4643 dev_kfree_skb_any(rxb
->skb
);
4647 pci_unmap_single(priv
->pci_dev
, rxb
->dma_addr
,
4648 IWL_RX_BUF_SIZE
, PCI_DMA_FROMDEVICE
);
4649 spin_lock_irqsave(&rxq
->lock
, flags
);
4650 list_add_tail(&rxb
->list
, &priv
->rxq
.rx_used
);
4651 spin_unlock_irqrestore(&rxq
->lock
, flags
);
4652 i
= (i
+ 1) & RX_QUEUE_MASK
;
4655 /* Backtrack one entry */
4657 iwl_rx_queue_restock(priv
);
4660 static int iwl_tx_queue_update_write_ptr(struct iwl_priv
*priv
,
4661 struct iwl_tx_queue
*txq
)
4665 int txq_id
= txq
->q
.id
;
4667 if (txq
->need_update
== 0)
4670 /* if we're trying to save power */
4671 if (test_bit(STATUS_POWER_PMI
, &priv
->status
)) {
4672 /* wake up nic if it's powered down ...
4673 * uCode will wake up, and interrupt us again, so next
4674 * time we'll skip this part. */
4675 reg
= iwl_read32(priv
, CSR_UCODE_DRV_GP1
);
4677 if (reg
& CSR_UCODE_DRV_GP1_BIT_MAC_SLEEP
) {
4678 IWL_DEBUG_INFO("Requesting wakeup, GP1 = 0x%x\n", reg
);
4679 iwl_set_bit(priv
, CSR_GP_CNTRL
,
4680 CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ
);
4684 /* restore this queue's parameters in nic hardware. */
4685 rc
= iwl_grab_nic_access(priv
);
4688 iwl_write_direct32(priv
, HBUS_TARG_WRPTR
,
4689 txq
->q
.write_ptr
| (txq_id
<< 8));
4690 iwl_release_nic_access(priv
);
4692 /* else not in power-save mode, uCode will never sleep when we're
4693 * trying to tx (during RFKILL, we're not trying to tx). */
4695 iwl_write32(priv
, HBUS_TARG_WRPTR
,
4696 txq
->q
.write_ptr
| (txq_id
<< 8));
4698 txq
->need_update
= 0;
4703 #ifdef CONFIG_IWL4965_DEBUG
4704 static void iwl_print_rx_config_cmd(struct iwl_rxon_cmd
*rxon
)
4706 DECLARE_MAC_BUF(mac
);
4708 IWL_DEBUG_RADIO("RX CONFIG:\n");
4709 iwl_print_hex_dump(IWL_DL_RADIO
, (u8
*) rxon
, sizeof(*rxon
));
4710 IWL_DEBUG_RADIO("u16 channel: 0x%x\n", le16_to_cpu(rxon
->channel
));
4711 IWL_DEBUG_RADIO("u32 flags: 0x%08X\n", le32_to_cpu(rxon
->flags
));
4712 IWL_DEBUG_RADIO("u32 filter_flags: 0x%08x\n",
4713 le32_to_cpu(rxon
->filter_flags
));
4714 IWL_DEBUG_RADIO("u8 dev_type: 0x%x\n", rxon
->dev_type
);
4715 IWL_DEBUG_RADIO("u8 ofdm_basic_rates: 0x%02x\n",
4716 rxon
->ofdm_basic_rates
);
4717 IWL_DEBUG_RADIO("u8 cck_basic_rates: 0x%02x\n", rxon
->cck_basic_rates
);
4718 IWL_DEBUG_RADIO("u8[6] node_addr: %s\n",
4719 print_mac(mac
, rxon
->node_addr
));
4720 IWL_DEBUG_RADIO("u8[6] bssid_addr: %s\n",
4721 print_mac(mac
, rxon
->bssid_addr
));
4722 IWL_DEBUG_RADIO("u16 assoc_id: 0x%x\n", le16_to_cpu(rxon
->assoc_id
));
4726 static void iwl_enable_interrupts(struct iwl_priv
*priv
)
4728 IWL_DEBUG_ISR("Enabling interrupts\n");
4729 set_bit(STATUS_INT_ENABLED
, &priv
->status
);
4730 iwl_write32(priv
, CSR_INT_MASK
, CSR_INI_SET_MASK
);
4733 static inline void iwl_disable_interrupts(struct iwl_priv
*priv
)
4735 clear_bit(STATUS_INT_ENABLED
, &priv
->status
);
4737 /* disable interrupts from uCode/NIC to host */
4738 iwl_write32(priv
, CSR_INT_MASK
, 0x00000000);
4740 /* acknowledge/clear/reset any interrupts still pending
4741 * from uCode or flow handler (Rx/Tx DMA) */
4742 iwl_write32(priv
, CSR_INT
, 0xffffffff);
4743 iwl_write32(priv
, CSR_FH_INT_STATUS
, 0xffffffff);
4744 IWL_DEBUG_ISR("Disabled interrupts\n");
4747 static const char *desc_lookup(int i
)
4755 return "BAD_CHECKSUM";
4757 return "NMI_INTERRUPT";
4761 return "FATAL_ERROR";
4767 #define ERROR_START_OFFSET (1 * sizeof(u32))
4768 #define ERROR_ELEM_SIZE (7 * sizeof(u32))
4770 static void iwl_dump_nic_error_log(struct iwl_priv
*priv
)
4773 u32 desc
, time
, count
, base
, data1
;
4774 u32 blink1
, blink2
, ilink1
, ilink2
;
4777 base
= le32_to_cpu(priv
->card_alive
.error_event_table_ptr
);
4779 if (!iwl_hw_valid_rtc_data_addr(base
)) {
4780 IWL_ERROR("Not valid error log pointer 0x%08X\n", base
);
4784 rc
= iwl_grab_nic_access(priv
);
4786 IWL_WARNING("Can not read from adapter at this time.\n");
4790 count
= iwl_read_targ_mem(priv
, base
);
4792 if (ERROR_START_OFFSET
<= count
* ERROR_ELEM_SIZE
) {
4793 IWL_ERROR("Start IWL Error Log Dump:\n");
4794 IWL_ERROR("Status: 0x%08lX, Config: %08X count: %d\n",
4795 priv
->status
, priv
->config
, count
);
4798 desc
= iwl_read_targ_mem(priv
, base
+ 1 * sizeof(u32
));
4799 blink1
= iwl_read_targ_mem(priv
, base
+ 3 * sizeof(u32
));
4800 blink2
= iwl_read_targ_mem(priv
, base
+ 4 * sizeof(u32
));
4801 ilink1
= iwl_read_targ_mem(priv
, base
+ 5 * sizeof(u32
));
4802 ilink2
= iwl_read_targ_mem(priv
, base
+ 6 * sizeof(u32
));
4803 data1
= iwl_read_targ_mem(priv
, base
+ 7 * sizeof(u32
));
4804 data2
= iwl_read_targ_mem(priv
, base
+ 8 * sizeof(u32
));
4805 line
= iwl_read_targ_mem(priv
, base
+ 9 * sizeof(u32
));
4806 time
= iwl_read_targ_mem(priv
, base
+ 11 * sizeof(u32
));
4808 IWL_ERROR("Desc Time "
4809 "data1 data2 line\n");
4810 IWL_ERROR("%-13s (#%d) %010u 0x%08X 0x%08X %u\n",
4811 desc_lookup(desc
), desc
, time
, data1
, data2
, line
);
4812 IWL_ERROR("blink1 blink2 ilink1 ilink2\n");
4813 IWL_ERROR("0x%05X 0x%05X 0x%05X 0x%05X\n", blink1
, blink2
,
4816 iwl_release_nic_access(priv
);
4819 #define EVENT_START_OFFSET (4 * sizeof(u32))
4822 * iwl_print_event_log - Dump error event log to syslog
4824 * NOTE: Must be called with iwl_grab_nic_access() already obtained!
4826 static void iwl_print_event_log(struct iwl_priv
*priv
, u32 start_idx
,
4827 u32 num_events
, u32 mode
)
4830 u32 base
; /* SRAM byte address of event log header */
4831 u32 event_size
; /* 2 u32s, or 3 u32s if timestamp recorded */
4832 u32 ptr
; /* SRAM byte address of log data */
4833 u32 ev
, time
, data
; /* event log data */
4835 if (num_events
== 0)
4838 base
= le32_to_cpu(priv
->card_alive
.log_event_table_ptr
);
4841 event_size
= 2 * sizeof(u32
);
4843 event_size
= 3 * sizeof(u32
);
4845 ptr
= base
+ EVENT_START_OFFSET
+ (start_idx
* event_size
);
4847 /* "time" is actually "data" for mode 0 (no timestamp).
4848 * place event id # at far right for easier visual parsing. */
4849 for (i
= 0; i
< num_events
; i
++) {
4850 ev
= iwl_read_targ_mem(priv
, ptr
);
4852 time
= iwl_read_targ_mem(priv
, ptr
);
4855 IWL_ERROR("0x%08x\t%04u\n", time
, ev
); /* data, ev */
4857 data
= iwl_read_targ_mem(priv
, ptr
);
4859 IWL_ERROR("%010u\t0x%08x\t%04u\n", time
, data
, ev
);
4864 static void iwl_dump_nic_event_log(struct iwl_priv
*priv
)
4867 u32 base
; /* SRAM byte address of event log header */
4868 u32 capacity
; /* event log capacity in # entries */
4869 u32 mode
; /* 0 - no timestamp, 1 - timestamp recorded */
4870 u32 num_wraps
; /* # times uCode wrapped to top of log */
4871 u32 next_entry
; /* index of next entry to be written by uCode */
4872 u32 size
; /* # entries that we'll print */
4874 base
= le32_to_cpu(priv
->card_alive
.log_event_table_ptr
);
4875 if (!iwl_hw_valid_rtc_data_addr(base
)) {
4876 IWL_ERROR("Invalid event log pointer 0x%08X\n", base
);
4880 rc
= iwl_grab_nic_access(priv
);
4882 IWL_WARNING("Can not read from adapter at this time.\n");
4886 /* event log header */
4887 capacity
= iwl_read_targ_mem(priv
, base
);
4888 mode
= iwl_read_targ_mem(priv
, base
+ (1 * sizeof(u32
)));
4889 num_wraps
= iwl_read_targ_mem(priv
, base
+ (2 * sizeof(u32
)));
4890 next_entry
= iwl_read_targ_mem(priv
, base
+ (3 * sizeof(u32
)));
4892 size
= num_wraps
? capacity
: next_entry
;
4894 /* bail out if nothing in log */
4896 IWL_ERROR("Start IWL Event Log Dump: nothing in log\n");
4897 iwl_release_nic_access(priv
);
4901 IWL_ERROR("Start IWL Event Log Dump: display count %d, wraps %d\n",
4904 /* if uCode has wrapped back to top of log, start at the oldest entry,
4905 * i.e the next one that uCode would fill. */
4907 iwl_print_event_log(priv
, next_entry
,
4908 capacity
- next_entry
, mode
);
4910 /* (then/else) start at top of log */
4911 iwl_print_event_log(priv
, 0, next_entry
, mode
);
4913 iwl_release_nic_access(priv
);
4917 * iwl_irq_handle_error - called for HW or SW error interrupt from card
4919 static void iwl_irq_handle_error(struct iwl_priv
*priv
)
4921 /* Set the FW error flag -- cleared on iwl_down */
4922 set_bit(STATUS_FW_ERROR
, &priv
->status
);
4924 /* Cancel currently queued command. */
4925 clear_bit(STATUS_HCMD_ACTIVE
, &priv
->status
);
4927 #ifdef CONFIG_IWL4965_DEBUG
4928 if (iwl_debug_level
& IWL_DL_FW_ERRORS
) {
4929 iwl_dump_nic_error_log(priv
);
4930 iwl_dump_nic_event_log(priv
);
4931 iwl_print_rx_config_cmd(&priv
->staging_rxon
);
4935 wake_up_interruptible(&priv
->wait_command_queue
);
4937 /* Keep the restart process from trying to send host
4938 * commands by clearing the INIT status bit */
4939 clear_bit(STATUS_READY
, &priv
->status
);
4941 if (!test_bit(STATUS_EXIT_PENDING
, &priv
->status
)) {
4942 IWL_DEBUG(IWL_DL_INFO
| IWL_DL_FW_ERRORS
,
4943 "Restarting adapter due to uCode error.\n");
4945 if (iwl_is_associated(priv
)) {
4946 memcpy(&priv
->recovery_rxon
, &priv
->active_rxon
,
4947 sizeof(priv
->recovery_rxon
));
4948 priv
->error_recovering
= 1;
4950 queue_work(priv
->workqueue
, &priv
->restart
);
4954 static void iwl_error_recovery(struct iwl_priv
*priv
)
4956 unsigned long flags
;
4958 memcpy(&priv
->staging_rxon
, &priv
->recovery_rxon
,
4959 sizeof(priv
->staging_rxon
));
4960 priv
->staging_rxon
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
4961 iwl_commit_rxon(priv
);
4963 iwl_rxon_add_station(priv
, priv
->bssid
, 1);
4965 spin_lock_irqsave(&priv
->lock
, flags
);
4966 priv
->assoc_id
= le16_to_cpu(priv
->staging_rxon
.assoc_id
);
4967 priv
->error_recovering
= 0;
4968 spin_unlock_irqrestore(&priv
->lock
, flags
);
4971 static void iwl_irq_tasklet(struct iwl_priv
*priv
)
4973 u32 inta
, handled
= 0;
4975 unsigned long flags
;
4976 #ifdef CONFIG_IWL4965_DEBUG
4980 spin_lock_irqsave(&priv
->lock
, flags
);
4982 /* Ack/clear/reset pending uCode interrupts.
4983 * Note: Some bits in CSR_INT are "OR" of bits in CSR_FH_INT_STATUS,
4984 * and will clear only when CSR_FH_INT_STATUS gets cleared. */
4985 inta
= iwl_read32(priv
, CSR_INT
);
4986 iwl_write32(priv
, CSR_INT
, inta
);
4988 /* Ack/clear/reset pending flow-handler (DMA) interrupts.
4989 * Any new interrupts that happen after this, either while we're
4990 * in this tasklet, or later, will show up in next ISR/tasklet. */
4991 inta_fh
= iwl_read32(priv
, CSR_FH_INT_STATUS
);
4992 iwl_write32(priv
, CSR_FH_INT_STATUS
, inta_fh
);
4994 #ifdef CONFIG_IWL4965_DEBUG
4995 if (iwl_debug_level
& IWL_DL_ISR
) {
4996 inta_mask
= iwl_read32(priv
, CSR_INT_MASK
); /* just for debug */
4997 IWL_DEBUG_ISR("inta 0x%08x, enabled 0x%08x, fh 0x%08x\n",
4998 inta
, inta_mask
, inta_fh
);
5002 /* Since CSR_INT and CSR_FH_INT_STATUS reads and clears are not
5003 * atomic, make sure that inta covers all the interrupts that
5004 * we've discovered, even if FH interrupt came in just after
5005 * reading CSR_INT. */
5006 if (inta_fh
& CSR_FH_INT_RX_MASK
)
5007 inta
|= CSR_INT_BIT_FH_RX
;
5008 if (inta_fh
& CSR_FH_INT_TX_MASK
)
5009 inta
|= CSR_INT_BIT_FH_TX
;
5011 /* Now service all interrupt bits discovered above. */
5012 if (inta
& CSR_INT_BIT_HW_ERR
) {
5013 IWL_ERROR("Microcode HW error detected. Restarting.\n");
5015 /* Tell the device to stop sending interrupts */
5016 iwl_disable_interrupts(priv
);
5018 iwl_irq_handle_error(priv
);
5020 handled
|= CSR_INT_BIT_HW_ERR
;
5022 spin_unlock_irqrestore(&priv
->lock
, flags
);
5027 #ifdef CONFIG_IWL4965_DEBUG
5028 if (iwl_debug_level
& (IWL_DL_ISR
)) {
5029 /* NIC fires this, but we don't use it, redundant with WAKEUP */
5030 if (inta
& CSR_INT_BIT_MAC_CLK_ACTV
)
5031 IWL_DEBUG_ISR("Microcode started or stopped.\n");
5033 /* Alive notification via Rx interrupt will do the real work */
5034 if (inta
& CSR_INT_BIT_ALIVE
)
5035 IWL_DEBUG_ISR("Alive interrupt\n");
5038 /* Safely ignore these bits for debug checks below */
5039 inta
&= ~(CSR_INT_BIT_MAC_CLK_ACTV
| CSR_INT_BIT_ALIVE
);
5041 /* HW RF KILL switch toggled (4965 only) */
5042 if (inta
& CSR_INT_BIT_RF_KILL
) {
5044 if (!(iwl_read32(priv
, CSR_GP_CNTRL
) &
5045 CSR_GP_CNTRL_REG_FLAG_HW_RF_KILL_SW
))
5048 IWL_DEBUG(IWL_DL_INFO
| IWL_DL_RF_KILL
| IWL_DL_ISR
,
5049 "RF_KILL bit toggled to %s.\n",
5050 hw_rf_kill
? "disable radio":"enable radio");
5052 /* Queue restart only if RF_KILL switch was set to "kill"
5053 * when we loaded driver, and is now set to "enable".
5054 * After we're Alive, RF_KILL gets handled by
5055 * iwl_rx_card_state_notif() */
5056 if (!hw_rf_kill
&& !test_bit(STATUS_ALIVE
, &priv
->status
)) {
5057 clear_bit(STATUS_RF_KILL_HW
, &priv
->status
);
5058 queue_work(priv
->workqueue
, &priv
->restart
);
5061 handled
|= CSR_INT_BIT_RF_KILL
;
5064 /* Chip got too hot and stopped itself (4965 only) */
5065 if (inta
& CSR_INT_BIT_CT_KILL
) {
5066 IWL_ERROR("Microcode CT kill error detected.\n");
5067 handled
|= CSR_INT_BIT_CT_KILL
;
5070 /* Error detected by uCode */
5071 if (inta
& CSR_INT_BIT_SW_ERR
) {
5072 IWL_ERROR("Microcode SW error detected. Restarting 0x%X.\n",
5074 iwl_irq_handle_error(priv
);
5075 handled
|= CSR_INT_BIT_SW_ERR
;
5078 /* uCode wakes up after power-down sleep */
5079 if (inta
& CSR_INT_BIT_WAKEUP
) {
5080 IWL_DEBUG_ISR("Wakeup interrupt\n");
5081 iwl_rx_queue_update_write_ptr(priv
, &priv
->rxq
);
5082 iwl_tx_queue_update_write_ptr(priv
, &priv
->txq
[0]);
5083 iwl_tx_queue_update_write_ptr(priv
, &priv
->txq
[1]);
5084 iwl_tx_queue_update_write_ptr(priv
, &priv
->txq
[2]);
5085 iwl_tx_queue_update_write_ptr(priv
, &priv
->txq
[3]);
5086 iwl_tx_queue_update_write_ptr(priv
, &priv
->txq
[4]);
5087 iwl_tx_queue_update_write_ptr(priv
, &priv
->txq
[5]);
5089 handled
|= CSR_INT_BIT_WAKEUP
;
5092 /* All uCode command responses, including Tx command responses,
5093 * Rx "responses" (frame-received notification), and other
5094 * notifications from uCode come through here*/
5095 if (inta
& (CSR_INT_BIT_FH_RX
| CSR_INT_BIT_SW_RX
)) {
5096 iwl_rx_handle(priv
);
5097 handled
|= (CSR_INT_BIT_FH_RX
| CSR_INT_BIT_SW_RX
);
5100 if (inta
& CSR_INT_BIT_FH_TX
) {
5101 IWL_DEBUG_ISR("Tx interrupt\n");
5102 handled
|= CSR_INT_BIT_FH_TX
;
5105 if (inta
& ~handled
)
5106 IWL_ERROR("Unhandled INTA bits 0x%08x\n", inta
& ~handled
);
5108 if (inta
& ~CSR_INI_SET_MASK
) {
5109 IWL_WARNING("Disabled INTA bits 0x%08x were pending\n",
5110 inta
& ~CSR_INI_SET_MASK
);
5111 IWL_WARNING(" with FH_INT = 0x%08x\n", inta_fh
);
5114 /* Re-enable all interrupts */
5115 iwl_enable_interrupts(priv
);
5117 #ifdef CONFIG_IWL4965_DEBUG
5118 if (iwl_debug_level
& (IWL_DL_ISR
)) {
5119 inta
= iwl_read32(priv
, CSR_INT
);
5120 inta_mask
= iwl_read32(priv
, CSR_INT_MASK
);
5121 inta_fh
= iwl_read32(priv
, CSR_FH_INT_STATUS
);
5122 IWL_DEBUG_ISR("End inta 0x%08x, enabled 0x%08x, fh 0x%08x, "
5123 "flags 0x%08lx\n", inta
, inta_mask
, inta_fh
, flags
);
5126 spin_unlock_irqrestore(&priv
->lock
, flags
);
5129 static irqreturn_t
iwl_isr(int irq
, void *data
)
5131 struct iwl_priv
*priv
= data
;
5132 u32 inta
, inta_mask
;
5137 spin_lock(&priv
->lock
);
5139 /* Disable (but don't clear!) interrupts here to avoid
5140 * back-to-back ISRs and sporadic interrupts from our NIC.
5141 * If we have something to service, the tasklet will re-enable ints.
5142 * If we *don't* have something, we'll re-enable before leaving here. */
5143 inta_mask
= iwl_read32(priv
, CSR_INT_MASK
); /* just for debug */
5144 iwl_write32(priv
, CSR_INT_MASK
, 0x00000000);
5146 /* Discover which interrupts are active/pending */
5147 inta
= iwl_read32(priv
, CSR_INT
);
5148 inta_fh
= iwl_read32(priv
, CSR_FH_INT_STATUS
);
5150 /* Ignore interrupt if there's nothing in NIC to service.
5151 * This may be due to IRQ shared with another device,
5152 * or due to sporadic interrupts thrown from our NIC. */
5153 if (!inta
&& !inta_fh
) {
5154 IWL_DEBUG_ISR("Ignore interrupt, inta == 0, inta_fh == 0\n");
5158 if ((inta
== 0xFFFFFFFF) || ((inta
& 0xFFFFFFF0) == 0xa5a5a5a0)) {
5159 /* Hardware disappeared. It might have already raised
5161 IWL_WARNING("HARDWARE GONE?? INTA == 0x%080x\n", inta
);
5165 IWL_DEBUG_ISR("ISR inta 0x%08x, enabled 0x%08x, fh 0x%08x\n",
5166 inta
, inta_mask
, inta_fh
);
5168 /* iwl_irq_tasklet() will service interrupts and re-enable them */
5169 tasklet_schedule(&priv
->irq_tasklet
);
5172 spin_unlock(&priv
->lock
);
5176 /* re-enable interrupts here since we don't have anything to service. */
5177 iwl_enable_interrupts(priv
);
5178 spin_unlock(&priv
->lock
);
5182 /************************** EEPROM BANDS ****************************
5184 * The iwl_eeprom_band definitions below provide the mapping from the
5185 * EEPROM contents to the specific channel number supported for each
5188 * For example, iwl_priv->eeprom.band_3_channels[4] from the band_3
5189 * definition below maps to physical channel 42 in the 5.2GHz spectrum.
5190 * The specific geography and calibration information for that channel
5191 * is contained in the eeprom map itself.
5193 * During init, we copy the eeprom information and channel map
5194 * information into priv->channel_info_24/52 and priv->channel_map_24/52
5196 * channel_map_24/52 provides the index in the channel_info array for a
5197 * given channel. We have to have two separate maps as there is channel
5198 * overlap with the 2.4GHz and 5.2GHz spectrum as seen in band_1 and
5201 * A value of 0xff stored in the channel_map indicates that the channel
5202 * is not supported by the hardware at all.
5204 * A value of 0xfe in the channel_map indicates that the channel is not
5205 * valid for Tx with the current hardware. This means that
5206 * while the system can tune and receive on a given channel, it may not
5207 * be able to associate or transmit any frames on that
5208 * channel. There is no corresponding channel information for that
5211 *********************************************************************/
5214 static const u8 iwl_eeprom_band_1
[14] = {
5215 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14
5219 static const u8 iwl_eeprom_band_2
[] = {
5220 183, 184, 185, 187, 188, 189, 192, 196, 7, 8, 11, 12, 16
5223 static const u8 iwl_eeprom_band_3
[] = { /* 5205-5320MHz */
5224 34, 36, 38, 40, 42, 44, 46, 48, 52, 56, 60, 64
5227 static const u8 iwl_eeprom_band_4
[] = { /* 5500-5700MHz */
5228 100, 104, 108, 112, 116, 120, 124, 128, 132, 136, 140
5231 static const u8 iwl_eeprom_band_5
[] = { /* 5725-5825MHz */
5232 145, 149, 153, 157, 161, 165
5235 static u8 iwl_eeprom_band_6
[] = { /* 2.4 FAT channel */
5239 static u8 iwl_eeprom_band_7
[] = { /* 5.2 FAT channel */
5240 36, 44, 52, 60, 100, 108, 116, 124, 132, 149, 157
5243 static void iwl_init_band_reference(const struct iwl_priv
*priv
, int band
,
5244 int *eeprom_ch_count
,
5245 const struct iwl_eeprom_channel
5247 const u8
**eeprom_ch_index
)
5250 case 1: /* 2.4GHz band */
5251 *eeprom_ch_count
= ARRAY_SIZE(iwl_eeprom_band_1
);
5252 *eeprom_ch_info
= priv
->eeprom
.band_1_channels
;
5253 *eeprom_ch_index
= iwl_eeprom_band_1
;
5255 case 2: /* 5.2GHz band */
5256 *eeprom_ch_count
= ARRAY_SIZE(iwl_eeprom_band_2
);
5257 *eeprom_ch_info
= priv
->eeprom
.band_2_channels
;
5258 *eeprom_ch_index
= iwl_eeprom_band_2
;
5260 case 3: /* 5.2GHz band */
5261 *eeprom_ch_count
= ARRAY_SIZE(iwl_eeprom_band_3
);
5262 *eeprom_ch_info
= priv
->eeprom
.band_3_channels
;
5263 *eeprom_ch_index
= iwl_eeprom_band_3
;
5265 case 4: /* 5.2GHz band */
5266 *eeprom_ch_count
= ARRAY_SIZE(iwl_eeprom_band_4
);
5267 *eeprom_ch_info
= priv
->eeprom
.band_4_channels
;
5268 *eeprom_ch_index
= iwl_eeprom_band_4
;
5270 case 5: /* 5.2GHz band */
5271 *eeprom_ch_count
= ARRAY_SIZE(iwl_eeprom_band_5
);
5272 *eeprom_ch_info
= priv
->eeprom
.band_5_channels
;
5273 *eeprom_ch_index
= iwl_eeprom_band_5
;
5276 *eeprom_ch_count
= ARRAY_SIZE(iwl_eeprom_band_6
);
5277 *eeprom_ch_info
= priv
->eeprom
.band_24_channels
;
5278 *eeprom_ch_index
= iwl_eeprom_band_6
;
5281 *eeprom_ch_count
= ARRAY_SIZE(iwl_eeprom_band_7
);
5282 *eeprom_ch_info
= priv
->eeprom
.band_52_channels
;
5283 *eeprom_ch_index
= iwl_eeprom_band_7
;
5291 const struct iwl_channel_info
*iwl_get_channel_info(const struct iwl_priv
*priv
,
5292 int phymode
, u16 channel
)
5297 case MODE_IEEE80211A
:
5298 for (i
= 14; i
< priv
->channel_count
; i
++) {
5299 if (priv
->channel_info
[i
].channel
== channel
)
5300 return &priv
->channel_info
[i
];
5304 case MODE_IEEE80211B
:
5305 case MODE_IEEE80211G
:
5306 if (channel
>= 1 && channel
<= 14)
5307 return &priv
->channel_info
[channel
- 1];
5315 #define CHECK_AND_PRINT(x) ((eeprom_ch_info[ch].flags & EEPROM_CHANNEL_##x) \
5318 static int iwl_init_channel_map(struct iwl_priv
*priv
)
5320 int eeprom_ch_count
= 0;
5321 const u8
*eeprom_ch_index
= NULL
;
5322 const struct iwl_eeprom_channel
*eeprom_ch_info
= NULL
;
5324 struct iwl_channel_info
*ch_info
;
5326 if (priv
->channel_count
) {
5327 IWL_DEBUG_INFO("Channel map already initialized.\n");
5331 if (priv
->eeprom
.version
< 0x2f) {
5332 IWL_WARNING("Unsupported EEPROM version: 0x%04X\n",
5333 priv
->eeprom
.version
);
5337 IWL_DEBUG_INFO("Initializing regulatory info from EEPROM\n");
5339 priv
->channel_count
=
5340 ARRAY_SIZE(iwl_eeprom_band_1
) +
5341 ARRAY_SIZE(iwl_eeprom_band_2
) +
5342 ARRAY_SIZE(iwl_eeprom_band_3
) +
5343 ARRAY_SIZE(iwl_eeprom_band_4
) +
5344 ARRAY_SIZE(iwl_eeprom_band_5
);
5346 IWL_DEBUG_INFO("Parsing data for %d channels.\n", priv
->channel_count
);
5348 priv
->channel_info
= kzalloc(sizeof(struct iwl_channel_info
) *
5349 priv
->channel_count
, GFP_KERNEL
);
5350 if (!priv
->channel_info
) {
5351 IWL_ERROR("Could not allocate channel_info\n");
5352 priv
->channel_count
= 0;
5356 ch_info
= priv
->channel_info
;
5358 /* Loop through the 5 EEPROM bands adding them in order to the
5359 * channel map we maintain (that contains additional information than
5360 * what just in the EEPROM) */
5361 for (band
= 1; band
<= 5; band
++) {
5363 iwl_init_band_reference(priv
, band
, &eeprom_ch_count
,
5364 &eeprom_ch_info
, &eeprom_ch_index
);
5366 /* Loop through each band adding each of the channels */
5367 for (ch
= 0; ch
< eeprom_ch_count
; ch
++) {
5368 ch_info
->channel
= eeprom_ch_index
[ch
];
5369 ch_info
->phymode
= (band
== 1) ? MODE_IEEE80211B
:
5372 /* permanently store EEPROM's channel regulatory flags
5373 * and max power in channel info database. */
5374 ch_info
->eeprom
= eeprom_ch_info
[ch
];
5376 /* Copy the run-time flags so they are there even on
5377 * invalid channels */
5378 ch_info
->flags
= eeprom_ch_info
[ch
].flags
;
5380 if (!(is_channel_valid(ch_info
))) {
5381 IWL_DEBUG_INFO("Ch. %d Flags %x [%sGHz] - "
5385 is_channel_a_band(ch_info
) ?
5391 /* Initialize regulatory-based run-time data */
5392 ch_info
->max_power_avg
= ch_info
->curr_txpow
=
5393 eeprom_ch_info
[ch
].max_power_avg
;
5394 ch_info
->scan_power
= eeprom_ch_info
[ch
].max_power_avg
;
5395 ch_info
->min_power
= 0;
5397 IWL_DEBUG_INFO("Ch. %d [%sGHz] %s%s%s%s%s%s(0x%02x"
5398 " %ddBm): Ad-Hoc %ssupported\n",
5400 is_channel_a_band(ch_info
) ?
5402 CHECK_AND_PRINT(IBSS
),
5403 CHECK_AND_PRINT(ACTIVE
),
5404 CHECK_AND_PRINT(RADAR
),
5405 CHECK_AND_PRINT(WIDE
),
5406 CHECK_AND_PRINT(NARROW
),
5407 CHECK_AND_PRINT(DFS
),
5408 eeprom_ch_info
[ch
].flags
,
5409 eeprom_ch_info
[ch
].max_power_avg
,
5410 ((eeprom_ch_info
[ch
].
5411 flags
& EEPROM_CHANNEL_IBSS
)
5412 && !(eeprom_ch_info
[ch
].
5413 flags
& EEPROM_CHANNEL_RADAR
))
5416 /* Set the user_txpower_limit to the highest power
5417 * supported by any channel */
5418 if (eeprom_ch_info
[ch
].max_power_avg
>
5419 priv
->user_txpower_limit
)
5420 priv
->user_txpower_limit
=
5421 eeprom_ch_info
[ch
].max_power_avg
;
5427 for (band
= 6; band
<= 7; band
++) {
5429 u8 fat_extension_chan
;
5431 iwl_init_band_reference(priv
, band
, &eeprom_ch_count
,
5432 &eeprom_ch_info
, &eeprom_ch_index
);
5434 phymode
= (band
== 6) ? MODE_IEEE80211B
: MODE_IEEE80211A
;
5435 /* Loop through each band adding each of the channels */
5436 for (ch
= 0; ch
< eeprom_ch_count
; ch
++) {
5439 ((eeprom_ch_index
[ch
] == 5) ||
5440 (eeprom_ch_index
[ch
] == 6) ||
5441 (eeprom_ch_index
[ch
] == 7)))
5442 fat_extension_chan
= HT_IE_EXT_CHANNEL_MAX
;
5444 fat_extension_chan
= HT_IE_EXT_CHANNEL_ABOVE
;
5446 iwl4965_set_fat_chan_info(priv
, phymode
,
5447 eeprom_ch_index
[ch
],
5448 &(eeprom_ch_info
[ch
]),
5449 fat_extension_chan
);
5451 iwl4965_set_fat_chan_info(priv
, phymode
,
5452 (eeprom_ch_index
[ch
] + 4),
5453 &(eeprom_ch_info
[ch
]),
5454 HT_IE_EXT_CHANNEL_BELOW
);
5461 /* For active scan, listen ACTIVE_DWELL_TIME (msec) on each channel after
5462 * sending probe req. This should be set long enough to hear probe responses
5463 * from more than one AP. */
5464 #define IWL_ACTIVE_DWELL_TIME_24 (20) /* all times in msec */
5465 #define IWL_ACTIVE_DWELL_TIME_52 (10)
5467 /* For faster active scanning, scan will move to the next channel if fewer than
5468 * PLCP_QUIET_THRESH packets are heard on this channel within
5469 * ACTIVE_QUIET_TIME after sending probe request. This shortens the dwell
5470 * time if it's a quiet channel (nothing responded to our probe, and there's
5471 * no other traffic).
5472 * Disable "quiet" feature by setting PLCP_QUIET_THRESH to 0. */
5473 #define IWL_PLCP_QUIET_THRESH __constant_cpu_to_le16(1) /* packets */
5474 #define IWL_ACTIVE_QUIET_TIME __constant_cpu_to_le16(5) /* msec */
5476 /* For passive scan, listen PASSIVE_DWELL_TIME (msec) on each channel.
5477 * Must be set longer than active dwell time.
5478 * For the most reliable scan, set > AP beacon interval (typically 100msec). */
5479 #define IWL_PASSIVE_DWELL_TIME_24 (20) /* all times in msec */
5480 #define IWL_PASSIVE_DWELL_TIME_52 (10)
5481 #define IWL_PASSIVE_DWELL_BASE (100)
5482 #define IWL_CHANNEL_TUNE_TIME 5
5484 static inline u16
iwl_get_active_dwell_time(struct iwl_priv
*priv
, int phymode
)
5486 if (phymode
== MODE_IEEE80211A
)
5487 return IWL_ACTIVE_DWELL_TIME_52
;
5489 return IWL_ACTIVE_DWELL_TIME_24
;
5492 static u16
iwl_get_passive_dwell_time(struct iwl_priv
*priv
, int phymode
)
5494 u16 active
= iwl_get_active_dwell_time(priv
, phymode
);
5495 u16 passive
= (phymode
!= MODE_IEEE80211A
) ?
5496 IWL_PASSIVE_DWELL_BASE
+ IWL_PASSIVE_DWELL_TIME_24
:
5497 IWL_PASSIVE_DWELL_BASE
+ IWL_PASSIVE_DWELL_TIME_52
;
5499 if (iwl_is_associated(priv
)) {
5500 /* If we're associated, we clamp the maximum passive
5501 * dwell time to be 98% of the beacon interval (minus
5502 * 2 * channel tune time) */
5503 passive
= priv
->beacon_int
;
5504 if ((passive
> IWL_PASSIVE_DWELL_BASE
) || !passive
)
5505 passive
= IWL_PASSIVE_DWELL_BASE
;
5506 passive
= (passive
* 98) / 100 - IWL_CHANNEL_TUNE_TIME
* 2;
5509 if (passive
<= active
)
5510 passive
= active
+ 1;
5515 static int iwl_get_channels_for_scan(struct iwl_priv
*priv
, int phymode
,
5516 u8 is_active
, u8 direct_mask
,
5517 struct iwl_scan_channel
*scan_ch
)
5519 const struct ieee80211_channel
*channels
= NULL
;
5520 const struct ieee80211_hw_mode
*hw_mode
;
5521 const struct iwl_channel_info
*ch_info
;
5522 u16 passive_dwell
= 0;
5523 u16 active_dwell
= 0;
5526 hw_mode
= iwl_get_hw_mode(priv
, phymode
);
5530 channels
= hw_mode
->channels
;
5532 active_dwell
= iwl_get_active_dwell_time(priv
, phymode
);
5533 passive_dwell
= iwl_get_passive_dwell_time(priv
, phymode
);
5535 for (i
= 0, added
= 0; i
< hw_mode
->num_channels
; i
++) {
5536 if (channels
[i
].chan
==
5537 le16_to_cpu(priv
->active_rxon
.channel
)) {
5538 if (iwl_is_associated(priv
)) {
5540 ("Skipping current channel %d\n",
5541 le16_to_cpu(priv
->active_rxon
.channel
));
5544 } else if (priv
->only_active_channel
)
5547 scan_ch
->channel
= channels
[i
].chan
;
5549 ch_info
= iwl_get_channel_info(priv
, phymode
, scan_ch
->channel
);
5550 if (!is_channel_valid(ch_info
)) {
5551 IWL_DEBUG_SCAN("Channel %d is INVALID for this SKU.\n",
5556 if (!is_active
|| is_channel_passive(ch_info
) ||
5557 !(channels
[i
].flag
& IEEE80211_CHAN_W_ACTIVE_SCAN
))
5558 scan_ch
->type
= 0; /* passive */
5560 scan_ch
->type
= 1; /* active */
5562 if (scan_ch
->type
& 1)
5563 scan_ch
->type
|= (direct_mask
<< 1);
5565 if (is_channel_narrow(ch_info
))
5566 scan_ch
->type
|= (1 << 7);
5568 scan_ch
->active_dwell
= cpu_to_le16(active_dwell
);
5569 scan_ch
->passive_dwell
= cpu_to_le16(passive_dwell
);
5571 /* Set power levels to defaults */
5572 scan_ch
->tpc
.dsp_atten
= 110;
5573 /* scan_pwr_info->tpc.dsp_atten; */
5575 /*scan_pwr_info->tpc.tx_gain; */
5576 if (phymode
== MODE_IEEE80211A
)
5577 scan_ch
->tpc
.tx_gain
= ((1 << 5) | (3 << 3)) | 3;
5579 scan_ch
->tpc
.tx_gain
= ((1 << 5) | (5 << 3));
5580 /* NOTE: if we were doing 6Mb OFDM for scans we'd use
5582 scan_ch->tpc.tx_gain = ((1<<5) | (2 << 3)) | 3;
5586 IWL_DEBUG_SCAN("Scanning %d [%s %d]\n",
5588 (scan_ch
->type
& 1) ? "ACTIVE" : "PASSIVE",
5589 (scan_ch
->type
& 1) ?
5590 active_dwell
: passive_dwell
);
5596 IWL_DEBUG_SCAN("total channels to scan %d \n", added
);
5600 static void iwl_reset_channel_flag(struct iwl_priv
*priv
)
5603 for (i
= 0; i
< 3; i
++) {
5604 struct ieee80211_hw_mode
*hw_mode
= (void *)&priv
->modes
[i
];
5605 for (j
= 0; j
< hw_mode
->num_channels
; j
++)
5606 hw_mode
->channels
[j
].flag
= hw_mode
->channels
[j
].val
;
5610 static void iwl_init_hw_rates(struct iwl_priv
*priv
,
5611 struct ieee80211_rate
*rates
)
5615 for (i
= 0; i
< IWL_RATE_COUNT
; i
++) {
5616 rates
[i
].rate
= iwl_rates
[i
].ieee
* 5;
5617 rates
[i
].val
= i
; /* Rate scaling will work on indexes */
5619 rates
[i
].flags
= IEEE80211_RATE_SUPPORTED
;
5620 /* Only OFDM have the bits-per-symbol set */
5621 if ((i
<= IWL_LAST_OFDM_RATE
) && (i
>= IWL_FIRST_OFDM_RATE
))
5622 rates
[i
].flags
|= IEEE80211_RATE_OFDM
;
5625 * If CCK 1M then set rate flag to CCK else CCK_2
5626 * which is CCK | PREAMBLE2
5628 rates
[i
].flags
|= (iwl_rates
[i
].plcp
== 10) ?
5629 IEEE80211_RATE_CCK
: IEEE80211_RATE_CCK_2
;
5632 /* Set up which ones are basic rates... */
5633 if (IWL_BASIC_RATES_MASK
& (1 << i
))
5634 rates
[i
].flags
|= IEEE80211_RATE_BASIC
;
5637 iwl4965_init_hw_rates(priv
, rates
);
5641 * iwl_init_geos - Initialize mac80211's geo/channel info based from eeprom
5643 static int iwl_init_geos(struct iwl_priv
*priv
)
5645 struct iwl_channel_info
*ch
;
5646 struct ieee80211_hw_mode
*modes
;
5647 struct ieee80211_channel
*channels
;
5648 struct ieee80211_channel
*geo_ch
;
5649 struct ieee80211_rate
*rates
;
5661 IWL_DEBUG_INFO("Geography modes already initialized.\n");
5662 set_bit(STATUS_GEO_CONFIGURED
, &priv
->status
);
5666 modes
= kzalloc(sizeof(struct ieee80211_hw_mode
) * mode_count
,
5671 channels
= kzalloc(sizeof(struct ieee80211_channel
) *
5672 priv
->channel_count
, GFP_KERNEL
);
5678 rates
= kzalloc((sizeof(struct ieee80211_rate
) * (IWL_MAX_RATES
+ 1)),
5691 /* 5.2GHz channels start after the 2.4GHz channels */
5692 modes
[A
].mode
= MODE_IEEE80211A
;
5693 modes
[A
].channels
= &channels
[ARRAY_SIZE(iwl_eeprom_band_1
)];
5694 modes
[A
].rates
= rates
;
5695 modes
[A
].num_rates
= 8; /* just OFDM */
5696 modes
[A
].rates
= &rates
[4];
5697 modes
[A
].num_channels
= 0;
5699 modes
[B
].mode
= MODE_IEEE80211B
;
5700 modes
[B
].channels
= channels
;
5701 modes
[B
].rates
= rates
;
5702 modes
[B
].num_rates
= 4; /* just CCK */
5703 modes
[B
].num_channels
= 0;
5705 modes
[G
].mode
= MODE_IEEE80211G
;
5706 modes
[G
].channels
= channels
;
5707 modes
[G
].rates
= rates
;
5708 modes
[G
].num_rates
= 12; /* OFDM & CCK */
5709 modes
[G
].num_channels
= 0;
5711 modes
[G_11N
].mode
= MODE_IEEE80211G
;
5712 modes
[G_11N
].channels
= channels
;
5713 modes
[G_11N
].num_rates
= 13; /* OFDM & CCK */
5714 modes
[G_11N
].rates
= rates
;
5715 modes
[G_11N
].num_channels
= 0;
5717 modes
[A_11N
].mode
= MODE_IEEE80211A
;
5718 modes
[A_11N
].channels
= &channels
[ARRAY_SIZE(iwl_eeprom_band_1
)];
5719 modes
[A_11N
].rates
= &rates
[4];
5720 modes
[A_11N
].num_rates
= 9; /* just OFDM */
5721 modes
[A_11N
].num_channels
= 0;
5723 priv
->ieee_channels
= channels
;
5724 priv
->ieee_rates
= rates
;
5726 iwl_init_hw_rates(priv
, rates
);
5728 for (i
= 0, geo_ch
= channels
; i
< priv
->channel_count
; i
++) {
5729 ch
= &priv
->channel_info
[i
];
5731 if (!is_channel_valid(ch
)) {
5732 IWL_DEBUG_INFO("Channel %d [%sGHz] is restricted -- "
5734 ch
->channel
, is_channel_a_band(ch
) ?
5739 if (is_channel_a_band(ch
)) {
5740 geo_ch
= &modes
[A
].channels
[modes
[A
].num_channels
++];
5741 modes
[A_11N
].num_channels
++;
5743 geo_ch
= &modes
[B
].channels
[modes
[B
].num_channels
++];
5744 modes
[G
].num_channels
++;
5745 modes
[G_11N
].num_channels
++;
5748 geo_ch
->freq
= ieee80211chan2mhz(ch
->channel
);
5749 geo_ch
->chan
= ch
->channel
;
5750 geo_ch
->power_level
= ch
->max_power_avg
;
5751 geo_ch
->antenna_max
= 0xff;
5753 if (is_channel_valid(ch
)) {
5754 geo_ch
->flag
= IEEE80211_CHAN_W_SCAN
;
5755 if (ch
->flags
& EEPROM_CHANNEL_IBSS
)
5756 geo_ch
->flag
|= IEEE80211_CHAN_W_IBSS
;
5758 if (ch
->flags
& EEPROM_CHANNEL_ACTIVE
)
5759 geo_ch
->flag
|= IEEE80211_CHAN_W_ACTIVE_SCAN
;
5761 if (ch
->flags
& EEPROM_CHANNEL_RADAR
)
5762 geo_ch
->flag
|= IEEE80211_CHAN_W_RADAR_DETECT
;
5764 if (ch
->max_power_avg
> priv
->max_channel_txpower_limit
)
5765 priv
->max_channel_txpower_limit
=
5769 geo_ch
->val
= geo_ch
->flag
;
5772 if ((modes
[A
].num_channels
== 0) && priv
->is_abg
) {
5773 printk(KERN_INFO DRV_NAME
5774 ": Incorrectly detected BG card as ABG. Please send "
5775 "your PCI ID 0x%04X:0x%04X to maintainer.\n",
5776 priv
->pci_dev
->device
, priv
->pci_dev
->subsystem_device
);
5780 printk(KERN_INFO DRV_NAME
5781 ": Tunable channels: %d 802.11bg, %d 802.11a channels\n",
5782 modes
[G
].num_channels
, modes
[A
].num_channels
);
5785 * NOTE: We register these in preference of order -- the
5786 * stack doesn't currently (as of 7.0.6 / Apr 24 '07) pick
5787 * a phymode based on rates or AP capabilities but seems to
5788 * configure it purely on if the channel being configured
5789 * is supported by a mode -- and the first match is taken
5792 if (modes
[G
].num_channels
)
5793 ieee80211_register_hwmode(priv
->hw
, &modes
[G
]);
5794 if (modes
[B
].num_channels
)
5795 ieee80211_register_hwmode(priv
->hw
, &modes
[B
]);
5796 if (modes
[A
].num_channels
)
5797 ieee80211_register_hwmode(priv
->hw
, &modes
[A
]);
5799 priv
->modes
= modes
;
5800 set_bit(STATUS_GEO_CONFIGURED
, &priv
->status
);
5805 /******************************************************************************
5807 * uCode download functions
5809 ******************************************************************************/
5811 static void iwl_dealloc_ucode_pci(struct iwl_priv
*priv
)
5813 if (priv
->ucode_code
.v_addr
!= NULL
) {
5814 pci_free_consistent(priv
->pci_dev
,
5815 priv
->ucode_code
.len
,
5816 priv
->ucode_code
.v_addr
,
5817 priv
->ucode_code
.p_addr
);
5818 priv
->ucode_code
.v_addr
= NULL
;
5820 if (priv
->ucode_data
.v_addr
!= NULL
) {
5821 pci_free_consistent(priv
->pci_dev
,
5822 priv
->ucode_data
.len
,
5823 priv
->ucode_data
.v_addr
,
5824 priv
->ucode_data
.p_addr
);
5825 priv
->ucode_data
.v_addr
= NULL
;
5827 if (priv
->ucode_data_backup
.v_addr
!= NULL
) {
5828 pci_free_consistent(priv
->pci_dev
,
5829 priv
->ucode_data_backup
.len
,
5830 priv
->ucode_data_backup
.v_addr
,
5831 priv
->ucode_data_backup
.p_addr
);
5832 priv
->ucode_data_backup
.v_addr
= NULL
;
5834 if (priv
->ucode_init
.v_addr
!= NULL
) {
5835 pci_free_consistent(priv
->pci_dev
,
5836 priv
->ucode_init
.len
,
5837 priv
->ucode_init
.v_addr
,
5838 priv
->ucode_init
.p_addr
);
5839 priv
->ucode_init
.v_addr
= NULL
;
5841 if (priv
->ucode_init_data
.v_addr
!= NULL
) {
5842 pci_free_consistent(priv
->pci_dev
,
5843 priv
->ucode_init_data
.len
,
5844 priv
->ucode_init_data
.v_addr
,
5845 priv
->ucode_init_data
.p_addr
);
5846 priv
->ucode_init_data
.v_addr
= NULL
;
5848 if (priv
->ucode_boot
.v_addr
!= NULL
) {
5849 pci_free_consistent(priv
->pci_dev
,
5850 priv
->ucode_boot
.len
,
5851 priv
->ucode_boot
.v_addr
,
5852 priv
->ucode_boot
.p_addr
);
5853 priv
->ucode_boot
.v_addr
= NULL
;
5858 * iwl_verify_inst_full - verify runtime uCode image in card vs. host,
5859 * looking at all data.
5861 static int iwl_verify_inst_full(struct iwl_priv
*priv
, __le32
* image
, u32 len
)
5868 IWL_DEBUG_INFO("ucode inst image size is %u\n", len
);
5870 rc
= iwl_grab_nic_access(priv
);
5874 iwl_write_direct32(priv
, HBUS_TARG_MEM_RADDR
, RTC_INST_LOWER_BOUND
);
5877 for (; len
> 0; len
-= sizeof(u32
), image
++) {
5878 /* read data comes through single port, auto-incr addr */
5879 /* NOTE: Use the debugless read so we don't flood kernel log
5880 * if IWL_DL_IO is set */
5881 val
= _iwl_read_direct32(priv
, HBUS_TARG_MEM_RDAT
);
5882 if (val
!= le32_to_cpu(*image
)) {
5883 IWL_ERROR("uCode INST section is invalid at "
5884 "offset 0x%x, is 0x%x, s/b 0x%x\n",
5885 save_len
- len
, val
, le32_to_cpu(*image
));
5893 iwl_release_nic_access(priv
);
5897 ("ucode image in INSTRUCTION memory is good\n");
5904 * iwl_verify_inst_sparse - verify runtime uCode image in card vs. host,
5905 * using sample data 100 bytes apart. If these sample points are good,
5906 * it's a pretty good bet that everything between them is good, too.
5908 static int iwl_verify_inst_sparse(struct iwl_priv
*priv
, __le32
*image
, u32 len
)
5915 IWL_DEBUG_INFO("ucode inst image size is %u\n", len
);
5917 rc
= iwl_grab_nic_access(priv
);
5921 for (i
= 0; i
< len
; i
+= 100, image
+= 100/sizeof(u32
)) {
5922 /* read data comes through single port, auto-incr addr */
5923 /* NOTE: Use the debugless read so we don't flood kernel log
5924 * if IWL_DL_IO is set */
5925 iwl_write_direct32(priv
, HBUS_TARG_MEM_RADDR
,
5926 i
+ RTC_INST_LOWER_BOUND
);
5927 val
= _iwl_read_direct32(priv
, HBUS_TARG_MEM_RDAT
);
5928 if (val
!= le32_to_cpu(*image
)) {
5929 #if 0 /* Enable this if you want to see details */
5930 IWL_ERROR("uCode INST section is invalid at "
5931 "offset 0x%x, is 0x%x, s/b 0x%x\n",
5941 iwl_release_nic_access(priv
);
5948 * iwl_verify_ucode - determine which instruction image is in SRAM,
5949 * and verify its contents
5951 static int iwl_verify_ucode(struct iwl_priv
*priv
)
5958 image
= (__le32
*)priv
->ucode_boot
.v_addr
;
5959 len
= priv
->ucode_boot
.len
;
5960 rc
= iwl_verify_inst_sparse(priv
, image
, len
);
5962 IWL_DEBUG_INFO("Bootstrap uCode is good in inst SRAM\n");
5966 /* Try initialize */
5967 image
= (__le32
*)priv
->ucode_init
.v_addr
;
5968 len
= priv
->ucode_init
.len
;
5969 rc
= iwl_verify_inst_sparse(priv
, image
, len
);
5971 IWL_DEBUG_INFO("Initialize uCode is good in inst SRAM\n");
5975 /* Try runtime/protocol */
5976 image
= (__le32
*)priv
->ucode_code
.v_addr
;
5977 len
= priv
->ucode_code
.len
;
5978 rc
= iwl_verify_inst_sparse(priv
, image
, len
);
5980 IWL_DEBUG_INFO("Runtime uCode is good in inst SRAM\n");
5984 IWL_ERROR("NO VALID UCODE IMAGE IN INSTRUCTION SRAM!!\n");
5986 /* Show first several data entries in instruction SRAM.
5987 * Selection of bootstrap image is arbitrary. */
5988 image
= (__le32
*)priv
->ucode_boot
.v_addr
;
5989 len
= priv
->ucode_boot
.len
;
5990 rc
= iwl_verify_inst_full(priv
, image
, len
);
5996 /* check contents of special bootstrap uCode SRAM */
5997 static int iwl_verify_bsm(struct iwl_priv
*priv
)
5999 __le32
*image
= priv
->ucode_boot
.v_addr
;
6000 u32 len
= priv
->ucode_boot
.len
;
6004 IWL_DEBUG_INFO("Begin verify bsm\n");
6006 /* verify BSM SRAM contents */
6007 val
= iwl_read_prph(priv
, BSM_WR_DWCOUNT_REG
);
6008 for (reg
= BSM_SRAM_LOWER_BOUND
;
6009 reg
< BSM_SRAM_LOWER_BOUND
+ len
;
6010 reg
+= sizeof(u32
), image
++) {
6011 val
= iwl_read_prph(priv
, reg
);
6012 if (val
!= le32_to_cpu(*image
)) {
6013 IWL_ERROR("BSM uCode verification failed at "
6014 "addr 0x%08X+%u (of %u), is 0x%x, s/b 0x%x\n",
6015 BSM_SRAM_LOWER_BOUND
,
6016 reg
- BSM_SRAM_LOWER_BOUND
, len
,
6017 val
, le32_to_cpu(*image
));
6022 IWL_DEBUG_INFO("BSM bootstrap uCode image OK\n");
6028 * iwl_load_bsm - Load bootstrap instructions
6032 * The Bootstrap State Machine (BSM) stores a short bootstrap uCode program
6033 * in special SRAM that does not power down during RFKILL. When powering back
6034 * up after power-saving sleeps (or during initial uCode load), the BSM loads
6035 * the bootstrap program into the on-board processor, and starts it.
6037 * The bootstrap program loads (via DMA) instructions and data for a new
6038 * program from host DRAM locations indicated by the host driver in the
6039 * BSM_DRAM_* registers. Once the new program is loaded, it starts
6042 * When initializing the NIC, the host driver points the BSM to the
6043 * "initialize" uCode image. This uCode sets up some internal data, then
6044 * notifies host via "initialize alive" that it is complete.
6046 * The host then replaces the BSM_DRAM_* pointer values to point to the
6047 * normal runtime uCode instructions and a backup uCode data cache buffer
6048 * (filled initially with starting data values for the on-board processor),
6049 * then triggers the "initialize" uCode to load and launch the runtime uCode,
6050 * which begins normal operation.
6052 * When doing a power-save shutdown, runtime uCode saves data SRAM into
6053 * the backup data cache in DRAM before SRAM is powered down.
6055 * When powering back up, the BSM loads the bootstrap program. This reloads
6056 * the runtime uCode instructions and the backup data cache into SRAM,
6057 * and re-launches the runtime uCode from where it left off.
6059 static int iwl_load_bsm(struct iwl_priv
*priv
)
6061 __le32
*image
= priv
->ucode_boot
.v_addr
;
6062 u32 len
= priv
->ucode_boot
.len
;
6072 IWL_DEBUG_INFO("Begin load bsm\n");
6074 /* make sure bootstrap program is no larger than BSM's SRAM size */
6075 if (len
> IWL_MAX_BSM_SIZE
)
6078 /* Tell bootstrap uCode where to find the "Initialize" uCode
6079 * in host DRAM ... bits 31:0 for 3945, bits 35:4 for 4965.
6080 * NOTE: iwl_initialize_alive_start() will replace these values,
6081 * after the "initialize" uCode has run, to point to
6082 * runtime/protocol instructions and backup data cache. */
6083 pinst
= priv
->ucode_init
.p_addr
>> 4;
6084 pdata
= priv
->ucode_init_data
.p_addr
>> 4;
6085 inst_len
= priv
->ucode_init
.len
;
6086 data_len
= priv
->ucode_init_data
.len
;
6088 rc
= iwl_grab_nic_access(priv
);
6092 iwl_write_prph(priv
, BSM_DRAM_INST_PTR_REG
, pinst
);
6093 iwl_write_prph(priv
, BSM_DRAM_DATA_PTR_REG
, pdata
);
6094 iwl_write_prph(priv
, BSM_DRAM_INST_BYTECOUNT_REG
, inst_len
);
6095 iwl_write_prph(priv
, BSM_DRAM_DATA_BYTECOUNT_REG
, data_len
);
6097 /* Fill BSM memory with bootstrap instructions */
6098 for (reg_offset
= BSM_SRAM_LOWER_BOUND
;
6099 reg_offset
< BSM_SRAM_LOWER_BOUND
+ len
;
6100 reg_offset
+= sizeof(u32
), image
++)
6101 _iwl_write_prph(priv
, reg_offset
,
6102 le32_to_cpu(*image
));
6104 rc
= iwl_verify_bsm(priv
);
6106 iwl_release_nic_access(priv
);
6110 /* Tell BSM to copy from BSM SRAM into instruction SRAM, when asked */
6111 iwl_write_prph(priv
, BSM_WR_MEM_SRC_REG
, 0x0);
6112 iwl_write_prph(priv
, BSM_WR_MEM_DST_REG
,
6113 RTC_INST_LOWER_BOUND
);
6114 iwl_write_prph(priv
, BSM_WR_DWCOUNT_REG
, len
/ sizeof(u32
));
6116 /* Load bootstrap code into instruction SRAM now,
6117 * to prepare to load "initialize" uCode */
6118 iwl_write_prph(priv
, BSM_WR_CTRL_REG
,
6119 BSM_WR_CTRL_REG_BIT_START
);
6121 /* Wait for load of bootstrap uCode to finish */
6122 for (i
= 0; i
< 100; i
++) {
6123 done
= iwl_read_prph(priv
, BSM_WR_CTRL_REG
);
6124 if (!(done
& BSM_WR_CTRL_REG_BIT_START
))
6129 IWL_DEBUG_INFO("BSM write complete, poll %d iterations\n", i
);
6131 IWL_ERROR("BSM write did not complete!\n");
6135 /* Enable future boot loads whenever power management unit triggers it
6136 * (e.g. when powering back up after power-save shutdown) */
6137 iwl_write_prph(priv
, BSM_WR_CTRL_REG
,
6138 BSM_WR_CTRL_REG_BIT_START_EN
);
6140 iwl_release_nic_access(priv
);
6145 static void iwl_nic_start(struct iwl_priv
*priv
)
6147 /* Remove all resets to allow NIC to operate */
6148 iwl_write32(priv
, CSR_RESET
, 0);
6152 * iwl_read_ucode - Read uCode images from disk file.
6154 * Copy into buffers for card to fetch via bus-mastering
6156 static int iwl_read_ucode(struct iwl_priv
*priv
)
6158 struct iwl_ucode
*ucode
;
6160 const struct firmware
*ucode_raw
;
6161 const char *name
= "iwlwifi-4965" IWL4965_UCODE_API
".ucode";
6164 u32 ver
, inst_size
, data_size
, init_size
, init_data_size
, boot_size
;
6166 /* Ask kernel firmware_class module to get the boot firmware off disk.
6167 * request_firmware() is synchronous, file is in memory on return. */
6168 rc
= request_firmware(&ucode_raw
, name
, &priv
->pci_dev
->dev
);
6170 IWL_ERROR("%s firmware file req failed: Reason %d\n", name
, rc
);
6174 IWL_DEBUG_INFO("Got firmware '%s' file (%zd bytes) from disk\n",
6175 name
, ucode_raw
->size
);
6177 /* Make sure that we got at least our header! */
6178 if (ucode_raw
->size
< sizeof(*ucode
)) {
6179 IWL_ERROR("File size way too small!\n");
6184 /* Data from ucode file: header followed by uCode images */
6185 ucode
= (void *)ucode_raw
->data
;
6187 ver
= le32_to_cpu(ucode
->ver
);
6188 inst_size
= le32_to_cpu(ucode
->inst_size
);
6189 data_size
= le32_to_cpu(ucode
->data_size
);
6190 init_size
= le32_to_cpu(ucode
->init_size
);
6191 init_data_size
= le32_to_cpu(ucode
->init_data_size
);
6192 boot_size
= le32_to_cpu(ucode
->boot_size
);
6194 IWL_DEBUG_INFO("f/w package hdr ucode version = 0x%x\n", ver
);
6195 IWL_DEBUG_INFO("f/w package hdr runtime inst size = %u\n",
6197 IWL_DEBUG_INFO("f/w package hdr runtime data size = %u\n",
6199 IWL_DEBUG_INFO("f/w package hdr init inst size = %u\n",
6201 IWL_DEBUG_INFO("f/w package hdr init data size = %u\n",
6203 IWL_DEBUG_INFO("f/w package hdr boot inst size = %u\n",
6206 /* Verify size of file vs. image size info in file's header */
6207 if (ucode_raw
->size
< sizeof(*ucode
) +
6208 inst_size
+ data_size
+ init_size
+
6209 init_data_size
+ boot_size
) {
6211 IWL_DEBUG_INFO("uCode file size %d too small\n",
6212 (int)ucode_raw
->size
);
6217 /* Verify that uCode images will fit in card's SRAM */
6218 if (inst_size
> IWL_MAX_INST_SIZE
) {
6219 IWL_DEBUG_INFO("uCode instr len %d too large to fit in card\n",
6225 if (data_size
> IWL_MAX_DATA_SIZE
) {
6226 IWL_DEBUG_INFO("uCode data len %d too large to fit in card\n",
6231 if (init_size
> IWL_MAX_INST_SIZE
) {
6233 ("uCode init instr len %d too large to fit in card\n",
6238 if (init_data_size
> IWL_MAX_DATA_SIZE
) {
6240 ("uCode init data len %d too large to fit in card\n",
6241 (int)init_data_size
);
6245 if (boot_size
> IWL_MAX_BSM_SIZE
) {
6247 ("uCode boot instr len %d too large to fit in bsm\n",
6253 /* Allocate ucode buffers for card's bus-master loading ... */
6255 /* Runtime instructions and 2 copies of data:
6256 * 1) unmodified from disk
6257 * 2) backup cache for save/restore during power-downs */
6258 priv
->ucode_code
.len
= inst_size
;
6259 priv
->ucode_code
.v_addr
=
6260 pci_alloc_consistent(priv
->pci_dev
,
6261 priv
->ucode_code
.len
,
6262 &(priv
->ucode_code
.p_addr
));
6264 priv
->ucode_data
.len
= data_size
;
6265 priv
->ucode_data
.v_addr
=
6266 pci_alloc_consistent(priv
->pci_dev
,
6267 priv
->ucode_data
.len
,
6268 &(priv
->ucode_data
.p_addr
));
6270 priv
->ucode_data_backup
.len
= data_size
;
6271 priv
->ucode_data_backup
.v_addr
=
6272 pci_alloc_consistent(priv
->pci_dev
,
6273 priv
->ucode_data_backup
.len
,
6274 &(priv
->ucode_data_backup
.p_addr
));
6277 /* Initialization instructions and data */
6278 priv
->ucode_init
.len
= init_size
;
6279 priv
->ucode_init
.v_addr
=
6280 pci_alloc_consistent(priv
->pci_dev
,
6281 priv
->ucode_init
.len
,
6282 &(priv
->ucode_init
.p_addr
));
6284 priv
->ucode_init_data
.len
= init_data_size
;
6285 priv
->ucode_init_data
.v_addr
=
6286 pci_alloc_consistent(priv
->pci_dev
,
6287 priv
->ucode_init_data
.len
,
6288 &(priv
->ucode_init_data
.p_addr
));
6290 /* Bootstrap (instructions only, no data) */
6291 priv
->ucode_boot
.len
= boot_size
;
6292 priv
->ucode_boot
.v_addr
=
6293 pci_alloc_consistent(priv
->pci_dev
,
6294 priv
->ucode_boot
.len
,
6295 &(priv
->ucode_boot
.p_addr
));
6297 if (!priv
->ucode_code
.v_addr
|| !priv
->ucode_data
.v_addr
||
6298 !priv
->ucode_init
.v_addr
|| !priv
->ucode_init_data
.v_addr
||
6299 !priv
->ucode_boot
.v_addr
|| !priv
->ucode_data_backup
.v_addr
)
6302 /* Copy images into buffers for card's bus-master reads ... */
6304 /* Runtime instructions (first block of data in file) */
6305 src
= &ucode
->data
[0];
6306 len
= priv
->ucode_code
.len
;
6307 IWL_DEBUG_INFO("Copying (but not loading) uCode instr len %d\n",
6309 memcpy(priv
->ucode_code
.v_addr
, src
, len
);
6310 IWL_DEBUG_INFO("uCode instr buf vaddr = 0x%p, paddr = 0x%08x\n",
6311 priv
->ucode_code
.v_addr
, (u32
)priv
->ucode_code
.p_addr
);
6313 /* Runtime data (2nd block)
6314 * NOTE: Copy into backup buffer will be done in iwl_up() */
6315 src
= &ucode
->data
[inst_size
];
6316 len
= priv
->ucode_data
.len
;
6317 IWL_DEBUG_INFO("Copying (but not loading) uCode data len %d\n",
6319 memcpy(priv
->ucode_data
.v_addr
, src
, len
);
6320 memcpy(priv
->ucode_data_backup
.v_addr
, src
, len
);
6322 /* Initialization instructions (3rd block) */
6324 src
= &ucode
->data
[inst_size
+ data_size
];
6325 len
= priv
->ucode_init
.len
;
6326 IWL_DEBUG_INFO("Copying (but not loading) init instr len %d\n",
6328 memcpy(priv
->ucode_init
.v_addr
, src
, len
);
6331 /* Initialization data (4th block) */
6332 if (init_data_size
) {
6333 src
= &ucode
->data
[inst_size
+ data_size
+ init_size
];
6334 len
= priv
->ucode_init_data
.len
;
6335 IWL_DEBUG_INFO("Copying (but not loading) init data len %d\n",
6337 memcpy(priv
->ucode_init_data
.v_addr
, src
, len
);
6340 /* Bootstrap instructions (5th block) */
6341 src
= &ucode
->data
[inst_size
+ data_size
+ init_size
+ init_data_size
];
6342 len
= priv
->ucode_boot
.len
;
6343 IWL_DEBUG_INFO("Copying (but not loading) boot instr len %d\n",
6345 memcpy(priv
->ucode_boot
.v_addr
, src
, len
);
6347 /* We have our copies now, allow OS release its copies */
6348 release_firmware(ucode_raw
);
6352 IWL_ERROR("failed to allocate pci memory\n");
6354 iwl_dealloc_ucode_pci(priv
);
6357 release_firmware(ucode_raw
);
6365 * iwl_set_ucode_ptrs - Set uCode address location
6367 * Tell initialization uCode where to find runtime uCode.
6369 * BSM registers initially contain pointers to initialization uCode.
6370 * We need to replace them to load runtime uCode inst and data,
6371 * and to save runtime data when powering down.
6373 static int iwl_set_ucode_ptrs(struct iwl_priv
*priv
)
6378 unsigned long flags
;
6380 /* bits 35:4 for 4965 */
6381 pinst
= priv
->ucode_code
.p_addr
>> 4;
6382 pdata
= priv
->ucode_data_backup
.p_addr
>> 4;
6384 spin_lock_irqsave(&priv
->lock
, flags
);
6385 rc
= iwl_grab_nic_access(priv
);
6387 spin_unlock_irqrestore(&priv
->lock
, flags
);
6391 /* Tell bootstrap uCode where to find image to load */
6392 iwl_write_prph(priv
, BSM_DRAM_INST_PTR_REG
, pinst
);
6393 iwl_write_prph(priv
, BSM_DRAM_DATA_PTR_REG
, pdata
);
6394 iwl_write_prph(priv
, BSM_DRAM_DATA_BYTECOUNT_REG
,
6395 priv
->ucode_data
.len
);
6397 /* Inst bytecount must be last to set up, bit 31 signals uCode
6398 * that all new ptr/size info is in place */
6399 iwl_write_prph(priv
, BSM_DRAM_INST_BYTECOUNT_REG
,
6400 priv
->ucode_code
.len
| BSM_DRAM_INST_LOAD
);
6402 iwl_release_nic_access(priv
);
6404 spin_unlock_irqrestore(&priv
->lock
, flags
);
6406 IWL_DEBUG_INFO("Runtime uCode pointers are set.\n");
6412 * iwl_init_alive_start - Called after REPLY_ALIVE notification received
6414 * Called after REPLY_ALIVE notification received from "initialize" uCode.
6416 * The 4965 "initialize" ALIVE reply contains calibration data for:
6417 * Voltage, temperature, and MIMO tx gain correction, now stored in priv
6418 * (3945 does not contain this data).
6420 * Tell "initialize" uCode to go ahead and load the runtime uCode.
6422 static void iwl_init_alive_start(struct iwl_priv
*priv
)
6424 /* Check alive response for "valid" sign from uCode */
6425 if (priv
->card_alive_init
.is_valid
!= UCODE_VALID_OK
) {
6426 /* We had an error bringing up the hardware, so take it
6427 * all the way back down so we can try again */
6428 IWL_DEBUG_INFO("Initialize Alive failed.\n");
6432 /* Bootstrap uCode has loaded initialize uCode ... verify inst image.
6433 * This is a paranoid check, because we would not have gotten the
6434 * "initialize" alive if code weren't properly loaded. */
6435 if (iwl_verify_ucode(priv
)) {
6436 /* Runtime instruction load was bad;
6437 * take it all the way back down so we can try again */
6438 IWL_DEBUG_INFO("Bad \"initialize\" uCode load.\n");
6442 /* Calculate temperature */
6443 priv
->temperature
= iwl4965_get_temperature(priv
);
6445 /* Send pointers to protocol/runtime uCode image ... init code will
6446 * load and launch runtime uCode, which will send us another "Alive"
6448 IWL_DEBUG_INFO("Initialization Alive received.\n");
6449 if (iwl_set_ucode_ptrs(priv
)) {
6450 /* Runtime instruction load won't happen;
6451 * take it all the way back down so we can try again */
6452 IWL_DEBUG_INFO("Couldn't set up uCode pointers.\n");
6458 queue_work(priv
->workqueue
, &priv
->restart
);
6463 * iwl_alive_start - called after REPLY_ALIVE notification received
6464 * from protocol/runtime uCode (initialization uCode's
6465 * Alive gets handled by iwl_init_alive_start()).
6467 static void iwl_alive_start(struct iwl_priv
*priv
)
6471 IWL_DEBUG_INFO("Runtime Alive received.\n");
6473 if (priv
->card_alive
.is_valid
!= UCODE_VALID_OK
) {
6474 /* We had an error bringing up the hardware, so take it
6475 * all the way back down so we can try again */
6476 IWL_DEBUG_INFO("Alive failed.\n");
6480 /* Initialize uCode has loaded Runtime uCode ... verify inst image.
6481 * This is a paranoid check, because we would not have gotten the
6482 * "runtime" alive if code weren't properly loaded. */
6483 if (iwl_verify_ucode(priv
)) {
6484 /* Runtime instruction load was bad;
6485 * take it all the way back down so we can try again */
6486 IWL_DEBUG_INFO("Bad runtime uCode load.\n");
6490 iwl_clear_stations_table(priv
);
6492 rc
= iwl4965_alive_notify(priv
);
6494 IWL_WARNING("Could not complete ALIVE transition [ntf]: %d\n",
6499 /* After the ALIVE response, we can process host commands */
6500 set_bit(STATUS_ALIVE
, &priv
->status
);
6502 /* Clear out the uCode error bit if it is set */
6503 clear_bit(STATUS_FW_ERROR
, &priv
->status
);
6505 rc
= iwl_init_channel_map(priv
);
6507 IWL_ERROR("initializing regulatory failed: %d\n", rc
);
6511 iwl_init_geos(priv
);
6513 if (iwl_is_rfkill(priv
))
6516 if (!priv
->mac80211_registered
) {
6517 /* Unlock so any user space entry points can call back into
6518 * the driver without a deadlock... */
6519 mutex_unlock(&priv
->mutex
);
6520 iwl_rate_control_register(priv
->hw
);
6521 rc
= ieee80211_register_hw(priv
->hw
);
6522 priv
->hw
->conf
.beacon_int
= 100;
6523 mutex_lock(&priv
->mutex
);
6526 iwl_rate_control_unregister(priv
->hw
);
6527 IWL_ERROR("Failed to register network "
6528 "device (error %d)\n", rc
);
6532 priv
->mac80211_registered
= 1;
6534 iwl_reset_channel_flag(priv
);
6536 ieee80211_start_queues(priv
->hw
);
6538 priv
->active_rate
= priv
->rates_mask
;
6539 priv
->active_rate_basic
= priv
->rates_mask
& IWL_BASIC_RATES_MASK
;
6541 iwl_send_power_mode(priv
, IWL_POWER_LEVEL(priv
->power_mode
));
6543 if (iwl_is_associated(priv
)) {
6544 struct iwl_rxon_cmd
*active_rxon
=
6545 (struct iwl_rxon_cmd
*)(&priv
->active_rxon
);
6547 memcpy(&priv
->staging_rxon
, &priv
->active_rxon
,
6548 sizeof(priv
->staging_rxon
));
6549 active_rxon
->filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
6551 /* Initialize our rx_config data */
6552 iwl_connection_init_rx_config(priv
);
6553 memcpy(priv
->staging_rxon
.node_addr
, priv
->mac_addr
, ETH_ALEN
);
6556 /* Configure BT coexistence */
6557 iwl_send_bt_config(priv
);
6559 /* Configure the adapter for unassociated operation */
6560 iwl_commit_rxon(priv
);
6562 /* At this point, the NIC is initialized and operational */
6563 priv
->notif_missed_beacons
= 0;
6564 set_bit(STATUS_READY
, &priv
->status
);
6566 iwl4965_rf_kill_ct_config(priv
);
6567 IWL_DEBUG_INFO("ALIVE processing complete.\n");
6569 if (priv
->error_recovering
)
6570 iwl_error_recovery(priv
);
6575 queue_work(priv
->workqueue
, &priv
->restart
);
6578 static void iwl_cancel_deferred_work(struct iwl_priv
*priv
);
6580 static void __iwl_down(struct iwl_priv
*priv
)
6582 unsigned long flags
;
6583 int exit_pending
= test_bit(STATUS_EXIT_PENDING
, &priv
->status
);
6584 struct ieee80211_conf
*conf
= NULL
;
6586 IWL_DEBUG_INFO(DRV_NAME
" is going down\n");
6588 conf
= ieee80211_get_hw_conf(priv
->hw
);
6591 set_bit(STATUS_EXIT_PENDING
, &priv
->status
);
6593 iwl_clear_stations_table(priv
);
6595 /* Unblock any waiting calls */
6596 wake_up_interruptible_all(&priv
->wait_command_queue
);
6598 /* Wipe out the EXIT_PENDING status bit if we are not actually
6599 * exiting the module */
6601 clear_bit(STATUS_EXIT_PENDING
, &priv
->status
);
6603 /* stop and reset the on-board processor */
6604 iwl_write32(priv
, CSR_RESET
, CSR_RESET_REG_FLAG_NEVO_RESET
);
6606 /* tell the device to stop sending interrupts */
6607 iwl_disable_interrupts(priv
);
6609 if (priv
->mac80211_registered
)
6610 ieee80211_stop_queues(priv
->hw
);
6612 /* If we have not previously called iwl_init() then
6613 * clear all bits but the RF Kill and SUSPEND bits and return */
6614 if (!iwl_is_init(priv
)) {
6615 priv
->status
= test_bit(STATUS_RF_KILL_HW
, &priv
->status
) <<
6617 test_bit(STATUS_RF_KILL_SW
, &priv
->status
) <<
6619 test_bit(STATUS_IN_SUSPEND
, &priv
->status
) <<
6624 /* ...otherwise clear out all the status bits but the RF Kill and
6625 * SUSPEND bits and continue taking the NIC down. */
6626 priv
->status
&= test_bit(STATUS_RF_KILL_HW
, &priv
->status
) <<
6628 test_bit(STATUS_RF_KILL_SW
, &priv
->status
) <<
6630 test_bit(STATUS_IN_SUSPEND
, &priv
->status
) <<
6632 test_bit(STATUS_FW_ERROR
, &priv
->status
) <<
6635 spin_lock_irqsave(&priv
->lock
, flags
);
6636 iwl_clear_bit(priv
, CSR_GP_CNTRL
, CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ
);
6637 spin_unlock_irqrestore(&priv
->lock
, flags
);
6639 iwl_hw_txq_ctx_stop(priv
);
6640 iwl_hw_rxq_stop(priv
);
6642 spin_lock_irqsave(&priv
->lock
, flags
);
6643 if (!iwl_grab_nic_access(priv
)) {
6644 iwl_write_prph(priv
, APMG_CLK_DIS_REG
,
6645 APMG_CLK_VAL_DMA_CLK_RQT
);
6646 iwl_release_nic_access(priv
);
6648 spin_unlock_irqrestore(&priv
->lock
, flags
);
6652 iwl_hw_nic_stop_master(priv
);
6653 iwl_set_bit(priv
, CSR_RESET
, CSR_RESET_REG_FLAG_SW_RESET
);
6654 iwl_hw_nic_reset(priv
);
6657 memset(&priv
->card_alive
, 0, sizeof(struct iwl_alive_resp
));
6659 if (priv
->ibss_beacon
)
6660 dev_kfree_skb(priv
->ibss_beacon
);
6661 priv
->ibss_beacon
= NULL
;
6663 /* clear out any free frames */
6664 iwl_clear_free_frames(priv
);
6667 static void iwl_down(struct iwl_priv
*priv
)
6669 mutex_lock(&priv
->mutex
);
6671 mutex_unlock(&priv
->mutex
);
6673 iwl_cancel_deferred_work(priv
);
6676 #define MAX_HW_RESTARTS 5
6678 static int __iwl_up(struct iwl_priv
*priv
)
6680 DECLARE_MAC_BUF(mac
);
6684 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
)) {
6685 IWL_WARNING("Exit pending; will not bring the NIC up\n");
6689 if (test_bit(STATUS_RF_KILL_SW
, &priv
->status
)) {
6690 IWL_WARNING("Radio disabled by SW RF kill (module "
6695 if (!priv
->ucode_data_backup
.v_addr
|| !priv
->ucode_data
.v_addr
) {
6696 IWL_ERROR("ucode not available for device bringup\n");
6700 iwl_write32(priv
, CSR_INT
, 0xFFFFFFFF);
6702 rc
= iwl_hw_nic_init(priv
);
6704 IWL_ERROR("Unable to int nic\n");
6708 /* make sure rfkill handshake bits are cleared */
6709 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
, CSR_UCODE_SW_BIT_RFKILL
);
6710 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
,
6711 CSR_UCODE_DRV_GP1_BIT_CMD_BLOCKED
);
6713 /* clear (again), then enable host interrupts */
6714 iwl_write32(priv
, CSR_INT
, 0xFFFFFFFF);
6715 iwl_enable_interrupts(priv
);
6717 /* really make sure rfkill handshake bits are cleared */
6718 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
, CSR_UCODE_SW_BIT_RFKILL
);
6719 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
, CSR_UCODE_SW_BIT_RFKILL
);
6721 /* Copy original ucode data image from disk into backup cache.
6722 * This will be used to initialize the on-board processor's
6723 * data SRAM for a clean start when the runtime program first loads. */
6724 memcpy(priv
->ucode_data_backup
.v_addr
, priv
->ucode_data
.v_addr
,
6725 priv
->ucode_data
.len
);
6727 /* If platform's RF_KILL switch is set to KILL,
6728 * wait for BIT_INT_RF_KILL interrupt before loading uCode
6729 * and getting things started */
6730 if (!(iwl_read32(priv
, CSR_GP_CNTRL
) &
6731 CSR_GP_CNTRL_REG_FLAG_HW_RF_KILL_SW
))
6734 if (test_bit(STATUS_RF_KILL_HW
, &priv
->status
) || hw_rf_kill
) {
6735 IWL_WARNING("Radio disabled by HW RF Kill switch\n");
6739 for (i
= 0; i
< MAX_HW_RESTARTS
; i
++) {
6741 iwl_clear_stations_table(priv
);
6743 /* load bootstrap state machine,
6744 * load bootstrap program into processor's memory,
6745 * prepare to load the "initialize" uCode */
6746 rc
= iwl_load_bsm(priv
);
6749 IWL_ERROR("Unable to set up bootstrap uCode: %d\n", rc
);
6753 /* start card; "initialize" will load runtime ucode */
6754 iwl_nic_start(priv
);
6756 /* MAC Address location in EEPROM same for 3945/4965 */
6757 get_eeprom_mac(priv
, priv
->mac_addr
);
6758 IWL_DEBUG_INFO("MAC address: %s\n",
6759 print_mac(mac
, priv
->mac_addr
));
6761 SET_IEEE80211_PERM_ADDR(priv
->hw
, priv
->mac_addr
);
6763 IWL_DEBUG_INFO(DRV_NAME
" is coming up\n");
6768 set_bit(STATUS_EXIT_PENDING
, &priv
->status
);
6771 /* tried to restart and config the device for as long as our
6772 * patience could withstand */
6773 IWL_ERROR("Unable to initialize device after %d attempts.\n", i
);
6778 /*****************************************************************************
6780 * Workqueue callbacks
6782 *****************************************************************************/
6784 static void iwl_bg_init_alive_start(struct work_struct
*data
)
6786 struct iwl_priv
*priv
=
6787 container_of(data
, struct iwl_priv
, init_alive_start
.work
);
6789 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
6792 mutex_lock(&priv
->mutex
);
6793 iwl_init_alive_start(priv
);
6794 mutex_unlock(&priv
->mutex
);
6797 static void iwl_bg_alive_start(struct work_struct
*data
)
6799 struct iwl_priv
*priv
=
6800 container_of(data
, struct iwl_priv
, alive_start
.work
);
6802 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
6805 mutex_lock(&priv
->mutex
);
6806 iwl_alive_start(priv
);
6807 mutex_unlock(&priv
->mutex
);
6810 static void iwl_bg_rf_kill(struct work_struct
*work
)
6812 struct iwl_priv
*priv
= container_of(work
, struct iwl_priv
, rf_kill
);
6814 wake_up_interruptible(&priv
->wait_command_queue
);
6816 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
6819 mutex_lock(&priv
->mutex
);
6821 if (!iwl_is_rfkill(priv
)) {
6822 IWL_DEBUG(IWL_DL_INFO
| IWL_DL_RF_KILL
,
6823 "HW and/or SW RF Kill no longer active, restarting "
6825 if (!test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
6826 queue_work(priv
->workqueue
, &priv
->restart
);
6829 if (!test_bit(STATUS_RF_KILL_HW
, &priv
->status
))
6830 IWL_DEBUG_RF_KILL("Can not turn radio back on - "
6831 "disabled by SW switch\n");
6833 IWL_WARNING("Radio Frequency Kill Switch is On:\n"
6834 "Kill switch must be turned off for "
6835 "wireless networking to work.\n");
6837 mutex_unlock(&priv
->mutex
);
6840 #define IWL_SCAN_CHECK_WATCHDOG (7 * HZ)
6842 static void iwl_bg_scan_check(struct work_struct
*data
)
6844 struct iwl_priv
*priv
=
6845 container_of(data
, struct iwl_priv
, scan_check
.work
);
6847 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
6850 mutex_lock(&priv
->mutex
);
6851 if (test_bit(STATUS_SCANNING
, &priv
->status
) ||
6852 test_bit(STATUS_SCAN_ABORTING
, &priv
->status
)) {
6853 IWL_DEBUG(IWL_DL_INFO
| IWL_DL_SCAN
,
6854 "Scan completion watchdog resetting adapter (%dms)\n",
6855 jiffies_to_msecs(IWL_SCAN_CHECK_WATCHDOG
));
6857 if (!test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
6858 iwl_send_scan_abort(priv
);
6860 mutex_unlock(&priv
->mutex
);
6863 static void iwl_bg_request_scan(struct work_struct
*data
)
6865 struct iwl_priv
*priv
=
6866 container_of(data
, struct iwl_priv
, request_scan
);
6867 struct iwl_host_cmd cmd
= {
6868 .id
= REPLY_SCAN_CMD
,
6869 .len
= sizeof(struct iwl_scan_cmd
),
6870 .meta
.flags
= CMD_SIZE_HUGE
,
6873 struct iwl_scan_cmd
*scan
;
6874 struct ieee80211_conf
*conf
= NULL
;
6878 conf
= ieee80211_get_hw_conf(priv
->hw
);
6880 mutex_lock(&priv
->mutex
);
6882 if (!iwl_is_ready(priv
)) {
6883 IWL_WARNING("request scan called when driver not ready.\n");
6887 /* Make sure the scan wasn't cancelled before this queued work
6888 * was given the chance to run... */
6889 if (!test_bit(STATUS_SCANNING
, &priv
->status
))
6892 /* This should never be called or scheduled if there is currently
6893 * a scan active in the hardware. */
6894 if (test_bit(STATUS_SCAN_HW
, &priv
->status
)) {
6895 IWL_DEBUG_INFO("Multiple concurrent scan requests in parallel. "
6896 "Ignoring second request.\n");
6901 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
)) {
6902 IWL_DEBUG_SCAN("Aborting scan due to device shutdown\n");
6906 if (test_bit(STATUS_SCAN_ABORTING
, &priv
->status
)) {
6907 IWL_DEBUG_HC("Scan request while abort pending. Queuing.\n");
6911 if (iwl_is_rfkill(priv
)) {
6912 IWL_DEBUG_HC("Aborting scan due to RF Kill activation\n");
6916 if (!test_bit(STATUS_READY
, &priv
->status
)) {
6917 IWL_DEBUG_HC("Scan request while uninitialized. Queuing.\n");
6921 if (!priv
->scan_bands
) {
6922 IWL_DEBUG_HC("Aborting scan due to no requested bands\n");
6927 priv
->scan
= kmalloc(sizeof(struct iwl_scan_cmd
) +
6928 IWL_MAX_SCAN_SIZE
, GFP_KERNEL
);
6935 memset(scan
, 0, sizeof(struct iwl_scan_cmd
) + IWL_MAX_SCAN_SIZE
);
6937 scan
->quiet_plcp_th
= IWL_PLCP_QUIET_THRESH
;
6938 scan
->quiet_time
= IWL_ACTIVE_QUIET_TIME
;
6940 if (iwl_is_associated(priv
)) {
6943 u32 suspend_time
= 100;
6944 u32 scan_suspend_time
= 100;
6945 unsigned long flags
;
6947 IWL_DEBUG_INFO("Scanning while associated...\n");
6949 spin_lock_irqsave(&priv
->lock
, flags
);
6950 interval
= priv
->beacon_int
;
6951 spin_unlock_irqrestore(&priv
->lock
, flags
);
6953 scan
->suspend_time
= 0;
6954 scan
->max_out_time
= cpu_to_le32(200 * 1024);
6956 interval
= suspend_time
;
6958 extra
= (suspend_time
/ interval
) << 22;
6959 scan_suspend_time
= (extra
|
6960 ((suspend_time
% interval
) * 1024));
6961 scan
->suspend_time
= cpu_to_le32(scan_suspend_time
);
6962 IWL_DEBUG_SCAN("suspend_time 0x%X beacon interval %d\n",
6963 scan_suspend_time
, interval
);
6966 /* We should add the ability for user to lock to PASSIVE ONLY */
6967 if (priv
->one_direct_scan
) {
6969 ("Kicking off one direct scan for '%s'\n",
6970 iwl_escape_essid(priv
->direct_ssid
,
6971 priv
->direct_ssid_len
));
6972 scan
->direct_scan
[0].id
= WLAN_EID_SSID
;
6973 scan
->direct_scan
[0].len
= priv
->direct_ssid_len
;
6974 memcpy(scan
->direct_scan
[0].ssid
,
6975 priv
->direct_ssid
, priv
->direct_ssid_len
);
6977 } else if (!iwl_is_associated(priv
) && priv
->essid_len
) {
6978 scan
->direct_scan
[0].id
= WLAN_EID_SSID
;
6979 scan
->direct_scan
[0].len
= priv
->essid_len
;
6980 memcpy(scan
->direct_scan
[0].ssid
, priv
->essid
, priv
->essid_len
);
6985 /* We don't build a direct scan probe request; the uCode will do
6986 * that based on the direct_mask added to each channel entry */
6987 scan
->tx_cmd
.len
= cpu_to_le16(
6988 iwl_fill_probe_req(priv
, (struct ieee80211_mgmt
*)scan
->data
,
6989 IWL_MAX_SCAN_SIZE
- sizeof(scan
), 0));
6990 scan
->tx_cmd
.tx_flags
= TX_CMD_FLG_SEQ_CTL_MSK
;
6991 scan
->tx_cmd
.sta_id
= priv
->hw_setting
.bcast_sta_id
;
6992 scan
->tx_cmd
.stop_time
.life_time
= TX_CMD_LIFE_TIME_INFINITE
;
6994 /* flags + rate selection */
6996 scan
->tx_cmd
.tx_flags
|= cpu_to_le32(0x200);
6998 switch (priv
->scan_bands
) {
7000 scan
->flags
= RXON_FLG_BAND_24G_MSK
| RXON_FLG_AUTO_DETECT_MSK
;
7001 scan
->tx_cmd
.rate_n_flags
=
7002 iwl_hw_set_rate_n_flags(IWL_RATE_1M_PLCP
,
7003 RATE_MCS_ANT_B_MSK
|RATE_MCS_CCK_MSK
);
7005 scan
->good_CRC_th
= 0;
7006 phymode
= MODE_IEEE80211G
;
7010 scan
->tx_cmd
.rate_n_flags
=
7011 iwl_hw_set_rate_n_flags(IWL_RATE_6M_PLCP
,
7012 RATE_MCS_ANT_B_MSK
);
7013 scan
->good_CRC_th
= IWL_GOOD_CRC_TH
;
7014 phymode
= MODE_IEEE80211A
;
7018 IWL_WARNING("Invalid scan band count\n");
7022 /* select Rx chains */
7024 /* Force use of chains B and C (0x6) for scan Rx.
7025 * Avoid A (0x1) because of its off-channel reception on A-band.
7026 * MIMO is not used here, but value is required to make uCode happy. */
7027 scan
->rx_chain
= RXON_RX_CHAIN_DRIVER_FORCE_MSK
|
7028 cpu_to_le16((0x7 << RXON_RX_CHAIN_VALID_POS
) |
7029 (0x6 << RXON_RX_CHAIN_FORCE_SEL_POS
) |
7030 (0x7 << RXON_RX_CHAIN_FORCE_MIMO_SEL_POS
));
7032 if (priv
->iw_mode
== IEEE80211_IF_TYPE_MNTR
)
7033 scan
->filter_flags
= RXON_FILTER_PROMISC_MSK
;
7037 ("Initiating direct scan for %s.\n",
7038 iwl_escape_essid(priv
->essid
, priv
->essid_len
));
7040 IWL_DEBUG_SCAN("Initiating indirect scan.\n");
7042 scan
->channel_count
=
7043 iwl_get_channels_for_scan(
7044 priv
, phymode
, 1, /* active */
7046 (void *)&scan
->data
[le16_to_cpu(scan
->tx_cmd
.len
)]);
7048 cmd
.len
+= le16_to_cpu(scan
->tx_cmd
.len
) +
7049 scan
->channel_count
* sizeof(struct iwl_scan_channel
);
7051 scan
->len
= cpu_to_le16(cmd
.len
);
7053 set_bit(STATUS_SCAN_HW
, &priv
->status
);
7054 rc
= iwl_send_cmd_sync(priv
, &cmd
);
7058 queue_delayed_work(priv
->workqueue
, &priv
->scan_check
,
7059 IWL_SCAN_CHECK_WATCHDOG
);
7061 mutex_unlock(&priv
->mutex
);
7065 /* inform mac80211 scan aborted */
7066 queue_work(priv
->workqueue
, &priv
->scan_completed
);
7067 mutex_unlock(&priv
->mutex
);
7070 static void iwl_bg_up(struct work_struct
*data
)
7072 struct iwl_priv
*priv
= container_of(data
, struct iwl_priv
, up
);
7074 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
7077 mutex_lock(&priv
->mutex
);
7079 mutex_unlock(&priv
->mutex
);
7082 static void iwl_bg_restart(struct work_struct
*data
)
7084 struct iwl_priv
*priv
= container_of(data
, struct iwl_priv
, restart
);
7086 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
7090 queue_work(priv
->workqueue
, &priv
->up
);
7093 static void iwl_bg_rx_replenish(struct work_struct
*data
)
7095 struct iwl_priv
*priv
=
7096 container_of(data
, struct iwl_priv
, rx_replenish
);
7098 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
7101 mutex_lock(&priv
->mutex
);
7102 iwl_rx_replenish(priv
);
7103 mutex_unlock(&priv
->mutex
);
7106 static void iwl_bg_post_associate(struct work_struct
*data
)
7108 struct iwl_priv
*priv
= container_of(data
, struct iwl_priv
,
7109 post_associate
.work
);
7112 struct ieee80211_conf
*conf
= NULL
;
7113 DECLARE_MAC_BUF(mac
);
7115 if (priv
->iw_mode
== IEEE80211_IF_TYPE_AP
) {
7116 IWL_ERROR("%s Should not be called in AP mode\n", __FUNCTION__
);
7120 IWL_DEBUG_ASSOC("Associated as %d to: %s\n",
7122 print_mac(mac
, priv
->active_rxon
.bssid_addr
));
7125 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
7128 mutex_lock(&priv
->mutex
);
7130 if (!priv
->interface_id
|| !priv
->is_open
) {
7131 mutex_unlock(&priv
->mutex
);
7134 iwl_scan_cancel_timeout(priv
, 200);
7136 conf
= ieee80211_get_hw_conf(priv
->hw
);
7138 priv
->staging_rxon
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
7139 iwl_commit_rxon(priv
);
7141 memset(&priv
->rxon_timing
, 0, sizeof(struct iwl_rxon_time_cmd
));
7142 iwl_setup_rxon_timing(priv
);
7143 rc
= iwl_send_cmd_pdu(priv
, REPLY_RXON_TIMING
,
7144 sizeof(priv
->rxon_timing
), &priv
->rxon_timing
);
7146 IWL_WARNING("REPLY_RXON_TIMING failed - "
7147 "Attempting to continue.\n");
7149 priv
->staging_rxon
.filter_flags
|= RXON_FILTER_ASSOC_MSK
;
7151 #ifdef CONFIG_IWL4965_HT
7152 if (priv
->is_ht_enabled
&& priv
->current_assoc_ht
.is_ht
)
7153 iwl4965_set_rxon_ht(priv
, &priv
->current_assoc_ht
);
7155 priv
->active_rate_ht
[0] = 0;
7156 priv
->active_rate_ht
[1] = 0;
7157 priv
->current_channel_width
= IWL_CHANNEL_WIDTH_20MHZ
;
7159 #endif /* CONFIG_IWL4965_HT*/
7160 iwl4965_set_rxon_chain(priv
);
7161 priv
->staging_rxon
.assoc_id
= cpu_to_le16(priv
->assoc_id
);
7163 IWL_DEBUG_ASSOC("assoc id %d beacon interval %d\n",
7164 priv
->assoc_id
, priv
->beacon_int
);
7166 if (priv
->assoc_capability
& WLAN_CAPABILITY_SHORT_PREAMBLE
)
7167 priv
->staging_rxon
.flags
|= RXON_FLG_SHORT_PREAMBLE_MSK
;
7169 priv
->staging_rxon
.flags
&= ~RXON_FLG_SHORT_PREAMBLE_MSK
;
7171 if (priv
->staging_rxon
.flags
& RXON_FLG_BAND_24G_MSK
) {
7172 if (priv
->assoc_capability
& WLAN_CAPABILITY_SHORT_SLOT_TIME
)
7173 priv
->staging_rxon
.flags
|= RXON_FLG_SHORT_SLOT_MSK
;
7175 priv
->staging_rxon
.flags
&= ~RXON_FLG_SHORT_SLOT_MSK
;
7177 if (priv
->iw_mode
== IEEE80211_IF_TYPE_IBSS
)
7178 priv
->staging_rxon
.flags
&= ~RXON_FLG_SHORT_SLOT_MSK
;
7182 iwl_commit_rxon(priv
);
7184 switch (priv
->iw_mode
) {
7185 case IEEE80211_IF_TYPE_STA
:
7186 iwl_rate_scale_init(priv
->hw
, IWL_AP_ID
);
7189 case IEEE80211_IF_TYPE_IBSS
:
7191 /* clear out the station table */
7192 iwl_clear_stations_table(priv
);
7194 iwl_rxon_add_station(priv
, BROADCAST_ADDR
, 0);
7195 iwl_rxon_add_station(priv
, priv
->bssid
, 0);
7196 iwl_rate_scale_init(priv
->hw
, IWL_STA_ID
);
7197 iwl_send_beacon_cmd(priv
);
7202 IWL_ERROR("%s Should not be called in %d mode\n",
7203 __FUNCTION__
, priv
->iw_mode
);
7207 iwl_sequence_reset(priv
);
7209 #ifdef CONFIG_IWL4965_SENSITIVITY
7210 /* Enable Rx differential gain and sensitivity calibrations */
7211 iwl4965_chain_noise_reset(priv
);
7212 priv
->start_calib
= 1;
7213 #endif /* CONFIG_IWL4965_SENSITIVITY */
7215 if (priv
->iw_mode
== IEEE80211_IF_TYPE_IBSS
)
7216 priv
->assoc_station_added
= 1;
7218 #ifdef CONFIG_IWL4965_QOS
7219 iwl_activate_qos(priv
, 0);
7220 #endif /* CONFIG_IWL4965_QOS */
7221 mutex_unlock(&priv
->mutex
);
7224 static void iwl_bg_abort_scan(struct work_struct
*work
)
7226 struct iwl_priv
*priv
= container_of(work
, struct iwl_priv
, abort_scan
);
7228 if (!iwl_is_ready(priv
))
7231 mutex_lock(&priv
->mutex
);
7233 set_bit(STATUS_SCAN_ABORTING
, &priv
->status
);
7234 iwl_send_scan_abort(priv
);
7236 mutex_unlock(&priv
->mutex
);
7239 static void iwl_bg_scan_completed(struct work_struct
*work
)
7241 struct iwl_priv
*priv
=
7242 container_of(work
, struct iwl_priv
, scan_completed
);
7244 IWL_DEBUG(IWL_DL_INFO
| IWL_DL_SCAN
, "SCAN complete scan\n");
7246 if (test_bit(STATUS_EXIT_PENDING
, &priv
->status
))
7249 ieee80211_scan_completed(priv
->hw
);
7251 /* Since setting the TXPOWER may have been deferred while
7252 * performing the scan, fire one off */
7253 mutex_lock(&priv
->mutex
);
7254 iwl_hw_reg_send_txpower(priv
);
7255 mutex_unlock(&priv
->mutex
);
7258 /*****************************************************************************
7260 * mac80211 entry point functions
7262 *****************************************************************************/
7264 static int iwl_mac_start(struct ieee80211_hw
*hw
)
7266 struct iwl_priv
*priv
= hw
->priv
;
7268 IWL_DEBUG_MAC80211("enter\n");
7270 /* we should be verifying the device is ready to be opened */
7271 mutex_lock(&priv
->mutex
);
7275 if (!iwl_is_rfkill(priv
))
7276 ieee80211_start_queues(priv
->hw
);
7278 mutex_unlock(&priv
->mutex
);
7279 IWL_DEBUG_MAC80211("leave\n");
7283 static void iwl_mac_stop(struct ieee80211_hw
*hw
)
7285 struct iwl_priv
*priv
= hw
->priv
;
7287 IWL_DEBUG_MAC80211("enter\n");
7290 mutex_lock(&priv
->mutex
);
7291 /* stop mac, cancel any scan request and clear
7292 * RXON_FILTER_ASSOC_MSK BIT
7295 iwl_scan_cancel_timeout(priv
, 100);
7296 cancel_delayed_work(&priv
->post_associate
);
7297 priv
->staging_rxon
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
7298 iwl_commit_rxon(priv
);
7299 mutex_unlock(&priv
->mutex
);
7301 IWL_DEBUG_MAC80211("leave\n");
7304 static int iwl_mac_tx(struct ieee80211_hw
*hw
, struct sk_buff
*skb
,
7305 struct ieee80211_tx_control
*ctl
)
7307 struct iwl_priv
*priv
= hw
->priv
;
7309 IWL_DEBUG_MAC80211("enter\n");
7311 if (priv
->iw_mode
== IEEE80211_IF_TYPE_MNTR
) {
7312 IWL_DEBUG_MAC80211("leave - monitor\n");
7316 IWL_DEBUG_TX("dev->xmit(%d bytes) at rate 0x%02x\n", skb
->len
,
7319 if (iwl_tx_skb(priv
, skb
, ctl
))
7320 dev_kfree_skb_any(skb
);
7322 IWL_DEBUG_MAC80211("leave\n");
7326 static int iwl_mac_add_interface(struct ieee80211_hw
*hw
,
7327 struct ieee80211_if_init_conf
*conf
)
7329 struct iwl_priv
*priv
= hw
->priv
;
7330 unsigned long flags
;
7331 DECLARE_MAC_BUF(mac
);
7333 IWL_DEBUG_MAC80211("enter: id %d, type %d\n", conf
->if_id
, conf
->type
);
7335 if (priv
->interface_id
) {
7336 IWL_DEBUG_MAC80211("leave - interface_id != 0\n");
7340 spin_lock_irqsave(&priv
->lock
, flags
);
7341 priv
->interface_id
= conf
->if_id
;
7343 spin_unlock_irqrestore(&priv
->lock
, flags
);
7345 mutex_lock(&priv
->mutex
);
7347 if (conf
->mac_addr
) {
7348 IWL_DEBUG_MAC80211("Set %s\n", print_mac(mac
, conf
->mac_addr
));
7349 memcpy(priv
->mac_addr
, conf
->mac_addr
, ETH_ALEN
);
7351 iwl_set_mode(priv
, conf
->type
);
7353 IWL_DEBUG_MAC80211("leave\n");
7354 mutex_unlock(&priv
->mutex
);
7360 * iwl_mac_config - mac80211 config callback
7362 * We ignore conf->flags & IEEE80211_CONF_SHORT_SLOT_TIME since it seems to
7363 * be set inappropriately and the driver currently sets the hardware up to
7364 * use it whenever needed.
7366 static int iwl_mac_config(struct ieee80211_hw
*hw
, struct ieee80211_conf
*conf
)
7368 struct iwl_priv
*priv
= hw
->priv
;
7369 const struct iwl_channel_info
*ch_info
;
7370 unsigned long flags
;
7372 mutex_lock(&priv
->mutex
);
7373 IWL_DEBUG_MAC80211("enter to channel %d\n", conf
->channel
);
7375 if (!iwl_is_ready(priv
)) {
7376 IWL_DEBUG_MAC80211("leave - not ready\n");
7377 mutex_unlock(&priv
->mutex
);
7381 /* TODO: Figure out how to get ieee80211_local->sta_scanning w/ only
7382 * what is exposed through include/ declarations */
7383 if (unlikely(!iwl_param_disable_hw_scan
&&
7384 test_bit(STATUS_SCANNING
, &priv
->status
))) {
7385 IWL_DEBUG_MAC80211("leave - scanning\n");
7386 mutex_unlock(&priv
->mutex
);
7390 spin_lock_irqsave(&priv
->lock
, flags
);
7392 ch_info
= iwl_get_channel_info(priv
, conf
->phymode
, conf
->channel
);
7393 if (!is_channel_valid(ch_info
)) {
7394 IWL_DEBUG_SCAN("Channel %d [%d] is INVALID for this SKU.\n",
7395 conf
->channel
, conf
->phymode
);
7396 IWL_DEBUG_MAC80211("leave - invalid channel\n");
7397 spin_unlock_irqrestore(&priv
->lock
, flags
);
7398 mutex_unlock(&priv
->mutex
);
7402 #ifdef CONFIG_IWL4965_HT
7403 /* if we are switching fron ht to 2.4 clear flags
7404 * from any ht related info since 2.4 does not
7406 if ((le16_to_cpu(priv
->staging_rxon
.channel
) != conf
->channel
)
7407 #ifdef IEEE80211_CONF_CHANNEL_SWITCH
7408 && !(conf
->flags
& IEEE80211_CONF_CHANNEL_SWITCH
)
7411 priv
->staging_rxon
.flags
= 0;
7412 #endif /* CONFIG_IWL4965_HT */
7414 iwl_set_rxon_channel(priv
, conf
->phymode
, conf
->channel
);
7416 iwl_set_flags_for_phymode(priv
, conf
->phymode
);
7418 /* The list of supported rates and rate mask can be different
7419 * for each phymode; since the phymode may have changed, reset
7420 * the rate mask to what mac80211 lists */
7423 spin_unlock_irqrestore(&priv
->lock
, flags
);
7425 #ifdef IEEE80211_CONF_CHANNEL_SWITCH
7426 if (conf
->flags
& IEEE80211_CONF_CHANNEL_SWITCH
) {
7427 iwl_hw_channel_switch(priv
, conf
->channel
);
7428 mutex_unlock(&priv
->mutex
);
7433 iwl_radio_kill_sw(priv
, !conf
->radio_enabled
);
7435 if (!conf
->radio_enabled
) {
7436 IWL_DEBUG_MAC80211("leave - radio disabled\n");
7437 mutex_unlock(&priv
->mutex
);
7441 if (iwl_is_rfkill(priv
)) {
7442 IWL_DEBUG_MAC80211("leave - RF kill\n");
7443 mutex_unlock(&priv
->mutex
);
7449 if (memcmp(&priv
->active_rxon
,
7450 &priv
->staging_rxon
, sizeof(priv
->staging_rxon
)))
7451 iwl_commit_rxon(priv
);
7453 IWL_DEBUG_INFO("No re-sending same RXON configuration.\n");
7455 IWL_DEBUG_MAC80211("leave\n");
7457 mutex_unlock(&priv
->mutex
);
7462 static void iwl_config_ap(struct iwl_priv
*priv
)
7466 if (priv
->status
& STATUS_EXIT_PENDING
)
7469 /* The following should be done only at AP bring up */
7470 if ((priv
->active_rxon
.filter_flags
& RXON_FILTER_ASSOC_MSK
) == 0) {
7472 /* RXON - unassoc (to set timing command) */
7473 priv
->staging_rxon
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
7474 iwl_commit_rxon(priv
);
7477 memset(&priv
->rxon_timing
, 0, sizeof(struct iwl_rxon_time_cmd
));
7478 iwl_setup_rxon_timing(priv
);
7479 rc
= iwl_send_cmd_pdu(priv
, REPLY_RXON_TIMING
,
7480 sizeof(priv
->rxon_timing
), &priv
->rxon_timing
);
7482 IWL_WARNING("REPLY_RXON_TIMING failed - "
7483 "Attempting to continue.\n");
7485 iwl4965_set_rxon_chain(priv
);
7487 /* FIXME: what should be the assoc_id for AP? */
7488 priv
->staging_rxon
.assoc_id
= cpu_to_le16(priv
->assoc_id
);
7489 if (priv
->assoc_capability
& WLAN_CAPABILITY_SHORT_PREAMBLE
)
7490 priv
->staging_rxon
.flags
|=
7491 RXON_FLG_SHORT_PREAMBLE_MSK
;
7493 priv
->staging_rxon
.flags
&=
7494 ~RXON_FLG_SHORT_PREAMBLE_MSK
;
7496 if (priv
->staging_rxon
.flags
& RXON_FLG_BAND_24G_MSK
) {
7497 if (priv
->assoc_capability
&
7498 WLAN_CAPABILITY_SHORT_SLOT_TIME
)
7499 priv
->staging_rxon
.flags
|=
7500 RXON_FLG_SHORT_SLOT_MSK
;
7502 priv
->staging_rxon
.flags
&=
7503 ~RXON_FLG_SHORT_SLOT_MSK
;
7505 if (priv
->iw_mode
== IEEE80211_IF_TYPE_IBSS
)
7506 priv
->staging_rxon
.flags
&=
7507 ~RXON_FLG_SHORT_SLOT_MSK
;
7509 /* restore RXON assoc */
7510 priv
->staging_rxon
.filter_flags
|= RXON_FILTER_ASSOC_MSK
;
7511 iwl_commit_rxon(priv
);
7512 #ifdef CONFIG_IWL4965_QOS
7513 iwl_activate_qos(priv
, 1);
7515 iwl_rxon_add_station(priv
, BROADCAST_ADDR
, 0);
7517 iwl_send_beacon_cmd(priv
);
7519 /* FIXME - we need to add code here to detect a totally new
7520 * configuration, reset the AP, unassoc, rxon timing, assoc,
7521 * clear sta table, add BCAST sta... */
7524 static int iwl_mac_config_interface(struct ieee80211_hw
*hw
, int if_id
,
7525 struct ieee80211_if_conf
*conf
)
7527 struct iwl_priv
*priv
= hw
->priv
;
7528 DECLARE_MAC_BUF(mac
);
7529 unsigned long flags
;
7535 if ((priv
->iw_mode
== IEEE80211_IF_TYPE_AP
) &&
7536 (!conf
->beacon
|| !conf
->ssid_len
)) {
7538 ("Leaving in AP mode because HostAPD is not ready.\n");
7542 mutex_lock(&priv
->mutex
);
7544 IWL_DEBUG_MAC80211("enter: interface id %d\n", if_id
);
7546 IWL_DEBUG_MAC80211("bssid: %s\n",
7547 print_mac(mac
, conf
->bssid
));
7550 * very dubious code was here; the probe filtering flag is never set:
7552 if (unlikely(test_bit(STATUS_SCANNING, &priv->status)) &&
7553 !(priv->hw->flags & IEEE80211_HW_NO_PROBE_FILTERING)) {
7555 if (unlikely(test_bit(STATUS_SCANNING
, &priv
->status
))) {
7556 IWL_DEBUG_MAC80211("leave - scanning\n");
7557 mutex_unlock(&priv
->mutex
);
7561 if (priv
->interface_id
!= if_id
) {
7562 IWL_DEBUG_MAC80211("leave - interface_id != if_id\n");
7563 mutex_unlock(&priv
->mutex
);
7567 if (priv
->iw_mode
== IEEE80211_IF_TYPE_AP
) {
7569 conf
->bssid
= priv
->mac_addr
;
7570 memcpy(priv
->bssid
, priv
->mac_addr
, ETH_ALEN
);
7571 IWL_DEBUG_MAC80211("bssid was set to: %s\n",
7572 print_mac(mac
, conf
->bssid
));
7574 if (priv
->ibss_beacon
)
7575 dev_kfree_skb(priv
->ibss_beacon
);
7577 priv
->ibss_beacon
= conf
->beacon
;
7580 if (conf
->bssid
&& !is_zero_ether_addr(conf
->bssid
) &&
7581 !is_multicast_ether_addr(conf
->bssid
)) {
7582 /* If there is currently a HW scan going on in the background
7583 * then we need to cancel it else the RXON below will fail. */
7584 if (iwl_scan_cancel_timeout(priv
, 100)) {
7585 IWL_WARNING("Aborted scan still in progress "
7587 IWL_DEBUG_MAC80211("leaving - scan abort failed.\n");
7588 mutex_unlock(&priv
->mutex
);
7591 memcpy(priv
->staging_rxon
.bssid_addr
, conf
->bssid
, ETH_ALEN
);
7593 /* TODO: Audit driver for usage of these members and see
7594 * if mac80211 deprecates them (priv->bssid looks like it
7595 * shouldn't be there, but I haven't scanned the IBSS code
7596 * to verify) - jpk */
7597 memcpy(priv
->bssid
, conf
->bssid
, ETH_ALEN
);
7599 if (priv
->iw_mode
== IEEE80211_IF_TYPE_AP
)
7600 iwl_config_ap(priv
);
7602 rc
= iwl_commit_rxon(priv
);
7603 if ((priv
->iw_mode
== IEEE80211_IF_TYPE_STA
) && rc
)
7604 iwl_rxon_add_station(
7605 priv
, priv
->active_rxon
.bssid_addr
, 1);
7609 iwl_scan_cancel_timeout(priv
, 100);
7610 priv
->staging_rxon
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
7611 iwl_commit_rxon(priv
);
7614 spin_lock_irqsave(&priv
->lock
, flags
);
7615 if (!conf
->ssid_len
)
7616 memset(priv
->essid
, 0, IW_ESSID_MAX_SIZE
);
7618 memcpy(priv
->essid
, conf
->ssid
, conf
->ssid_len
);
7620 priv
->essid_len
= conf
->ssid_len
;
7621 spin_unlock_irqrestore(&priv
->lock
, flags
);
7623 IWL_DEBUG_MAC80211("leave\n");
7624 mutex_unlock(&priv
->mutex
);
7629 static void iwl_configure_filter(struct ieee80211_hw
*hw
,
7630 unsigned int changed_flags
,
7631 unsigned int *total_flags
,
7632 int mc_count
, struct dev_addr_list
*mc_list
)
7636 * see also iwl_connection_init_rx_config
7641 static void iwl_mac_remove_interface(struct ieee80211_hw
*hw
,
7642 struct ieee80211_if_init_conf
*conf
)
7644 struct iwl_priv
*priv
= hw
->priv
;
7646 IWL_DEBUG_MAC80211("enter\n");
7648 mutex_lock(&priv
->mutex
);
7650 iwl_scan_cancel_timeout(priv
, 100);
7651 cancel_delayed_work(&priv
->post_associate
);
7652 priv
->staging_rxon
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
7653 iwl_commit_rxon(priv
);
7655 if (priv
->interface_id
== conf
->if_id
) {
7656 priv
->interface_id
= 0;
7657 memset(priv
->bssid
, 0, ETH_ALEN
);
7658 memset(priv
->essid
, 0, IW_ESSID_MAX_SIZE
);
7659 priv
->essid_len
= 0;
7661 mutex_unlock(&priv
->mutex
);
7663 IWL_DEBUG_MAC80211("leave\n");
7666 static void iwl_mac_erp_ie_changed(struct ieee80211_hw
*hw
,
7667 u8 changes
, int cts_protection
, int preamble
)
7670 struct iwl_priv
*priv
= hw
->priv
;
7672 if (changes
& IEEE80211_ERP_CHANGE_PREAMBLE
) {
7673 if (preamble
== WLAN_ERP_PREAMBLE_SHORT
)
7674 priv
->staging_rxon
.flags
|= RXON_FLG_SHORT_PREAMBLE_MSK
;
7676 priv
->staging_rxon
.flags
&= ~RXON_FLG_SHORT_PREAMBLE_MSK
;
7679 if (changes
& IEEE80211_ERP_CHANGE_PROTECTION
) {
7681 priv
->staging_rxon
.flags
|= RXON_FLG_TGG_PROTECT_MSK
;
7683 priv
->staging_rxon
.flags
&= ~RXON_FLG_TGG_PROTECT_MSK
;
7686 if (iwl_is_associated(priv
))
7687 iwl_send_rxon_assoc(priv
);
7690 #define IWL_DELAY_NEXT_SCAN (HZ*2)
7691 static int iwl_mac_hw_scan(struct ieee80211_hw
*hw
, u8
*ssid
, size_t len
)
7694 unsigned long flags
;
7695 struct iwl_priv
*priv
= hw
->priv
;
7697 IWL_DEBUG_MAC80211("enter\n");
7699 mutex_lock(&priv
->mutex
);
7700 spin_lock_irqsave(&priv
->lock
, flags
);
7702 if (!iwl_is_ready_rf(priv
)) {
7704 IWL_DEBUG_MAC80211("leave - not ready or exit pending\n");
7708 if (priv
->iw_mode
== IEEE80211_IF_TYPE_AP
) { /* APs don't scan */
7710 IWL_ERROR("ERROR: APs don't scan\n");
7714 /* if we just finished scan ask for delay */
7715 if (priv
->last_scan_jiffies
&&
7716 time_after(priv
->last_scan_jiffies
+ IWL_DELAY_NEXT_SCAN
,
7722 IWL_DEBUG_SCAN("direct scan for "
7724 iwl_escape_essid(ssid
, len
), (int)len
);
7726 priv
->one_direct_scan
= 1;
7727 priv
->direct_ssid_len
= (u8
)
7728 min((u8
) len
, (u8
) IW_ESSID_MAX_SIZE
);
7729 memcpy(priv
->direct_ssid
, ssid
, priv
->direct_ssid_len
);
7731 priv
->one_direct_scan
= 0;
7733 rc
= iwl_scan_initiate(priv
);
7735 IWL_DEBUG_MAC80211("leave\n");
7738 spin_unlock_irqrestore(&priv
->lock
, flags
);
7739 mutex_unlock(&priv
->mutex
);
7744 static int iwl_mac_set_key(struct ieee80211_hw
*hw
, enum set_key_cmd cmd
,
7745 const u8
*local_addr
, const u8
*addr
,
7746 struct ieee80211_key_conf
*key
)
7748 struct iwl_priv
*priv
= hw
->priv
;
7749 DECLARE_MAC_BUF(mac
);
7753 IWL_DEBUG_MAC80211("enter\n");
7755 if (!iwl_param_hwcrypto
) {
7756 IWL_DEBUG_MAC80211("leave - hwcrypto disabled\n");
7760 if (is_zero_ether_addr(addr
))
7761 /* only support pairwise keys */
7764 sta_id
= iwl_hw_find_station(priv
, addr
);
7765 if (sta_id
== IWL_INVALID_STATION
) {
7766 IWL_DEBUG_MAC80211("leave - %s not in station map.\n",
7767 print_mac(mac
, addr
));
7771 mutex_lock(&priv
->mutex
);
7773 iwl_scan_cancel_timeout(priv
, 100);
7777 rc
= iwl_update_sta_key_info(priv
, key
, sta_id
);
7779 iwl_set_rxon_hwcrypto(priv
, 1);
7780 iwl_commit_rxon(priv
);
7781 key
->hw_key_idx
= sta_id
;
7782 IWL_DEBUG_MAC80211("set_key success, using hwcrypto\n");
7783 key
->flags
|= IEEE80211_KEY_FLAG_GENERATE_IV
;
7787 rc
= iwl_clear_sta_key_info(priv
, sta_id
);
7789 iwl_set_rxon_hwcrypto(priv
, 0);
7790 iwl_commit_rxon(priv
);
7791 IWL_DEBUG_MAC80211("disable hwcrypto key\n");
7798 IWL_DEBUG_MAC80211("leave\n");
7799 mutex_unlock(&priv
->mutex
);
7804 static int iwl_mac_conf_tx(struct ieee80211_hw
*hw
, int queue
,
7805 const struct ieee80211_tx_queue_params
*params
)
7807 struct iwl_priv
*priv
= hw
->priv
;
7808 #ifdef CONFIG_IWL4965_QOS
7809 unsigned long flags
;
7811 #endif /* CONFIG_IWL_QOS */
7813 IWL_DEBUG_MAC80211("enter\n");
7815 if (!iwl_is_ready_rf(priv
)) {
7816 IWL_DEBUG_MAC80211("leave - RF not ready\n");
7820 if (queue
>= AC_NUM
) {
7821 IWL_DEBUG_MAC80211("leave - queue >= AC_NUM %d\n", queue
);
7825 #ifdef CONFIG_IWL4965_QOS
7826 if (!priv
->qos_data
.qos_enable
) {
7827 priv
->qos_data
.qos_active
= 0;
7828 IWL_DEBUG_MAC80211("leave - qos not enabled\n");
7831 q
= AC_NUM
- 1 - queue
;
7833 spin_lock_irqsave(&priv
->lock
, flags
);
7835 priv
->qos_data
.def_qos_parm
.ac
[q
].cw_min
= cpu_to_le16(params
->cw_min
);
7836 priv
->qos_data
.def_qos_parm
.ac
[q
].cw_max
= cpu_to_le16(params
->cw_max
);
7837 priv
->qos_data
.def_qos_parm
.ac
[q
].aifsn
= params
->aifs
;
7838 priv
->qos_data
.def_qos_parm
.ac
[q
].edca_txop
=
7839 cpu_to_le16((params
->burst_time
* 100));
7841 priv
->qos_data
.def_qos_parm
.ac
[q
].reserved1
= 0;
7842 priv
->qos_data
.qos_active
= 1;
7844 spin_unlock_irqrestore(&priv
->lock
, flags
);
7846 mutex_lock(&priv
->mutex
);
7847 if (priv
->iw_mode
== IEEE80211_IF_TYPE_AP
)
7848 iwl_activate_qos(priv
, 1);
7849 else if (priv
->assoc_id
&& iwl_is_associated(priv
))
7850 iwl_activate_qos(priv
, 0);
7852 mutex_unlock(&priv
->mutex
);
7854 #endif /*CONFIG_IWL4965_QOS */
7856 IWL_DEBUG_MAC80211("leave\n");
7860 static int iwl_mac_get_tx_stats(struct ieee80211_hw
*hw
,
7861 struct ieee80211_tx_queue_stats
*stats
)
7863 struct iwl_priv
*priv
= hw
->priv
;
7865 struct iwl_tx_queue
*txq
;
7866 struct iwl_queue
*q
;
7867 unsigned long flags
;
7869 IWL_DEBUG_MAC80211("enter\n");
7871 if (!iwl_is_ready_rf(priv
)) {
7872 IWL_DEBUG_MAC80211("leave - RF not ready\n");
7876 spin_lock_irqsave(&priv
->lock
, flags
);
7878 for (i
= 0; i
< AC_NUM
; i
++) {
7879 txq
= &priv
->txq
[i
];
7881 avail
= iwl_queue_space(q
);
7883 stats
->data
[i
].len
= q
->n_window
- avail
;
7884 stats
->data
[i
].limit
= q
->n_window
- q
->high_mark
;
7885 stats
->data
[i
].count
= q
->n_window
;
7888 spin_unlock_irqrestore(&priv
->lock
, flags
);
7890 IWL_DEBUG_MAC80211("leave\n");
7895 static int iwl_mac_get_stats(struct ieee80211_hw
*hw
,
7896 struct ieee80211_low_level_stats
*stats
)
7898 IWL_DEBUG_MAC80211("enter\n");
7899 IWL_DEBUG_MAC80211("leave\n");
7904 static u64
iwl_mac_get_tsf(struct ieee80211_hw
*hw
)
7906 IWL_DEBUG_MAC80211("enter\n");
7907 IWL_DEBUG_MAC80211("leave\n");
7912 static void iwl_mac_reset_tsf(struct ieee80211_hw
*hw
)
7914 struct iwl_priv
*priv
= hw
->priv
;
7915 unsigned long flags
;
7917 mutex_lock(&priv
->mutex
);
7918 IWL_DEBUG_MAC80211("enter\n");
7920 priv
->lq_mngr
.lq_ready
= 0;
7921 #ifdef CONFIG_IWL4965_HT
7922 spin_lock_irqsave(&priv
->lock
, flags
);
7923 memset(&priv
->current_assoc_ht
, 0, sizeof(struct sta_ht_info
));
7924 spin_unlock_irqrestore(&priv
->lock
, flags
);
7925 #ifdef CONFIG_IWL4965_HT_AGG
7926 /* if (priv->lq_mngr.agg_ctrl.granted_ba)
7927 iwl4965_turn_off_agg(priv, TID_ALL_SPECIFIED);*/
7929 memset(&(priv
->lq_mngr
.agg_ctrl
), 0, sizeof(struct iwl_agg_control
));
7930 priv
->lq_mngr
.agg_ctrl
.tid_traffic_load_threshold
= 10;
7931 priv
->lq_mngr
.agg_ctrl
.ba_timeout
= 5000;
7932 priv
->lq_mngr
.agg_ctrl
.auto_agg
= 1;
7934 if (priv
->lq_mngr
.agg_ctrl
.auto_agg
)
7935 priv
->lq_mngr
.agg_ctrl
.requested_ba
= TID_ALL_ENABLED
;
7936 #endif /*CONFIG_IWL4965_HT_AGG */
7937 #endif /* CONFIG_IWL4965_HT */
7939 #ifdef CONFIG_IWL4965_QOS
7940 iwl_reset_qos(priv
);
7943 cancel_delayed_work(&priv
->post_associate
);
7945 spin_lock_irqsave(&priv
->lock
, flags
);
7947 priv
->assoc_capability
= 0;
7948 priv
->call_post_assoc_from_beacon
= 0;
7949 priv
->assoc_station_added
= 0;
7951 /* new association get rid of ibss beacon skb */
7952 if (priv
->ibss_beacon
)
7953 dev_kfree_skb(priv
->ibss_beacon
);
7955 priv
->ibss_beacon
= NULL
;
7957 priv
->beacon_int
= priv
->hw
->conf
.beacon_int
;
7958 priv
->timestamp1
= 0;
7959 priv
->timestamp0
= 0;
7960 if ((priv
->iw_mode
== IEEE80211_IF_TYPE_STA
))
7961 priv
->beacon_int
= 0;
7963 spin_unlock_irqrestore(&priv
->lock
, flags
);
7965 /* we are restarting association process
7966 * clear RXON_FILTER_ASSOC_MSK bit
7968 if (priv
->iw_mode
!= IEEE80211_IF_TYPE_AP
) {
7969 iwl_scan_cancel_timeout(priv
, 100);
7970 priv
->staging_rxon
.filter_flags
&= ~RXON_FILTER_ASSOC_MSK
;
7971 iwl_commit_rxon(priv
);
7974 /* Per mac80211.h: This is only used in IBSS mode... */
7975 if (priv
->iw_mode
!= IEEE80211_IF_TYPE_IBSS
) {
7977 IWL_DEBUG_MAC80211("leave - not in IBSS\n");
7978 mutex_unlock(&priv
->mutex
);
7982 if (!iwl_is_ready_rf(priv
)) {
7983 IWL_DEBUG_MAC80211("leave - not ready\n");
7984 mutex_unlock(&priv
->mutex
);
7988 priv
->only_active_channel
= 0;
7992 mutex_unlock(&priv
->mutex
);
7994 IWL_DEBUG_MAC80211("leave\n");
7998 static int iwl_mac_beacon_update(struct ieee80211_hw
*hw
, struct sk_buff
*skb
,
7999 struct ieee80211_tx_control
*control
)
8001 struct iwl_priv
*priv
= hw
->priv
;
8002 unsigned long flags
;
8004 mutex_lock(&priv
->mutex
);
8005 IWL_DEBUG_MAC80211("enter\n");
8007 if (!iwl_is_ready_rf(priv
)) {
8008 IWL_DEBUG_MAC80211("leave - RF not ready\n");
8009 mutex_unlock(&priv
->mutex
);
8013 if (priv
->iw_mode
!= IEEE80211_IF_TYPE_IBSS
) {
8014 IWL_DEBUG_MAC80211("leave - not IBSS\n");
8015 mutex_unlock(&priv
->mutex
);
8019 spin_lock_irqsave(&priv
->lock
, flags
);
8021 if (priv
->ibss_beacon
)
8022 dev_kfree_skb(priv
->ibss_beacon
);
8024 priv
->ibss_beacon
= skb
;
8028 IWL_DEBUG_MAC80211("leave\n");
8029 spin_unlock_irqrestore(&priv
->lock
, flags
);
8031 #ifdef CONFIG_IWL4965_QOS
8032 iwl_reset_qos(priv
);
8035 queue_work(priv
->workqueue
, &priv
->post_associate
.work
);
8037 mutex_unlock(&priv
->mutex
);
8042 #ifdef CONFIG_IWL4965_HT
8045 u16 advanced_coding_cap
:1;
8046 u16 supported_chan_width_set
:1;
8047 u16 mimo_power_save_mode
:2;
8053 u16 beam_forming
:1;
8055 u16 maximal_amsdu_size
:1;
8056 u16 cck_mode_at_40MHz
:1;
8057 u16 psmp_support
:1;
8058 u16 stbc_ctrl_frame_support
:1;
8059 u16 sig_txop_protection_support
:1;
8062 } __attribute__ ((packed
));
8064 union ht_param_info
{
8066 u8 max_rx_ampdu_factor
:2;
8071 } __attribute__ ((packed
));
8073 union ht_exra_param_info
{
8075 u8 ext_chan_offset
:2;
8076 u8 tx_chan_width
:1;
8078 u8 controlled_access_only
:1;
8079 u8 service_interval_granularity
:3;
8082 } __attribute__ ((packed
));
8084 union ht_operation_mode
{
8091 } __attribute__ ((packed
));
8094 static int sta_ht_info_init(struct ieee80211_ht_capability
*ht_cap
,
8095 struct ieee80211_ht_additional_info
*ht_extra
,
8096 struct sta_ht_info
*ht_info_ap
,
8097 struct sta_ht_info
*ht_info
)
8099 union ht_cap_info cap
;
8100 union ht_operation_mode op_mode
;
8101 union ht_param_info param_info
;
8102 union ht_exra_param_info extra_param_info
;
8104 IWL_DEBUG_MAC80211("enter: \n");
8107 IWL_DEBUG_MAC80211("leave: ht_info is NULL\n");
8112 cap
.val
= (u16
) le16_to_cpu(ht_cap
->capabilities_info
);
8113 param_info
.val
= ht_cap
->mac_ht_params_info
;
8116 ht_info
->sgf
|= 0x1;
8118 ht_info
->sgf
|= 0x2;
8119 ht_info
->is_green_field
= cap
.green_field
;
8120 ht_info
->max_amsdu_size
= cap
.maximal_amsdu_size
;
8121 ht_info
->supported_chan_width
= cap
.supported_chan_width_set
;
8122 ht_info
->tx_mimo_ps_mode
= cap
.mimo_power_save_mode
;
8123 memcpy(ht_info
->supp_rates
, ht_cap
->supported_mcs_set
, 16);
8125 ht_info
->ampdu_factor
= param_info
.max_rx_ampdu_factor
;
8126 ht_info
->mpdu_density
= param_info
.mpdu_density
;
8128 IWL_DEBUG_MAC80211("SISO mask 0x%X MIMO mask 0x%X \n",
8129 ht_cap
->supported_mcs_set
[0],
8130 ht_cap
->supported_mcs_set
[1]);
8133 ht_info
->control_channel
= ht_info_ap
->control_channel
;
8134 ht_info
->extension_chan_offset
=
8135 ht_info_ap
->extension_chan_offset
;
8136 ht_info
->tx_chan_width
= ht_info_ap
->tx_chan_width
;
8137 ht_info
->operating_mode
= ht_info_ap
->operating_mode
;
8141 extra_param_info
.val
= ht_extra
->ht_param
;
8142 ht_info
->control_channel
= ht_extra
->control_chan
;
8143 ht_info
->extension_chan_offset
=
8144 extra_param_info
.ext_chan_offset
;
8145 ht_info
->tx_chan_width
= extra_param_info
.tx_chan_width
;
8147 le16_to_cpu(ht_extra
->operation_mode
);
8148 ht_info
->operating_mode
= op_mode
.op_mode
;
8149 IWL_DEBUG_MAC80211("control channel %d\n",
8150 ht_extra
->control_chan
);
8155 IWL_DEBUG_MAC80211("leave\n");
8159 static int iwl_mac_conf_ht(struct ieee80211_hw
*hw
,
8160 struct ieee80211_ht_capability
*ht_cap
,
8161 struct ieee80211_ht_additional_info
*ht_extra
)
8163 struct iwl_priv
*priv
= hw
->priv
;
8166 IWL_DEBUG_MAC80211("enter: \n");
8168 rs
= sta_ht_info_init(ht_cap
, ht_extra
, NULL
, &priv
->current_assoc_ht
);
8169 iwl4965_set_rxon_chain(priv
);
8171 if (priv
&& priv
->assoc_id
&&
8172 (priv
->iw_mode
== IEEE80211_IF_TYPE_STA
)) {
8173 unsigned long flags
;
8175 spin_lock_irqsave(&priv
->lock
, flags
);
8176 if (priv
->beacon_int
)
8177 queue_work(priv
->workqueue
, &priv
->post_associate
.work
);
8179 priv
->call_post_assoc_from_beacon
= 1;
8180 spin_unlock_irqrestore(&priv
->lock
, flags
);
8183 IWL_DEBUG_MAC80211("leave: control channel %d\n",
8184 ht_extra
->control_chan
);
8189 static void iwl_set_ht_capab(struct ieee80211_hw
*hw
,
8190 struct ieee80211_ht_capability
*ht_cap
,
8193 union ht_cap_info cap
;
8194 union ht_param_info param_info
;
8196 memset(&cap
, 0, sizeof(union ht_cap_info
));
8197 memset(¶m_info
, 0, sizeof(union ht_param_info
));
8199 cap
.maximal_amsdu_size
= HT_IE_MAX_AMSDU_SIZE_4K
;
8200 cap
.green_field
= 1;
8203 cap
.supported_chan_width_set
= use_wide_chan
;
8204 cap
.mimo_power_save_mode
= 0x3;
8206 param_info
.max_rx_ampdu_factor
= CFG_HT_RX_AMPDU_FACTOR_DEF
;
8207 param_info
.mpdu_density
= CFG_HT_MPDU_DENSITY_DEF
;
8208 ht_cap
->capabilities_info
= (__le16
) cpu_to_le16(cap
.val
);
8209 ht_cap
->mac_ht_params_info
= (u8
) param_info
.val
;
8211 ht_cap
->supported_mcs_set
[0] = 0xff;
8212 ht_cap
->supported_mcs_set
[1] = 0xff;
8213 ht_cap
->supported_mcs_set
[4] =
8214 (cap
.supported_chan_width_set
) ? 0x1: 0x0;
8217 static void iwl_mac_get_ht_capab(struct ieee80211_hw
*hw
,
8218 struct ieee80211_ht_capability
*ht_cap
)
8220 u8 use_wide_channel
= 1;
8221 struct iwl_priv
*priv
= hw
->priv
;
8223 IWL_DEBUG_MAC80211("enter: \n");
8224 if (priv
->channel_width
!= IWL_CHANNEL_WIDTH_40MHZ
)
8225 use_wide_channel
= 0;
8227 /* no fat tx allowed on 2.4GHZ */
8228 if (priv
->phymode
!= MODE_IEEE80211A
)
8229 use_wide_channel
= 0;
8231 iwl_set_ht_capab(hw
, ht_cap
, use_wide_channel
);
8232 IWL_DEBUG_MAC80211("leave: \n");
8234 #endif /*CONFIG_IWL4965_HT*/
8236 /*****************************************************************************
8240 *****************************************************************************/
8242 #ifdef CONFIG_IWL4965_DEBUG
8245 * The following adds a new attribute to the sysfs representation
8246 * of this device driver (i.e. a new file in /sys/bus/pci/drivers/iwl/)
8247 * used for controlling the debug level.
8249 * See the level definitions in iwl for details.
8252 static ssize_t
show_debug_level(struct device_driver
*d
, char *buf
)
8254 return sprintf(buf
, "0x%08X\n", iwl_debug_level
);
8256 static ssize_t
store_debug_level(struct device_driver
*d
,
8257 const char *buf
, size_t count
)
8259 char *p
= (char *)buf
;
8262 val
= simple_strtoul(p
, &p
, 0);
8264 printk(KERN_INFO DRV_NAME
8265 ": %s is not in hex or decimal form.\n", buf
);
8267 iwl_debug_level
= val
;
8269 return strnlen(buf
, count
);
8272 static DRIVER_ATTR(debug_level
, S_IWUSR
| S_IRUGO
,
8273 show_debug_level
, store_debug_level
);
8275 #endif /* CONFIG_IWL4965_DEBUG */
8277 static ssize_t
show_rf_kill(struct device
*d
,
8278 struct device_attribute
*attr
, char *buf
)
8281 * 0 - RF kill not enabled
8282 * 1 - SW based RF kill active (sysfs)
8283 * 2 - HW based RF kill active
8284 * 3 - Both HW and SW based RF kill active
8286 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8287 int val
= (test_bit(STATUS_RF_KILL_SW
, &priv
->status
) ? 0x1 : 0x0) |
8288 (test_bit(STATUS_RF_KILL_HW
, &priv
->status
) ? 0x2 : 0x0);
8290 return sprintf(buf
, "%i\n", val
);
8293 static ssize_t
store_rf_kill(struct device
*d
,
8294 struct device_attribute
*attr
,
8295 const char *buf
, size_t count
)
8297 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8299 mutex_lock(&priv
->mutex
);
8300 iwl_radio_kill_sw(priv
, buf
[0] == '1');
8301 mutex_unlock(&priv
->mutex
);
8306 static DEVICE_ATTR(rf_kill
, S_IWUSR
| S_IRUGO
, show_rf_kill
, store_rf_kill
);
8308 static ssize_t
show_temperature(struct device
*d
,
8309 struct device_attribute
*attr
, char *buf
)
8311 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8313 if (!iwl_is_alive(priv
))
8316 return sprintf(buf
, "%d\n", iwl_hw_get_temperature(priv
));
8319 static DEVICE_ATTR(temperature
, S_IRUGO
, show_temperature
, NULL
);
8321 static ssize_t
show_rs_window(struct device
*d
,
8322 struct device_attribute
*attr
,
8325 struct iwl_priv
*priv
= d
->driver_data
;
8326 return iwl_fill_rs_info(priv
->hw
, buf
, IWL_AP_ID
);
8328 static DEVICE_ATTR(rs_window
, S_IRUGO
, show_rs_window
, NULL
);
8330 static ssize_t
show_tx_power(struct device
*d
,
8331 struct device_attribute
*attr
, char *buf
)
8333 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8334 return sprintf(buf
, "%d\n", priv
->user_txpower_limit
);
8337 static ssize_t
store_tx_power(struct device
*d
,
8338 struct device_attribute
*attr
,
8339 const char *buf
, size_t count
)
8341 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8342 char *p
= (char *)buf
;
8345 val
= simple_strtoul(p
, &p
, 10);
8347 printk(KERN_INFO DRV_NAME
8348 ": %s is not in decimal form.\n", buf
);
8350 iwl_hw_reg_set_txpower(priv
, val
);
8355 static DEVICE_ATTR(tx_power
, S_IWUSR
| S_IRUGO
, show_tx_power
, store_tx_power
);
8357 static ssize_t
show_flags(struct device
*d
,
8358 struct device_attribute
*attr
, char *buf
)
8360 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8362 return sprintf(buf
, "0x%04X\n", priv
->active_rxon
.flags
);
8365 static ssize_t
store_flags(struct device
*d
,
8366 struct device_attribute
*attr
,
8367 const char *buf
, size_t count
)
8369 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8370 u32 flags
= simple_strtoul(buf
, NULL
, 0);
8372 mutex_lock(&priv
->mutex
);
8373 if (le32_to_cpu(priv
->staging_rxon
.flags
) != flags
) {
8374 /* Cancel any currently running scans... */
8375 if (iwl_scan_cancel_timeout(priv
, 100))
8376 IWL_WARNING("Could not cancel scan.\n");
8378 IWL_DEBUG_INFO("Committing rxon.flags = 0x%04X\n",
8380 priv
->staging_rxon
.flags
= cpu_to_le32(flags
);
8381 iwl_commit_rxon(priv
);
8384 mutex_unlock(&priv
->mutex
);
8389 static DEVICE_ATTR(flags
, S_IWUSR
| S_IRUGO
, show_flags
, store_flags
);
8391 static ssize_t
show_filter_flags(struct device
*d
,
8392 struct device_attribute
*attr
, char *buf
)
8394 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8396 return sprintf(buf
, "0x%04X\n",
8397 le32_to_cpu(priv
->active_rxon
.filter_flags
));
8400 static ssize_t
store_filter_flags(struct device
*d
,
8401 struct device_attribute
*attr
,
8402 const char *buf
, size_t count
)
8404 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8405 u32 filter_flags
= simple_strtoul(buf
, NULL
, 0);
8407 mutex_lock(&priv
->mutex
);
8408 if (le32_to_cpu(priv
->staging_rxon
.filter_flags
) != filter_flags
) {
8409 /* Cancel any currently running scans... */
8410 if (iwl_scan_cancel_timeout(priv
, 100))
8411 IWL_WARNING("Could not cancel scan.\n");
8413 IWL_DEBUG_INFO("Committing rxon.filter_flags = "
8414 "0x%04X\n", filter_flags
);
8415 priv
->staging_rxon
.filter_flags
=
8416 cpu_to_le32(filter_flags
);
8417 iwl_commit_rxon(priv
);
8420 mutex_unlock(&priv
->mutex
);
8425 static DEVICE_ATTR(filter_flags
, S_IWUSR
| S_IRUGO
, show_filter_flags
,
8426 store_filter_flags
);
8428 static ssize_t
show_tune(struct device
*d
,
8429 struct device_attribute
*attr
, char *buf
)
8431 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8433 return sprintf(buf
, "0x%04X\n",
8434 (priv
->phymode
<< 8) |
8435 le16_to_cpu(priv
->active_rxon
.channel
));
8438 static void iwl_set_flags_for_phymode(struct iwl_priv
*priv
, u8 phymode
);
8440 static ssize_t
store_tune(struct device
*d
,
8441 struct device_attribute
*attr
,
8442 const char *buf
, size_t count
)
8444 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8445 char *p
= (char *)buf
;
8446 u16 tune
= simple_strtoul(p
, &p
, 0);
8447 u8 phymode
= (tune
>> 8) & 0xff;
8448 u16 channel
= tune
& 0xff;
8450 IWL_DEBUG_INFO("Tune request to:%d channel:%d\n", phymode
, channel
);
8452 mutex_lock(&priv
->mutex
);
8453 if ((le16_to_cpu(priv
->staging_rxon
.channel
) != channel
) ||
8454 (priv
->phymode
!= phymode
)) {
8455 const struct iwl_channel_info
*ch_info
;
8457 ch_info
= iwl_get_channel_info(priv
, phymode
, channel
);
8459 IWL_WARNING("Requested invalid phymode/channel "
8460 "combination: %d %d\n", phymode
, channel
);
8461 mutex_unlock(&priv
->mutex
);
8465 /* Cancel any currently running scans... */
8466 if (iwl_scan_cancel_timeout(priv
, 100))
8467 IWL_WARNING("Could not cancel scan.\n");
8469 IWL_DEBUG_INFO("Committing phymode and "
8470 "rxon.channel = %d %d\n",
8473 iwl_set_rxon_channel(priv
, phymode
, channel
);
8474 iwl_set_flags_for_phymode(priv
, phymode
);
8477 iwl_commit_rxon(priv
);
8480 mutex_unlock(&priv
->mutex
);
8485 static DEVICE_ATTR(tune
, S_IWUSR
| S_IRUGO
, show_tune
, store_tune
);
8487 #ifdef CONFIG_IWL4965_SPECTRUM_MEASUREMENT
8489 static ssize_t
show_measurement(struct device
*d
,
8490 struct device_attribute
*attr
, char *buf
)
8492 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8493 struct iwl_spectrum_notification measure_report
;
8494 u32 size
= sizeof(measure_report
), len
= 0, ofs
= 0;
8495 u8
*data
= (u8
*) & measure_report
;
8496 unsigned long flags
;
8498 spin_lock_irqsave(&priv
->lock
, flags
);
8499 if (!(priv
->measurement_status
& MEASUREMENT_READY
)) {
8500 spin_unlock_irqrestore(&priv
->lock
, flags
);
8503 memcpy(&measure_report
, &priv
->measure_report
, size
);
8504 priv
->measurement_status
= 0;
8505 spin_unlock_irqrestore(&priv
->lock
, flags
);
8507 while (size
&& (PAGE_SIZE
- len
)) {
8508 hex_dump_to_buffer(data
+ ofs
, size
, 16, 1, buf
+ len
,
8509 PAGE_SIZE
- len
, 1);
8511 if (PAGE_SIZE
- len
)
8515 size
-= min(size
, 16U);
8521 static ssize_t
store_measurement(struct device
*d
,
8522 struct device_attribute
*attr
,
8523 const char *buf
, size_t count
)
8525 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8526 struct ieee80211_measurement_params params
= {
8527 .channel
= le16_to_cpu(priv
->active_rxon
.channel
),
8528 .start_time
= cpu_to_le64(priv
->last_tsf
),
8529 .duration
= cpu_to_le16(1),
8531 u8 type
= IWL_MEASURE_BASIC
;
8537 strncpy(buffer
, buf
, min(sizeof(buffer
), count
));
8538 channel
= simple_strtoul(p
, NULL
, 0);
8540 params
.channel
= channel
;
8543 while (*p
&& *p
!= ' ')
8546 type
= simple_strtoul(p
+ 1, NULL
, 0);
8549 IWL_DEBUG_INFO("Invoking measurement of type %d on "
8550 "channel %d (for '%s')\n", type
, params
.channel
, buf
);
8551 iwl_get_measurement(priv
, ¶ms
, type
);
8556 static DEVICE_ATTR(measurement
, S_IRUSR
| S_IWUSR
,
8557 show_measurement
, store_measurement
);
8558 #endif /* CONFIG_IWL4965_SPECTRUM_MEASUREMENT */
8560 static ssize_t
store_retry_rate(struct device
*d
,
8561 struct device_attribute
*attr
,
8562 const char *buf
, size_t count
)
8564 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8566 priv
->retry_rate
= simple_strtoul(buf
, NULL
, 0);
8567 if (priv
->retry_rate
<= 0)
8568 priv
->retry_rate
= 1;
8573 static ssize_t
show_retry_rate(struct device
*d
,
8574 struct device_attribute
*attr
, char *buf
)
8576 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8577 return sprintf(buf
, "%d", priv
->retry_rate
);
8580 static DEVICE_ATTR(retry_rate
, S_IWUSR
| S_IRUSR
, show_retry_rate
,
8583 static ssize_t
store_power_level(struct device
*d
,
8584 struct device_attribute
*attr
,
8585 const char *buf
, size_t count
)
8587 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8591 mode
= simple_strtoul(buf
, NULL
, 0);
8592 mutex_lock(&priv
->mutex
);
8594 if (!iwl_is_ready(priv
)) {
8599 if ((mode
< 1) || (mode
> IWL_POWER_LIMIT
) || (mode
== IWL_POWER_AC
))
8600 mode
= IWL_POWER_AC
;
8602 mode
|= IWL_POWER_ENABLED
;
8604 if (mode
!= priv
->power_mode
) {
8605 rc
= iwl_send_power_mode(priv
, IWL_POWER_LEVEL(mode
));
8607 IWL_DEBUG_MAC80211("failed setting power mode.\n");
8610 priv
->power_mode
= mode
;
8616 mutex_unlock(&priv
->mutex
);
8620 #define MAX_WX_STRING 80
8622 /* Values are in microsecond */
8623 static const s32 timeout_duration
[] = {
8630 static const s32 period_duration
[] = {
8638 static ssize_t
show_power_level(struct device
*d
,
8639 struct device_attribute
*attr
, char *buf
)
8641 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8642 int level
= IWL_POWER_LEVEL(priv
->power_mode
);
8645 p
+= sprintf(p
, "%d ", level
);
8647 case IWL_POWER_MODE_CAM
:
8649 p
+= sprintf(p
, "(AC)");
8651 case IWL_POWER_BATTERY
:
8652 p
+= sprintf(p
, "(BATTERY)");
8656 "(Timeout %dms, Period %dms)",
8657 timeout_duration
[level
- 1] / 1000,
8658 period_duration
[level
- 1] / 1000);
8661 if (!(priv
->power_mode
& IWL_POWER_ENABLED
))
8662 p
+= sprintf(p
, " OFF\n");
8664 p
+= sprintf(p
, " \n");
8666 return (p
- buf
+ 1);
8670 static DEVICE_ATTR(power_level
, S_IWUSR
| S_IRUSR
, show_power_level
,
8673 static ssize_t
show_channels(struct device
*d
,
8674 struct device_attribute
*attr
, char *buf
)
8676 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8678 struct ieee80211_channel
*channels
= NULL
;
8679 const struct ieee80211_hw_mode
*hw_mode
= NULL
;
8682 if (!iwl_is_ready(priv
))
8685 hw_mode
= iwl_get_hw_mode(priv
, MODE_IEEE80211G
);
8687 hw_mode
= iwl_get_hw_mode(priv
, MODE_IEEE80211B
);
8689 channels
= hw_mode
->channels
;
8690 count
= hw_mode
->num_channels
;
8695 "Displaying %d channels in 2.4GHz band "
8696 "(802.11bg):\n", count
);
8698 for (i
= 0; i
< count
; i
++)
8699 len
+= sprintf(&buf
[len
], "%d: %ddBm: BSS%s%s, %s.\n",
8701 channels
[i
].power_level
,
8703 flag
& IEEE80211_CHAN_W_RADAR_DETECT
?
8704 " (IEEE 802.11h required)" : "",
8705 (!(channels
[i
].flag
& IEEE80211_CHAN_W_IBSS
)
8708 IEEE80211_CHAN_W_RADAR_DETECT
)) ? "" :
8711 flag
& IEEE80211_CHAN_W_ACTIVE_SCAN
?
8712 "active/passive" : "passive only");
8714 hw_mode
= iwl_get_hw_mode(priv
, MODE_IEEE80211A
);
8716 channels
= hw_mode
->channels
;
8717 count
= hw_mode
->num_channels
;
8723 len
+= sprintf(&buf
[len
], "Displaying %d channels in 5.2GHz band "
8724 "(802.11a):\n", count
);
8726 for (i
= 0; i
< count
; i
++)
8727 len
+= sprintf(&buf
[len
], "%d: %ddBm: BSS%s%s, %s.\n",
8729 channels
[i
].power_level
,
8731 flag
& IEEE80211_CHAN_W_RADAR_DETECT
?
8732 " (IEEE 802.11h required)" : "",
8733 (!(channels
[i
].flag
& IEEE80211_CHAN_W_IBSS
)
8736 IEEE80211_CHAN_W_RADAR_DETECT
)) ? "" :
8739 flag
& IEEE80211_CHAN_W_ACTIVE_SCAN
?
8740 "active/passive" : "passive only");
8745 static DEVICE_ATTR(channels
, S_IRUSR
, show_channels
, NULL
);
8747 static ssize_t
show_statistics(struct device
*d
,
8748 struct device_attribute
*attr
, char *buf
)
8750 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8751 u32 size
= sizeof(struct iwl_notif_statistics
);
8752 u32 len
= 0, ofs
= 0;
8753 u8
*data
= (u8
*) & priv
->statistics
;
8756 if (!iwl_is_alive(priv
))
8759 mutex_lock(&priv
->mutex
);
8760 rc
= iwl_send_statistics_request(priv
);
8761 mutex_unlock(&priv
->mutex
);
8765 "Error sending statistics request: 0x%08X\n", rc
);
8769 while (size
&& (PAGE_SIZE
- len
)) {
8770 hex_dump_to_buffer(data
+ ofs
, size
, 16, 1, buf
+ len
,
8771 PAGE_SIZE
- len
, 1);
8773 if (PAGE_SIZE
- len
)
8777 size
-= min(size
, 16U);
8783 static DEVICE_ATTR(statistics
, S_IRUGO
, show_statistics
, NULL
);
8785 static ssize_t
show_antenna(struct device
*d
,
8786 struct device_attribute
*attr
, char *buf
)
8788 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8790 if (!iwl_is_alive(priv
))
8793 return sprintf(buf
, "%d\n", priv
->antenna
);
8796 static ssize_t
store_antenna(struct device
*d
,
8797 struct device_attribute
*attr
,
8798 const char *buf
, size_t count
)
8801 struct iwl_priv
*priv
= dev_get_drvdata(d
);
8806 if (sscanf(buf
, "%1i", &ant
) != 1) {
8807 IWL_DEBUG_INFO("not in hex or decimal form.\n");
8811 if ((ant
>= 0) && (ant
<= 2)) {
8812 IWL_DEBUG_INFO("Setting antenna select to %d.\n", ant
);
8813 priv
->antenna
= (enum iwl_antenna
)ant
;
8815 IWL_DEBUG_INFO("Bad antenna select value %d.\n", ant
);
8821 static DEVICE_ATTR(antenna
, S_IWUSR
| S_IRUGO
, show_antenna
, store_antenna
);
8823 static ssize_t
show_status(struct device
*d
,
8824 struct device_attribute
*attr
, char *buf
)
8826 struct iwl_priv
*priv
= (struct iwl_priv
*)d
->driver_data
;
8827 if (!iwl_is_alive(priv
))
8829 return sprintf(buf
, "0x%08x\n", (int)priv
->status
);
8832 static DEVICE_ATTR(status
, S_IRUGO
, show_status
, NULL
);
8834 static ssize_t
dump_error_log(struct device
*d
,
8835 struct device_attribute
*attr
,
8836 const char *buf
, size_t count
)
8838 char *p
= (char *)buf
;
8841 iwl_dump_nic_error_log((struct iwl_priv
*)d
->driver_data
);
8843 return strnlen(buf
, count
);
8846 static DEVICE_ATTR(dump_errors
, S_IWUSR
, NULL
, dump_error_log
);
8848 static ssize_t
dump_event_log(struct device
*d
,
8849 struct device_attribute
*attr
,
8850 const char *buf
, size_t count
)
8852 char *p
= (char *)buf
;
8855 iwl_dump_nic_event_log((struct iwl_priv
*)d
->driver_data
);
8857 return strnlen(buf
, count
);
8860 static DEVICE_ATTR(dump_events
, S_IWUSR
, NULL
, dump_event_log
);
8862 /*****************************************************************************
8864 * driver setup and teardown
8866 *****************************************************************************/
8868 static void iwl_setup_deferred_work(struct iwl_priv
*priv
)
8870 priv
->workqueue
= create_workqueue(DRV_NAME
);
8872 init_waitqueue_head(&priv
->wait_command_queue
);
8874 INIT_WORK(&priv
->up
, iwl_bg_up
);
8875 INIT_WORK(&priv
->restart
, iwl_bg_restart
);
8876 INIT_WORK(&priv
->rx_replenish
, iwl_bg_rx_replenish
);
8877 INIT_WORK(&priv
->scan_completed
, iwl_bg_scan_completed
);
8878 INIT_WORK(&priv
->request_scan
, iwl_bg_request_scan
);
8879 INIT_WORK(&priv
->abort_scan
, iwl_bg_abort_scan
);
8880 INIT_WORK(&priv
->rf_kill
, iwl_bg_rf_kill
);
8881 INIT_WORK(&priv
->beacon_update
, iwl_bg_beacon_update
);
8882 INIT_DELAYED_WORK(&priv
->post_associate
, iwl_bg_post_associate
);
8883 INIT_DELAYED_WORK(&priv
->init_alive_start
, iwl_bg_init_alive_start
);
8884 INIT_DELAYED_WORK(&priv
->alive_start
, iwl_bg_alive_start
);
8885 INIT_DELAYED_WORK(&priv
->scan_check
, iwl_bg_scan_check
);
8887 iwl_hw_setup_deferred_work(priv
);
8889 tasklet_init(&priv
->irq_tasklet
, (void (*)(unsigned long))
8890 iwl_irq_tasklet
, (unsigned long)priv
);
8893 static void iwl_cancel_deferred_work(struct iwl_priv
*priv
)
8895 iwl_hw_cancel_deferred_work(priv
);
8897 cancel_delayed_work_sync(&priv
->init_alive_start
);
8898 cancel_delayed_work(&priv
->scan_check
);
8899 cancel_delayed_work(&priv
->alive_start
);
8900 cancel_delayed_work(&priv
->post_associate
);
8901 cancel_work_sync(&priv
->beacon_update
);
8904 static struct attribute
*iwl_sysfs_entries
[] = {
8905 &dev_attr_antenna
.attr
,
8906 &dev_attr_channels
.attr
,
8907 &dev_attr_dump_errors
.attr
,
8908 &dev_attr_dump_events
.attr
,
8909 &dev_attr_flags
.attr
,
8910 &dev_attr_filter_flags
.attr
,
8911 #ifdef CONFIG_IWL4965_SPECTRUM_MEASUREMENT
8912 &dev_attr_measurement
.attr
,
8914 &dev_attr_power_level
.attr
,
8915 &dev_attr_retry_rate
.attr
,
8916 &dev_attr_rf_kill
.attr
,
8917 &dev_attr_rs_window
.attr
,
8918 &dev_attr_statistics
.attr
,
8919 &dev_attr_status
.attr
,
8920 &dev_attr_temperature
.attr
,
8921 &dev_attr_tune
.attr
,
8922 &dev_attr_tx_power
.attr
,
8927 static struct attribute_group iwl_attribute_group
= {
8928 .name
= NULL
, /* put in device directory */
8929 .attrs
= iwl_sysfs_entries
,
8932 static struct ieee80211_ops iwl_hw_ops
= {
8934 .start
= iwl_mac_start
,
8935 .stop
= iwl_mac_stop
,
8936 .add_interface
= iwl_mac_add_interface
,
8937 .remove_interface
= iwl_mac_remove_interface
,
8938 .config
= iwl_mac_config
,
8939 .config_interface
= iwl_mac_config_interface
,
8940 .configure_filter
= iwl_configure_filter
,
8941 .set_key
= iwl_mac_set_key
,
8942 .get_stats
= iwl_mac_get_stats
,
8943 .get_tx_stats
= iwl_mac_get_tx_stats
,
8944 .conf_tx
= iwl_mac_conf_tx
,
8945 .get_tsf
= iwl_mac_get_tsf
,
8946 .reset_tsf
= iwl_mac_reset_tsf
,
8947 .beacon_update
= iwl_mac_beacon_update
,
8948 .erp_ie_changed
= iwl_mac_erp_ie_changed
,
8949 #ifdef CONFIG_IWL4965_HT
8950 .conf_ht
= iwl_mac_conf_ht
,
8951 .get_ht_capab
= iwl_mac_get_ht_capab
,
8952 #ifdef CONFIG_IWL4965_HT_AGG
8953 .ht_tx_agg_start
= iwl_mac_ht_tx_agg_start
,
8954 .ht_tx_agg_stop
= iwl_mac_ht_tx_agg_stop
,
8955 .ht_rx_agg_start
= iwl_mac_ht_rx_agg_start
,
8956 .ht_rx_agg_stop
= iwl_mac_ht_rx_agg_stop
,
8957 #endif /* CONFIG_IWL4965_HT_AGG */
8958 #endif /* CONFIG_IWL4965_HT */
8959 .hw_scan
= iwl_mac_hw_scan
8962 static int iwl_pci_probe(struct pci_dev
*pdev
, const struct pci_device_id
*ent
)
8965 struct iwl_priv
*priv
;
8966 struct ieee80211_hw
*hw
;
8969 if (iwl_param_disable_hw_scan
) {
8970 IWL_DEBUG_INFO("Disabling hw_scan\n");
8971 iwl_hw_ops
.hw_scan
= NULL
;
8974 if ((iwl_param_queues_num
> IWL_MAX_NUM_QUEUES
) ||
8975 (iwl_param_queues_num
< IWL_MIN_NUM_QUEUES
)) {
8976 IWL_ERROR("invalid queues_num, should be between %d and %d\n",
8977 IWL_MIN_NUM_QUEUES
, IWL_MAX_NUM_QUEUES
);
8982 /* mac80211 allocates memory for this device instance, including
8983 * space for this driver's private structure */
8984 hw
= ieee80211_alloc_hw(sizeof(struct iwl_priv
), &iwl_hw_ops
);
8986 IWL_ERROR("Can not allocate network device\n");
8990 SET_IEEE80211_DEV(hw
, &pdev
->dev
);
8992 hw
->rate_control_algorithm
= "iwl-4965-rs";
8994 IWL_DEBUG_INFO("*** LOAD DRIVER ***\n");
8998 priv
->pci_dev
= pdev
;
8999 priv
->antenna
= (enum iwl_antenna
)iwl_param_antenna
;
9000 #ifdef CONFIG_IWL4965_DEBUG
9001 iwl_debug_level
= iwl_param_debug
;
9002 atomic_set(&priv
->restrict_refcnt
, 0);
9004 priv
->retry_rate
= 1;
9006 priv
->ibss_beacon
= NULL
;
9008 /* Tell mac80211 and its clients (e.g. Wireless Extensions)
9009 * the range of signal quality values that we'll provide.
9010 * Negative values for level/noise indicate that we'll provide dBm.
9011 * For WE, at least, non-0 values here *enable* display of values
9012 * in app (iwconfig). */
9013 hw
->max_rssi
= -20; /* signal level, negative indicates dBm */
9014 hw
->max_noise
= -20; /* noise level, negative indicates dBm */
9015 hw
->max_signal
= 100; /* link quality indication (%) */
9017 /* Tell mac80211 our Tx characteristics */
9018 hw
->flags
= IEEE80211_HW_HOST_GEN_BEACON_TEMPLATE
;
9021 #ifdef CONFIG_IWL4965_HT
9022 #ifdef CONFIG_IWL4965_HT_AGG
9024 #endif /* CONFIG_IWL4965_HT_AGG */
9025 #endif /* CONFIG_IWL4965_HT */
9027 spin_lock_init(&priv
->lock
);
9028 spin_lock_init(&priv
->power_data
.lock
);
9029 spin_lock_init(&priv
->sta_lock
);
9030 spin_lock_init(&priv
->hcmd_lock
);
9031 spin_lock_init(&priv
->lq_mngr
.lock
);
9033 for (i
= 0; i
< IWL_IBSS_MAC_HASH_SIZE
; i
++)
9034 INIT_LIST_HEAD(&priv
->ibss_mac_hash
[i
]);
9036 INIT_LIST_HEAD(&priv
->free_frames
);
9038 mutex_init(&priv
->mutex
);
9039 if (pci_enable_device(pdev
)) {
9041 goto out_ieee80211_free_hw
;
9044 pci_set_master(pdev
);
9046 iwl_clear_stations_table(priv
);
9048 priv
->data_retry_limit
= -1;
9049 priv
->ieee_channels
= NULL
;
9050 priv
->ieee_rates
= NULL
;
9053 err
= pci_set_dma_mask(pdev
, DMA_32BIT_MASK
);
9055 err
= pci_set_consistent_dma_mask(pdev
, DMA_32BIT_MASK
);
9057 printk(KERN_WARNING DRV_NAME
": No suitable DMA available.\n");
9058 goto out_pci_disable_device
;
9061 pci_set_drvdata(pdev
, priv
);
9062 err
= pci_request_regions(pdev
, DRV_NAME
);
9064 goto out_pci_disable_device
;
9065 /* We disable the RETRY_TIMEOUT register (0x41) to keep
9066 * PCI Tx retries from interfering with C3 CPU state */
9067 pci_write_config_byte(pdev
, 0x41, 0x00);
9068 priv
->hw_base
= pci_iomap(pdev
, 0, 0);
9069 if (!priv
->hw_base
) {
9071 goto out_pci_release_regions
;
9074 IWL_DEBUG_INFO("pci_resource_len = 0x%08llx\n",
9075 (unsigned long long) pci_resource_len(pdev
, 0));
9076 IWL_DEBUG_INFO("pci_resource_base = %p\n", priv
->hw_base
);
9078 /* Initialize module parameter values here */
9080 if (iwl_param_disable
) {
9081 set_bit(STATUS_RF_KILL_SW
, &priv
->status
);
9082 IWL_DEBUG_INFO("Radio disabled.\n");
9085 priv
->iw_mode
= IEEE80211_IF_TYPE_STA
;
9088 priv
->use_ant_b_for_management_frame
= 1; /* start with ant B */
9089 priv
->is_ht_enabled
= 1;
9090 priv
->channel_width
= IWL_CHANNEL_WIDTH_40MHZ
;
9091 priv
->valid_antenna
= 0x7; /* assume all 3 connected */
9092 priv
->ps_mode
= IWL_MIMO_PS_NONE
;
9094 iwl4965_set_rxon_chain(priv
);
9096 printk(KERN_INFO DRV_NAME
9097 ": Detected Intel Wireless WiFi Link 4965AGN\n");
9099 /* Device-specific setup */
9100 if (iwl_hw_set_hw_setting(priv
)) {
9101 IWL_ERROR("failed to set hw settings\n");
9102 mutex_unlock(&priv
->mutex
);
9106 #ifdef CONFIG_IWL4965_QOS
9107 if (iwl_param_qos_enable
)
9108 priv
->qos_data
.qos_enable
= 1;
9110 iwl_reset_qos(priv
);
9112 priv
->qos_data
.qos_active
= 0;
9113 priv
->qos_data
.qos_cap
.val
= 0;
9114 #endif /* CONFIG_IWL4965_QOS */
9116 iwl_set_rxon_channel(priv
, MODE_IEEE80211G
, 6);
9117 iwl_setup_deferred_work(priv
);
9118 iwl_setup_rx_handlers(priv
);
9120 priv
->rates_mask
= IWL_RATES_MASK
;
9121 /* If power management is turned on, default to AC mode */
9122 priv
->power_mode
= IWL_POWER_AC
;
9123 priv
->user_txpower_limit
= IWL_DEFAULT_TX_POWER
;
9125 pci_enable_msi(pdev
);
9127 err
= request_irq(pdev
->irq
, iwl_isr
, IRQF_SHARED
, DRV_NAME
, priv
);
9129 IWL_ERROR("Error allocating IRQ %d\n", pdev
->irq
);
9130 goto out_disable_msi
;
9133 mutex_lock(&priv
->mutex
);
9135 err
= sysfs_create_group(&pdev
->dev
.kobj
, &iwl_attribute_group
);
9137 IWL_ERROR("failed to create sysfs device attributes\n");
9138 mutex_unlock(&priv
->mutex
);
9139 goto out_release_irq
;
9142 /* fetch ucode file from disk, alloc and copy to bus-master buffers ...
9143 * ucode filename and max sizes are card-specific. */
9144 err
= iwl_read_ucode(priv
);
9146 IWL_ERROR("Could not read microcode: %d\n", err
);
9147 mutex_unlock(&priv
->mutex
);
9151 mutex_unlock(&priv
->mutex
);
9153 IWL_DEBUG_INFO("Queueing UP work.\n");
9155 queue_work(priv
->workqueue
, &priv
->up
);
9160 iwl_dealloc_ucode_pci(priv
);
9162 sysfs_remove_group(&pdev
->dev
.kobj
, &iwl_attribute_group
);
9165 free_irq(pdev
->irq
, priv
);
9168 pci_disable_msi(pdev
);
9169 destroy_workqueue(priv
->workqueue
);
9170 priv
->workqueue
= NULL
;
9171 iwl_unset_hw_setting(priv
);
9174 pci_iounmap(pdev
, priv
->hw_base
);
9175 out_pci_release_regions
:
9176 pci_release_regions(pdev
);
9177 out_pci_disable_device
:
9178 pci_disable_device(pdev
);
9179 pci_set_drvdata(pdev
, NULL
);
9180 out_ieee80211_free_hw
:
9181 ieee80211_free_hw(priv
->hw
);
9186 static void iwl_pci_remove(struct pci_dev
*pdev
)
9188 struct iwl_priv
*priv
= pci_get_drvdata(pdev
);
9189 struct list_head
*p
, *q
;
9195 IWL_DEBUG_INFO("*** UNLOAD DRIVER ***\n");
9197 set_bit(STATUS_EXIT_PENDING
, &priv
->status
);
9201 /* Free MAC hash list for ADHOC */
9202 for (i
= 0; i
< IWL_IBSS_MAC_HASH_SIZE
; i
++) {
9203 list_for_each_safe(p
, q
, &priv
->ibss_mac_hash
[i
]) {
9205 kfree(list_entry(p
, struct iwl_ibss_seq
, list
));
9209 sysfs_remove_group(&pdev
->dev
.kobj
, &iwl_attribute_group
);
9211 iwl_dealloc_ucode_pci(priv
);
9214 iwl_rx_queue_free(priv
, &priv
->rxq
);
9215 iwl_hw_txq_ctx_free(priv
);
9217 iwl_unset_hw_setting(priv
);
9218 iwl_clear_stations_table(priv
);
9220 if (priv
->mac80211_registered
) {
9221 ieee80211_unregister_hw(priv
->hw
);
9222 iwl_rate_control_unregister(priv
->hw
);
9225 /*netif_stop_queue(dev); */
9226 flush_workqueue(priv
->workqueue
);
9228 /* ieee80211_unregister_hw calls iwl_mac_stop, which flushes
9229 * priv->workqueue... so we can't take down the workqueue
9231 destroy_workqueue(priv
->workqueue
);
9232 priv
->workqueue
= NULL
;
9234 free_irq(pdev
->irq
, priv
);
9235 pci_disable_msi(pdev
);
9236 pci_iounmap(pdev
, priv
->hw_base
);
9237 pci_release_regions(pdev
);
9238 pci_disable_device(pdev
);
9239 pci_set_drvdata(pdev
, NULL
);
9241 kfree(priv
->channel_info
);
9243 kfree(priv
->ieee_channels
);
9244 kfree(priv
->ieee_rates
);
9246 if (priv
->ibss_beacon
)
9247 dev_kfree_skb(priv
->ibss_beacon
);
9249 ieee80211_free_hw(priv
->hw
);
9254 static int iwl_pci_suspend(struct pci_dev
*pdev
, pm_message_t state
)
9256 struct iwl_priv
*priv
= pci_get_drvdata(pdev
);
9258 set_bit(STATUS_IN_SUSPEND
, &priv
->status
);
9260 /* Take down the device; powers it off, etc. */
9263 if (priv
->mac80211_registered
)
9264 ieee80211_stop_queues(priv
->hw
);
9266 pci_save_state(pdev
);
9267 pci_disable_device(pdev
);
9268 pci_set_power_state(pdev
, PCI_D3hot
);
9273 static void iwl_resume(struct iwl_priv
*priv
)
9275 unsigned long flags
;
9277 /* The following it a temporary work around due to the
9278 * suspend / resume not fully initializing the NIC correctly.
9279 * Without all of the following, resume will not attempt to take
9280 * down the NIC (it shouldn't really need to) and will just try
9281 * and bring the NIC back up. However that fails during the
9282 * ucode verification process. This then causes iwl_down to be
9283 * called *after* iwl_hw_nic_init() has succeeded -- which
9284 * then lets the next init sequence succeed. So, we've
9285 * replicated all of that NIC init code here... */
9287 iwl_write32(priv
, CSR_INT
, 0xFFFFFFFF);
9289 iwl_hw_nic_init(priv
);
9291 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
, CSR_UCODE_SW_BIT_RFKILL
);
9292 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
,
9293 CSR_UCODE_DRV_GP1_BIT_CMD_BLOCKED
);
9294 iwl_write32(priv
, CSR_INT
, 0xFFFFFFFF);
9295 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
, CSR_UCODE_SW_BIT_RFKILL
);
9296 iwl_write32(priv
, CSR_UCODE_DRV_GP1_CLR
, CSR_UCODE_SW_BIT_RFKILL
);
9298 /* tell the device to stop sending interrupts */
9299 iwl_disable_interrupts(priv
);
9301 spin_lock_irqsave(&priv
->lock
, flags
);
9302 iwl_clear_bit(priv
, CSR_GP_CNTRL
, CSR_GP_CNTRL_REG_FLAG_MAC_ACCESS_REQ
);
9304 if (!iwl_grab_nic_access(priv
)) {
9305 iwl_write_prph(priv
, APMG_CLK_DIS_REG
,
9306 APMG_CLK_VAL_DMA_CLK_RQT
);
9307 iwl_release_nic_access(priv
);
9309 spin_unlock_irqrestore(&priv
->lock
, flags
);
9313 iwl_hw_nic_reset(priv
);
9315 /* Bring the device back up */
9316 clear_bit(STATUS_IN_SUSPEND
, &priv
->status
);
9317 queue_work(priv
->workqueue
, &priv
->up
);
9320 static int iwl_pci_resume(struct pci_dev
*pdev
)
9322 struct iwl_priv
*priv
= pci_get_drvdata(pdev
);
9325 printk(KERN_INFO
"Coming out of suspend...\n");
9327 pci_set_power_state(pdev
, PCI_D0
);
9328 err
= pci_enable_device(pdev
);
9329 pci_restore_state(pdev
);
9332 * Suspend/Resume resets the PCI configuration space, so we have to
9333 * re-disable the RETRY_TIMEOUT register (0x41) to keep PCI Tx retries
9334 * from interfering with C3 CPU state. pci_restore_state won't help
9335 * here since it only restores the first 64 bytes pci config header.
9337 pci_write_config_byte(pdev
, 0x41, 0x00);
9344 #endif /* CONFIG_PM */
9346 /*****************************************************************************
9348 * driver and module entry point
9350 *****************************************************************************/
9352 static struct pci_driver iwl_driver
= {
9354 .id_table
= iwl_hw_card_ids
,
9355 .probe
= iwl_pci_probe
,
9356 .remove
= __devexit_p(iwl_pci_remove
),
9358 .suspend
= iwl_pci_suspend
,
9359 .resume
= iwl_pci_resume
,
9363 static int __init
iwl_init(void)
9367 printk(KERN_INFO DRV_NAME
": " DRV_DESCRIPTION
", " DRV_VERSION
"\n");
9368 printk(KERN_INFO DRV_NAME
": " DRV_COPYRIGHT
"\n");
9369 ret
= pci_register_driver(&iwl_driver
);
9371 IWL_ERROR("Unable to initialize PCI module\n");
9374 #ifdef CONFIG_IWL4965_DEBUG
9375 ret
= driver_create_file(&iwl_driver
.driver
, &driver_attr_debug_level
);
9377 IWL_ERROR("Unable to create driver sysfs file\n");
9378 pci_unregister_driver(&iwl_driver
);
9386 static void __exit
iwl_exit(void)
9388 #ifdef CONFIG_IWL4965_DEBUG
9389 driver_remove_file(&iwl_driver
.driver
, &driver_attr_debug_level
);
9391 pci_unregister_driver(&iwl_driver
);
9394 module_param_named(antenna
, iwl_param_antenna
, int, 0444);
9395 MODULE_PARM_DESC(antenna
, "select antenna (1=Main, 2=Aux, default 0 [both])");
9396 module_param_named(disable
, iwl_param_disable
, int, 0444);
9397 MODULE_PARM_DESC(disable
, "manually disable the radio (default 0 [radio on])");
9398 module_param_named(hwcrypto
, iwl_param_hwcrypto
, int, 0444);
9399 MODULE_PARM_DESC(hwcrypto
,
9400 "using hardware crypto engine (default 0 [software])\n");
9401 module_param_named(debug
, iwl_param_debug
, int, 0444);
9402 MODULE_PARM_DESC(debug
, "debug output mask");
9403 module_param_named(disable_hw_scan
, iwl_param_disable_hw_scan
, int, 0444);
9404 MODULE_PARM_DESC(disable_hw_scan
, "disable hardware scanning (default 0)");
9406 module_param_named(queues_num
, iwl_param_queues_num
, int, 0444);
9407 MODULE_PARM_DESC(queues_num
, "number of hw queues.");
9410 module_param_named(qos_enable
, iwl_param_qos_enable
, int, 0444);
9411 MODULE_PARM_DESC(qos_enable
, "enable all QoS functionality");
9413 module_exit(iwl_exit
);
9414 module_init(iwl_init
);