1 /******************************************************************************
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
8 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
10 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of version 2 of the GNU General Public License as
14 * published by the Free Software Foundation.
16 * This program is distributed in the hope that it will be useful, but
17 * WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
19 * General Public License for more details.
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, write to the Free Software
23 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
26 * The full GNU General Public License is included in this distribution
27 * in the file called COPYING.
29 * Contact Information:
30 * Intel Linux Wireless <linuxwifi@intel.com>
31 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
35 * Copyright(c) 2012 - 2014 Intel Corporation. All rights reserved.
36 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
37 * Copyright(c) 2016 - 2017 Intel Deutschland GmbH
38 * All rights reserved.
40 * Redistribution and use in source and binary forms, with or without
41 * modification, are permitted provided that the following conditions
44 * * Redistributions of source code must retain the above copyright
45 * notice, this list of conditions and the following disclaimer.
46 * * Redistributions in binary form must reproduce the above copyright
47 * notice, this list of conditions and the following disclaimer in
48 * the documentation and/or other materials provided with the
50 * * Neither the name Intel Corporation nor the names of its
51 * contributors may be used to endorse or promote products derived
52 * from this software without specific prior written permission.
54 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
55 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
56 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
57 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
58 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
59 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
60 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
61 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
62 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
63 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
64 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
66 *****************************************************************************/
67 #include <linux/firmware.h>
68 #include <linux/rtnetlink.h>
69 #include "iwl-trans.h"
72 #include "iwl-eeprom-parse.h"
73 #include "iwl-eeprom-read.h"
74 #include "iwl-nvm-parse.h"
78 /* Default NVM size to read */
79 #define IWL_NVM_DEFAULT_CHUNK_SIZE (2*1024)
80 #define IWL_MAX_NVM_SECTION_SIZE 0x1b58
81 #define IWL_MAX_EXT_NVM_SECTION_SIZE 0x1ffc
83 #define NVM_WRITE_OPCODE 1
84 #define NVM_READ_OPCODE 0
86 /* load nvm chunk response */
88 READ_NVM_CHUNK_SUCCEED
= 0,
89 READ_NVM_CHUNK_NOT_VALID_ADDRESS
= 1
93 * prepare the NVM host command w/ the pointers to the nvm buffer
96 static int iwl_nvm_write_chunk(struct iwl_mvm
*mvm
, u16 section
,
97 u16 offset
, u16 length
, const u8
*data
)
99 struct iwl_nvm_access_cmd nvm_access_cmd
= {
100 .offset
= cpu_to_le16(offset
),
101 .length
= cpu_to_le16(length
),
102 .type
= cpu_to_le16(section
),
103 .op_code
= NVM_WRITE_OPCODE
,
105 struct iwl_host_cmd cmd
= {
106 .id
= NVM_ACCESS_CMD
,
107 .len
= { sizeof(struct iwl_nvm_access_cmd
), length
},
108 .flags
= CMD_WANT_SKB
| CMD_SEND_IN_RFKILL
,
109 .data
= { &nvm_access_cmd
, data
},
110 /* data may come from vmalloc, so use _DUP */
111 .dataflags
= { 0, IWL_HCMD_DFL_DUP
},
113 struct iwl_rx_packet
*pkt
;
114 struct iwl_nvm_access_resp
*nvm_resp
;
117 ret
= iwl_mvm_send_cmd(mvm
, &cmd
);
122 /* Extract & check NVM write response */
123 nvm_resp
= (void *)pkt
->data
;
124 if (le16_to_cpu(nvm_resp
->status
) != READ_NVM_CHUNK_SUCCEED
) {
126 "NVM access write command failed for section %u (status = 0x%x)\n",
127 section
, le16_to_cpu(nvm_resp
->status
));
135 static int iwl_nvm_read_chunk(struct iwl_mvm
*mvm
, u16 section
,
136 u16 offset
, u16 length
, u8
*data
)
138 struct iwl_nvm_access_cmd nvm_access_cmd
= {
139 .offset
= cpu_to_le16(offset
),
140 .length
= cpu_to_le16(length
),
141 .type
= cpu_to_le16(section
),
142 .op_code
= NVM_READ_OPCODE
,
144 struct iwl_nvm_access_resp
*nvm_resp
;
145 struct iwl_rx_packet
*pkt
;
146 struct iwl_host_cmd cmd
= {
147 .id
= NVM_ACCESS_CMD
,
148 .flags
= CMD_WANT_SKB
| CMD_SEND_IN_RFKILL
,
149 .data
= { &nvm_access_cmd
, },
151 int ret
, bytes_read
, offset_read
;
154 cmd
.len
[0] = sizeof(struct iwl_nvm_access_cmd
);
156 ret
= iwl_mvm_send_cmd(mvm
, &cmd
);
162 /* Extract NVM response */
163 nvm_resp
= (void *)pkt
->data
;
164 ret
= le16_to_cpu(nvm_resp
->status
);
165 bytes_read
= le16_to_cpu(nvm_resp
->length
);
166 offset_read
= le16_to_cpu(nvm_resp
->offset
);
167 resp_data
= nvm_resp
->data
;
170 (ret
== READ_NVM_CHUNK_NOT_VALID_ADDRESS
)) {
172 * meaning of NOT_VALID_ADDRESS:
173 * driver try to read chunk from address that is
174 * multiple of 2K and got an error since addr is empty.
175 * meaning of (offset != 0): driver already
176 * read valid data from another chunk so this case
179 IWL_DEBUG_EEPROM(mvm
->trans
->dev
,
180 "NVM access command failed on offset 0x%x since that section size is multiple 2K\n",
184 IWL_DEBUG_EEPROM(mvm
->trans
->dev
,
185 "NVM access command failed with status %d (device: %s)\n",
186 ret
, mvm
->cfg
->name
);
192 if (offset_read
!= offset
) {
193 IWL_ERR(mvm
, "NVM ACCESS response with invalid offset %d\n",
199 /* Write data to NVM */
200 memcpy(data
+ offset
, resp_data
, bytes_read
);
208 static int iwl_nvm_write_section(struct iwl_mvm
*mvm
, u16 section
,
209 const u8
*data
, u16 length
)
213 /* copy data in chunks of 2k (and remainder if any) */
215 while (offset
< length
) {
218 chunk_size
= min(IWL_NVM_DEFAULT_CHUNK_SIZE
,
221 ret
= iwl_nvm_write_chunk(mvm
, section
, offset
,
222 chunk_size
, data
+ offset
);
226 offset
+= chunk_size
;
232 static void iwl_mvm_nvm_fixups(struct iwl_mvm
*mvm
, unsigned int section
,
233 u8
*data
, unsigned int len
)
235 #define IWL_4165_DEVICE_ID 0x5501
236 #define NVM_SKU_CAP_MIMO_DISABLE BIT(5)
238 if (section
== NVM_SECTION_TYPE_PHY_SKU
&&
239 mvm
->trans
->hw_id
== IWL_4165_DEVICE_ID
&& data
&& len
>= 5 &&
240 (data
[4] & NVM_SKU_CAP_MIMO_DISABLE
))
241 /* OTP 0x52 bug work around: it's a 1x1 device */
242 data
[3] = ANT_B
| (ANT_B
<< 4);
246 * Reads an NVM section completely.
247 * NICs prior to 7000 family doesn't have a real NVM, but just read
248 * section 0 which is the EEPROM. Because the EEPROM reading is unlimited
249 * by uCode, we need to manually check in this case that we don't
250 * overflow and try to read more than the EEPROM size.
251 * For 7000 family NICs, we supply the maximal size we can read, and
252 * the uCode fills the response with as much data as we can,
253 * without overflowing, so no check is needed.
255 static int iwl_nvm_read_section(struct iwl_mvm
*mvm
, u16 section
,
256 u8
*data
, u32 size_read
)
258 u16 length
, offset
= 0;
261 /* Set nvm section read length */
262 length
= IWL_NVM_DEFAULT_CHUNK_SIZE
;
266 /* Read the NVM until exhausted (reading less than requested) */
267 while (ret
== length
) {
268 /* Check no memory assumptions fail and cause an overflow */
269 if ((size_read
+ offset
+ length
) >
270 mvm
->cfg
->base_params
->eeprom_size
) {
271 IWL_ERR(mvm
, "EEPROM size is too small for NVM\n");
275 ret
= iwl_nvm_read_chunk(mvm
, section
, offset
, length
, data
);
277 IWL_DEBUG_EEPROM(mvm
->trans
->dev
,
278 "Cannot read NVM from section %d offset %d, length %d\n",
279 section
, offset
, length
);
285 iwl_mvm_nvm_fixups(mvm
, section
, data
, offset
);
287 IWL_DEBUG_EEPROM(mvm
->trans
->dev
,
288 "NVM section %d read completed\n", section
);
292 static struct iwl_nvm_data
*
293 iwl_parse_nvm_sections(struct iwl_mvm
*mvm
)
295 struct iwl_nvm_section
*sections
= mvm
->nvm_sections
;
297 const __le16
*sw
, *calib
, *regulatory
, *mac_override
, *phy_sku
;
301 /* Checking for required sections */
302 if (mvm
->trans
->cfg
->nvm_type
== IWL_NVM
) {
303 if (!mvm
->nvm_sections
[NVM_SECTION_TYPE_SW
].data
||
304 !mvm
->nvm_sections
[mvm
->cfg
->nvm_hw_section_num
].data
) {
305 IWL_ERR(mvm
, "Can't parse empty OTP/NVM sections\n");
309 if (mvm
->trans
->cfg
->nvm_type
== IWL_NVM_SDP
)
310 regulatory_type
= NVM_SECTION_TYPE_REGULATORY_SDP
;
312 regulatory_type
= NVM_SECTION_TYPE_REGULATORY
;
314 /* SW and REGULATORY sections are mandatory */
315 if (!mvm
->nvm_sections
[NVM_SECTION_TYPE_SW
].data
||
316 !mvm
->nvm_sections
[regulatory_type
].data
) {
318 "Can't parse empty family 8000 OTP/NVM sections\n");
321 /* MAC_OVERRIDE or at least HW section must exist */
322 if (!mvm
->nvm_sections
[mvm
->cfg
->nvm_hw_section_num
].data
&&
323 !mvm
->nvm_sections
[NVM_SECTION_TYPE_MAC_OVERRIDE
].data
) {
325 "Can't parse mac_address, empty sections\n");
329 /* PHY_SKU section is mandatory in B0 */
330 if (!mvm
->nvm_sections
[NVM_SECTION_TYPE_PHY_SKU
].data
) {
332 "Can't parse phy_sku in B0, empty sections\n");
337 hw
= (const __be16
*)sections
[mvm
->cfg
->nvm_hw_section_num
].data
;
338 sw
= (const __le16
*)sections
[NVM_SECTION_TYPE_SW
].data
;
339 calib
= (const __le16
*)sections
[NVM_SECTION_TYPE_CALIBRATION
].data
;
341 (const __le16
*)sections
[NVM_SECTION_TYPE_MAC_OVERRIDE
].data
;
342 phy_sku
= (const __le16
*)sections
[NVM_SECTION_TYPE_PHY_SKU
].data
;
344 regulatory
= mvm
->trans
->cfg
->nvm_type
== IWL_NVM_SDP
?
345 (const __le16
*)sections
[NVM_SECTION_TYPE_REGULATORY_SDP
].data
:
346 (const __le16
*)sections
[NVM_SECTION_TYPE_REGULATORY
].data
;
348 lar_enabled
= !iwlwifi_mod_params
.lar_disable
&&
349 fw_has_capa(&mvm
->fw
->ucode_capa
,
350 IWL_UCODE_TLV_CAPA_LAR_SUPPORT
);
352 return iwl_parse_nvm_data(mvm
->trans
, mvm
->cfg
, hw
, sw
, calib
,
353 regulatory
, mac_override
, phy_sku
,
354 mvm
->fw
->valid_tx_ant
, mvm
->fw
->valid_rx_ant
,
358 #define MAX_NVM_FILE_LEN 16384
361 * Reads external NVM from a file into mvm->nvm_sections
363 * HOW TO CREATE THE NVM FILE FORMAT:
364 * ------------------------------
365 * 1. create hex file, format:
370 * rev - 6 bit (word1)
371 * len - 10 bit (word1)
373 * rsv - 12 bit (word2)
375 * 2. flip 8bits with 8 bits per line to get the right NVM file format
377 * 3. create binary file from the hex file
379 * 4. save as "iNVM_xxx.bin" under /lib/firmware
381 int iwl_mvm_read_external_nvm(struct iwl_mvm
*mvm
)
383 int ret
, section_size
;
385 const struct firmware
*fw_entry
;
393 int max_section_size
;
394 const __le32
*dword_buff
;
396 #define NVM_WORD1_LEN(x) (8 * (x & 0x03FF))
397 #define NVM_WORD2_ID(x) (x >> 12)
398 #define EXT_NVM_WORD2_LEN(x) (2 * (((x) & 0xFF) << 8 | (x) >> 8))
399 #define EXT_NVM_WORD1_ID(x) ((x) >> 4)
400 #define NVM_HEADER_0 (0x2A504C54)
401 #define NVM_HEADER_1 (0x4E564D2A)
402 #define NVM_HEADER_SIZE (4 * sizeof(u32))
404 IWL_DEBUG_EEPROM(mvm
->trans
->dev
, "Read from external NVM\n");
406 /* Maximal size depends on NVM version */
407 if (mvm
->trans
->cfg
->nvm_type
!= IWL_NVM_EXT
)
408 max_section_size
= IWL_MAX_NVM_SECTION_SIZE
;
410 max_section_size
= IWL_MAX_EXT_NVM_SECTION_SIZE
;
413 * Obtain NVM image via request_firmware. Since we already used
414 * request_firmware_nowait() for the firmware binary load and only
415 * get here after that we assume the NVM request can be satisfied
418 ret
= request_firmware(&fw_entry
, mvm
->nvm_file_name
,
421 IWL_ERR(mvm
, "ERROR: %s isn't available %d\n",
422 mvm
->nvm_file_name
, ret
);
426 IWL_INFO(mvm
, "Loaded NVM file %s (%zu bytes)\n",
427 mvm
->nvm_file_name
, fw_entry
->size
);
429 if (fw_entry
->size
> MAX_NVM_FILE_LEN
) {
430 IWL_ERR(mvm
, "NVM file too large\n");
435 eof
= fw_entry
->data
+ fw_entry
->size
;
436 dword_buff
= (__le32
*)fw_entry
->data
;
438 /* some NVM file will contain a header.
439 * The header is identified by 2 dwords header as follow:
440 * dword[0] = 0x2A504C54
441 * dword[1] = 0x4E564D2A
443 * This header must be skipped when providing the NVM data to the FW.
445 if (fw_entry
->size
> NVM_HEADER_SIZE
&&
446 dword_buff
[0] == cpu_to_le32(NVM_HEADER_0
) &&
447 dword_buff
[1] == cpu_to_le32(NVM_HEADER_1
)) {
448 file_sec
= (void *)(fw_entry
->data
+ NVM_HEADER_SIZE
);
449 IWL_INFO(mvm
, "NVM Version %08X\n", le32_to_cpu(dword_buff
[2]));
450 IWL_INFO(mvm
, "NVM Manufacturing date %08X\n",
451 le32_to_cpu(dword_buff
[3]));
453 /* nvm file validation, dword_buff[2] holds the file version */
454 if (mvm
->trans
->cfg
->device_family
== IWL_DEVICE_FAMILY_8000
&&
455 CSR_HW_REV_STEP(mvm
->trans
->hw_rev
) == SILICON_C_STEP
&&
456 le32_to_cpu(dword_buff
[2]) < 0xE4A) {
461 file_sec
= (void *)fw_entry
->data
;
465 if (file_sec
->data
> eof
) {
467 "ERROR - NVM file too short for section header\n");
472 /* check for EOF marker */
473 if (!file_sec
->word1
&& !file_sec
->word2
) {
478 if (mvm
->trans
->cfg
->nvm_type
!= IWL_NVM_EXT
) {
480 2 * NVM_WORD1_LEN(le16_to_cpu(file_sec
->word1
));
481 section_id
= NVM_WORD2_ID(le16_to_cpu(file_sec
->word2
));
483 section_size
= 2 * EXT_NVM_WORD2_LEN(
484 le16_to_cpu(file_sec
->word2
));
485 section_id
= EXT_NVM_WORD1_ID(
486 le16_to_cpu(file_sec
->word1
));
489 if (section_size
> max_section_size
) {
490 IWL_ERR(mvm
, "ERROR - section too large (%d)\n",
497 IWL_ERR(mvm
, "ERROR - section empty\n");
502 if (file_sec
->data
+ section_size
> eof
) {
504 "ERROR - NVM file too short for section (%d bytes)\n",
510 if (WARN(section_id
>= NVM_MAX_NUM_SECTIONS
,
511 "Invalid NVM section ID %d\n", section_id
)) {
516 temp
= kmemdup(file_sec
->data
, section_size
, GFP_KERNEL
);
522 iwl_mvm_nvm_fixups(mvm
, section_id
, temp
, section_size
);
524 kfree(mvm
->nvm_sections
[section_id
].data
);
525 mvm
->nvm_sections
[section_id
].data
= temp
;
526 mvm
->nvm_sections
[section_id
].length
= section_size
;
528 /* advance to the next section */
529 file_sec
= (void *)(file_sec
->data
+ section_size
);
532 release_firmware(fw_entry
);
536 /* Loads the NVM data stored in mvm->nvm_sections into the NIC */
537 int iwl_mvm_load_nvm_to_nic(struct iwl_mvm
*mvm
)
540 struct iwl_nvm_section
*sections
= mvm
->nvm_sections
;
542 IWL_DEBUG_EEPROM(mvm
->trans
->dev
, "'Write to NVM\n");
544 for (i
= 0; i
< ARRAY_SIZE(mvm
->nvm_sections
); i
++) {
545 if (!mvm
->nvm_sections
[i
].data
|| !mvm
->nvm_sections
[i
].length
)
547 ret
= iwl_nvm_write_section(mvm
, i
, sections
[i
].data
,
550 IWL_ERR(mvm
, "iwl_mvm_send_cmd failed: %d\n", ret
);
557 int iwl_nvm_init(struct iwl_mvm
*mvm
)
561 u8
*nvm_buffer
, *temp
;
562 const char *nvm_file_C
= mvm
->cfg
->default_nvm_file_C_step
;
564 if (WARN_ON_ONCE(mvm
->cfg
->nvm_hw_section_num
>= NVM_MAX_NUM_SECTIONS
))
567 /* load NVM values from nic */
568 /* Read From FW NVM */
569 IWL_DEBUG_EEPROM(mvm
->trans
->dev
, "Read from NVM\n");
571 nvm_buffer
= kmalloc(mvm
->cfg
->base_params
->eeprom_size
,
575 for (section
= 0; section
< NVM_MAX_NUM_SECTIONS
; section
++) {
576 /* we override the constness for initial read */
577 ret
= iwl_nvm_read_section(mvm
, section
, nvm_buffer
,
582 temp
= kmemdup(nvm_buffer
, ret
, GFP_KERNEL
);
588 iwl_mvm_nvm_fixups(mvm
, section
, temp
, ret
);
590 mvm
->nvm_sections
[section
].data
= temp
;
591 mvm
->nvm_sections
[section
].length
= ret
;
593 #ifdef CONFIG_IWLWIFI_DEBUGFS
595 case NVM_SECTION_TYPE_SW
:
596 mvm
->nvm_sw_blob
.data
= temp
;
597 mvm
->nvm_sw_blob
.size
= ret
;
599 case NVM_SECTION_TYPE_CALIBRATION
:
600 mvm
->nvm_calib_blob
.data
= temp
;
601 mvm
->nvm_calib_blob
.size
= ret
;
603 case NVM_SECTION_TYPE_PRODUCTION
:
604 mvm
->nvm_prod_blob
.data
= temp
;
605 mvm
->nvm_prod_blob
.size
= ret
;
607 case NVM_SECTION_TYPE_PHY_SKU
:
608 mvm
->nvm_phy_sku_blob
.data
= temp
;
609 mvm
->nvm_phy_sku_blob
.size
= ret
;
612 if (section
== mvm
->cfg
->nvm_hw_section_num
) {
613 mvm
->nvm_hw_blob
.data
= temp
;
614 mvm
->nvm_hw_blob
.size
= ret
;
621 IWL_ERR(mvm
, "OTP is blank\n");
624 /* Only if PNVM selected in the mod param - load external NVM */
625 if (mvm
->nvm_file_name
) {
626 /* read External NVM file from the mod param */
627 ret
= iwl_mvm_read_external_nvm(mvm
);
629 mvm
->nvm_file_name
= nvm_file_C
;
631 if ((ret
== -EFAULT
|| ret
== -ENOENT
) &&
632 mvm
->nvm_file_name
) {
633 /* in case nvm file was failed try again */
634 ret
= iwl_mvm_read_external_nvm(mvm
);
643 /* parse the relevant nvm sections */
644 mvm
->nvm_data
= iwl_parse_nvm_sections(mvm
);
647 IWL_DEBUG_EEPROM(mvm
->trans
->dev
, "nvm version = %x\n",
648 mvm
->nvm_data
->nvm_version
);
653 struct iwl_mcc_update_resp
*
654 iwl_mvm_update_mcc(struct iwl_mvm
*mvm
, const char *alpha2
,
655 enum iwl_mcc_source src_id
)
657 struct iwl_mcc_update_cmd mcc_update_cmd
= {
658 .mcc
= cpu_to_le16(alpha2
[0] << 8 | alpha2
[1]),
659 .source_id
= (u8
)src_id
,
661 struct iwl_mcc_update_resp
*resp_cp
;
662 struct iwl_rx_packet
*pkt
;
663 struct iwl_host_cmd cmd
= {
664 .id
= MCC_UPDATE_CMD
,
665 .flags
= CMD_WANT_SKB
,
666 .data
= { &mcc_update_cmd
},
671 int resp_len
, n_channels
;
673 bool resp_v2
= fw_has_capa(&mvm
->fw
->ucode_capa
,
674 IWL_UCODE_TLV_CAPA_LAR_SUPPORT_V2
);
676 if (WARN_ON_ONCE(!iwl_mvm_is_lar_supported(mvm
)))
677 return ERR_PTR(-EOPNOTSUPP
);
679 cmd
.len
[0] = sizeof(struct iwl_mcc_update_cmd
);
681 cmd
.len
[0] = sizeof(struct iwl_mcc_update_cmd_v1
);
683 IWL_DEBUG_LAR(mvm
, "send MCC update to FW with '%c%c' src = %d\n",
684 alpha2
[0], alpha2
[1], src_id
);
686 ret
= iwl_mvm_send_cmd(mvm
, &cmd
);
692 /* Extract MCC response */
694 struct iwl_mcc_update_resp
*mcc_resp
= (void *)pkt
->data
;
696 n_channels
= __le32_to_cpu(mcc_resp
->n_channels
);
697 resp_len
= sizeof(struct iwl_mcc_update_resp
) +
698 n_channels
* sizeof(__le32
);
699 resp_cp
= kmemdup(mcc_resp
, resp_len
, GFP_KERNEL
);
701 resp_cp
= ERR_PTR(-ENOMEM
);
705 struct iwl_mcc_update_resp_v1
*mcc_resp_v1
= (void *)pkt
->data
;
707 n_channels
= __le32_to_cpu(mcc_resp_v1
->n_channels
);
708 resp_len
= sizeof(struct iwl_mcc_update_resp
) +
709 n_channels
* sizeof(__le32
);
710 resp_cp
= kzalloc(resp_len
, GFP_KERNEL
);
712 resp_cp
= ERR_PTR(-ENOMEM
);
716 resp_cp
->status
= mcc_resp_v1
->status
;
717 resp_cp
->mcc
= mcc_resp_v1
->mcc
;
718 resp_cp
->cap
= mcc_resp_v1
->cap
;
719 resp_cp
->source_id
= mcc_resp_v1
->source_id
;
720 resp_cp
->n_channels
= mcc_resp_v1
->n_channels
;
721 memcpy(resp_cp
->channels
, mcc_resp_v1
->channels
,
722 n_channels
* sizeof(__le32
));
725 status
= le32_to_cpu(resp_cp
->status
);
727 mcc
= le16_to_cpu(resp_cp
->mcc
);
729 /* W/A for a FW/NVM issue - returns 0x00 for the world domain */
731 mcc
= 0x3030; /* "00" - world */
732 resp_cp
->mcc
= cpu_to_le16(mcc
);
736 "MCC response status: 0x%x. new MCC: 0x%x ('%c%c') n_chans: %d\n",
737 status
, mcc
, mcc
>> 8, mcc
& 0xff, n_channels
);
744 int iwl_mvm_init_mcc(struct iwl_mvm
*mvm
)
749 struct ieee80211_regdomain
*regd
;
752 if (mvm
->cfg
->nvm_type
== IWL_NVM_EXT
) {
753 tlv_lar
= fw_has_capa(&mvm
->fw
->ucode_capa
,
754 IWL_UCODE_TLV_CAPA_LAR_SUPPORT
);
755 nvm_lar
= mvm
->nvm_data
->lar_enabled
;
756 if (tlv_lar
!= nvm_lar
)
758 "Conflict between TLV & NVM regarding enabling LAR (TLV = %s NVM =%s)\n",
759 tlv_lar
? "enabled" : "disabled",
760 nvm_lar
? "enabled" : "disabled");
763 if (!iwl_mvm_is_lar_supported(mvm
))
767 * try to replay the last set MCC to FW. If it doesn't exist,
768 * queue an update to cfg80211 to retrieve the default alpha2 from FW.
770 retval
= iwl_mvm_init_fw_regd(mvm
);
771 if (retval
!= -ENOENT
)
775 * Driver regulatory hint for initial update, this also informs the
776 * firmware we support wifi location updates.
777 * Disallow scans that might crash the FW while the LAR regdomain
780 mvm
->lar_regdom_set
= false;
782 regd
= iwl_mvm_get_current_regdomain(mvm
, NULL
);
783 if (IS_ERR_OR_NULL(regd
))
786 if (iwl_mvm_is_wifi_mcc_supported(mvm
) &&
787 !iwl_acpi_get_mcc(mvm
->dev
, mcc
)) {
789 regd
= iwl_mvm_get_regdomain(mvm
->hw
->wiphy
, mcc
,
790 MCC_SOURCE_BIOS
, NULL
);
791 if (IS_ERR_OR_NULL(regd
))
795 retval
= regulatory_set_wiphy_regd_sync_rtnl(mvm
->hw
->wiphy
, regd
);
800 void iwl_mvm_rx_chub_update_mcc(struct iwl_mvm
*mvm
,
801 struct iwl_rx_cmd_buffer
*rxb
)
803 struct iwl_rx_packet
*pkt
= rxb_addr(rxb
);
804 struct iwl_mcc_chub_notif
*notif
= (void *)pkt
->data
;
805 enum iwl_mcc_source src
;
807 struct ieee80211_regdomain
*regd
;
809 lockdep_assert_held(&mvm
->mutex
);
811 if (iwl_mvm_is_vif_assoc(mvm
) && notif
->source_id
== MCC_SOURCE_WIFI
) {
812 IWL_DEBUG_LAR(mvm
, "Ignore mcc update while associated\n");
816 if (WARN_ON_ONCE(!iwl_mvm_is_lar_supported(mvm
)))
819 mcc
[0] = le16_to_cpu(notif
->mcc
) >> 8;
820 mcc
[1] = le16_to_cpu(notif
->mcc
) & 0xff;
822 src
= notif
->source_id
;
825 "RX: received chub update mcc cmd (mcc '%s' src %d)\n",
827 regd
= iwl_mvm_get_regdomain(mvm
->hw
->wiphy
, mcc
, src
, NULL
);
828 if (IS_ERR_OR_NULL(regd
))
831 regulatory_set_wiphy_regd(mvm
->hw
->wiphy
, regd
);