]> git.proxmox.com Git - mirror_ubuntu-hirsute-kernel.git/blob - drivers/net/ethernet/emulex/benet/be_cmds.c
Revert "be2net: disable bh with spin_lock in be_process_mcc"
[mirror_ubuntu-hirsute-kernel.git] / drivers / net / ethernet / emulex / benet / be_cmds.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Copyright (C) 2005 - 2016 Broadcom
4 * All rights reserved.
5 *
6 * Contact Information:
7 * linux-drivers@emulex.com
8 *
9 * Emulex
10 * 3333 Susan Street
11 * Costa Mesa, CA 92626
12 */
13
14 #include <linux/module.h>
15 #include "be.h"
16 #include "be_cmds.h"
17
18 const char * const be_misconfig_evt_port_state[] = {
19 "Physical Link is functional",
20 "Optics faulted/incorrectly installed/not installed - Reseat optics. If issue not resolved, replace.",
21 "Optics of two types installed – Remove one optic or install matching pair of optics.",
22 "Incompatible optics – Replace with compatible optics for card to function.",
23 "Unqualified optics – Replace with Avago optics for Warranty and Technical Support.",
24 "Uncertified optics – Replace with Avago-certified optics to enable link operation."
25 };
26
27 static char *be_port_misconfig_evt_severity[] = {
28 "KERN_WARN",
29 "KERN_INFO",
30 "KERN_ERR",
31 "KERN_WARN"
32 };
33
34 static char *phy_state_oper_desc[] = {
35 "Link is non-operational",
36 "Link is operational",
37 ""
38 };
39
40 static struct be_cmd_priv_map cmd_priv_map[] = {
41 {
42 OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG,
43 CMD_SUBSYSTEM_ETH,
44 BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
45 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
46 },
47 {
48 OPCODE_COMMON_GET_FLOW_CONTROL,
49 CMD_SUBSYSTEM_COMMON,
50 BE_PRIV_LNKQUERY | BE_PRIV_VHADM |
51 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
52 },
53 {
54 OPCODE_COMMON_SET_FLOW_CONTROL,
55 CMD_SUBSYSTEM_COMMON,
56 BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
57 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
58 },
59 {
60 OPCODE_ETH_GET_PPORT_STATS,
61 CMD_SUBSYSTEM_ETH,
62 BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
63 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
64 },
65 {
66 OPCODE_COMMON_GET_PHY_DETAILS,
67 CMD_SUBSYSTEM_COMMON,
68 BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
69 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
70 },
71 {
72 OPCODE_LOWLEVEL_HOST_DDR_DMA,
73 CMD_SUBSYSTEM_LOWLEVEL,
74 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
75 },
76 {
77 OPCODE_LOWLEVEL_LOOPBACK_TEST,
78 CMD_SUBSYSTEM_LOWLEVEL,
79 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
80 },
81 {
82 OPCODE_LOWLEVEL_SET_LOOPBACK_MODE,
83 CMD_SUBSYSTEM_LOWLEVEL,
84 BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
85 },
86 {
87 OPCODE_COMMON_SET_HSW_CONFIG,
88 CMD_SUBSYSTEM_COMMON,
89 BE_PRIV_DEVCFG | BE_PRIV_VHADM |
90 BE_PRIV_DEVSEC
91 },
92 {
93 OPCODE_COMMON_GET_EXT_FAT_CAPABILITIES,
94 CMD_SUBSYSTEM_COMMON,
95 BE_PRIV_DEVCFG
96 }
97 };
98
99 static bool be_cmd_allowed(struct be_adapter *adapter, u8 opcode, u8 subsystem)
100 {
101 int i;
102 int num_entries = ARRAY_SIZE(cmd_priv_map);
103 u32 cmd_privileges = adapter->cmd_privileges;
104
105 for (i = 0; i < num_entries; i++)
106 if (opcode == cmd_priv_map[i].opcode &&
107 subsystem == cmd_priv_map[i].subsystem)
108 if (!(cmd_privileges & cmd_priv_map[i].priv_mask))
109 return false;
110
111 return true;
112 }
113
114 static inline void *embedded_payload(struct be_mcc_wrb *wrb)
115 {
116 return wrb->payload.embedded_payload;
117 }
118
119 static int be_mcc_notify(struct be_adapter *adapter)
120 {
121 struct be_queue_info *mccq = &adapter->mcc_obj.q;
122 u32 val = 0;
123
124 if (be_check_error(adapter, BE_ERROR_ANY))
125 return -EIO;
126
127 val |= mccq->id & DB_MCCQ_RING_ID_MASK;
128 val |= 1 << DB_MCCQ_NUM_POSTED_SHIFT;
129
130 wmb();
131 iowrite32(val, adapter->db + DB_MCCQ_OFFSET);
132
133 return 0;
134 }
135
136 /* To check if valid bit is set, check the entire word as we don't know
137 * the endianness of the data (old entry is host endian while a new entry is
138 * little endian) */
139 static inline bool be_mcc_compl_is_new(struct be_mcc_compl *compl)
140 {
141 u32 flags;
142
143 if (compl->flags != 0) {
144 flags = le32_to_cpu(compl->flags);
145 if (flags & CQE_FLAGS_VALID_MASK) {
146 compl->flags = flags;
147 return true;
148 }
149 }
150 return false;
151 }
152
153 /* Need to reset the entire word that houses the valid bit */
154 static inline void be_mcc_compl_use(struct be_mcc_compl *compl)
155 {
156 compl->flags = 0;
157 }
158
159 static struct be_cmd_resp_hdr *be_decode_resp_hdr(u32 tag0, u32 tag1)
160 {
161 unsigned long addr;
162
163 addr = tag1;
164 addr = ((addr << 16) << 16) | tag0;
165 return (void *)addr;
166 }
167
168 static bool be_skip_err_log(u8 opcode, u16 base_status, u16 addl_status)
169 {
170 if (base_status == MCC_STATUS_NOT_SUPPORTED ||
171 base_status == MCC_STATUS_ILLEGAL_REQUEST ||
172 addl_status == MCC_ADDL_STATUS_TOO_MANY_INTERFACES ||
173 addl_status == MCC_ADDL_STATUS_INSUFFICIENT_VLANS ||
174 (opcode == OPCODE_COMMON_WRITE_FLASHROM &&
175 (base_status == MCC_STATUS_ILLEGAL_FIELD ||
176 addl_status == MCC_ADDL_STATUS_FLASH_IMAGE_CRC_MISMATCH)))
177 return true;
178 else
179 return false;
180 }
181
182 /* Place holder for all the async MCC cmds wherein the caller is not in a busy
183 * loop (has not issued be_mcc_notify_wait())
184 */
185 static void be_async_cmd_process(struct be_adapter *adapter,
186 struct be_mcc_compl *compl,
187 struct be_cmd_resp_hdr *resp_hdr)
188 {
189 enum mcc_base_status base_status = base_status(compl->status);
190 u8 opcode = 0, subsystem = 0;
191
192 if (resp_hdr) {
193 opcode = resp_hdr->opcode;
194 subsystem = resp_hdr->subsystem;
195 }
196
197 if (opcode == OPCODE_LOWLEVEL_LOOPBACK_TEST &&
198 subsystem == CMD_SUBSYSTEM_LOWLEVEL) {
199 complete(&adapter->et_cmd_compl);
200 return;
201 }
202
203 if (opcode == OPCODE_LOWLEVEL_SET_LOOPBACK_MODE &&
204 subsystem == CMD_SUBSYSTEM_LOWLEVEL) {
205 complete(&adapter->et_cmd_compl);
206 return;
207 }
208
209 if ((opcode == OPCODE_COMMON_WRITE_FLASHROM ||
210 opcode == OPCODE_COMMON_WRITE_OBJECT) &&
211 subsystem == CMD_SUBSYSTEM_COMMON) {
212 adapter->flash_status = compl->status;
213 complete(&adapter->et_cmd_compl);
214 return;
215 }
216
217 if ((opcode == OPCODE_ETH_GET_STATISTICS ||
218 opcode == OPCODE_ETH_GET_PPORT_STATS) &&
219 subsystem == CMD_SUBSYSTEM_ETH &&
220 base_status == MCC_STATUS_SUCCESS) {
221 be_parse_stats(adapter);
222 adapter->stats_cmd_sent = false;
223 return;
224 }
225
226 if (opcode == OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES &&
227 subsystem == CMD_SUBSYSTEM_COMMON) {
228 if (base_status == MCC_STATUS_SUCCESS) {
229 struct be_cmd_resp_get_cntl_addnl_attribs *resp =
230 (void *)resp_hdr;
231 adapter->hwmon_info.be_on_die_temp =
232 resp->on_die_temperature;
233 } else {
234 adapter->be_get_temp_freq = 0;
235 adapter->hwmon_info.be_on_die_temp =
236 BE_INVALID_DIE_TEMP;
237 }
238 return;
239 }
240 }
241
242 static int be_mcc_compl_process(struct be_adapter *adapter,
243 struct be_mcc_compl *compl)
244 {
245 enum mcc_base_status base_status;
246 enum mcc_addl_status addl_status;
247 struct be_cmd_resp_hdr *resp_hdr;
248 u8 opcode = 0, subsystem = 0;
249
250 /* Just swap the status to host endian; mcc tag is opaquely copied
251 * from mcc_wrb */
252 be_dws_le_to_cpu(compl, 4);
253
254 base_status = base_status(compl->status);
255 addl_status = addl_status(compl->status);
256
257 resp_hdr = be_decode_resp_hdr(compl->tag0, compl->tag1);
258 if (resp_hdr) {
259 opcode = resp_hdr->opcode;
260 subsystem = resp_hdr->subsystem;
261 }
262
263 be_async_cmd_process(adapter, compl, resp_hdr);
264
265 if (base_status != MCC_STATUS_SUCCESS &&
266 !be_skip_err_log(opcode, base_status, addl_status)) {
267 if (base_status == MCC_STATUS_UNAUTHORIZED_REQUEST ||
268 addl_status == MCC_ADDL_STATUS_INSUFFICIENT_PRIVILEGES) {
269 dev_warn(&adapter->pdev->dev,
270 "VF is not privileged to issue opcode %d-%d\n",
271 opcode, subsystem);
272 } else {
273 dev_err(&adapter->pdev->dev,
274 "opcode %d-%d failed:status %d-%d\n",
275 opcode, subsystem, base_status, addl_status);
276 }
277 }
278 return compl->status;
279 }
280
281 /* Link state evt is a string of bytes; no need for endian swapping */
282 static void be_async_link_state_process(struct be_adapter *adapter,
283 struct be_mcc_compl *compl)
284 {
285 struct be_async_event_link_state *evt =
286 (struct be_async_event_link_state *)compl;
287
288 /* When link status changes, link speed must be re-queried from FW */
289 adapter->phy.link_speed = -1;
290
291 /* On BEx the FW does not send a separate link status
292 * notification for physical and logical link.
293 * On other chips just process the logical link
294 * status notification
295 */
296 if (!BEx_chip(adapter) &&
297 !(evt->port_link_status & LOGICAL_LINK_STATUS_MASK))
298 return;
299
300 /* For the initial link status do not rely on the ASYNC event as
301 * it may not be received in some cases.
302 */
303 if (adapter->flags & BE_FLAGS_LINK_STATUS_INIT)
304 be_link_status_update(adapter,
305 evt->port_link_status & LINK_STATUS_MASK);
306 }
307
308 static void be_async_port_misconfig_event_process(struct be_adapter *adapter,
309 struct be_mcc_compl *compl)
310 {
311 struct be_async_event_misconfig_port *evt =
312 (struct be_async_event_misconfig_port *)compl;
313 u32 sfp_misconfig_evt_word1 = le32_to_cpu(evt->event_data_word1);
314 u32 sfp_misconfig_evt_word2 = le32_to_cpu(evt->event_data_word2);
315 u8 phy_oper_state = PHY_STATE_OPER_MSG_NONE;
316 struct device *dev = &adapter->pdev->dev;
317 u8 msg_severity = DEFAULT_MSG_SEVERITY;
318 u8 phy_state_info;
319 u8 new_phy_state;
320
321 new_phy_state =
322 (sfp_misconfig_evt_word1 >> (adapter->hba_port_num * 8)) & 0xff;
323
324 if (new_phy_state == adapter->phy_state)
325 return;
326
327 adapter->phy_state = new_phy_state;
328
329 /* for older fw that doesn't populate link effect data */
330 if (!sfp_misconfig_evt_word2)
331 goto log_message;
332
333 phy_state_info =
334 (sfp_misconfig_evt_word2 >> (adapter->hba_port_num * 8)) & 0xff;
335
336 if (phy_state_info & PHY_STATE_INFO_VALID) {
337 msg_severity = (phy_state_info & PHY_STATE_MSG_SEVERITY) >> 1;
338
339 if (be_phy_unqualified(new_phy_state))
340 phy_oper_state = (phy_state_info & PHY_STATE_OPER);
341 }
342
343 log_message:
344 /* Log an error message that would allow a user to determine
345 * whether the SFPs have an issue
346 */
347 if (be_phy_state_unknown(new_phy_state))
348 dev_printk(be_port_misconfig_evt_severity[msg_severity], dev,
349 "Port %c: Unrecognized Optics state: 0x%x. %s",
350 adapter->port_name,
351 new_phy_state,
352 phy_state_oper_desc[phy_oper_state]);
353 else
354 dev_printk(be_port_misconfig_evt_severity[msg_severity], dev,
355 "Port %c: %s %s",
356 adapter->port_name,
357 be_misconfig_evt_port_state[new_phy_state],
358 phy_state_oper_desc[phy_oper_state]);
359
360 /* Log Vendor name and part no. if a misconfigured SFP is detected */
361 if (be_phy_misconfigured(new_phy_state))
362 adapter->flags |= BE_FLAGS_PHY_MISCONFIGURED;
363 }
364
365 /* Grp5 CoS Priority evt */
366 static void be_async_grp5_cos_priority_process(struct be_adapter *adapter,
367 struct be_mcc_compl *compl)
368 {
369 struct be_async_event_grp5_cos_priority *evt =
370 (struct be_async_event_grp5_cos_priority *)compl;
371
372 if (evt->valid) {
373 adapter->vlan_prio_bmap = evt->available_priority_bmap;
374 adapter->recommended_prio_bits =
375 evt->reco_default_priority << VLAN_PRIO_SHIFT;
376 }
377 }
378
379 /* Grp5 QOS Speed evt: qos_link_speed is in units of 10 Mbps */
380 static void be_async_grp5_qos_speed_process(struct be_adapter *adapter,
381 struct be_mcc_compl *compl)
382 {
383 struct be_async_event_grp5_qos_link_speed *evt =
384 (struct be_async_event_grp5_qos_link_speed *)compl;
385
386 if (adapter->phy.link_speed >= 0 &&
387 evt->physical_port == adapter->port_num)
388 adapter->phy.link_speed = le16_to_cpu(evt->qos_link_speed) * 10;
389 }
390
391 /*Grp5 PVID evt*/
392 static void be_async_grp5_pvid_state_process(struct be_adapter *adapter,
393 struct be_mcc_compl *compl)
394 {
395 struct be_async_event_grp5_pvid_state *evt =
396 (struct be_async_event_grp5_pvid_state *)compl;
397
398 if (evt->enabled) {
399 adapter->pvid = le16_to_cpu(evt->tag) & VLAN_VID_MASK;
400 dev_info(&adapter->pdev->dev, "LPVID: %d\n", adapter->pvid);
401 } else {
402 adapter->pvid = 0;
403 }
404 }
405
406 #define MGMT_ENABLE_MASK 0x4
407 static void be_async_grp5_fw_control_process(struct be_adapter *adapter,
408 struct be_mcc_compl *compl)
409 {
410 struct be_async_fw_control *evt = (struct be_async_fw_control *)compl;
411 u32 evt_dw1 = le32_to_cpu(evt->event_data_word1);
412
413 if (evt_dw1 & MGMT_ENABLE_MASK) {
414 adapter->flags |= BE_FLAGS_OS2BMC;
415 adapter->bmc_filt_mask = le32_to_cpu(evt->event_data_word2);
416 } else {
417 adapter->flags &= ~BE_FLAGS_OS2BMC;
418 }
419 }
420
421 static void be_async_grp5_evt_process(struct be_adapter *adapter,
422 struct be_mcc_compl *compl)
423 {
424 u8 event_type = (compl->flags >> ASYNC_EVENT_TYPE_SHIFT) &
425 ASYNC_EVENT_TYPE_MASK;
426
427 switch (event_type) {
428 case ASYNC_EVENT_COS_PRIORITY:
429 be_async_grp5_cos_priority_process(adapter, compl);
430 break;
431 case ASYNC_EVENT_QOS_SPEED:
432 be_async_grp5_qos_speed_process(adapter, compl);
433 break;
434 case ASYNC_EVENT_PVID_STATE:
435 be_async_grp5_pvid_state_process(adapter, compl);
436 break;
437 /* Async event to disable/enable os2bmc and/or mac-learning */
438 case ASYNC_EVENT_FW_CONTROL:
439 be_async_grp5_fw_control_process(adapter, compl);
440 break;
441 default:
442 break;
443 }
444 }
445
446 static void be_async_dbg_evt_process(struct be_adapter *adapter,
447 struct be_mcc_compl *cmp)
448 {
449 u8 event_type = 0;
450 struct be_async_event_qnq *evt = (struct be_async_event_qnq *)cmp;
451
452 event_type = (cmp->flags >> ASYNC_EVENT_TYPE_SHIFT) &
453 ASYNC_EVENT_TYPE_MASK;
454
455 switch (event_type) {
456 case ASYNC_DEBUG_EVENT_TYPE_QNQ:
457 if (evt->valid)
458 adapter->qnq_vid = le16_to_cpu(evt->vlan_tag);
459 adapter->flags |= BE_FLAGS_QNQ_ASYNC_EVT_RCVD;
460 break;
461 default:
462 dev_warn(&adapter->pdev->dev, "Unknown debug event 0x%x!\n",
463 event_type);
464 break;
465 }
466 }
467
468 static void be_async_sliport_evt_process(struct be_adapter *adapter,
469 struct be_mcc_compl *cmp)
470 {
471 u8 event_type = (cmp->flags >> ASYNC_EVENT_TYPE_SHIFT) &
472 ASYNC_EVENT_TYPE_MASK;
473
474 if (event_type == ASYNC_EVENT_PORT_MISCONFIG)
475 be_async_port_misconfig_event_process(adapter, cmp);
476 }
477
478 static inline bool is_link_state_evt(u32 flags)
479 {
480 return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
481 ASYNC_EVENT_CODE_LINK_STATE;
482 }
483
484 static inline bool is_grp5_evt(u32 flags)
485 {
486 return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
487 ASYNC_EVENT_CODE_GRP_5;
488 }
489
490 static inline bool is_dbg_evt(u32 flags)
491 {
492 return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
493 ASYNC_EVENT_CODE_QNQ;
494 }
495
496 static inline bool is_sliport_evt(u32 flags)
497 {
498 return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
499 ASYNC_EVENT_CODE_SLIPORT;
500 }
501
502 static void be_mcc_event_process(struct be_adapter *adapter,
503 struct be_mcc_compl *compl)
504 {
505 if (is_link_state_evt(compl->flags))
506 be_async_link_state_process(adapter, compl);
507 else if (is_grp5_evt(compl->flags))
508 be_async_grp5_evt_process(adapter, compl);
509 else if (is_dbg_evt(compl->flags))
510 be_async_dbg_evt_process(adapter, compl);
511 else if (is_sliport_evt(compl->flags))
512 be_async_sliport_evt_process(adapter, compl);
513 }
514
515 static struct be_mcc_compl *be_mcc_compl_get(struct be_adapter *adapter)
516 {
517 struct be_queue_info *mcc_cq = &adapter->mcc_obj.cq;
518 struct be_mcc_compl *compl = queue_tail_node(mcc_cq);
519
520 if (be_mcc_compl_is_new(compl)) {
521 queue_tail_inc(mcc_cq);
522 return compl;
523 }
524 return NULL;
525 }
526
527 void be_async_mcc_enable(struct be_adapter *adapter)
528 {
529 spin_lock_bh(&adapter->mcc_cq_lock);
530
531 be_cq_notify(adapter, adapter->mcc_obj.cq.id, true, 0);
532 adapter->mcc_obj.rearm_cq = true;
533
534 spin_unlock_bh(&adapter->mcc_cq_lock);
535 }
536
537 void be_async_mcc_disable(struct be_adapter *adapter)
538 {
539 spin_lock_bh(&adapter->mcc_cq_lock);
540
541 adapter->mcc_obj.rearm_cq = false;
542 be_cq_notify(adapter, adapter->mcc_obj.cq.id, false, 0);
543
544 spin_unlock_bh(&adapter->mcc_cq_lock);
545 }
546
547 int be_process_mcc(struct be_adapter *adapter)
548 {
549 struct be_mcc_compl *compl;
550 int num = 0, status = 0;
551 struct be_mcc_obj *mcc_obj = &adapter->mcc_obj;
552
553 spin_lock(&adapter->mcc_cq_lock);
554
555 while ((compl = be_mcc_compl_get(adapter))) {
556 if (compl->flags & CQE_FLAGS_ASYNC_MASK) {
557 be_mcc_event_process(adapter, compl);
558 } else if (compl->flags & CQE_FLAGS_COMPLETED_MASK) {
559 status = be_mcc_compl_process(adapter, compl);
560 atomic_dec(&mcc_obj->q.used);
561 }
562 be_mcc_compl_use(compl);
563 num++;
564 }
565
566 if (num)
567 be_cq_notify(adapter, mcc_obj->cq.id, mcc_obj->rearm_cq, num);
568
569 spin_unlock(&adapter->mcc_cq_lock);
570 return status;
571 }
572
573 /* Wait till no more pending mcc requests are present */
574 static int be_mcc_wait_compl(struct be_adapter *adapter)
575 {
576 #define mcc_timeout 12000 /* 12s timeout */
577 int i, status = 0;
578 struct be_mcc_obj *mcc_obj = &adapter->mcc_obj;
579
580 for (i = 0; i < mcc_timeout; i++) {
581 if (be_check_error(adapter, BE_ERROR_ANY))
582 return -EIO;
583
584 local_bh_disable();
585 status = be_process_mcc(adapter);
586 local_bh_enable();
587
588 if (atomic_read(&mcc_obj->q.used) == 0)
589 break;
590 usleep_range(500, 1000);
591 }
592 if (i == mcc_timeout) {
593 dev_err(&adapter->pdev->dev, "FW not responding\n");
594 be_set_error(adapter, BE_ERROR_FW);
595 return -EIO;
596 }
597 return status;
598 }
599
600 /* Notify MCC requests and wait for completion */
601 static int be_mcc_notify_wait(struct be_adapter *adapter)
602 {
603 int status;
604 struct be_mcc_wrb *wrb;
605 struct be_mcc_obj *mcc_obj = &adapter->mcc_obj;
606 u32 index = mcc_obj->q.head;
607 struct be_cmd_resp_hdr *resp;
608
609 index_dec(&index, mcc_obj->q.len);
610 wrb = queue_index_node(&mcc_obj->q, index);
611
612 resp = be_decode_resp_hdr(wrb->tag0, wrb->tag1);
613
614 status = be_mcc_notify(adapter);
615 if (status)
616 goto out;
617
618 status = be_mcc_wait_compl(adapter);
619 if (status == -EIO)
620 goto out;
621
622 status = (resp->base_status |
623 ((resp->addl_status & CQE_ADDL_STATUS_MASK) <<
624 CQE_ADDL_STATUS_SHIFT));
625 out:
626 return status;
627 }
628
629 static int be_mbox_db_ready_wait(struct be_adapter *adapter, void __iomem *db)
630 {
631 int msecs = 0;
632 u32 ready;
633
634 do {
635 if (be_check_error(adapter, BE_ERROR_ANY))
636 return -EIO;
637
638 ready = ioread32(db);
639 if (ready == 0xffffffff)
640 return -1;
641
642 ready &= MPU_MAILBOX_DB_RDY_MASK;
643 if (ready)
644 break;
645
646 if (msecs > 4000) {
647 dev_err(&adapter->pdev->dev, "FW not responding\n");
648 be_set_error(adapter, BE_ERROR_FW);
649 be_detect_error(adapter);
650 return -1;
651 }
652
653 msleep(1);
654 msecs++;
655 } while (true);
656
657 return 0;
658 }
659
660 /*
661 * Insert the mailbox address into the doorbell in two steps
662 * Polls on the mbox doorbell till a command completion (or a timeout) occurs
663 */
664 static int be_mbox_notify_wait(struct be_adapter *adapter)
665 {
666 int status;
667 u32 val = 0;
668 void __iomem *db = adapter->db + MPU_MAILBOX_DB_OFFSET;
669 struct be_dma_mem *mbox_mem = &adapter->mbox_mem;
670 struct be_mcc_mailbox *mbox = mbox_mem->va;
671 struct be_mcc_compl *compl = &mbox->compl;
672
673 /* wait for ready to be set */
674 status = be_mbox_db_ready_wait(adapter, db);
675 if (status != 0)
676 return status;
677
678 val |= MPU_MAILBOX_DB_HI_MASK;
679 /* at bits 2 - 31 place mbox dma addr msb bits 34 - 63 */
680 val |= (upper_32_bits(mbox_mem->dma) >> 2) << 2;
681 iowrite32(val, db);
682
683 /* wait for ready to be set */
684 status = be_mbox_db_ready_wait(adapter, db);
685 if (status != 0)
686 return status;
687
688 val = 0;
689 /* at bits 2 - 31 place mbox dma addr lsb bits 4 - 33 */
690 val |= (u32)(mbox_mem->dma >> 4) << 2;
691 iowrite32(val, db);
692
693 status = be_mbox_db_ready_wait(adapter, db);
694 if (status != 0)
695 return status;
696
697 /* A cq entry has been made now */
698 if (be_mcc_compl_is_new(compl)) {
699 status = be_mcc_compl_process(adapter, &mbox->compl);
700 be_mcc_compl_use(compl);
701 if (status)
702 return status;
703 } else {
704 dev_err(&adapter->pdev->dev, "invalid mailbox completion\n");
705 return -1;
706 }
707 return 0;
708 }
709
710 u16 be_POST_stage_get(struct be_adapter *adapter)
711 {
712 u32 sem;
713
714 if (BEx_chip(adapter))
715 sem = ioread32(adapter->csr + SLIPORT_SEMAPHORE_OFFSET_BEx);
716 else
717 pci_read_config_dword(adapter->pdev,
718 SLIPORT_SEMAPHORE_OFFSET_SH, &sem);
719
720 return sem & POST_STAGE_MASK;
721 }
722
723 static int lancer_wait_ready(struct be_adapter *adapter)
724 {
725 #define SLIPORT_READY_TIMEOUT 30
726 u32 sliport_status;
727 int i;
728
729 for (i = 0; i < SLIPORT_READY_TIMEOUT; i++) {
730 sliport_status = ioread32(adapter->db + SLIPORT_STATUS_OFFSET);
731 if (sliport_status & SLIPORT_STATUS_RDY_MASK)
732 return 0;
733
734 if (sliport_status & SLIPORT_STATUS_ERR_MASK &&
735 !(sliport_status & SLIPORT_STATUS_RN_MASK))
736 return -EIO;
737
738 msleep(1000);
739 }
740
741 return sliport_status ? : -1;
742 }
743
744 int be_fw_wait_ready(struct be_adapter *adapter)
745 {
746 u16 stage;
747 int status, timeout = 0;
748 struct device *dev = &adapter->pdev->dev;
749
750 if (lancer_chip(adapter)) {
751 status = lancer_wait_ready(adapter);
752 if (status) {
753 stage = status;
754 goto err;
755 }
756 return 0;
757 }
758
759 do {
760 /* There's no means to poll POST state on BE2/3 VFs */
761 if (BEx_chip(adapter) && be_virtfn(adapter))
762 return 0;
763
764 stage = be_POST_stage_get(adapter);
765 if (stage == POST_STAGE_ARMFW_RDY)
766 return 0;
767
768 dev_info(dev, "Waiting for POST, %ds elapsed\n", timeout);
769 if (msleep_interruptible(2000)) {
770 dev_err(dev, "Waiting for POST aborted\n");
771 return -EINTR;
772 }
773 timeout += 2;
774 } while (timeout < 60);
775
776 err:
777 dev_err(dev, "POST timeout; stage=%#x\n", stage);
778 return -ETIMEDOUT;
779 }
780
781 static inline struct be_sge *nonembedded_sgl(struct be_mcc_wrb *wrb)
782 {
783 return &wrb->payload.sgl[0];
784 }
785
786 static inline void fill_wrb_tags(struct be_mcc_wrb *wrb, unsigned long addr)
787 {
788 wrb->tag0 = addr & 0xFFFFFFFF;
789 wrb->tag1 = upper_32_bits(addr);
790 }
791
792 /* Don't touch the hdr after it's prepared */
793 /* mem will be NULL for embedded commands */
794 static void be_wrb_cmd_hdr_prepare(struct be_cmd_req_hdr *req_hdr,
795 u8 subsystem, u8 opcode, int cmd_len,
796 struct be_mcc_wrb *wrb,
797 struct be_dma_mem *mem)
798 {
799 struct be_sge *sge;
800
801 req_hdr->opcode = opcode;
802 req_hdr->subsystem = subsystem;
803 req_hdr->request_length = cpu_to_le32(cmd_len - sizeof(*req_hdr));
804 req_hdr->version = 0;
805 fill_wrb_tags(wrb, (ulong) req_hdr);
806 wrb->payload_length = cmd_len;
807 if (mem) {
808 wrb->embedded |= (1 & MCC_WRB_SGE_CNT_MASK) <<
809 MCC_WRB_SGE_CNT_SHIFT;
810 sge = nonembedded_sgl(wrb);
811 sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma));
812 sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF);
813 sge->len = cpu_to_le32(mem->size);
814 } else
815 wrb->embedded |= MCC_WRB_EMBEDDED_MASK;
816 be_dws_cpu_to_le(wrb, 8);
817 }
818
819 static void be_cmd_page_addrs_prepare(struct phys_addr *pages, u32 max_pages,
820 struct be_dma_mem *mem)
821 {
822 int i, buf_pages = min(PAGES_4K_SPANNED(mem->va, mem->size), max_pages);
823 u64 dma = (u64)mem->dma;
824
825 for (i = 0; i < buf_pages; i++) {
826 pages[i].lo = cpu_to_le32(dma & 0xFFFFFFFF);
827 pages[i].hi = cpu_to_le32(upper_32_bits(dma));
828 dma += PAGE_SIZE_4K;
829 }
830 }
831
832 static inline struct be_mcc_wrb *wrb_from_mbox(struct be_adapter *adapter)
833 {
834 struct be_dma_mem *mbox_mem = &adapter->mbox_mem;
835 struct be_mcc_wrb *wrb
836 = &((struct be_mcc_mailbox *)(mbox_mem->va))->wrb;
837 memset(wrb, 0, sizeof(*wrb));
838 return wrb;
839 }
840
841 static struct be_mcc_wrb *wrb_from_mccq(struct be_adapter *adapter)
842 {
843 struct be_queue_info *mccq = &adapter->mcc_obj.q;
844 struct be_mcc_wrb *wrb;
845
846 if (!mccq->created)
847 return NULL;
848
849 if (atomic_read(&mccq->used) >= mccq->len)
850 return NULL;
851
852 wrb = queue_head_node(mccq);
853 queue_head_inc(mccq);
854 atomic_inc(&mccq->used);
855 memset(wrb, 0, sizeof(*wrb));
856 return wrb;
857 }
858
859 static bool use_mcc(struct be_adapter *adapter)
860 {
861 return adapter->mcc_obj.q.created;
862 }
863
864 /* Must be used only in process context */
865 static int be_cmd_lock(struct be_adapter *adapter)
866 {
867 if (use_mcc(adapter)) {
868 mutex_lock(&adapter->mcc_lock);
869 return 0;
870 } else {
871 return mutex_lock_interruptible(&adapter->mbox_lock);
872 }
873 }
874
875 /* Must be used only in process context */
876 static void be_cmd_unlock(struct be_adapter *adapter)
877 {
878 if (use_mcc(adapter))
879 return mutex_unlock(&adapter->mcc_lock);
880 else
881 return mutex_unlock(&adapter->mbox_lock);
882 }
883
884 static struct be_mcc_wrb *be_cmd_copy(struct be_adapter *adapter,
885 struct be_mcc_wrb *wrb)
886 {
887 struct be_mcc_wrb *dest_wrb;
888
889 if (use_mcc(adapter)) {
890 dest_wrb = wrb_from_mccq(adapter);
891 if (!dest_wrb)
892 return NULL;
893 } else {
894 dest_wrb = wrb_from_mbox(adapter);
895 }
896
897 memcpy(dest_wrb, wrb, sizeof(*wrb));
898 if (wrb->embedded & cpu_to_le32(MCC_WRB_EMBEDDED_MASK))
899 fill_wrb_tags(dest_wrb, (ulong) embedded_payload(wrb));
900
901 return dest_wrb;
902 }
903
904 /* Must be used only in process context */
905 static int be_cmd_notify_wait(struct be_adapter *adapter,
906 struct be_mcc_wrb *wrb)
907 {
908 struct be_mcc_wrb *dest_wrb;
909 int status;
910
911 status = be_cmd_lock(adapter);
912 if (status)
913 return status;
914
915 dest_wrb = be_cmd_copy(adapter, wrb);
916 if (!dest_wrb) {
917 status = -EBUSY;
918 goto unlock;
919 }
920
921 if (use_mcc(adapter))
922 status = be_mcc_notify_wait(adapter);
923 else
924 status = be_mbox_notify_wait(adapter);
925
926 if (!status)
927 memcpy(wrb, dest_wrb, sizeof(*wrb));
928
929 unlock:
930 be_cmd_unlock(adapter);
931 return status;
932 }
933
934 /* Tell fw we're about to start firing cmds by writing a
935 * special pattern across the wrb hdr; uses mbox
936 */
937 int be_cmd_fw_init(struct be_adapter *adapter)
938 {
939 u8 *wrb;
940 int status;
941
942 if (lancer_chip(adapter))
943 return 0;
944
945 if (mutex_lock_interruptible(&adapter->mbox_lock))
946 return -1;
947
948 wrb = (u8 *)wrb_from_mbox(adapter);
949 *wrb++ = 0xFF;
950 *wrb++ = 0x12;
951 *wrb++ = 0x34;
952 *wrb++ = 0xFF;
953 *wrb++ = 0xFF;
954 *wrb++ = 0x56;
955 *wrb++ = 0x78;
956 *wrb = 0xFF;
957
958 status = be_mbox_notify_wait(adapter);
959
960 mutex_unlock(&adapter->mbox_lock);
961 return status;
962 }
963
964 /* Tell fw we're done with firing cmds by writing a
965 * special pattern across the wrb hdr; uses mbox
966 */
967 int be_cmd_fw_clean(struct be_adapter *adapter)
968 {
969 u8 *wrb;
970 int status;
971
972 if (lancer_chip(adapter))
973 return 0;
974
975 if (mutex_lock_interruptible(&adapter->mbox_lock))
976 return -1;
977
978 wrb = (u8 *)wrb_from_mbox(adapter);
979 *wrb++ = 0xFF;
980 *wrb++ = 0xAA;
981 *wrb++ = 0xBB;
982 *wrb++ = 0xFF;
983 *wrb++ = 0xFF;
984 *wrb++ = 0xCC;
985 *wrb++ = 0xDD;
986 *wrb = 0xFF;
987
988 status = be_mbox_notify_wait(adapter);
989
990 mutex_unlock(&adapter->mbox_lock);
991 return status;
992 }
993
994 int be_cmd_eq_create(struct be_adapter *adapter, struct be_eq_obj *eqo)
995 {
996 struct be_mcc_wrb *wrb;
997 struct be_cmd_req_eq_create *req;
998 struct be_dma_mem *q_mem = &eqo->q.dma_mem;
999 int status, ver = 0;
1000
1001 if (mutex_lock_interruptible(&adapter->mbox_lock))
1002 return -1;
1003
1004 wrb = wrb_from_mbox(adapter);
1005 req = embedded_payload(wrb);
1006
1007 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1008 OPCODE_COMMON_EQ_CREATE, sizeof(*req), wrb,
1009 NULL);
1010
1011 /* Support for EQ_CREATEv2 available only SH-R onwards */
1012 if (!(BEx_chip(adapter) || lancer_chip(adapter)))
1013 ver = 2;
1014
1015 req->hdr.version = ver;
1016 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1017
1018 AMAP_SET_BITS(struct amap_eq_context, valid, req->context, 1);
1019 /* 4byte eqe*/
1020 AMAP_SET_BITS(struct amap_eq_context, size, req->context, 0);
1021 AMAP_SET_BITS(struct amap_eq_context, count, req->context,
1022 __ilog2_u32(eqo->q.len / 256));
1023 be_dws_cpu_to_le(req->context, sizeof(req->context));
1024
1025 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1026
1027 status = be_mbox_notify_wait(adapter);
1028 if (!status) {
1029 struct be_cmd_resp_eq_create *resp = embedded_payload(wrb);
1030
1031 eqo->q.id = le16_to_cpu(resp->eq_id);
1032 eqo->msix_idx =
1033 (ver == 2) ? le16_to_cpu(resp->msix_idx) : eqo->idx;
1034 eqo->q.created = true;
1035 }
1036
1037 mutex_unlock(&adapter->mbox_lock);
1038 return status;
1039 }
1040
1041 /* Use MCC */
1042 int be_cmd_mac_addr_query(struct be_adapter *adapter, u8 *mac_addr,
1043 bool permanent, u32 if_handle, u32 pmac_id)
1044 {
1045 struct be_mcc_wrb *wrb;
1046 struct be_cmd_req_mac_query *req;
1047 int status;
1048
1049 mutex_lock(&adapter->mcc_lock);
1050
1051 wrb = wrb_from_mccq(adapter);
1052 if (!wrb) {
1053 status = -EBUSY;
1054 goto err;
1055 }
1056 req = embedded_payload(wrb);
1057
1058 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1059 OPCODE_COMMON_NTWK_MAC_QUERY, sizeof(*req), wrb,
1060 NULL);
1061 req->type = MAC_ADDRESS_TYPE_NETWORK;
1062 if (permanent) {
1063 req->permanent = 1;
1064 } else {
1065 req->if_id = cpu_to_le16((u16)if_handle);
1066 req->pmac_id = cpu_to_le32(pmac_id);
1067 req->permanent = 0;
1068 }
1069
1070 status = be_mcc_notify_wait(adapter);
1071 if (!status) {
1072 struct be_cmd_resp_mac_query *resp = embedded_payload(wrb);
1073
1074 memcpy(mac_addr, resp->mac.addr, ETH_ALEN);
1075 }
1076
1077 err:
1078 mutex_unlock(&adapter->mcc_lock);
1079 return status;
1080 }
1081
1082 /* Uses synchronous MCCQ */
1083 int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr,
1084 u32 if_id, u32 *pmac_id, u32 domain)
1085 {
1086 struct be_mcc_wrb *wrb;
1087 struct be_cmd_req_pmac_add *req;
1088 int status;
1089
1090 mutex_lock(&adapter->mcc_lock);
1091
1092 wrb = wrb_from_mccq(adapter);
1093 if (!wrb) {
1094 status = -EBUSY;
1095 goto err;
1096 }
1097 req = embedded_payload(wrb);
1098
1099 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1100 OPCODE_COMMON_NTWK_PMAC_ADD, sizeof(*req), wrb,
1101 NULL);
1102
1103 req->hdr.domain = domain;
1104 req->if_id = cpu_to_le32(if_id);
1105 memcpy(req->mac_address, mac_addr, ETH_ALEN);
1106
1107 status = be_mcc_notify_wait(adapter);
1108 if (!status) {
1109 struct be_cmd_resp_pmac_add *resp = embedded_payload(wrb);
1110
1111 *pmac_id = le32_to_cpu(resp->pmac_id);
1112 }
1113
1114 err:
1115 mutex_unlock(&adapter->mcc_lock);
1116
1117 if (base_status(status) == MCC_STATUS_UNAUTHORIZED_REQUEST)
1118 status = -EPERM;
1119
1120 return status;
1121 }
1122
1123 /* Uses synchronous MCCQ */
1124 int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, int pmac_id, u32 dom)
1125 {
1126 struct be_mcc_wrb *wrb;
1127 struct be_cmd_req_pmac_del *req;
1128 int status;
1129
1130 if (pmac_id == -1)
1131 return 0;
1132
1133 mutex_lock(&adapter->mcc_lock);
1134
1135 wrb = wrb_from_mccq(adapter);
1136 if (!wrb) {
1137 status = -EBUSY;
1138 goto err;
1139 }
1140 req = embedded_payload(wrb);
1141
1142 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1143 OPCODE_COMMON_NTWK_PMAC_DEL, sizeof(*req),
1144 wrb, NULL);
1145
1146 req->hdr.domain = dom;
1147 req->if_id = cpu_to_le32(if_id);
1148 req->pmac_id = cpu_to_le32(pmac_id);
1149
1150 status = be_mcc_notify_wait(adapter);
1151
1152 err:
1153 mutex_unlock(&adapter->mcc_lock);
1154 return status;
1155 }
1156
1157 /* Uses Mbox */
1158 int be_cmd_cq_create(struct be_adapter *adapter, struct be_queue_info *cq,
1159 struct be_queue_info *eq, bool no_delay, int coalesce_wm)
1160 {
1161 struct be_mcc_wrb *wrb;
1162 struct be_cmd_req_cq_create *req;
1163 struct be_dma_mem *q_mem = &cq->dma_mem;
1164 void *ctxt;
1165 int status;
1166
1167 if (mutex_lock_interruptible(&adapter->mbox_lock))
1168 return -1;
1169
1170 wrb = wrb_from_mbox(adapter);
1171 req = embedded_payload(wrb);
1172 ctxt = &req->context;
1173
1174 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1175 OPCODE_COMMON_CQ_CREATE, sizeof(*req), wrb,
1176 NULL);
1177
1178 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1179
1180 if (BEx_chip(adapter)) {
1181 AMAP_SET_BITS(struct amap_cq_context_be, coalescwm, ctxt,
1182 coalesce_wm);
1183 AMAP_SET_BITS(struct amap_cq_context_be, nodelay,
1184 ctxt, no_delay);
1185 AMAP_SET_BITS(struct amap_cq_context_be, count, ctxt,
1186 __ilog2_u32(cq->len / 256));
1187 AMAP_SET_BITS(struct amap_cq_context_be, valid, ctxt, 1);
1188 AMAP_SET_BITS(struct amap_cq_context_be, eventable, ctxt, 1);
1189 AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id);
1190 } else {
1191 req->hdr.version = 2;
1192 req->page_size = 1; /* 1 for 4K */
1193
1194 /* coalesce-wm field in this cmd is not relevant to Lancer.
1195 * Lancer uses COMMON_MODIFY_CQ to set this field
1196 */
1197 if (!lancer_chip(adapter))
1198 AMAP_SET_BITS(struct amap_cq_context_v2, coalescwm,
1199 ctxt, coalesce_wm);
1200 AMAP_SET_BITS(struct amap_cq_context_v2, nodelay, ctxt,
1201 no_delay);
1202 AMAP_SET_BITS(struct amap_cq_context_v2, count, ctxt,
1203 __ilog2_u32(cq->len / 256));
1204 AMAP_SET_BITS(struct amap_cq_context_v2, valid, ctxt, 1);
1205 AMAP_SET_BITS(struct amap_cq_context_v2, eventable, ctxt, 1);
1206 AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id);
1207 }
1208
1209 be_dws_cpu_to_le(ctxt, sizeof(req->context));
1210
1211 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1212
1213 status = be_mbox_notify_wait(adapter);
1214 if (!status) {
1215 struct be_cmd_resp_cq_create *resp = embedded_payload(wrb);
1216
1217 cq->id = le16_to_cpu(resp->cq_id);
1218 cq->created = true;
1219 }
1220
1221 mutex_unlock(&adapter->mbox_lock);
1222
1223 return status;
1224 }
1225
1226 static u32 be_encoded_q_len(int q_len)
1227 {
1228 u32 len_encoded = fls(q_len); /* log2(len) + 1 */
1229
1230 if (len_encoded == 16)
1231 len_encoded = 0;
1232 return len_encoded;
1233 }
1234
1235 static int be_cmd_mccq_ext_create(struct be_adapter *adapter,
1236 struct be_queue_info *mccq,
1237 struct be_queue_info *cq)
1238 {
1239 struct be_mcc_wrb *wrb;
1240 struct be_cmd_req_mcc_ext_create *req;
1241 struct be_dma_mem *q_mem = &mccq->dma_mem;
1242 void *ctxt;
1243 int status;
1244
1245 if (mutex_lock_interruptible(&adapter->mbox_lock))
1246 return -1;
1247
1248 wrb = wrb_from_mbox(adapter);
1249 req = embedded_payload(wrb);
1250 ctxt = &req->context;
1251
1252 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1253 OPCODE_COMMON_MCC_CREATE_EXT, sizeof(*req), wrb,
1254 NULL);
1255
1256 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1257 if (BEx_chip(adapter)) {
1258 AMAP_SET_BITS(struct amap_mcc_context_be, valid, ctxt, 1);
1259 AMAP_SET_BITS(struct amap_mcc_context_be, ring_size, ctxt,
1260 be_encoded_q_len(mccq->len));
1261 AMAP_SET_BITS(struct amap_mcc_context_be, cq_id, ctxt, cq->id);
1262 } else {
1263 req->hdr.version = 1;
1264 req->cq_id = cpu_to_le16(cq->id);
1265
1266 AMAP_SET_BITS(struct amap_mcc_context_v1, ring_size, ctxt,
1267 be_encoded_q_len(mccq->len));
1268 AMAP_SET_BITS(struct amap_mcc_context_v1, valid, ctxt, 1);
1269 AMAP_SET_BITS(struct amap_mcc_context_v1, async_cq_id,
1270 ctxt, cq->id);
1271 AMAP_SET_BITS(struct amap_mcc_context_v1, async_cq_valid,
1272 ctxt, 1);
1273 }
1274
1275 /* Subscribe to Link State, Sliport Event and Group 5 Events
1276 * (bits 1, 5 and 17 set)
1277 */
1278 req->async_event_bitmap[0] =
1279 cpu_to_le32(BIT(ASYNC_EVENT_CODE_LINK_STATE) |
1280 BIT(ASYNC_EVENT_CODE_GRP_5) |
1281 BIT(ASYNC_EVENT_CODE_QNQ) |
1282 BIT(ASYNC_EVENT_CODE_SLIPORT));
1283
1284 be_dws_cpu_to_le(ctxt, sizeof(req->context));
1285
1286 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1287
1288 status = be_mbox_notify_wait(adapter);
1289 if (!status) {
1290 struct be_cmd_resp_mcc_create *resp = embedded_payload(wrb);
1291
1292 mccq->id = le16_to_cpu(resp->id);
1293 mccq->created = true;
1294 }
1295 mutex_unlock(&adapter->mbox_lock);
1296
1297 return status;
1298 }
1299
1300 static int be_cmd_mccq_org_create(struct be_adapter *adapter,
1301 struct be_queue_info *mccq,
1302 struct be_queue_info *cq)
1303 {
1304 struct be_mcc_wrb *wrb;
1305 struct be_cmd_req_mcc_create *req;
1306 struct be_dma_mem *q_mem = &mccq->dma_mem;
1307 void *ctxt;
1308 int status;
1309
1310 if (mutex_lock_interruptible(&adapter->mbox_lock))
1311 return -1;
1312
1313 wrb = wrb_from_mbox(adapter);
1314 req = embedded_payload(wrb);
1315 ctxt = &req->context;
1316
1317 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1318 OPCODE_COMMON_MCC_CREATE, sizeof(*req), wrb,
1319 NULL);
1320
1321 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1322
1323 AMAP_SET_BITS(struct amap_mcc_context_be, valid, ctxt, 1);
1324 AMAP_SET_BITS(struct amap_mcc_context_be, ring_size, ctxt,
1325 be_encoded_q_len(mccq->len));
1326 AMAP_SET_BITS(struct amap_mcc_context_be, cq_id, ctxt, cq->id);
1327
1328 be_dws_cpu_to_le(ctxt, sizeof(req->context));
1329
1330 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1331
1332 status = be_mbox_notify_wait(adapter);
1333 if (!status) {
1334 struct be_cmd_resp_mcc_create *resp = embedded_payload(wrb);
1335
1336 mccq->id = le16_to_cpu(resp->id);
1337 mccq->created = true;
1338 }
1339
1340 mutex_unlock(&adapter->mbox_lock);
1341 return status;
1342 }
1343
1344 int be_cmd_mccq_create(struct be_adapter *adapter,
1345 struct be_queue_info *mccq, struct be_queue_info *cq)
1346 {
1347 int status;
1348
1349 status = be_cmd_mccq_ext_create(adapter, mccq, cq);
1350 if (status && BEx_chip(adapter)) {
1351 dev_warn(&adapter->pdev->dev, "Upgrade to F/W ver 2.102.235.0 "
1352 "or newer to avoid conflicting priorities between NIC "
1353 "and FCoE traffic");
1354 status = be_cmd_mccq_org_create(adapter, mccq, cq);
1355 }
1356 return status;
1357 }
1358
1359 int be_cmd_txq_create(struct be_adapter *adapter, struct be_tx_obj *txo)
1360 {
1361 struct be_mcc_wrb wrb = {0};
1362 struct be_cmd_req_eth_tx_create *req;
1363 struct be_queue_info *txq = &txo->q;
1364 struct be_queue_info *cq = &txo->cq;
1365 struct be_dma_mem *q_mem = &txq->dma_mem;
1366 int status, ver = 0;
1367
1368 req = embedded_payload(&wrb);
1369 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1370 OPCODE_ETH_TX_CREATE, sizeof(*req), &wrb, NULL);
1371
1372 if (lancer_chip(adapter)) {
1373 req->hdr.version = 1;
1374 } else if (BEx_chip(adapter)) {
1375 if (adapter->function_caps & BE_FUNCTION_CAPS_SUPER_NIC)
1376 req->hdr.version = 2;
1377 } else { /* For SH */
1378 req->hdr.version = 2;
1379 }
1380
1381 if (req->hdr.version > 0)
1382 req->if_id = cpu_to_le16(adapter->if_handle);
1383 req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size);
1384 req->ulp_num = BE_ULP1_NUM;
1385 req->type = BE_ETH_TX_RING_TYPE_STANDARD;
1386 req->cq_id = cpu_to_le16(cq->id);
1387 req->queue_size = be_encoded_q_len(txq->len);
1388 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1389 ver = req->hdr.version;
1390
1391 status = be_cmd_notify_wait(adapter, &wrb);
1392 if (!status) {
1393 struct be_cmd_resp_eth_tx_create *resp = embedded_payload(&wrb);
1394
1395 txq->id = le16_to_cpu(resp->cid);
1396 if (ver == 2)
1397 txo->db_offset = le32_to_cpu(resp->db_offset);
1398 else
1399 txo->db_offset = DB_TXULP1_OFFSET;
1400 txq->created = true;
1401 }
1402
1403 return status;
1404 }
1405
1406 /* Uses MCC */
1407 int be_cmd_rxq_create(struct be_adapter *adapter,
1408 struct be_queue_info *rxq, u16 cq_id, u16 frag_size,
1409 u32 if_id, u32 rss, u8 *rss_id)
1410 {
1411 struct be_mcc_wrb *wrb;
1412 struct be_cmd_req_eth_rx_create *req;
1413 struct be_dma_mem *q_mem = &rxq->dma_mem;
1414 int status;
1415
1416 mutex_lock(&adapter->mcc_lock);
1417
1418 wrb = wrb_from_mccq(adapter);
1419 if (!wrb) {
1420 status = -EBUSY;
1421 goto err;
1422 }
1423 req = embedded_payload(wrb);
1424
1425 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1426 OPCODE_ETH_RX_CREATE, sizeof(*req), wrb, NULL);
1427
1428 req->cq_id = cpu_to_le16(cq_id);
1429 req->frag_size = fls(frag_size) - 1;
1430 req->num_pages = 2;
1431 be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1432 req->interface_id = cpu_to_le32(if_id);
1433 req->max_frame_size = cpu_to_le16(BE_MAX_JUMBO_FRAME_SIZE);
1434 req->rss_queue = cpu_to_le32(rss);
1435
1436 status = be_mcc_notify_wait(adapter);
1437 if (!status) {
1438 struct be_cmd_resp_eth_rx_create *resp = embedded_payload(wrb);
1439
1440 rxq->id = le16_to_cpu(resp->id);
1441 rxq->created = true;
1442 *rss_id = resp->rss_id;
1443 }
1444
1445 err:
1446 mutex_unlock(&adapter->mcc_lock);
1447 return status;
1448 }
1449
1450 /* Generic destroyer function for all types of queues
1451 * Uses Mbox
1452 */
1453 int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q,
1454 int queue_type)
1455 {
1456 struct be_mcc_wrb *wrb;
1457 struct be_cmd_req_q_destroy *req;
1458 u8 subsys = 0, opcode = 0;
1459 int status;
1460
1461 if (mutex_lock_interruptible(&adapter->mbox_lock))
1462 return -1;
1463
1464 wrb = wrb_from_mbox(adapter);
1465 req = embedded_payload(wrb);
1466
1467 switch (queue_type) {
1468 case QTYPE_EQ:
1469 subsys = CMD_SUBSYSTEM_COMMON;
1470 opcode = OPCODE_COMMON_EQ_DESTROY;
1471 break;
1472 case QTYPE_CQ:
1473 subsys = CMD_SUBSYSTEM_COMMON;
1474 opcode = OPCODE_COMMON_CQ_DESTROY;
1475 break;
1476 case QTYPE_TXQ:
1477 subsys = CMD_SUBSYSTEM_ETH;
1478 opcode = OPCODE_ETH_TX_DESTROY;
1479 break;
1480 case QTYPE_RXQ:
1481 subsys = CMD_SUBSYSTEM_ETH;
1482 opcode = OPCODE_ETH_RX_DESTROY;
1483 break;
1484 case QTYPE_MCCQ:
1485 subsys = CMD_SUBSYSTEM_COMMON;
1486 opcode = OPCODE_COMMON_MCC_DESTROY;
1487 break;
1488 default:
1489 BUG();
1490 }
1491
1492 be_wrb_cmd_hdr_prepare(&req->hdr, subsys, opcode, sizeof(*req), wrb,
1493 NULL);
1494 req->id = cpu_to_le16(q->id);
1495
1496 status = be_mbox_notify_wait(adapter);
1497 q->created = false;
1498
1499 mutex_unlock(&adapter->mbox_lock);
1500 return status;
1501 }
1502
1503 /* Uses MCC */
1504 int be_cmd_rxq_destroy(struct be_adapter *adapter, struct be_queue_info *q)
1505 {
1506 struct be_mcc_wrb *wrb;
1507 struct be_cmd_req_q_destroy *req;
1508 int status;
1509
1510 mutex_lock(&adapter->mcc_lock);
1511
1512 wrb = wrb_from_mccq(adapter);
1513 if (!wrb) {
1514 status = -EBUSY;
1515 goto err;
1516 }
1517 req = embedded_payload(wrb);
1518
1519 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1520 OPCODE_ETH_RX_DESTROY, sizeof(*req), wrb, NULL);
1521 req->id = cpu_to_le16(q->id);
1522
1523 status = be_mcc_notify_wait(adapter);
1524 q->created = false;
1525
1526 err:
1527 mutex_unlock(&adapter->mcc_lock);
1528 return status;
1529 }
1530
1531 /* Create an rx filtering policy configuration on an i/f
1532 * Will use MBOX only if MCCQ has not been created.
1533 */
1534 int be_cmd_if_create(struct be_adapter *adapter, u32 cap_flags, u32 en_flags,
1535 u32 *if_handle, u32 domain)
1536 {
1537 struct be_mcc_wrb wrb = {0};
1538 struct be_cmd_req_if_create *req;
1539 int status;
1540
1541 req = embedded_payload(&wrb);
1542 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1543 OPCODE_COMMON_NTWK_INTERFACE_CREATE,
1544 sizeof(*req), &wrb, NULL);
1545 req->hdr.domain = domain;
1546 req->capability_flags = cpu_to_le32(cap_flags);
1547 req->enable_flags = cpu_to_le32(en_flags);
1548 req->pmac_invalid = true;
1549
1550 status = be_cmd_notify_wait(adapter, &wrb);
1551 if (!status) {
1552 struct be_cmd_resp_if_create *resp = embedded_payload(&wrb);
1553
1554 *if_handle = le32_to_cpu(resp->interface_id);
1555
1556 /* Hack to retrieve VF's pmac-id on BE3 */
1557 if (BE3_chip(adapter) && be_virtfn(adapter))
1558 adapter->pmac_id[0] = le32_to_cpu(resp->pmac_id);
1559 }
1560 return status;
1561 }
1562
1563 /* Uses MCCQ if available else MBOX */
1564 int be_cmd_if_destroy(struct be_adapter *adapter, int interface_id, u32 domain)
1565 {
1566 struct be_mcc_wrb wrb = {0};
1567 struct be_cmd_req_if_destroy *req;
1568 int status;
1569
1570 if (interface_id == -1)
1571 return 0;
1572
1573 req = embedded_payload(&wrb);
1574
1575 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1576 OPCODE_COMMON_NTWK_INTERFACE_DESTROY,
1577 sizeof(*req), &wrb, NULL);
1578 req->hdr.domain = domain;
1579 req->interface_id = cpu_to_le32(interface_id);
1580
1581 status = be_cmd_notify_wait(adapter, &wrb);
1582 return status;
1583 }
1584
1585 /* Get stats is a non embedded command: the request is not embedded inside
1586 * WRB but is a separate dma memory block
1587 * Uses asynchronous MCC
1588 */
1589 int be_cmd_get_stats(struct be_adapter *adapter, struct be_dma_mem *nonemb_cmd)
1590 {
1591 struct be_mcc_wrb *wrb;
1592 struct be_cmd_req_hdr *hdr;
1593 int status = 0;
1594
1595 mutex_lock(&adapter->mcc_lock);
1596
1597 wrb = wrb_from_mccq(adapter);
1598 if (!wrb) {
1599 status = -EBUSY;
1600 goto err;
1601 }
1602 hdr = nonemb_cmd->va;
1603
1604 be_wrb_cmd_hdr_prepare(hdr, CMD_SUBSYSTEM_ETH,
1605 OPCODE_ETH_GET_STATISTICS, nonemb_cmd->size, wrb,
1606 nonemb_cmd);
1607
1608 /* version 1 of the cmd is not supported only by BE2 */
1609 if (BE2_chip(adapter))
1610 hdr->version = 0;
1611 if (BE3_chip(adapter) || lancer_chip(adapter))
1612 hdr->version = 1;
1613 else
1614 hdr->version = 2;
1615
1616 status = be_mcc_notify(adapter);
1617 if (status)
1618 goto err;
1619
1620 adapter->stats_cmd_sent = true;
1621
1622 err:
1623 mutex_unlock(&adapter->mcc_lock);
1624 return status;
1625 }
1626
1627 /* Lancer Stats */
1628 int lancer_cmd_get_pport_stats(struct be_adapter *adapter,
1629 struct be_dma_mem *nonemb_cmd)
1630 {
1631 struct be_mcc_wrb *wrb;
1632 struct lancer_cmd_req_pport_stats *req;
1633 int status = 0;
1634
1635 if (!be_cmd_allowed(adapter, OPCODE_ETH_GET_PPORT_STATS,
1636 CMD_SUBSYSTEM_ETH))
1637 return -EPERM;
1638
1639 mutex_lock(&adapter->mcc_lock);
1640
1641 wrb = wrb_from_mccq(adapter);
1642 if (!wrb) {
1643 status = -EBUSY;
1644 goto err;
1645 }
1646 req = nonemb_cmd->va;
1647
1648 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1649 OPCODE_ETH_GET_PPORT_STATS, nonemb_cmd->size,
1650 wrb, nonemb_cmd);
1651
1652 req->cmd_params.params.pport_num = cpu_to_le16(adapter->hba_port_num);
1653 req->cmd_params.params.reset_stats = 0;
1654
1655 status = be_mcc_notify(adapter);
1656 if (status)
1657 goto err;
1658
1659 adapter->stats_cmd_sent = true;
1660
1661 err:
1662 mutex_unlock(&adapter->mcc_lock);
1663 return status;
1664 }
1665
1666 static int be_mac_to_link_speed(int mac_speed)
1667 {
1668 switch (mac_speed) {
1669 case PHY_LINK_SPEED_ZERO:
1670 return 0;
1671 case PHY_LINK_SPEED_10MBPS:
1672 return 10;
1673 case PHY_LINK_SPEED_100MBPS:
1674 return 100;
1675 case PHY_LINK_SPEED_1GBPS:
1676 return 1000;
1677 case PHY_LINK_SPEED_10GBPS:
1678 return 10000;
1679 case PHY_LINK_SPEED_20GBPS:
1680 return 20000;
1681 case PHY_LINK_SPEED_25GBPS:
1682 return 25000;
1683 case PHY_LINK_SPEED_40GBPS:
1684 return 40000;
1685 }
1686 return 0;
1687 }
1688
1689 /* Uses synchronous mcc
1690 * Returns link_speed in Mbps
1691 */
1692 int be_cmd_link_status_query(struct be_adapter *adapter, u16 *link_speed,
1693 u8 *link_status, u32 dom)
1694 {
1695 struct be_mcc_wrb *wrb;
1696 struct be_cmd_req_link_status *req;
1697 int status;
1698
1699 mutex_lock(&adapter->mcc_lock);
1700
1701 if (link_status)
1702 *link_status = LINK_DOWN;
1703
1704 wrb = wrb_from_mccq(adapter);
1705 if (!wrb) {
1706 status = -EBUSY;
1707 goto err;
1708 }
1709 req = embedded_payload(wrb);
1710
1711 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1712 OPCODE_COMMON_NTWK_LINK_STATUS_QUERY,
1713 sizeof(*req), wrb, NULL);
1714
1715 /* version 1 of the cmd is not supported only by BE2 */
1716 if (!BE2_chip(adapter))
1717 req->hdr.version = 1;
1718
1719 req->hdr.domain = dom;
1720
1721 status = be_mcc_notify_wait(adapter);
1722 if (!status) {
1723 struct be_cmd_resp_link_status *resp = embedded_payload(wrb);
1724
1725 if (link_speed) {
1726 *link_speed = resp->link_speed ?
1727 le16_to_cpu(resp->link_speed) * 10 :
1728 be_mac_to_link_speed(resp->mac_speed);
1729
1730 if (!resp->logical_link_status)
1731 *link_speed = 0;
1732 }
1733 if (link_status)
1734 *link_status = resp->logical_link_status;
1735 }
1736
1737 err:
1738 mutex_unlock(&adapter->mcc_lock);
1739 return status;
1740 }
1741
1742 /* Uses synchronous mcc */
1743 int be_cmd_get_die_temperature(struct be_adapter *adapter)
1744 {
1745 struct be_mcc_wrb *wrb;
1746 struct be_cmd_req_get_cntl_addnl_attribs *req;
1747 int status = 0;
1748
1749 mutex_lock(&adapter->mcc_lock);
1750
1751 wrb = wrb_from_mccq(adapter);
1752 if (!wrb) {
1753 status = -EBUSY;
1754 goto err;
1755 }
1756 req = embedded_payload(wrb);
1757
1758 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1759 OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES,
1760 sizeof(*req), wrb, NULL);
1761
1762 status = be_mcc_notify(adapter);
1763 err:
1764 mutex_unlock(&adapter->mcc_lock);
1765 return status;
1766 }
1767
1768 /* Uses synchronous mcc */
1769 int be_cmd_get_fat_dump_len(struct be_adapter *adapter, u32 *dump_size)
1770 {
1771 struct be_mcc_wrb wrb = {0};
1772 struct be_cmd_req_get_fat *req;
1773 int status;
1774
1775 req = embedded_payload(&wrb);
1776
1777 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1778 OPCODE_COMMON_MANAGE_FAT, sizeof(*req),
1779 &wrb, NULL);
1780 req->fat_operation = cpu_to_le32(QUERY_FAT);
1781 status = be_cmd_notify_wait(adapter, &wrb);
1782 if (!status) {
1783 struct be_cmd_resp_get_fat *resp = embedded_payload(&wrb);
1784
1785 if (dump_size && resp->log_size)
1786 *dump_size = le32_to_cpu(resp->log_size) -
1787 sizeof(u32);
1788 }
1789 return status;
1790 }
1791
1792 int be_cmd_get_fat_dump(struct be_adapter *adapter, u32 buf_len, void *buf)
1793 {
1794 struct be_dma_mem get_fat_cmd;
1795 struct be_mcc_wrb *wrb;
1796 struct be_cmd_req_get_fat *req;
1797 u32 offset = 0, total_size, buf_size,
1798 log_offset = sizeof(u32), payload_len;
1799 int status;
1800
1801 if (buf_len == 0)
1802 return 0;
1803
1804 total_size = buf_len;
1805
1806 get_fat_cmd.size = sizeof(struct be_cmd_req_get_fat) + 60*1024;
1807 get_fat_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
1808 get_fat_cmd.size,
1809 &get_fat_cmd.dma, GFP_ATOMIC);
1810 if (!get_fat_cmd.va)
1811 return -ENOMEM;
1812
1813 mutex_lock(&adapter->mcc_lock);
1814
1815 while (total_size) {
1816 buf_size = min(total_size, (u32)60*1024);
1817 total_size -= buf_size;
1818
1819 wrb = wrb_from_mccq(adapter);
1820 if (!wrb) {
1821 status = -EBUSY;
1822 goto err;
1823 }
1824 req = get_fat_cmd.va;
1825
1826 payload_len = sizeof(struct be_cmd_req_get_fat) + buf_size;
1827 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1828 OPCODE_COMMON_MANAGE_FAT, payload_len,
1829 wrb, &get_fat_cmd);
1830
1831 req->fat_operation = cpu_to_le32(RETRIEVE_FAT);
1832 req->read_log_offset = cpu_to_le32(log_offset);
1833 req->read_log_length = cpu_to_le32(buf_size);
1834 req->data_buffer_size = cpu_to_le32(buf_size);
1835
1836 status = be_mcc_notify_wait(adapter);
1837 if (!status) {
1838 struct be_cmd_resp_get_fat *resp = get_fat_cmd.va;
1839
1840 memcpy(buf + offset,
1841 resp->data_buffer,
1842 le32_to_cpu(resp->read_log_length));
1843 } else {
1844 dev_err(&adapter->pdev->dev, "FAT Table Retrieve error\n");
1845 goto err;
1846 }
1847 offset += buf_size;
1848 log_offset += buf_size;
1849 }
1850 err:
1851 dma_free_coherent(&adapter->pdev->dev, get_fat_cmd.size,
1852 get_fat_cmd.va, get_fat_cmd.dma);
1853 mutex_unlock(&adapter->mcc_lock);
1854 return status;
1855 }
1856
1857 /* Uses synchronous mcc */
1858 int be_cmd_get_fw_ver(struct be_adapter *adapter)
1859 {
1860 struct be_mcc_wrb *wrb;
1861 struct be_cmd_req_get_fw_version *req;
1862 int status;
1863
1864 mutex_lock(&adapter->mcc_lock);
1865
1866 wrb = wrb_from_mccq(adapter);
1867 if (!wrb) {
1868 status = -EBUSY;
1869 goto err;
1870 }
1871
1872 req = embedded_payload(wrb);
1873
1874 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1875 OPCODE_COMMON_GET_FW_VERSION, sizeof(*req), wrb,
1876 NULL);
1877 status = be_mcc_notify_wait(adapter);
1878 if (!status) {
1879 struct be_cmd_resp_get_fw_version *resp = embedded_payload(wrb);
1880
1881 strlcpy(adapter->fw_ver, resp->firmware_version_string,
1882 sizeof(adapter->fw_ver));
1883 strlcpy(adapter->fw_on_flash, resp->fw_on_flash_version_string,
1884 sizeof(adapter->fw_on_flash));
1885 }
1886 err:
1887 mutex_unlock(&adapter->mcc_lock);
1888 return status;
1889 }
1890
1891 /* set the EQ delay interval of an EQ to specified value
1892 * Uses async mcc
1893 */
1894 static int __be_cmd_modify_eqd(struct be_adapter *adapter,
1895 struct be_set_eqd *set_eqd, int num)
1896 {
1897 struct be_mcc_wrb *wrb;
1898 struct be_cmd_req_modify_eq_delay *req;
1899 int status = 0, i;
1900
1901 mutex_lock(&adapter->mcc_lock);
1902
1903 wrb = wrb_from_mccq(adapter);
1904 if (!wrb) {
1905 status = -EBUSY;
1906 goto err;
1907 }
1908 req = embedded_payload(wrb);
1909
1910 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1911 OPCODE_COMMON_MODIFY_EQ_DELAY, sizeof(*req), wrb,
1912 NULL);
1913
1914 req->num_eq = cpu_to_le32(num);
1915 for (i = 0; i < num; i++) {
1916 req->set_eqd[i].eq_id = cpu_to_le32(set_eqd[i].eq_id);
1917 req->set_eqd[i].phase = 0;
1918 req->set_eqd[i].delay_multiplier =
1919 cpu_to_le32(set_eqd[i].delay_multiplier);
1920 }
1921
1922 status = be_mcc_notify(adapter);
1923 err:
1924 mutex_unlock(&adapter->mcc_lock);
1925 return status;
1926 }
1927
1928 int be_cmd_modify_eqd(struct be_adapter *adapter, struct be_set_eqd *set_eqd,
1929 int num)
1930 {
1931 int num_eqs, i = 0;
1932
1933 while (num) {
1934 num_eqs = min(num, 8);
1935 __be_cmd_modify_eqd(adapter, &set_eqd[i], num_eqs);
1936 i += num_eqs;
1937 num -= num_eqs;
1938 }
1939
1940 return 0;
1941 }
1942
1943 /* Uses sycnhronous mcc */
1944 int be_cmd_vlan_config(struct be_adapter *adapter, u32 if_id, u16 *vtag_array,
1945 u32 num, u32 domain)
1946 {
1947 struct be_mcc_wrb *wrb;
1948 struct be_cmd_req_vlan_config *req;
1949 int status;
1950
1951 mutex_lock(&adapter->mcc_lock);
1952
1953 wrb = wrb_from_mccq(adapter);
1954 if (!wrb) {
1955 status = -EBUSY;
1956 goto err;
1957 }
1958 req = embedded_payload(wrb);
1959
1960 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1961 OPCODE_COMMON_NTWK_VLAN_CONFIG, sizeof(*req),
1962 wrb, NULL);
1963 req->hdr.domain = domain;
1964
1965 req->interface_id = if_id;
1966 req->untagged = BE_IF_FLAGS_UNTAGGED & be_if_cap_flags(adapter) ? 1 : 0;
1967 req->num_vlan = num;
1968 memcpy(req->normal_vlan, vtag_array,
1969 req->num_vlan * sizeof(vtag_array[0]));
1970
1971 status = be_mcc_notify_wait(adapter);
1972 err:
1973 mutex_unlock(&adapter->mcc_lock);
1974 return status;
1975 }
1976
1977 static int __be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 value)
1978 {
1979 struct be_mcc_wrb *wrb;
1980 struct be_dma_mem *mem = &adapter->rx_filter;
1981 struct be_cmd_req_rx_filter *req = mem->va;
1982 int status;
1983
1984 mutex_lock(&adapter->mcc_lock);
1985
1986 wrb = wrb_from_mccq(adapter);
1987 if (!wrb) {
1988 status = -EBUSY;
1989 goto err;
1990 }
1991 memset(req, 0, sizeof(*req));
1992 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1993 OPCODE_COMMON_NTWK_RX_FILTER, sizeof(*req),
1994 wrb, mem);
1995
1996 req->if_id = cpu_to_le32(adapter->if_handle);
1997 req->if_flags_mask = cpu_to_le32(flags);
1998 req->if_flags = (value == ON) ? req->if_flags_mask : 0;
1999
2000 if (flags & BE_IF_FLAGS_MULTICAST) {
2001 int i;
2002
2003 /* Reset mcast promisc mode if already set by setting mask
2004 * and not setting flags field
2005 */
2006 req->if_flags_mask |=
2007 cpu_to_le32(BE_IF_FLAGS_MCAST_PROMISCUOUS &
2008 be_if_cap_flags(adapter));
2009 req->mcast_num = cpu_to_le32(adapter->mc_count);
2010 for (i = 0; i < adapter->mc_count; i++)
2011 ether_addr_copy(req->mcast_mac[i].byte,
2012 adapter->mc_list[i].mac);
2013 }
2014
2015 status = be_mcc_notify_wait(adapter);
2016 err:
2017 mutex_unlock(&adapter->mcc_lock);
2018 return status;
2019 }
2020
2021 int be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 value)
2022 {
2023 struct device *dev = &adapter->pdev->dev;
2024
2025 if ((flags & be_if_cap_flags(adapter)) != flags) {
2026 dev_warn(dev, "Cannot set rx filter flags 0x%x\n", flags);
2027 dev_warn(dev, "Interface is capable of 0x%x flags only\n",
2028 be_if_cap_flags(adapter));
2029 }
2030 flags &= be_if_cap_flags(adapter);
2031 if (!flags)
2032 return -ENOTSUPP;
2033
2034 return __be_cmd_rx_filter(adapter, flags, value);
2035 }
2036
2037 /* Uses synchrounous mcc */
2038 int be_cmd_set_flow_control(struct be_adapter *adapter, u32 tx_fc, u32 rx_fc)
2039 {
2040 struct be_mcc_wrb *wrb;
2041 struct be_cmd_req_set_flow_control *req;
2042 int status;
2043
2044 if (!be_cmd_allowed(adapter, OPCODE_COMMON_SET_FLOW_CONTROL,
2045 CMD_SUBSYSTEM_COMMON))
2046 return -EPERM;
2047
2048 mutex_lock(&adapter->mcc_lock);
2049
2050 wrb = wrb_from_mccq(adapter);
2051 if (!wrb) {
2052 status = -EBUSY;
2053 goto err;
2054 }
2055 req = embedded_payload(wrb);
2056
2057 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2058 OPCODE_COMMON_SET_FLOW_CONTROL, sizeof(*req),
2059 wrb, NULL);
2060
2061 req->hdr.version = 1;
2062 req->tx_flow_control = cpu_to_le16((u16)tx_fc);
2063 req->rx_flow_control = cpu_to_le16((u16)rx_fc);
2064
2065 status = be_mcc_notify_wait(adapter);
2066
2067 err:
2068 mutex_unlock(&adapter->mcc_lock);
2069
2070 if (base_status(status) == MCC_STATUS_FEATURE_NOT_SUPPORTED)
2071 return -EOPNOTSUPP;
2072
2073 return status;
2074 }
2075
2076 /* Uses sycn mcc */
2077 int be_cmd_get_flow_control(struct be_adapter *adapter, u32 *tx_fc, u32 *rx_fc)
2078 {
2079 struct be_mcc_wrb *wrb;
2080 struct be_cmd_req_get_flow_control *req;
2081 int status;
2082
2083 if (!be_cmd_allowed(adapter, OPCODE_COMMON_GET_FLOW_CONTROL,
2084 CMD_SUBSYSTEM_COMMON))
2085 return -EPERM;
2086
2087 mutex_lock(&adapter->mcc_lock);
2088
2089 wrb = wrb_from_mccq(adapter);
2090 if (!wrb) {
2091 status = -EBUSY;
2092 goto err;
2093 }
2094 req = embedded_payload(wrb);
2095
2096 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2097 OPCODE_COMMON_GET_FLOW_CONTROL, sizeof(*req),
2098 wrb, NULL);
2099
2100 status = be_mcc_notify_wait(adapter);
2101 if (!status) {
2102 struct be_cmd_resp_get_flow_control *resp =
2103 embedded_payload(wrb);
2104
2105 *tx_fc = le16_to_cpu(resp->tx_flow_control);
2106 *rx_fc = le16_to_cpu(resp->rx_flow_control);
2107 }
2108
2109 err:
2110 mutex_unlock(&adapter->mcc_lock);
2111 return status;
2112 }
2113
2114 /* Uses mbox */
2115 int be_cmd_query_fw_cfg(struct be_adapter *adapter)
2116 {
2117 struct be_mcc_wrb *wrb;
2118 struct be_cmd_req_query_fw_cfg *req;
2119 int status;
2120
2121 if (mutex_lock_interruptible(&adapter->mbox_lock))
2122 return -1;
2123
2124 wrb = wrb_from_mbox(adapter);
2125 req = embedded_payload(wrb);
2126
2127 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2128 OPCODE_COMMON_QUERY_FIRMWARE_CONFIG,
2129 sizeof(*req), wrb, NULL);
2130
2131 status = be_mbox_notify_wait(adapter);
2132 if (!status) {
2133 struct be_cmd_resp_query_fw_cfg *resp = embedded_payload(wrb);
2134
2135 adapter->port_num = le32_to_cpu(resp->phys_port);
2136 adapter->function_mode = le32_to_cpu(resp->function_mode);
2137 adapter->function_caps = le32_to_cpu(resp->function_caps);
2138 adapter->asic_rev = le32_to_cpu(resp->asic_revision) & 0xFF;
2139 dev_info(&adapter->pdev->dev,
2140 "FW config: function_mode=0x%x, function_caps=0x%x\n",
2141 adapter->function_mode, adapter->function_caps);
2142 }
2143
2144 mutex_unlock(&adapter->mbox_lock);
2145 return status;
2146 }
2147
2148 /* Uses mbox */
2149 int be_cmd_reset_function(struct be_adapter *adapter)
2150 {
2151 struct be_mcc_wrb *wrb;
2152 struct be_cmd_req_hdr *req;
2153 int status;
2154
2155 if (lancer_chip(adapter)) {
2156 iowrite32(SLI_PORT_CONTROL_IP_MASK,
2157 adapter->db + SLIPORT_CONTROL_OFFSET);
2158 status = lancer_wait_ready(adapter);
2159 if (status)
2160 dev_err(&adapter->pdev->dev,
2161 "Adapter in non recoverable error\n");
2162 return status;
2163 }
2164
2165 if (mutex_lock_interruptible(&adapter->mbox_lock))
2166 return -1;
2167
2168 wrb = wrb_from_mbox(adapter);
2169 req = embedded_payload(wrb);
2170
2171 be_wrb_cmd_hdr_prepare(req, CMD_SUBSYSTEM_COMMON,
2172 OPCODE_COMMON_FUNCTION_RESET, sizeof(*req), wrb,
2173 NULL);
2174
2175 status = be_mbox_notify_wait(adapter);
2176
2177 mutex_unlock(&adapter->mbox_lock);
2178 return status;
2179 }
2180
2181 int be_cmd_rss_config(struct be_adapter *adapter, u8 *rsstable,
2182 u32 rss_hash_opts, u16 table_size, const u8 *rss_hkey)
2183 {
2184 struct be_mcc_wrb *wrb;
2185 struct be_cmd_req_rss_config *req;
2186 int status;
2187
2188 if (!(be_if_cap_flags(adapter) & BE_IF_FLAGS_RSS))
2189 return 0;
2190
2191 mutex_lock(&adapter->mcc_lock);
2192
2193 wrb = wrb_from_mccq(adapter);
2194 if (!wrb) {
2195 status = -EBUSY;
2196 goto err;
2197 }
2198 req = embedded_payload(wrb);
2199
2200 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
2201 OPCODE_ETH_RSS_CONFIG, sizeof(*req), wrb, NULL);
2202
2203 req->if_id = cpu_to_le32(adapter->if_handle);
2204 req->enable_rss = cpu_to_le16(rss_hash_opts);
2205 req->cpu_table_size_log2 = cpu_to_le16(fls(table_size) - 1);
2206
2207 if (!BEx_chip(adapter))
2208 req->hdr.version = 1;
2209
2210 memcpy(req->cpu_table, rsstable, table_size);
2211 memcpy(req->hash, rss_hkey, RSS_HASH_KEY_LEN);
2212 be_dws_cpu_to_le(req->hash, sizeof(req->hash));
2213
2214 status = be_mcc_notify_wait(adapter);
2215 err:
2216 mutex_unlock(&adapter->mcc_lock);
2217 return status;
2218 }
2219
2220 /* Uses sync mcc */
2221 int be_cmd_set_beacon_state(struct be_adapter *adapter, u8 port_num,
2222 u8 bcn, u8 sts, u8 state)
2223 {
2224 struct be_mcc_wrb *wrb;
2225 struct be_cmd_req_enable_disable_beacon *req;
2226 int status;
2227
2228 mutex_lock(&adapter->mcc_lock);
2229
2230 wrb = wrb_from_mccq(adapter);
2231 if (!wrb) {
2232 status = -EBUSY;
2233 goto err;
2234 }
2235 req = embedded_payload(wrb);
2236
2237 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2238 OPCODE_COMMON_ENABLE_DISABLE_BEACON,
2239 sizeof(*req), wrb, NULL);
2240
2241 req->port_num = port_num;
2242 req->beacon_state = state;
2243 req->beacon_duration = bcn;
2244 req->status_duration = sts;
2245
2246 status = be_mcc_notify_wait(adapter);
2247
2248 err:
2249 mutex_unlock(&adapter->mcc_lock);
2250 return status;
2251 }
2252
2253 /* Uses sync mcc */
2254 int be_cmd_get_beacon_state(struct be_adapter *adapter, u8 port_num, u32 *state)
2255 {
2256 struct be_mcc_wrb *wrb;
2257 struct be_cmd_req_get_beacon_state *req;
2258 int status;
2259
2260 mutex_lock(&adapter->mcc_lock);
2261
2262 wrb = wrb_from_mccq(adapter);
2263 if (!wrb) {
2264 status = -EBUSY;
2265 goto err;
2266 }
2267 req = embedded_payload(wrb);
2268
2269 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2270 OPCODE_COMMON_GET_BEACON_STATE, sizeof(*req),
2271 wrb, NULL);
2272
2273 req->port_num = port_num;
2274
2275 status = be_mcc_notify_wait(adapter);
2276 if (!status) {
2277 struct be_cmd_resp_get_beacon_state *resp =
2278 embedded_payload(wrb);
2279
2280 *state = resp->beacon_state;
2281 }
2282
2283 err:
2284 mutex_unlock(&adapter->mcc_lock);
2285 return status;
2286 }
2287
2288 /* Uses sync mcc */
2289 int be_cmd_read_port_transceiver_data(struct be_adapter *adapter,
2290 u8 page_num, u8 *data)
2291 {
2292 struct be_dma_mem cmd;
2293 struct be_mcc_wrb *wrb;
2294 struct be_cmd_req_port_type *req;
2295 int status;
2296
2297 if (page_num > TR_PAGE_A2)
2298 return -EINVAL;
2299
2300 cmd.size = sizeof(struct be_cmd_resp_port_type);
2301 cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
2302 GFP_ATOMIC);
2303 if (!cmd.va) {
2304 dev_err(&adapter->pdev->dev, "Memory allocation failed\n");
2305 return -ENOMEM;
2306 }
2307
2308 mutex_lock(&adapter->mcc_lock);
2309
2310 wrb = wrb_from_mccq(adapter);
2311 if (!wrb) {
2312 status = -EBUSY;
2313 goto err;
2314 }
2315 req = cmd.va;
2316
2317 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2318 OPCODE_COMMON_READ_TRANSRECV_DATA,
2319 cmd.size, wrb, &cmd);
2320
2321 req->port = cpu_to_le32(adapter->hba_port_num);
2322 req->page_num = cpu_to_le32(page_num);
2323 status = be_mcc_notify_wait(adapter);
2324 if (!status) {
2325 struct be_cmd_resp_port_type *resp = cmd.va;
2326
2327 memcpy(data, resp->page_data, PAGE_DATA_LEN);
2328 }
2329 err:
2330 mutex_unlock(&adapter->mcc_lock);
2331 dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va, cmd.dma);
2332 return status;
2333 }
2334
2335 static int lancer_cmd_write_object(struct be_adapter *adapter,
2336 struct be_dma_mem *cmd, u32 data_size,
2337 u32 data_offset, const char *obj_name,
2338 u32 *data_written, u8 *change_status,
2339 u8 *addn_status)
2340 {
2341 struct be_mcc_wrb *wrb;
2342 struct lancer_cmd_req_write_object *req;
2343 struct lancer_cmd_resp_write_object *resp;
2344 void *ctxt = NULL;
2345 int status;
2346
2347 mutex_lock(&adapter->mcc_lock);
2348 adapter->flash_status = 0;
2349
2350 wrb = wrb_from_mccq(adapter);
2351 if (!wrb) {
2352 status = -EBUSY;
2353 goto err_unlock;
2354 }
2355
2356 req = embedded_payload(wrb);
2357
2358 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2359 OPCODE_COMMON_WRITE_OBJECT,
2360 sizeof(struct lancer_cmd_req_write_object), wrb,
2361 NULL);
2362
2363 ctxt = &req->context;
2364 AMAP_SET_BITS(struct amap_lancer_write_obj_context,
2365 write_length, ctxt, data_size);
2366
2367 if (data_size == 0)
2368 AMAP_SET_BITS(struct amap_lancer_write_obj_context,
2369 eof, ctxt, 1);
2370 else
2371 AMAP_SET_BITS(struct amap_lancer_write_obj_context,
2372 eof, ctxt, 0);
2373
2374 be_dws_cpu_to_le(ctxt, sizeof(req->context));
2375 req->write_offset = cpu_to_le32(data_offset);
2376 strlcpy(req->object_name, obj_name, sizeof(req->object_name));
2377 req->descriptor_count = cpu_to_le32(1);
2378 req->buf_len = cpu_to_le32(data_size);
2379 req->addr_low = cpu_to_le32((cmd->dma +
2380 sizeof(struct lancer_cmd_req_write_object))
2381 & 0xFFFFFFFF);
2382 req->addr_high = cpu_to_le32(upper_32_bits(cmd->dma +
2383 sizeof(struct lancer_cmd_req_write_object)));
2384
2385 status = be_mcc_notify(adapter);
2386 if (status)
2387 goto err_unlock;
2388
2389 mutex_unlock(&adapter->mcc_lock);
2390
2391 if (!wait_for_completion_timeout(&adapter->et_cmd_compl,
2392 msecs_to_jiffies(60000)))
2393 status = -ETIMEDOUT;
2394 else
2395 status = adapter->flash_status;
2396
2397 resp = embedded_payload(wrb);
2398 if (!status) {
2399 *data_written = le32_to_cpu(resp->actual_write_len);
2400 *change_status = resp->change_status;
2401 } else {
2402 *addn_status = resp->additional_status;
2403 }
2404
2405 return status;
2406
2407 err_unlock:
2408 mutex_unlock(&adapter->mcc_lock);
2409 return status;
2410 }
2411
2412 int be_cmd_query_cable_type(struct be_adapter *adapter)
2413 {
2414 u8 page_data[PAGE_DATA_LEN];
2415 int status;
2416
2417 status = be_cmd_read_port_transceiver_data(adapter, TR_PAGE_A0,
2418 page_data);
2419 if (!status) {
2420 switch (adapter->phy.interface_type) {
2421 case PHY_TYPE_QSFP:
2422 adapter->phy.cable_type =
2423 page_data[QSFP_PLUS_CABLE_TYPE_OFFSET];
2424 break;
2425 case PHY_TYPE_SFP_PLUS_10GB:
2426 adapter->phy.cable_type =
2427 page_data[SFP_PLUS_CABLE_TYPE_OFFSET];
2428 break;
2429 default:
2430 adapter->phy.cable_type = 0;
2431 break;
2432 }
2433 }
2434 return status;
2435 }
2436
2437 int be_cmd_query_sfp_info(struct be_adapter *adapter)
2438 {
2439 u8 page_data[PAGE_DATA_LEN];
2440 int status;
2441
2442 status = be_cmd_read_port_transceiver_data(adapter, TR_PAGE_A0,
2443 page_data);
2444 if (!status) {
2445 strlcpy(adapter->phy.vendor_name, page_data +
2446 SFP_VENDOR_NAME_OFFSET, SFP_VENDOR_NAME_LEN - 1);
2447 strlcpy(adapter->phy.vendor_pn,
2448 page_data + SFP_VENDOR_PN_OFFSET,
2449 SFP_VENDOR_NAME_LEN - 1);
2450 }
2451
2452 return status;
2453 }
2454
2455 static int lancer_cmd_delete_object(struct be_adapter *adapter,
2456 const char *obj_name)
2457 {
2458 struct lancer_cmd_req_delete_object *req;
2459 struct be_mcc_wrb *wrb;
2460 int status;
2461
2462 mutex_lock(&adapter->mcc_lock);
2463
2464 wrb = wrb_from_mccq(adapter);
2465 if (!wrb) {
2466 status = -EBUSY;
2467 goto err;
2468 }
2469
2470 req = embedded_payload(wrb);
2471
2472 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2473 OPCODE_COMMON_DELETE_OBJECT,
2474 sizeof(*req), wrb, NULL);
2475
2476 strlcpy(req->object_name, obj_name, sizeof(req->object_name));
2477
2478 status = be_mcc_notify_wait(adapter);
2479 err:
2480 mutex_unlock(&adapter->mcc_lock);
2481 return status;
2482 }
2483
2484 int lancer_cmd_read_object(struct be_adapter *adapter, struct be_dma_mem *cmd,
2485 u32 data_size, u32 data_offset, const char *obj_name,
2486 u32 *data_read, u32 *eof, u8 *addn_status)
2487 {
2488 struct be_mcc_wrb *wrb;
2489 struct lancer_cmd_req_read_object *req;
2490 struct lancer_cmd_resp_read_object *resp;
2491 int status;
2492
2493 mutex_lock(&adapter->mcc_lock);
2494
2495 wrb = wrb_from_mccq(adapter);
2496 if (!wrb) {
2497 status = -EBUSY;
2498 goto err_unlock;
2499 }
2500
2501 req = embedded_payload(wrb);
2502
2503 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2504 OPCODE_COMMON_READ_OBJECT,
2505 sizeof(struct lancer_cmd_req_read_object), wrb,
2506 NULL);
2507
2508 req->desired_read_len = cpu_to_le32(data_size);
2509 req->read_offset = cpu_to_le32(data_offset);
2510 strcpy(req->object_name, obj_name);
2511 req->descriptor_count = cpu_to_le32(1);
2512 req->buf_len = cpu_to_le32(data_size);
2513 req->addr_low = cpu_to_le32((cmd->dma & 0xFFFFFFFF));
2514 req->addr_high = cpu_to_le32(upper_32_bits(cmd->dma));
2515
2516 status = be_mcc_notify_wait(adapter);
2517
2518 resp = embedded_payload(wrb);
2519 if (!status) {
2520 *data_read = le32_to_cpu(resp->actual_read_len);
2521 *eof = le32_to_cpu(resp->eof);
2522 } else {
2523 *addn_status = resp->additional_status;
2524 }
2525
2526 err_unlock:
2527 mutex_unlock(&adapter->mcc_lock);
2528 return status;
2529 }
2530
2531 static int be_cmd_write_flashrom(struct be_adapter *adapter,
2532 struct be_dma_mem *cmd, u32 flash_type,
2533 u32 flash_opcode, u32 img_offset, u32 buf_size)
2534 {
2535 struct be_mcc_wrb *wrb;
2536 struct be_cmd_write_flashrom *req;
2537 int status;
2538
2539 mutex_lock(&adapter->mcc_lock);
2540 adapter->flash_status = 0;
2541
2542 wrb = wrb_from_mccq(adapter);
2543 if (!wrb) {
2544 status = -EBUSY;
2545 goto err_unlock;
2546 }
2547 req = cmd->va;
2548
2549 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2550 OPCODE_COMMON_WRITE_FLASHROM, cmd->size, wrb,
2551 cmd);
2552
2553 req->params.op_type = cpu_to_le32(flash_type);
2554 if (flash_type == OPTYPE_OFFSET_SPECIFIED)
2555 req->params.offset = cpu_to_le32(img_offset);
2556
2557 req->params.op_code = cpu_to_le32(flash_opcode);
2558 req->params.data_buf_size = cpu_to_le32(buf_size);
2559
2560 status = be_mcc_notify(adapter);
2561 if (status)
2562 goto err_unlock;
2563
2564 mutex_unlock(&adapter->mcc_lock);
2565
2566 if (!wait_for_completion_timeout(&adapter->et_cmd_compl,
2567 msecs_to_jiffies(40000)))
2568 status = -ETIMEDOUT;
2569 else
2570 status = adapter->flash_status;
2571
2572 return status;
2573
2574 err_unlock:
2575 mutex_unlock(&adapter->mcc_lock);
2576 return status;
2577 }
2578
2579 static int be_cmd_get_flash_crc(struct be_adapter *adapter, u8 *flashed_crc,
2580 u16 img_optype, u32 img_offset, u32 crc_offset)
2581 {
2582 struct be_cmd_read_flash_crc *req;
2583 struct be_mcc_wrb *wrb;
2584 int status;
2585
2586 mutex_lock(&adapter->mcc_lock);
2587
2588 wrb = wrb_from_mccq(adapter);
2589 if (!wrb) {
2590 status = -EBUSY;
2591 goto err;
2592 }
2593 req = embedded_payload(wrb);
2594
2595 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2596 OPCODE_COMMON_READ_FLASHROM, sizeof(*req),
2597 wrb, NULL);
2598
2599 req->params.op_type = cpu_to_le32(img_optype);
2600 if (img_optype == OPTYPE_OFFSET_SPECIFIED)
2601 req->params.offset = cpu_to_le32(img_offset + crc_offset);
2602 else
2603 req->params.offset = cpu_to_le32(crc_offset);
2604
2605 req->params.op_code = cpu_to_le32(FLASHROM_OPER_REPORT);
2606 req->params.data_buf_size = cpu_to_le32(0x4);
2607
2608 status = be_mcc_notify_wait(adapter);
2609 if (!status)
2610 memcpy(flashed_crc, req->crc, 4);
2611
2612 err:
2613 mutex_unlock(&adapter->mcc_lock);
2614 return status;
2615 }
2616
2617 static char flash_cookie[2][16] = {"*** SE FLAS", "H DIRECTORY *** "};
2618
2619 static bool phy_flashing_required(struct be_adapter *adapter)
2620 {
2621 return (adapter->phy.phy_type == PHY_TYPE_TN_8022 &&
2622 adapter->phy.interface_type == PHY_TYPE_BASET_10GB);
2623 }
2624
2625 static bool is_comp_in_ufi(struct be_adapter *adapter,
2626 struct flash_section_info *fsec, int type)
2627 {
2628 int i = 0, img_type = 0;
2629 struct flash_section_info_g2 *fsec_g2 = NULL;
2630
2631 if (BE2_chip(adapter))
2632 fsec_g2 = (struct flash_section_info_g2 *)fsec;
2633
2634 for (i = 0; i < MAX_FLASH_COMP; i++) {
2635 if (fsec_g2)
2636 img_type = le32_to_cpu(fsec_g2->fsec_entry[i].type);
2637 else
2638 img_type = le32_to_cpu(fsec->fsec_entry[i].type);
2639
2640 if (img_type == type)
2641 return true;
2642 }
2643 return false;
2644 }
2645
2646 static struct flash_section_info *get_fsec_info(struct be_adapter *adapter,
2647 int header_size,
2648 const struct firmware *fw)
2649 {
2650 struct flash_section_info *fsec = NULL;
2651 const u8 *p = fw->data;
2652
2653 p += header_size;
2654 while (p < (fw->data + fw->size)) {
2655 fsec = (struct flash_section_info *)p;
2656 if (!memcmp(flash_cookie, fsec->cookie, sizeof(flash_cookie)))
2657 return fsec;
2658 p += 32;
2659 }
2660 return NULL;
2661 }
2662
2663 static int be_check_flash_crc(struct be_adapter *adapter, const u8 *p,
2664 u32 img_offset, u32 img_size, int hdr_size,
2665 u16 img_optype, bool *crc_match)
2666 {
2667 u32 crc_offset;
2668 int status;
2669 u8 crc[4];
2670
2671 status = be_cmd_get_flash_crc(adapter, crc, img_optype, img_offset,
2672 img_size - 4);
2673 if (status)
2674 return status;
2675
2676 crc_offset = hdr_size + img_offset + img_size - 4;
2677
2678 /* Skip flashing, if crc of flashed region matches */
2679 if (!memcmp(crc, p + crc_offset, 4))
2680 *crc_match = true;
2681 else
2682 *crc_match = false;
2683
2684 return status;
2685 }
2686
2687 static int be_flash(struct be_adapter *adapter, const u8 *img,
2688 struct be_dma_mem *flash_cmd, int optype, int img_size,
2689 u32 img_offset)
2690 {
2691 u32 flash_op, num_bytes, total_bytes = img_size, bytes_sent = 0;
2692 struct be_cmd_write_flashrom *req = flash_cmd->va;
2693 int status;
2694
2695 while (total_bytes) {
2696 num_bytes = min_t(u32, 32 * 1024, total_bytes);
2697
2698 total_bytes -= num_bytes;
2699
2700 if (!total_bytes) {
2701 if (optype == OPTYPE_PHY_FW)
2702 flash_op = FLASHROM_OPER_PHY_FLASH;
2703 else
2704 flash_op = FLASHROM_OPER_FLASH;
2705 } else {
2706 if (optype == OPTYPE_PHY_FW)
2707 flash_op = FLASHROM_OPER_PHY_SAVE;
2708 else
2709 flash_op = FLASHROM_OPER_SAVE;
2710 }
2711
2712 memcpy(req->data_buf, img, num_bytes);
2713 img += num_bytes;
2714 status = be_cmd_write_flashrom(adapter, flash_cmd, optype,
2715 flash_op, img_offset +
2716 bytes_sent, num_bytes);
2717 if (base_status(status) == MCC_STATUS_ILLEGAL_REQUEST &&
2718 optype == OPTYPE_PHY_FW)
2719 break;
2720 else if (status)
2721 return status;
2722
2723 bytes_sent += num_bytes;
2724 }
2725 return 0;
2726 }
2727
2728 #define NCSI_UPDATE_LOG "NCSI section update is not supported in FW ver %s\n"
2729 static bool be_fw_ncsi_supported(char *ver)
2730 {
2731 int v1[4] = {3, 102, 148, 0}; /* Min ver that supports NCSI FW */
2732 int v2[4];
2733 int i;
2734
2735 if (sscanf(ver, "%d.%d.%d.%d", &v2[0], &v2[1], &v2[2], &v2[3]) != 4)
2736 return false;
2737
2738 for (i = 0; i < 4; i++) {
2739 if (v1[i] < v2[i])
2740 return true;
2741 else if (v1[i] > v2[i])
2742 return false;
2743 }
2744
2745 return true;
2746 }
2747
2748 /* For BE2, BE3 and BE3-R */
2749 static int be_flash_BEx(struct be_adapter *adapter,
2750 const struct firmware *fw,
2751 struct be_dma_mem *flash_cmd, int num_of_images)
2752 {
2753 int img_hdrs_size = (num_of_images * sizeof(struct image_hdr));
2754 struct device *dev = &adapter->pdev->dev;
2755 struct flash_section_info *fsec = NULL;
2756 int status, i, filehdr_size, num_comp;
2757 const struct flash_comp *pflashcomp;
2758 bool crc_match;
2759 const u8 *p;
2760
2761 static const struct flash_comp gen3_flash_types[] = {
2762 { BE3_ISCSI_PRIMARY_IMAGE_START, OPTYPE_ISCSI_ACTIVE,
2763 BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_ISCSI},
2764 { BE3_REDBOOT_START, OPTYPE_REDBOOT,
2765 BE3_REDBOOT_COMP_MAX_SIZE, IMAGE_BOOT_CODE},
2766 { BE3_ISCSI_BIOS_START, OPTYPE_BIOS,
2767 BE3_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_ISCSI},
2768 { BE3_PXE_BIOS_START, OPTYPE_PXE_BIOS,
2769 BE3_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_PXE},
2770 { BE3_FCOE_BIOS_START, OPTYPE_FCOE_BIOS,
2771 BE3_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_FCOE},
2772 { BE3_ISCSI_BACKUP_IMAGE_START, OPTYPE_ISCSI_BACKUP,
2773 BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_ISCSI},
2774 { BE3_FCOE_PRIMARY_IMAGE_START, OPTYPE_FCOE_FW_ACTIVE,
2775 BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_FCOE},
2776 { BE3_FCOE_BACKUP_IMAGE_START, OPTYPE_FCOE_FW_BACKUP,
2777 BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_FCOE},
2778 { BE3_NCSI_START, OPTYPE_NCSI_FW,
2779 BE3_NCSI_COMP_MAX_SIZE, IMAGE_NCSI},
2780 { BE3_PHY_FW_START, OPTYPE_PHY_FW,
2781 BE3_PHY_FW_COMP_MAX_SIZE, IMAGE_FIRMWARE_PHY}
2782 };
2783
2784 static const struct flash_comp gen2_flash_types[] = {
2785 { BE2_ISCSI_PRIMARY_IMAGE_START, OPTYPE_ISCSI_ACTIVE,
2786 BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_ISCSI},
2787 { BE2_REDBOOT_START, OPTYPE_REDBOOT,
2788 BE2_REDBOOT_COMP_MAX_SIZE, IMAGE_BOOT_CODE},
2789 { BE2_ISCSI_BIOS_START, OPTYPE_BIOS,
2790 BE2_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_ISCSI},
2791 { BE2_PXE_BIOS_START, OPTYPE_PXE_BIOS,
2792 BE2_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_PXE},
2793 { BE2_FCOE_BIOS_START, OPTYPE_FCOE_BIOS,
2794 BE2_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_FCOE},
2795 { BE2_ISCSI_BACKUP_IMAGE_START, OPTYPE_ISCSI_BACKUP,
2796 BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_ISCSI},
2797 { BE2_FCOE_PRIMARY_IMAGE_START, OPTYPE_FCOE_FW_ACTIVE,
2798 BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_FCOE},
2799 { BE2_FCOE_BACKUP_IMAGE_START, OPTYPE_FCOE_FW_BACKUP,
2800 BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_FCOE}
2801 };
2802
2803 if (BE3_chip(adapter)) {
2804 pflashcomp = gen3_flash_types;
2805 filehdr_size = sizeof(struct flash_file_hdr_g3);
2806 num_comp = ARRAY_SIZE(gen3_flash_types);
2807 } else {
2808 pflashcomp = gen2_flash_types;
2809 filehdr_size = sizeof(struct flash_file_hdr_g2);
2810 num_comp = ARRAY_SIZE(gen2_flash_types);
2811 img_hdrs_size = 0;
2812 }
2813
2814 /* Get flash section info*/
2815 fsec = get_fsec_info(adapter, filehdr_size + img_hdrs_size, fw);
2816 if (!fsec) {
2817 dev_err(dev, "Invalid Cookie. FW image may be corrupted\n");
2818 return -1;
2819 }
2820 for (i = 0; i < num_comp; i++) {
2821 if (!is_comp_in_ufi(adapter, fsec, pflashcomp[i].img_type))
2822 continue;
2823
2824 if ((pflashcomp[i].optype == OPTYPE_NCSI_FW) &&
2825 !be_fw_ncsi_supported(adapter->fw_ver)) {
2826 dev_info(dev, NCSI_UPDATE_LOG, adapter->fw_ver);
2827 continue;
2828 }
2829
2830 if (pflashcomp[i].optype == OPTYPE_PHY_FW &&
2831 !phy_flashing_required(adapter))
2832 continue;
2833
2834 if (pflashcomp[i].optype == OPTYPE_REDBOOT) {
2835 status = be_check_flash_crc(adapter, fw->data,
2836 pflashcomp[i].offset,
2837 pflashcomp[i].size,
2838 filehdr_size +
2839 img_hdrs_size,
2840 OPTYPE_REDBOOT, &crc_match);
2841 if (status) {
2842 dev_err(dev,
2843 "Could not get CRC for 0x%x region\n",
2844 pflashcomp[i].optype);
2845 continue;
2846 }
2847
2848 if (crc_match)
2849 continue;
2850 }
2851
2852 p = fw->data + filehdr_size + pflashcomp[i].offset +
2853 img_hdrs_size;
2854 if (p + pflashcomp[i].size > fw->data + fw->size)
2855 return -1;
2856
2857 status = be_flash(adapter, p, flash_cmd, pflashcomp[i].optype,
2858 pflashcomp[i].size, 0);
2859 if (status) {
2860 dev_err(dev, "Flashing section type 0x%x failed\n",
2861 pflashcomp[i].img_type);
2862 return status;
2863 }
2864 }
2865 return 0;
2866 }
2867
2868 static u16 be_get_img_optype(struct flash_section_entry fsec_entry)
2869 {
2870 u32 img_type = le32_to_cpu(fsec_entry.type);
2871 u16 img_optype = le16_to_cpu(fsec_entry.optype);
2872
2873 if (img_optype != 0xFFFF)
2874 return img_optype;
2875
2876 switch (img_type) {
2877 case IMAGE_FIRMWARE_ISCSI:
2878 img_optype = OPTYPE_ISCSI_ACTIVE;
2879 break;
2880 case IMAGE_BOOT_CODE:
2881 img_optype = OPTYPE_REDBOOT;
2882 break;
2883 case IMAGE_OPTION_ROM_ISCSI:
2884 img_optype = OPTYPE_BIOS;
2885 break;
2886 case IMAGE_OPTION_ROM_PXE:
2887 img_optype = OPTYPE_PXE_BIOS;
2888 break;
2889 case IMAGE_OPTION_ROM_FCOE:
2890 img_optype = OPTYPE_FCOE_BIOS;
2891 break;
2892 case IMAGE_FIRMWARE_BACKUP_ISCSI:
2893 img_optype = OPTYPE_ISCSI_BACKUP;
2894 break;
2895 case IMAGE_NCSI:
2896 img_optype = OPTYPE_NCSI_FW;
2897 break;
2898 case IMAGE_FLASHISM_JUMPVECTOR:
2899 img_optype = OPTYPE_FLASHISM_JUMPVECTOR;
2900 break;
2901 case IMAGE_FIRMWARE_PHY:
2902 img_optype = OPTYPE_SH_PHY_FW;
2903 break;
2904 case IMAGE_REDBOOT_DIR:
2905 img_optype = OPTYPE_REDBOOT_DIR;
2906 break;
2907 case IMAGE_REDBOOT_CONFIG:
2908 img_optype = OPTYPE_REDBOOT_CONFIG;
2909 break;
2910 case IMAGE_UFI_DIR:
2911 img_optype = OPTYPE_UFI_DIR;
2912 break;
2913 default:
2914 break;
2915 }
2916
2917 return img_optype;
2918 }
2919
2920 static int be_flash_skyhawk(struct be_adapter *adapter,
2921 const struct firmware *fw,
2922 struct be_dma_mem *flash_cmd, int num_of_images)
2923 {
2924 int img_hdrs_size = num_of_images * sizeof(struct image_hdr);
2925 bool crc_match, old_fw_img, flash_offset_support = true;
2926 struct device *dev = &adapter->pdev->dev;
2927 struct flash_section_info *fsec = NULL;
2928 u32 img_offset, img_size, img_type;
2929 u16 img_optype, flash_optype;
2930 int status, i, filehdr_size;
2931 const u8 *p;
2932
2933 filehdr_size = sizeof(struct flash_file_hdr_g3);
2934 fsec = get_fsec_info(adapter, filehdr_size + img_hdrs_size, fw);
2935 if (!fsec) {
2936 dev_err(dev, "Invalid Cookie. FW image may be corrupted\n");
2937 return -EINVAL;
2938 }
2939
2940 retry_flash:
2941 for (i = 0; i < le32_to_cpu(fsec->fsec_hdr.num_images); i++) {
2942 img_offset = le32_to_cpu(fsec->fsec_entry[i].offset);
2943 img_size = le32_to_cpu(fsec->fsec_entry[i].pad_size);
2944 img_type = le32_to_cpu(fsec->fsec_entry[i].type);
2945 img_optype = be_get_img_optype(fsec->fsec_entry[i]);
2946 old_fw_img = fsec->fsec_entry[i].optype == 0xFFFF;
2947
2948 if (img_optype == 0xFFFF)
2949 continue;
2950
2951 if (flash_offset_support)
2952 flash_optype = OPTYPE_OFFSET_SPECIFIED;
2953 else
2954 flash_optype = img_optype;
2955
2956 /* Don't bother verifying CRC if an old FW image is being
2957 * flashed
2958 */
2959 if (old_fw_img)
2960 goto flash;
2961
2962 status = be_check_flash_crc(adapter, fw->data, img_offset,
2963 img_size, filehdr_size +
2964 img_hdrs_size, flash_optype,
2965 &crc_match);
2966 if (base_status(status) == MCC_STATUS_ILLEGAL_REQUEST ||
2967 base_status(status) == MCC_STATUS_ILLEGAL_FIELD) {
2968 /* The current FW image on the card does not support
2969 * OFFSET based flashing. Retry using older mechanism
2970 * of OPTYPE based flashing
2971 */
2972 if (flash_optype == OPTYPE_OFFSET_SPECIFIED) {
2973 flash_offset_support = false;
2974 goto retry_flash;
2975 }
2976
2977 /* The current FW image on the card does not recognize
2978 * the new FLASH op_type. The FW download is partially
2979 * complete. Reboot the server now to enable FW image
2980 * to recognize the new FLASH op_type. To complete the
2981 * remaining process, download the same FW again after
2982 * the reboot.
2983 */
2984 dev_err(dev, "Flash incomplete. Reset the server\n");
2985 dev_err(dev, "Download FW image again after reset\n");
2986 return -EAGAIN;
2987 } else if (status) {
2988 dev_err(dev, "Could not get CRC for 0x%x region\n",
2989 img_optype);
2990 return -EFAULT;
2991 }
2992
2993 if (crc_match)
2994 continue;
2995
2996 flash:
2997 p = fw->data + filehdr_size + img_offset + img_hdrs_size;
2998 if (p + img_size > fw->data + fw->size)
2999 return -1;
3000
3001 status = be_flash(adapter, p, flash_cmd, flash_optype, img_size,
3002 img_offset);
3003
3004 /* The current FW image on the card does not support OFFSET
3005 * based flashing. Retry using older mechanism of OPTYPE based
3006 * flashing
3007 */
3008 if (base_status(status) == MCC_STATUS_ILLEGAL_FIELD &&
3009 flash_optype == OPTYPE_OFFSET_SPECIFIED) {
3010 flash_offset_support = false;
3011 goto retry_flash;
3012 }
3013
3014 /* For old FW images ignore ILLEGAL_FIELD error or errors on
3015 * UFI_DIR region
3016 */
3017 if (old_fw_img &&
3018 (base_status(status) == MCC_STATUS_ILLEGAL_FIELD ||
3019 (img_optype == OPTYPE_UFI_DIR &&
3020 base_status(status) == MCC_STATUS_FAILED))) {
3021 continue;
3022 } else if (status) {
3023 dev_err(dev, "Flashing section type 0x%x failed\n",
3024 img_type);
3025
3026 switch (addl_status(status)) {
3027 case MCC_ADDL_STATUS_MISSING_SIGNATURE:
3028 dev_err(dev,
3029 "Digital signature missing in FW\n");
3030 return -EINVAL;
3031 case MCC_ADDL_STATUS_INVALID_SIGNATURE:
3032 dev_err(dev,
3033 "Invalid digital signature in FW\n");
3034 return -EINVAL;
3035 default:
3036 return -EFAULT;
3037 }
3038 }
3039 }
3040 return 0;
3041 }
3042
3043 int lancer_fw_download(struct be_adapter *adapter,
3044 const struct firmware *fw)
3045 {
3046 struct device *dev = &adapter->pdev->dev;
3047 struct be_dma_mem flash_cmd;
3048 const u8 *data_ptr = NULL;
3049 u8 *dest_image_ptr = NULL;
3050 size_t image_size = 0;
3051 u32 chunk_size = 0;
3052 u32 data_written = 0;
3053 u32 offset = 0;
3054 int status = 0;
3055 u8 add_status = 0;
3056 u8 change_status;
3057
3058 if (!IS_ALIGNED(fw->size, sizeof(u32))) {
3059 dev_err(dev, "FW image size should be multiple of 4\n");
3060 return -EINVAL;
3061 }
3062
3063 flash_cmd.size = sizeof(struct lancer_cmd_req_write_object)
3064 + LANCER_FW_DOWNLOAD_CHUNK;
3065 flash_cmd.va = dma_alloc_coherent(dev, flash_cmd.size, &flash_cmd.dma,
3066 GFP_KERNEL);
3067 if (!flash_cmd.va)
3068 return -ENOMEM;
3069
3070 dest_image_ptr = flash_cmd.va +
3071 sizeof(struct lancer_cmd_req_write_object);
3072 image_size = fw->size;
3073 data_ptr = fw->data;
3074
3075 while (image_size) {
3076 chunk_size = min_t(u32, image_size, LANCER_FW_DOWNLOAD_CHUNK);
3077
3078 /* Copy the image chunk content. */
3079 memcpy(dest_image_ptr, data_ptr, chunk_size);
3080
3081 status = lancer_cmd_write_object(adapter, &flash_cmd,
3082 chunk_size, offset,
3083 LANCER_FW_DOWNLOAD_LOCATION,
3084 &data_written, &change_status,
3085 &add_status);
3086 if (status)
3087 break;
3088
3089 offset += data_written;
3090 data_ptr += data_written;
3091 image_size -= data_written;
3092 }
3093
3094 if (!status) {
3095 /* Commit the FW written */
3096 status = lancer_cmd_write_object(adapter, &flash_cmd,
3097 0, offset,
3098 LANCER_FW_DOWNLOAD_LOCATION,
3099 &data_written, &change_status,
3100 &add_status);
3101 }
3102
3103 dma_free_coherent(dev, flash_cmd.size, flash_cmd.va, flash_cmd.dma);
3104 if (status) {
3105 dev_err(dev, "Firmware load error\n");
3106 return be_cmd_status(status);
3107 }
3108
3109 dev_info(dev, "Firmware flashed successfully\n");
3110
3111 if (change_status == LANCER_FW_RESET_NEEDED) {
3112 dev_info(dev, "Resetting adapter to activate new FW\n");
3113 status = lancer_physdev_ctrl(adapter,
3114 PHYSDEV_CONTROL_FW_RESET_MASK);
3115 if (status) {
3116 dev_err(dev, "Adapter busy, could not reset FW\n");
3117 dev_err(dev, "Reboot server to activate new FW\n");
3118 }
3119 } else if (change_status != LANCER_NO_RESET_NEEDED) {
3120 dev_info(dev, "Reboot server to activate new FW\n");
3121 }
3122
3123 return 0;
3124 }
3125
3126 /* Check if the flash image file is compatible with the adapter that
3127 * is being flashed.
3128 */
3129 static bool be_check_ufi_compatibility(struct be_adapter *adapter,
3130 struct flash_file_hdr_g3 *fhdr)
3131 {
3132 if (!fhdr) {
3133 dev_err(&adapter->pdev->dev, "Invalid FW UFI file");
3134 return false;
3135 }
3136
3137 /* First letter of the build version is used to identify
3138 * which chip this image file is meant for.
3139 */
3140 switch (fhdr->build[0]) {
3141 case BLD_STR_UFI_TYPE_SH:
3142 if (!skyhawk_chip(adapter))
3143 return false;
3144 break;
3145 case BLD_STR_UFI_TYPE_BE3:
3146 if (!BE3_chip(adapter))
3147 return false;
3148 break;
3149 case BLD_STR_UFI_TYPE_BE2:
3150 if (!BE2_chip(adapter))
3151 return false;
3152 break;
3153 default:
3154 return false;
3155 }
3156
3157 /* In BE3 FW images the "asic_type_rev" field doesn't track the
3158 * asic_rev of the chips it is compatible with.
3159 * When asic_type_rev is 0 the image is compatible only with
3160 * pre-BE3-R chips (asic_rev < 0x10)
3161 */
3162 if (BEx_chip(adapter) && fhdr->asic_type_rev == 0)
3163 return adapter->asic_rev < 0x10;
3164 else
3165 return (fhdr->asic_type_rev >= adapter->asic_rev);
3166 }
3167
3168 int be_fw_download(struct be_adapter *adapter, const struct firmware *fw)
3169 {
3170 struct device *dev = &adapter->pdev->dev;
3171 struct flash_file_hdr_g3 *fhdr3;
3172 struct image_hdr *img_hdr_ptr;
3173 int status = 0, i, num_imgs;
3174 struct be_dma_mem flash_cmd;
3175
3176 fhdr3 = (struct flash_file_hdr_g3 *)fw->data;
3177 if (!be_check_ufi_compatibility(adapter, fhdr3)) {
3178 dev_err(dev, "Flash image is not compatible with adapter\n");
3179 return -EINVAL;
3180 }
3181
3182 flash_cmd.size = sizeof(struct be_cmd_write_flashrom);
3183 flash_cmd.va = dma_alloc_coherent(dev, flash_cmd.size, &flash_cmd.dma,
3184 GFP_KERNEL);
3185 if (!flash_cmd.va)
3186 return -ENOMEM;
3187
3188 num_imgs = le32_to_cpu(fhdr3->num_imgs);
3189 for (i = 0; i < num_imgs; i++) {
3190 img_hdr_ptr = (struct image_hdr *)(fw->data +
3191 (sizeof(struct flash_file_hdr_g3) +
3192 i * sizeof(struct image_hdr)));
3193 if (!BE2_chip(adapter) &&
3194 le32_to_cpu(img_hdr_ptr->imageid) != 1)
3195 continue;
3196
3197 if (skyhawk_chip(adapter))
3198 status = be_flash_skyhawk(adapter, fw, &flash_cmd,
3199 num_imgs);
3200 else
3201 status = be_flash_BEx(adapter, fw, &flash_cmd,
3202 num_imgs);
3203 }
3204
3205 dma_free_coherent(dev, flash_cmd.size, flash_cmd.va, flash_cmd.dma);
3206 if (!status)
3207 dev_info(dev, "Firmware flashed successfully\n");
3208
3209 return status;
3210 }
3211
3212 int be_cmd_enable_magic_wol(struct be_adapter *adapter, u8 *mac,
3213 struct be_dma_mem *nonemb_cmd)
3214 {
3215 struct be_mcc_wrb *wrb;
3216 struct be_cmd_req_acpi_wol_magic_config *req;
3217 int status;
3218
3219 mutex_lock(&adapter->mcc_lock);
3220
3221 wrb = wrb_from_mccq(adapter);
3222 if (!wrb) {
3223 status = -EBUSY;
3224 goto err;
3225 }
3226 req = nonemb_cmd->va;
3227
3228 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
3229 OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG, sizeof(*req),
3230 wrb, nonemb_cmd);
3231 memcpy(req->magic_mac, mac, ETH_ALEN);
3232
3233 status = be_mcc_notify_wait(adapter);
3234
3235 err:
3236 mutex_unlock(&adapter->mcc_lock);
3237 return status;
3238 }
3239
3240 int be_cmd_set_loopback(struct be_adapter *adapter, u8 port_num,
3241 u8 loopback_type, u8 enable)
3242 {
3243 struct be_mcc_wrb *wrb;
3244 struct be_cmd_req_set_lmode *req;
3245 int status;
3246
3247 if (!be_cmd_allowed(adapter, OPCODE_LOWLEVEL_SET_LOOPBACK_MODE,
3248 CMD_SUBSYSTEM_LOWLEVEL))
3249 return -EPERM;
3250
3251 mutex_lock(&adapter->mcc_lock);
3252
3253 wrb = wrb_from_mccq(adapter);
3254 if (!wrb) {
3255 status = -EBUSY;
3256 goto err_unlock;
3257 }
3258
3259 req = embedded_payload(wrb);
3260
3261 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL,
3262 OPCODE_LOWLEVEL_SET_LOOPBACK_MODE, sizeof(*req),
3263 wrb, NULL);
3264
3265 req->src_port = port_num;
3266 req->dest_port = port_num;
3267 req->loopback_type = loopback_type;
3268 req->loopback_state = enable;
3269
3270 status = be_mcc_notify(adapter);
3271 if (status)
3272 goto err_unlock;
3273
3274 mutex_unlock(&adapter->mcc_lock);
3275
3276 if (!wait_for_completion_timeout(&adapter->et_cmd_compl,
3277 msecs_to_jiffies(SET_LB_MODE_TIMEOUT)))
3278 status = -ETIMEDOUT;
3279
3280 return status;
3281
3282 err_unlock:
3283 mutex_unlock(&adapter->mcc_lock);
3284 return status;
3285 }
3286
3287 int be_cmd_loopback_test(struct be_adapter *adapter, u32 port_num,
3288 u32 loopback_type, u32 pkt_size, u32 num_pkts,
3289 u64 pattern)
3290 {
3291 struct be_mcc_wrb *wrb;
3292 struct be_cmd_req_loopback_test *req;
3293 struct be_cmd_resp_loopback_test *resp;
3294 int status;
3295
3296 if (!be_cmd_allowed(adapter, OPCODE_LOWLEVEL_LOOPBACK_TEST,
3297 CMD_SUBSYSTEM_LOWLEVEL))
3298 return -EPERM;
3299
3300 mutex_lock(&adapter->mcc_lock);
3301
3302 wrb = wrb_from_mccq(adapter);
3303 if (!wrb) {
3304 status = -EBUSY;
3305 goto err;
3306 }
3307
3308 req = embedded_payload(wrb);
3309
3310 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL,
3311 OPCODE_LOWLEVEL_LOOPBACK_TEST, sizeof(*req), wrb,
3312 NULL);
3313
3314 req->hdr.timeout = cpu_to_le32(15);
3315 req->pattern = cpu_to_le64(pattern);
3316 req->src_port = cpu_to_le32(port_num);
3317 req->dest_port = cpu_to_le32(port_num);
3318 req->pkt_size = cpu_to_le32(pkt_size);
3319 req->num_pkts = cpu_to_le32(num_pkts);
3320 req->loopback_type = cpu_to_le32(loopback_type);
3321
3322 status = be_mcc_notify(adapter);
3323 if (status)
3324 goto err;
3325
3326 mutex_unlock(&adapter->mcc_lock);
3327
3328 wait_for_completion(&adapter->et_cmd_compl);
3329 resp = embedded_payload(wrb);
3330 status = le32_to_cpu(resp->status);
3331
3332 return status;
3333 err:
3334 mutex_unlock(&adapter->mcc_lock);
3335 return status;
3336 }
3337
3338 int be_cmd_ddr_dma_test(struct be_adapter *adapter, u64 pattern,
3339 u32 byte_cnt, struct be_dma_mem *cmd)
3340 {
3341 struct be_mcc_wrb *wrb;
3342 struct be_cmd_req_ddrdma_test *req;
3343 int status;
3344 int i, j = 0;
3345
3346 if (!be_cmd_allowed(adapter, OPCODE_LOWLEVEL_HOST_DDR_DMA,
3347 CMD_SUBSYSTEM_LOWLEVEL))
3348 return -EPERM;
3349
3350 mutex_lock(&adapter->mcc_lock);
3351
3352 wrb = wrb_from_mccq(adapter);
3353 if (!wrb) {
3354 status = -EBUSY;
3355 goto err;
3356 }
3357 req = cmd->va;
3358 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL,
3359 OPCODE_LOWLEVEL_HOST_DDR_DMA, cmd->size, wrb,
3360 cmd);
3361
3362 req->pattern = cpu_to_le64(pattern);
3363 req->byte_count = cpu_to_le32(byte_cnt);
3364 for (i = 0; i < byte_cnt; i++) {
3365 req->snd_buff[i] = (u8)(pattern >> (j*8));
3366 j++;
3367 if (j > 7)
3368 j = 0;
3369 }
3370
3371 status = be_mcc_notify_wait(adapter);
3372
3373 if (!status) {
3374 struct be_cmd_resp_ddrdma_test *resp;
3375
3376 resp = cmd->va;
3377 if ((memcmp(resp->rcv_buff, req->snd_buff, byte_cnt) != 0) ||
3378 resp->snd_err) {
3379 status = -1;
3380 }
3381 }
3382
3383 err:
3384 mutex_unlock(&adapter->mcc_lock);
3385 return status;
3386 }
3387
3388 int be_cmd_get_seeprom_data(struct be_adapter *adapter,
3389 struct be_dma_mem *nonemb_cmd)
3390 {
3391 struct be_mcc_wrb *wrb;
3392 struct be_cmd_req_seeprom_read *req;
3393 int status;
3394
3395 mutex_lock(&adapter->mcc_lock);
3396
3397 wrb = wrb_from_mccq(adapter);
3398 if (!wrb) {
3399 status = -EBUSY;
3400 goto err;
3401 }
3402 req = nonemb_cmd->va;
3403
3404 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3405 OPCODE_COMMON_SEEPROM_READ, sizeof(*req), wrb,
3406 nonemb_cmd);
3407
3408 status = be_mcc_notify_wait(adapter);
3409
3410 err:
3411 mutex_unlock(&adapter->mcc_lock);
3412 return status;
3413 }
3414
3415 int be_cmd_get_phy_info(struct be_adapter *adapter)
3416 {
3417 struct be_mcc_wrb *wrb;
3418 struct be_cmd_req_get_phy_info *req;
3419 struct be_dma_mem cmd;
3420 int status;
3421
3422 if (!be_cmd_allowed(adapter, OPCODE_COMMON_GET_PHY_DETAILS,
3423 CMD_SUBSYSTEM_COMMON))
3424 return -EPERM;
3425
3426 mutex_lock(&adapter->mcc_lock);
3427
3428 wrb = wrb_from_mccq(adapter);
3429 if (!wrb) {
3430 status = -EBUSY;
3431 goto err;
3432 }
3433 cmd.size = sizeof(struct be_cmd_req_get_phy_info);
3434 cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
3435 GFP_ATOMIC);
3436 if (!cmd.va) {
3437 dev_err(&adapter->pdev->dev, "Memory alloc failure\n");
3438 status = -ENOMEM;
3439 goto err;
3440 }
3441
3442 req = cmd.va;
3443
3444 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3445 OPCODE_COMMON_GET_PHY_DETAILS, sizeof(*req),
3446 wrb, &cmd);
3447
3448 status = be_mcc_notify_wait(adapter);
3449 if (!status) {
3450 struct be_phy_info *resp_phy_info =
3451 cmd.va + sizeof(struct be_cmd_req_hdr);
3452
3453 adapter->phy.phy_type = le16_to_cpu(resp_phy_info->phy_type);
3454 adapter->phy.interface_type =
3455 le16_to_cpu(resp_phy_info->interface_type);
3456 adapter->phy.auto_speeds_supported =
3457 le16_to_cpu(resp_phy_info->auto_speeds_supported);
3458 adapter->phy.fixed_speeds_supported =
3459 le16_to_cpu(resp_phy_info->fixed_speeds_supported);
3460 adapter->phy.misc_params =
3461 le32_to_cpu(resp_phy_info->misc_params);
3462
3463 if (BE2_chip(adapter)) {
3464 adapter->phy.fixed_speeds_supported =
3465 BE_SUPPORTED_SPEED_10GBPS |
3466 BE_SUPPORTED_SPEED_1GBPS;
3467 }
3468 }
3469 dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va, cmd.dma);
3470 err:
3471 mutex_unlock(&adapter->mcc_lock);
3472 return status;
3473 }
3474
3475 static int be_cmd_set_qos(struct be_adapter *adapter, u32 bps, u32 domain)
3476 {
3477 struct be_mcc_wrb *wrb;
3478 struct be_cmd_req_set_qos *req;
3479 int status;
3480
3481 mutex_lock(&adapter->mcc_lock);
3482
3483 wrb = wrb_from_mccq(adapter);
3484 if (!wrb) {
3485 status = -EBUSY;
3486 goto err;
3487 }
3488
3489 req = embedded_payload(wrb);
3490
3491 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3492 OPCODE_COMMON_SET_QOS, sizeof(*req), wrb, NULL);
3493
3494 req->hdr.domain = domain;
3495 req->valid_bits = cpu_to_le32(BE_QOS_BITS_NIC);
3496 req->max_bps_nic = cpu_to_le32(bps);
3497
3498 status = be_mcc_notify_wait(adapter);
3499
3500 err:
3501 mutex_unlock(&adapter->mcc_lock);
3502 return status;
3503 }
3504
3505 int be_cmd_get_cntl_attributes(struct be_adapter *adapter)
3506 {
3507 struct be_mcc_wrb *wrb;
3508 struct be_cmd_req_cntl_attribs *req;
3509 struct be_cmd_resp_cntl_attribs *resp;
3510 int status, i;
3511 int payload_len = max(sizeof(*req), sizeof(*resp));
3512 struct mgmt_controller_attrib *attribs;
3513 struct be_dma_mem attribs_cmd;
3514 u32 *serial_num;
3515
3516 if (mutex_lock_interruptible(&adapter->mbox_lock))
3517 return -1;
3518
3519 memset(&attribs_cmd, 0, sizeof(struct be_dma_mem));
3520 attribs_cmd.size = sizeof(struct be_cmd_resp_cntl_attribs);
3521 attribs_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
3522 attribs_cmd.size,
3523 &attribs_cmd.dma, GFP_ATOMIC);
3524 if (!attribs_cmd.va) {
3525 dev_err(&adapter->pdev->dev, "Memory allocation failure\n");
3526 status = -ENOMEM;
3527 goto err;
3528 }
3529
3530 wrb = wrb_from_mbox(adapter);
3531 if (!wrb) {
3532 status = -EBUSY;
3533 goto err;
3534 }
3535 req = attribs_cmd.va;
3536
3537 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3538 OPCODE_COMMON_GET_CNTL_ATTRIBUTES, payload_len,
3539 wrb, &attribs_cmd);
3540
3541 status = be_mbox_notify_wait(adapter);
3542 if (!status) {
3543 attribs = attribs_cmd.va + sizeof(struct be_cmd_resp_hdr);
3544 adapter->hba_port_num = attribs->hba_attribs.phy_port;
3545 serial_num = attribs->hba_attribs.controller_serial_number;
3546 for (i = 0; i < CNTL_SERIAL_NUM_WORDS; i++)
3547 adapter->serial_num[i] = le32_to_cpu(serial_num[i]) &
3548 (BIT_MASK(16) - 1);
3549 /* For BEx, since GET_FUNC_CONFIG command is not
3550 * supported, we read funcnum here as a workaround.
3551 */
3552 if (BEx_chip(adapter))
3553 adapter->pf_num = attribs->hba_attribs.pci_funcnum;
3554 }
3555
3556 err:
3557 mutex_unlock(&adapter->mbox_lock);
3558 if (attribs_cmd.va)
3559 dma_free_coherent(&adapter->pdev->dev, attribs_cmd.size,
3560 attribs_cmd.va, attribs_cmd.dma);
3561 return status;
3562 }
3563
3564 /* Uses mbox */
3565 int be_cmd_req_native_mode(struct be_adapter *adapter)
3566 {
3567 struct be_mcc_wrb *wrb;
3568 struct be_cmd_req_set_func_cap *req;
3569 int status;
3570
3571 if (mutex_lock_interruptible(&adapter->mbox_lock))
3572 return -1;
3573
3574 wrb = wrb_from_mbox(adapter);
3575 if (!wrb) {
3576 status = -EBUSY;
3577 goto err;
3578 }
3579
3580 req = embedded_payload(wrb);
3581
3582 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3583 OPCODE_COMMON_SET_DRIVER_FUNCTION_CAP,
3584 sizeof(*req), wrb, NULL);
3585
3586 req->valid_cap_flags = cpu_to_le32(CAPABILITY_SW_TIMESTAMPS |
3587 CAPABILITY_BE3_NATIVE_ERX_API);
3588 req->cap_flags = cpu_to_le32(CAPABILITY_BE3_NATIVE_ERX_API);
3589
3590 status = be_mbox_notify_wait(adapter);
3591 if (!status) {
3592 struct be_cmd_resp_set_func_cap *resp = embedded_payload(wrb);
3593
3594 adapter->be3_native = le32_to_cpu(resp->cap_flags) &
3595 CAPABILITY_BE3_NATIVE_ERX_API;
3596 if (!adapter->be3_native)
3597 dev_warn(&adapter->pdev->dev,
3598 "adapter not in advanced mode\n");
3599 }
3600 err:
3601 mutex_unlock(&adapter->mbox_lock);
3602 return status;
3603 }
3604
3605 /* Get privilege(s) for a function */
3606 int be_cmd_get_fn_privileges(struct be_adapter *adapter, u32 *privilege,
3607 u32 domain)
3608 {
3609 struct be_mcc_wrb *wrb;
3610 struct be_cmd_req_get_fn_privileges *req;
3611 int status;
3612
3613 mutex_lock(&adapter->mcc_lock);
3614
3615 wrb = wrb_from_mccq(adapter);
3616 if (!wrb) {
3617 status = -EBUSY;
3618 goto err;
3619 }
3620
3621 req = embedded_payload(wrb);
3622
3623 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3624 OPCODE_COMMON_GET_FN_PRIVILEGES, sizeof(*req),
3625 wrb, NULL);
3626
3627 req->hdr.domain = domain;
3628
3629 status = be_mcc_notify_wait(adapter);
3630 if (!status) {
3631 struct be_cmd_resp_get_fn_privileges *resp =
3632 embedded_payload(wrb);
3633
3634 *privilege = le32_to_cpu(resp->privilege_mask);
3635
3636 /* In UMC mode FW does not return right privileges.
3637 * Override with correct privilege equivalent to PF.
3638 */
3639 if (BEx_chip(adapter) && be_is_mc(adapter) &&
3640 be_physfn(adapter))
3641 *privilege = MAX_PRIVILEGES;
3642 }
3643
3644 err:
3645 mutex_unlock(&adapter->mcc_lock);
3646 return status;
3647 }
3648
3649 /* Set privilege(s) for a function */
3650 int be_cmd_set_fn_privileges(struct be_adapter *adapter, u32 privileges,
3651 u32 domain)
3652 {
3653 struct be_mcc_wrb *wrb;
3654 struct be_cmd_req_set_fn_privileges *req;
3655 int status;
3656
3657 mutex_lock(&adapter->mcc_lock);
3658
3659 wrb = wrb_from_mccq(adapter);
3660 if (!wrb) {
3661 status = -EBUSY;
3662 goto err;
3663 }
3664
3665 req = embedded_payload(wrb);
3666 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3667 OPCODE_COMMON_SET_FN_PRIVILEGES, sizeof(*req),
3668 wrb, NULL);
3669 req->hdr.domain = domain;
3670 if (lancer_chip(adapter))
3671 req->privileges_lancer = cpu_to_le32(privileges);
3672 else
3673 req->privileges = cpu_to_le32(privileges);
3674
3675 status = be_mcc_notify_wait(adapter);
3676 err:
3677 mutex_unlock(&adapter->mcc_lock);
3678 return status;
3679 }
3680
3681 /* pmac_id_valid: true => pmac_id is supplied and MAC address is requested.
3682 * pmac_id_valid: false => pmac_id or MAC address is requested.
3683 * If pmac_id is returned, pmac_id_valid is returned as true
3684 */
3685 int be_cmd_get_mac_from_list(struct be_adapter *adapter, u8 *mac,
3686 bool *pmac_id_valid, u32 *pmac_id, u32 if_handle,
3687 u8 domain)
3688 {
3689 struct be_mcc_wrb *wrb;
3690 struct be_cmd_req_get_mac_list *req;
3691 int status;
3692 int mac_count;
3693 struct be_dma_mem get_mac_list_cmd;
3694 int i;
3695
3696 memset(&get_mac_list_cmd, 0, sizeof(struct be_dma_mem));
3697 get_mac_list_cmd.size = sizeof(struct be_cmd_resp_get_mac_list);
3698 get_mac_list_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
3699 get_mac_list_cmd.size,
3700 &get_mac_list_cmd.dma,
3701 GFP_ATOMIC);
3702
3703 if (!get_mac_list_cmd.va) {
3704 dev_err(&adapter->pdev->dev,
3705 "Memory allocation failure during GET_MAC_LIST\n");
3706 return -ENOMEM;
3707 }
3708
3709 mutex_lock(&adapter->mcc_lock);
3710
3711 wrb = wrb_from_mccq(adapter);
3712 if (!wrb) {
3713 status = -EBUSY;
3714 goto out;
3715 }
3716
3717 req = get_mac_list_cmd.va;
3718
3719 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3720 OPCODE_COMMON_GET_MAC_LIST,
3721 get_mac_list_cmd.size, wrb, &get_mac_list_cmd);
3722 req->hdr.domain = domain;
3723 req->mac_type = MAC_ADDRESS_TYPE_NETWORK;
3724 if (*pmac_id_valid) {
3725 req->mac_id = cpu_to_le32(*pmac_id);
3726 req->iface_id = cpu_to_le16(if_handle);
3727 req->perm_override = 0;
3728 } else {
3729 req->perm_override = 1;
3730 }
3731
3732 status = be_mcc_notify_wait(adapter);
3733 if (!status) {
3734 struct be_cmd_resp_get_mac_list *resp =
3735 get_mac_list_cmd.va;
3736
3737 if (*pmac_id_valid) {
3738 memcpy(mac, resp->macid_macaddr.mac_addr_id.macaddr,
3739 ETH_ALEN);
3740 goto out;
3741 }
3742
3743 mac_count = resp->true_mac_count + resp->pseudo_mac_count;
3744 /* Mac list returned could contain one or more active mac_ids
3745 * or one or more true or pseudo permanent mac addresses.
3746 * If an active mac_id is present, return first active mac_id
3747 * found.
3748 */
3749 for (i = 0; i < mac_count; i++) {
3750 struct get_list_macaddr *mac_entry;
3751 u16 mac_addr_size;
3752 u32 mac_id;
3753
3754 mac_entry = &resp->macaddr_list[i];
3755 mac_addr_size = le16_to_cpu(mac_entry->mac_addr_size);
3756 /* mac_id is a 32 bit value and mac_addr size
3757 * is 6 bytes
3758 */
3759 if (mac_addr_size == sizeof(u32)) {
3760 *pmac_id_valid = true;
3761 mac_id = mac_entry->mac_addr_id.s_mac_id.mac_id;
3762 *pmac_id = le32_to_cpu(mac_id);
3763 goto out;
3764 }
3765 }
3766 /* If no active mac_id found, return first mac addr */
3767 *pmac_id_valid = false;
3768 memcpy(mac, resp->macaddr_list[0].mac_addr_id.macaddr,
3769 ETH_ALEN);
3770 }
3771
3772 out:
3773 mutex_unlock(&adapter->mcc_lock);
3774 dma_free_coherent(&adapter->pdev->dev, get_mac_list_cmd.size,
3775 get_mac_list_cmd.va, get_mac_list_cmd.dma);
3776 return status;
3777 }
3778
3779 int be_cmd_get_active_mac(struct be_adapter *adapter, u32 curr_pmac_id,
3780 u8 *mac, u32 if_handle, bool active, u32 domain)
3781 {
3782 if (!active)
3783 be_cmd_get_mac_from_list(adapter, mac, &active, &curr_pmac_id,
3784 if_handle, domain);
3785 if (BEx_chip(adapter))
3786 return be_cmd_mac_addr_query(adapter, mac, false,
3787 if_handle, curr_pmac_id);
3788 else
3789 /* Fetch the MAC address using pmac_id */
3790 return be_cmd_get_mac_from_list(adapter, mac, &active,
3791 &curr_pmac_id,
3792 if_handle, domain);
3793 }
3794
3795 int be_cmd_get_perm_mac(struct be_adapter *adapter, u8 *mac)
3796 {
3797 int status;
3798 bool pmac_valid = false;
3799
3800 eth_zero_addr(mac);
3801
3802 if (BEx_chip(adapter)) {
3803 if (be_physfn(adapter))
3804 status = be_cmd_mac_addr_query(adapter, mac, true, 0,
3805 0);
3806 else
3807 status = be_cmd_mac_addr_query(adapter, mac, false,
3808 adapter->if_handle, 0);
3809 } else {
3810 status = be_cmd_get_mac_from_list(adapter, mac, &pmac_valid,
3811 NULL, adapter->if_handle, 0);
3812 }
3813
3814 return status;
3815 }
3816
3817 /* Uses synchronous MCCQ */
3818 int be_cmd_set_mac_list(struct be_adapter *adapter, u8 *mac_array,
3819 u8 mac_count, u32 domain)
3820 {
3821 struct be_mcc_wrb *wrb;
3822 struct be_cmd_req_set_mac_list *req;
3823 int status;
3824 struct be_dma_mem cmd;
3825
3826 memset(&cmd, 0, sizeof(struct be_dma_mem));
3827 cmd.size = sizeof(struct be_cmd_req_set_mac_list);
3828 cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
3829 GFP_KERNEL);
3830 if (!cmd.va)
3831 return -ENOMEM;
3832
3833 mutex_lock(&adapter->mcc_lock);
3834
3835 wrb = wrb_from_mccq(adapter);
3836 if (!wrb) {
3837 status = -EBUSY;
3838 goto err;
3839 }
3840
3841 req = cmd.va;
3842 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3843 OPCODE_COMMON_SET_MAC_LIST, sizeof(*req),
3844 wrb, &cmd);
3845
3846 req->hdr.domain = domain;
3847 req->mac_count = mac_count;
3848 if (mac_count)
3849 memcpy(req->mac, mac_array, ETH_ALEN*mac_count);
3850
3851 status = be_mcc_notify_wait(adapter);
3852
3853 err:
3854 dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va, cmd.dma);
3855 mutex_unlock(&adapter->mcc_lock);
3856 return status;
3857 }
3858
3859 /* Wrapper to delete any active MACs and provision the new mac.
3860 * Changes to MAC_LIST are allowed iff none of the MAC addresses in the
3861 * current list are active.
3862 */
3863 int be_cmd_set_mac(struct be_adapter *adapter, u8 *mac, int if_id, u32 dom)
3864 {
3865 bool active_mac = false;
3866 u8 old_mac[ETH_ALEN];
3867 u32 pmac_id;
3868 int status;
3869
3870 status = be_cmd_get_mac_from_list(adapter, old_mac, &active_mac,
3871 &pmac_id, if_id, dom);
3872
3873 if (!status && active_mac)
3874 be_cmd_pmac_del(adapter, if_id, pmac_id, dom);
3875
3876 return be_cmd_set_mac_list(adapter, mac, mac ? 1 : 0, dom);
3877 }
3878
3879 int be_cmd_set_hsw_config(struct be_adapter *adapter, u16 pvid,
3880 u32 domain, u16 intf_id, u16 hsw_mode, u8 spoofchk)
3881 {
3882 struct be_mcc_wrb *wrb;
3883 struct be_cmd_req_set_hsw_config *req;
3884 void *ctxt;
3885 int status;
3886
3887 if (!be_cmd_allowed(adapter, OPCODE_COMMON_SET_HSW_CONFIG,
3888 CMD_SUBSYSTEM_COMMON))
3889 return -EPERM;
3890
3891 mutex_lock(&adapter->mcc_lock);
3892
3893 wrb = wrb_from_mccq(adapter);
3894 if (!wrb) {
3895 status = -EBUSY;
3896 goto err;
3897 }
3898
3899 req = embedded_payload(wrb);
3900 ctxt = &req->context;
3901
3902 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3903 OPCODE_COMMON_SET_HSW_CONFIG, sizeof(*req), wrb,
3904 NULL);
3905
3906 req->hdr.domain = domain;
3907 AMAP_SET_BITS(struct amap_set_hsw_context, interface_id, ctxt, intf_id);
3908 if (pvid) {
3909 AMAP_SET_BITS(struct amap_set_hsw_context, pvid_valid, ctxt, 1);
3910 AMAP_SET_BITS(struct amap_set_hsw_context, pvid, ctxt, pvid);
3911 }
3912 if (hsw_mode) {
3913 AMAP_SET_BITS(struct amap_set_hsw_context, interface_id,
3914 ctxt, adapter->hba_port_num);
3915 AMAP_SET_BITS(struct amap_set_hsw_context, pport, ctxt, 1);
3916 AMAP_SET_BITS(struct amap_set_hsw_context, port_fwd_type,
3917 ctxt, hsw_mode);
3918 }
3919
3920 /* Enable/disable both mac and vlan spoof checking */
3921 if (!BEx_chip(adapter) && spoofchk) {
3922 AMAP_SET_BITS(struct amap_set_hsw_context, mac_spoofchk,
3923 ctxt, spoofchk);
3924 AMAP_SET_BITS(struct amap_set_hsw_context, vlan_spoofchk,
3925 ctxt, spoofchk);
3926 }
3927
3928 be_dws_cpu_to_le(req->context, sizeof(req->context));
3929 status = be_mcc_notify_wait(adapter);
3930
3931 err:
3932 mutex_unlock(&adapter->mcc_lock);
3933 return status;
3934 }
3935
3936 /* Get Hyper switch config */
3937 int be_cmd_get_hsw_config(struct be_adapter *adapter, u16 *pvid,
3938 u32 domain, u16 intf_id, u8 *mode, bool *spoofchk)
3939 {
3940 struct be_mcc_wrb *wrb;
3941 struct be_cmd_req_get_hsw_config *req;
3942 void *ctxt;
3943 int status;
3944 u16 vid;
3945
3946 mutex_lock(&adapter->mcc_lock);
3947
3948 wrb = wrb_from_mccq(adapter);
3949 if (!wrb) {
3950 status = -EBUSY;
3951 goto err;
3952 }
3953
3954 req = embedded_payload(wrb);
3955 ctxt = &req->context;
3956
3957 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3958 OPCODE_COMMON_GET_HSW_CONFIG, sizeof(*req), wrb,
3959 NULL);
3960
3961 req->hdr.domain = domain;
3962 AMAP_SET_BITS(struct amap_get_hsw_req_context, interface_id,
3963 ctxt, intf_id);
3964 AMAP_SET_BITS(struct amap_get_hsw_req_context, pvid_valid, ctxt, 1);
3965
3966 if (!BEx_chip(adapter) && mode) {
3967 AMAP_SET_BITS(struct amap_get_hsw_req_context, interface_id,
3968 ctxt, adapter->hba_port_num);
3969 AMAP_SET_BITS(struct amap_get_hsw_req_context, pport, ctxt, 1);
3970 }
3971 be_dws_cpu_to_le(req->context, sizeof(req->context));
3972
3973 status = be_mcc_notify_wait(adapter);
3974 if (!status) {
3975 struct be_cmd_resp_get_hsw_config *resp =
3976 embedded_payload(wrb);
3977
3978 be_dws_le_to_cpu(&resp->context, sizeof(resp->context));
3979 vid = AMAP_GET_BITS(struct amap_get_hsw_resp_context,
3980 pvid, &resp->context);
3981 if (pvid)
3982 *pvid = le16_to_cpu(vid);
3983 if (mode)
3984 *mode = AMAP_GET_BITS(struct amap_get_hsw_resp_context,
3985 port_fwd_type, &resp->context);
3986 if (spoofchk)
3987 *spoofchk =
3988 AMAP_GET_BITS(struct amap_get_hsw_resp_context,
3989 spoofchk, &resp->context);
3990 }
3991
3992 err:
3993 mutex_unlock(&adapter->mcc_lock);
3994 return status;
3995 }
3996
3997 static bool be_is_wol_excluded(struct be_adapter *adapter)
3998 {
3999 struct pci_dev *pdev = adapter->pdev;
4000
4001 if (be_virtfn(adapter))
4002 return true;
4003
4004 switch (pdev->subsystem_device) {
4005 case OC_SUBSYS_DEVICE_ID1:
4006 case OC_SUBSYS_DEVICE_ID2:
4007 case OC_SUBSYS_DEVICE_ID3:
4008 case OC_SUBSYS_DEVICE_ID4:
4009 return true;
4010 default:
4011 return false;
4012 }
4013 }
4014
4015 int be_cmd_get_acpi_wol_cap(struct be_adapter *adapter)
4016 {
4017 struct be_mcc_wrb *wrb;
4018 struct be_cmd_req_acpi_wol_magic_config_v1 *req;
4019 int status = 0;
4020 struct be_dma_mem cmd;
4021
4022 if (!be_cmd_allowed(adapter, OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG,
4023 CMD_SUBSYSTEM_ETH))
4024 return -EPERM;
4025
4026 if (be_is_wol_excluded(adapter))
4027 return status;
4028
4029 if (mutex_lock_interruptible(&adapter->mbox_lock))
4030 return -1;
4031
4032 memset(&cmd, 0, sizeof(struct be_dma_mem));
4033 cmd.size = sizeof(struct be_cmd_resp_acpi_wol_magic_config_v1);
4034 cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
4035 GFP_ATOMIC);
4036 if (!cmd.va) {
4037 dev_err(&adapter->pdev->dev, "Memory allocation failure\n");
4038 status = -ENOMEM;
4039 goto err;
4040 }
4041
4042 wrb = wrb_from_mbox(adapter);
4043 if (!wrb) {
4044 status = -EBUSY;
4045 goto err;
4046 }
4047
4048 req = cmd.va;
4049
4050 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
4051 OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG,
4052 sizeof(*req), wrb, &cmd);
4053
4054 req->hdr.version = 1;
4055 req->query_options = BE_GET_WOL_CAP;
4056
4057 status = be_mbox_notify_wait(adapter);
4058 if (!status) {
4059 struct be_cmd_resp_acpi_wol_magic_config_v1 *resp;
4060
4061 resp = (struct be_cmd_resp_acpi_wol_magic_config_v1 *)cmd.va;
4062
4063 adapter->wol_cap = resp->wol_settings;
4064
4065 /* Non-zero macaddr indicates WOL is enabled */
4066 if (adapter->wol_cap & BE_WOL_CAP &&
4067 !is_zero_ether_addr(resp->magic_mac))
4068 adapter->wol_en = true;
4069 }
4070 err:
4071 mutex_unlock(&adapter->mbox_lock);
4072 if (cmd.va)
4073 dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va,
4074 cmd.dma);
4075 return status;
4076
4077 }
4078
4079 int be_cmd_set_fw_log_level(struct be_adapter *adapter, u32 level)
4080 {
4081 struct be_dma_mem extfat_cmd;
4082 struct be_fat_conf_params *cfgs;
4083 int status;
4084 int i, j;
4085
4086 memset(&extfat_cmd, 0, sizeof(struct be_dma_mem));
4087 extfat_cmd.size = sizeof(struct be_cmd_resp_get_ext_fat_caps);
4088 extfat_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
4089 extfat_cmd.size, &extfat_cmd.dma,
4090 GFP_ATOMIC);
4091 if (!extfat_cmd.va)
4092 return -ENOMEM;
4093
4094 status = be_cmd_get_ext_fat_capabilites(adapter, &extfat_cmd);
4095 if (status)
4096 goto err;
4097
4098 cfgs = (struct be_fat_conf_params *)
4099 (extfat_cmd.va + sizeof(struct be_cmd_resp_hdr));
4100 for (i = 0; i < le32_to_cpu(cfgs->num_modules); i++) {
4101 u32 num_modes = le32_to_cpu(cfgs->module[i].num_modes);
4102
4103 for (j = 0; j < num_modes; j++) {
4104 if (cfgs->module[i].trace_lvl[j].mode == MODE_UART)
4105 cfgs->module[i].trace_lvl[j].dbg_lvl =
4106 cpu_to_le32(level);
4107 }
4108 }
4109
4110 status = be_cmd_set_ext_fat_capabilites(adapter, &extfat_cmd, cfgs);
4111 err:
4112 dma_free_coherent(&adapter->pdev->dev, extfat_cmd.size, extfat_cmd.va,
4113 extfat_cmd.dma);
4114 return status;
4115 }
4116
4117 int be_cmd_get_fw_log_level(struct be_adapter *adapter)
4118 {
4119 struct be_dma_mem extfat_cmd;
4120 struct be_fat_conf_params *cfgs;
4121 int status, j;
4122 int level = 0;
4123
4124 memset(&extfat_cmd, 0, sizeof(struct be_dma_mem));
4125 extfat_cmd.size = sizeof(struct be_cmd_resp_get_ext_fat_caps);
4126 extfat_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
4127 extfat_cmd.size, &extfat_cmd.dma,
4128 GFP_ATOMIC);
4129
4130 if (!extfat_cmd.va) {
4131 dev_err(&adapter->pdev->dev, "%s: Memory allocation failure\n",
4132 __func__);
4133 goto err;
4134 }
4135
4136 status = be_cmd_get_ext_fat_capabilites(adapter, &extfat_cmd);
4137 if (!status) {
4138 cfgs = (struct be_fat_conf_params *)(extfat_cmd.va +
4139 sizeof(struct be_cmd_resp_hdr));
4140
4141 for (j = 0; j < le32_to_cpu(cfgs->module[0].num_modes); j++) {
4142 if (cfgs->module[0].trace_lvl[j].mode == MODE_UART)
4143 level = cfgs->module[0].trace_lvl[j].dbg_lvl;
4144 }
4145 }
4146 dma_free_coherent(&adapter->pdev->dev, extfat_cmd.size, extfat_cmd.va,
4147 extfat_cmd.dma);
4148 err:
4149 return level;
4150 }
4151
4152 int be_cmd_get_ext_fat_capabilites(struct be_adapter *adapter,
4153 struct be_dma_mem *cmd)
4154 {
4155 struct be_mcc_wrb *wrb;
4156 struct be_cmd_req_get_ext_fat_caps *req;
4157 int status;
4158
4159 if (!be_cmd_allowed(adapter, OPCODE_COMMON_GET_EXT_FAT_CAPABILITIES,
4160 CMD_SUBSYSTEM_COMMON))
4161 return -EPERM;
4162
4163 if (mutex_lock_interruptible(&adapter->mbox_lock))
4164 return -1;
4165
4166 wrb = wrb_from_mbox(adapter);
4167 if (!wrb) {
4168 status = -EBUSY;
4169 goto err;
4170 }
4171
4172 req = cmd->va;
4173 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4174 OPCODE_COMMON_GET_EXT_FAT_CAPABILITIES,
4175 cmd->size, wrb, cmd);
4176 req->parameter_type = cpu_to_le32(1);
4177
4178 status = be_mbox_notify_wait(adapter);
4179 err:
4180 mutex_unlock(&adapter->mbox_lock);
4181 return status;
4182 }
4183
4184 int be_cmd_set_ext_fat_capabilites(struct be_adapter *adapter,
4185 struct be_dma_mem *cmd,
4186 struct be_fat_conf_params *configs)
4187 {
4188 struct be_mcc_wrb *wrb;
4189 struct be_cmd_req_set_ext_fat_caps *req;
4190 int status;
4191
4192 mutex_lock(&adapter->mcc_lock);
4193
4194 wrb = wrb_from_mccq(adapter);
4195 if (!wrb) {
4196 status = -EBUSY;
4197 goto err;
4198 }
4199
4200 req = cmd->va;
4201 memcpy(&req->set_params, configs, sizeof(struct be_fat_conf_params));
4202 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4203 OPCODE_COMMON_SET_EXT_FAT_CAPABILITIES,
4204 cmd->size, wrb, cmd);
4205
4206 status = be_mcc_notify_wait(adapter);
4207 err:
4208 mutex_unlock(&adapter->mcc_lock);
4209 return status;
4210 }
4211
4212 int be_cmd_query_port_name(struct be_adapter *adapter)
4213 {
4214 struct be_cmd_req_get_port_name *req;
4215 struct be_mcc_wrb *wrb;
4216 int status;
4217
4218 if (mutex_lock_interruptible(&adapter->mbox_lock))
4219 return -1;
4220
4221 wrb = wrb_from_mbox(adapter);
4222 req = embedded_payload(wrb);
4223
4224 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4225 OPCODE_COMMON_GET_PORT_NAME, sizeof(*req), wrb,
4226 NULL);
4227 if (!BEx_chip(adapter))
4228 req->hdr.version = 1;
4229
4230 status = be_mbox_notify_wait(adapter);
4231 if (!status) {
4232 struct be_cmd_resp_get_port_name *resp = embedded_payload(wrb);
4233
4234 adapter->port_name = resp->port_name[adapter->hba_port_num];
4235 } else {
4236 adapter->port_name = adapter->hba_port_num + '0';
4237 }
4238
4239 mutex_unlock(&adapter->mbox_lock);
4240 return status;
4241 }
4242
4243 /* When more than 1 NIC descriptor is present in the descriptor list,
4244 * the caller must specify the pf_num to obtain the NIC descriptor
4245 * corresponding to its pci function.
4246 * get_vft must be true when the caller wants the VF-template desc of the
4247 * PF-pool.
4248 * The pf_num should be set to PF_NUM_IGNORE when the caller knows
4249 * that only it's NIC descriptor is present in the descriptor list.
4250 */
4251 static struct be_nic_res_desc *be_get_nic_desc(u8 *buf, u32 desc_count,
4252 bool get_vft, u8 pf_num)
4253 {
4254 struct be_res_desc_hdr *hdr = (struct be_res_desc_hdr *)buf;
4255 struct be_nic_res_desc *nic;
4256 int i;
4257
4258 for (i = 0; i < desc_count; i++) {
4259 if (hdr->desc_type == NIC_RESOURCE_DESC_TYPE_V0 ||
4260 hdr->desc_type == NIC_RESOURCE_DESC_TYPE_V1) {
4261 nic = (struct be_nic_res_desc *)hdr;
4262
4263 if ((pf_num == PF_NUM_IGNORE ||
4264 nic->pf_num == pf_num) &&
4265 (!get_vft || nic->flags & BIT(VFT_SHIFT)))
4266 return nic;
4267 }
4268 hdr->desc_len = hdr->desc_len ? : RESOURCE_DESC_SIZE_V0;
4269 hdr = (void *)hdr + hdr->desc_len;
4270 }
4271 return NULL;
4272 }
4273
4274 static struct be_nic_res_desc *be_get_vft_desc(u8 *buf, u32 desc_count,
4275 u8 pf_num)
4276 {
4277 return be_get_nic_desc(buf, desc_count, true, pf_num);
4278 }
4279
4280 static struct be_nic_res_desc *be_get_func_nic_desc(u8 *buf, u32 desc_count,
4281 u8 pf_num)
4282 {
4283 return be_get_nic_desc(buf, desc_count, false, pf_num);
4284 }
4285
4286 static struct be_pcie_res_desc *be_get_pcie_desc(u8 *buf, u32 desc_count,
4287 u8 pf_num)
4288 {
4289 struct be_res_desc_hdr *hdr = (struct be_res_desc_hdr *)buf;
4290 struct be_pcie_res_desc *pcie;
4291 int i;
4292
4293 for (i = 0; i < desc_count; i++) {
4294 if (hdr->desc_type == PCIE_RESOURCE_DESC_TYPE_V0 ||
4295 hdr->desc_type == PCIE_RESOURCE_DESC_TYPE_V1) {
4296 pcie = (struct be_pcie_res_desc *)hdr;
4297 if (pcie->pf_num == pf_num)
4298 return pcie;
4299 }
4300
4301 hdr->desc_len = hdr->desc_len ? : RESOURCE_DESC_SIZE_V0;
4302 hdr = (void *)hdr + hdr->desc_len;
4303 }
4304 return NULL;
4305 }
4306
4307 static struct be_port_res_desc *be_get_port_desc(u8 *buf, u32 desc_count)
4308 {
4309 struct be_res_desc_hdr *hdr = (struct be_res_desc_hdr *)buf;
4310 int i;
4311
4312 for (i = 0; i < desc_count; i++) {
4313 if (hdr->desc_type == PORT_RESOURCE_DESC_TYPE_V1)
4314 return (struct be_port_res_desc *)hdr;
4315
4316 hdr->desc_len = hdr->desc_len ? : RESOURCE_DESC_SIZE_V0;
4317 hdr = (void *)hdr + hdr->desc_len;
4318 }
4319 return NULL;
4320 }
4321
4322 static void be_copy_nic_desc(struct be_resources *res,
4323 struct be_nic_res_desc *desc)
4324 {
4325 res->max_uc_mac = le16_to_cpu(desc->unicast_mac_count);
4326 res->max_vlans = le16_to_cpu(desc->vlan_count);
4327 res->max_mcast_mac = le16_to_cpu(desc->mcast_mac_count);
4328 res->max_tx_qs = le16_to_cpu(desc->txq_count);
4329 res->max_rss_qs = le16_to_cpu(desc->rssq_count);
4330 res->max_rx_qs = le16_to_cpu(desc->rq_count);
4331 res->max_evt_qs = le16_to_cpu(desc->eq_count);
4332 res->max_cq_count = le16_to_cpu(desc->cq_count);
4333 res->max_iface_count = le16_to_cpu(desc->iface_count);
4334 res->max_mcc_count = le16_to_cpu(desc->mcc_count);
4335 /* Clear flags that driver is not interested in */
4336 res->if_cap_flags = le32_to_cpu(desc->cap_flags) &
4337 BE_IF_CAP_FLAGS_WANT;
4338 }
4339
4340 /* Uses Mbox */
4341 int be_cmd_get_func_config(struct be_adapter *adapter, struct be_resources *res)
4342 {
4343 struct be_mcc_wrb *wrb;
4344 struct be_cmd_req_get_func_config *req;
4345 int status;
4346 struct be_dma_mem cmd;
4347
4348 if (mutex_lock_interruptible(&adapter->mbox_lock))
4349 return -1;
4350
4351 memset(&cmd, 0, sizeof(struct be_dma_mem));
4352 cmd.size = sizeof(struct be_cmd_resp_get_func_config);
4353 cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
4354 GFP_ATOMIC);
4355 if (!cmd.va) {
4356 dev_err(&adapter->pdev->dev, "Memory alloc failure\n");
4357 status = -ENOMEM;
4358 goto err;
4359 }
4360
4361 wrb = wrb_from_mbox(adapter);
4362 if (!wrb) {
4363 status = -EBUSY;
4364 goto err;
4365 }
4366
4367 req = cmd.va;
4368
4369 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4370 OPCODE_COMMON_GET_FUNC_CONFIG,
4371 cmd.size, wrb, &cmd);
4372
4373 if (skyhawk_chip(adapter))
4374 req->hdr.version = 1;
4375
4376 status = be_mbox_notify_wait(adapter);
4377 if (!status) {
4378 struct be_cmd_resp_get_func_config *resp = cmd.va;
4379 u32 desc_count = le32_to_cpu(resp->desc_count);
4380 struct be_nic_res_desc *desc;
4381
4382 /* GET_FUNC_CONFIG returns resource descriptors of the
4383 * current function only. So, pf_num should be set to
4384 * PF_NUM_IGNORE.
4385 */
4386 desc = be_get_func_nic_desc(resp->func_param, desc_count,
4387 PF_NUM_IGNORE);
4388 if (!desc) {
4389 status = -EINVAL;
4390 goto err;
4391 }
4392
4393 /* Store pf_num & vf_num for later use in GET_PROFILE_CONFIG */
4394 adapter->pf_num = desc->pf_num;
4395 adapter->vf_num = desc->vf_num;
4396
4397 if (res)
4398 be_copy_nic_desc(res, desc);
4399 }
4400 err:
4401 mutex_unlock(&adapter->mbox_lock);
4402 if (cmd.va)
4403 dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va,
4404 cmd.dma);
4405 return status;
4406 }
4407
4408 /* This routine returns a list of all the NIC PF_nums in the adapter */
4409 static u16 be_get_nic_pf_num_list(u8 *buf, u32 desc_count, u16 *nic_pf_nums)
4410 {
4411 struct be_res_desc_hdr *hdr = (struct be_res_desc_hdr *)buf;
4412 struct be_pcie_res_desc *pcie = NULL;
4413 int i;
4414 u16 nic_pf_count = 0;
4415
4416 for (i = 0; i < desc_count; i++) {
4417 if (hdr->desc_type == PCIE_RESOURCE_DESC_TYPE_V0 ||
4418 hdr->desc_type == PCIE_RESOURCE_DESC_TYPE_V1) {
4419 pcie = (struct be_pcie_res_desc *)hdr;
4420 if (pcie->pf_state && (pcie->pf_type == MISSION_NIC ||
4421 pcie->pf_type == MISSION_RDMA)) {
4422 nic_pf_nums[nic_pf_count++] = pcie->pf_num;
4423 }
4424 }
4425
4426 hdr->desc_len = hdr->desc_len ? : RESOURCE_DESC_SIZE_V0;
4427 hdr = (void *)hdr + hdr->desc_len;
4428 }
4429 return nic_pf_count;
4430 }
4431
4432 /* Will use MBOX only if MCCQ has not been created */
4433 int be_cmd_get_profile_config(struct be_adapter *adapter,
4434 struct be_resources *res,
4435 struct be_port_resources *port_res,
4436 u8 profile_type, u8 query, u8 domain)
4437 {
4438 struct be_cmd_resp_get_profile_config *resp;
4439 struct be_cmd_req_get_profile_config *req;
4440 struct be_nic_res_desc *vf_res;
4441 struct be_pcie_res_desc *pcie;
4442 struct be_port_res_desc *port;
4443 struct be_nic_res_desc *nic;
4444 struct be_mcc_wrb wrb = {0};
4445 struct be_dma_mem cmd;
4446 u16 desc_count;
4447 int status;
4448
4449 memset(&cmd, 0, sizeof(struct be_dma_mem));
4450 cmd.size = sizeof(struct be_cmd_resp_get_profile_config);
4451 cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
4452 GFP_ATOMIC);
4453 if (!cmd.va)
4454 return -ENOMEM;
4455
4456 req = cmd.va;
4457 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4458 OPCODE_COMMON_GET_PROFILE_CONFIG,
4459 cmd.size, &wrb, &cmd);
4460
4461 if (!lancer_chip(adapter))
4462 req->hdr.version = 1;
4463 req->type = profile_type;
4464 req->hdr.domain = domain;
4465
4466 /* When QUERY_MODIFIABLE_FIELDS_TYPE bit is set, cmd returns the
4467 * descriptors with all bits set to "1" for the fields which can be
4468 * modified using SET_PROFILE_CONFIG cmd.
4469 */
4470 if (query == RESOURCE_MODIFIABLE)
4471 req->type |= QUERY_MODIFIABLE_FIELDS_TYPE;
4472
4473 status = be_cmd_notify_wait(adapter, &wrb);
4474 if (status)
4475 goto err;
4476
4477 resp = cmd.va;
4478 desc_count = le16_to_cpu(resp->desc_count);
4479
4480 if (port_res) {
4481 u16 nic_pf_cnt = 0, i;
4482 u16 nic_pf_num_list[MAX_NIC_FUNCS];
4483
4484 nic_pf_cnt = be_get_nic_pf_num_list(resp->func_param,
4485 desc_count,
4486 nic_pf_num_list);
4487
4488 for (i = 0; i < nic_pf_cnt; i++) {
4489 nic = be_get_func_nic_desc(resp->func_param, desc_count,
4490 nic_pf_num_list[i]);
4491 if (nic->link_param == adapter->port_num) {
4492 port_res->nic_pfs++;
4493 pcie = be_get_pcie_desc(resp->func_param,
4494 desc_count,
4495 nic_pf_num_list[i]);
4496 port_res->max_vfs += le16_to_cpu(pcie->num_vfs);
4497 }
4498 }
4499 goto err;
4500 }
4501
4502 pcie = be_get_pcie_desc(resp->func_param, desc_count,
4503 adapter->pf_num);
4504 if (pcie)
4505 res->max_vfs = le16_to_cpu(pcie->num_vfs);
4506
4507 port = be_get_port_desc(resp->func_param, desc_count);
4508 if (port)
4509 adapter->mc_type = port->mc_type;
4510
4511 nic = be_get_func_nic_desc(resp->func_param, desc_count,
4512 adapter->pf_num);
4513 if (nic)
4514 be_copy_nic_desc(res, nic);
4515
4516 vf_res = be_get_vft_desc(resp->func_param, desc_count,
4517 adapter->pf_num);
4518 if (vf_res)
4519 res->vf_if_cap_flags = vf_res->cap_flags;
4520 err:
4521 if (cmd.va)
4522 dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va,
4523 cmd.dma);
4524 return status;
4525 }
4526
4527 /* Will use MBOX only if MCCQ has not been created */
4528 static int be_cmd_set_profile_config(struct be_adapter *adapter, void *desc,
4529 int size, int count, u8 version, u8 domain)
4530 {
4531 struct be_cmd_req_set_profile_config *req;
4532 struct be_mcc_wrb wrb = {0};
4533 struct be_dma_mem cmd;
4534 int status;
4535
4536 memset(&cmd, 0, sizeof(struct be_dma_mem));
4537 cmd.size = sizeof(struct be_cmd_req_set_profile_config);
4538 cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
4539 GFP_ATOMIC);
4540 if (!cmd.va)
4541 return -ENOMEM;
4542
4543 req = cmd.va;
4544 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4545 OPCODE_COMMON_SET_PROFILE_CONFIG, cmd.size,
4546 &wrb, &cmd);
4547 req->hdr.version = version;
4548 req->hdr.domain = domain;
4549 req->desc_count = cpu_to_le32(count);
4550 memcpy(req->desc, desc, size);
4551
4552 status = be_cmd_notify_wait(adapter, &wrb);
4553
4554 if (cmd.va)
4555 dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va,
4556 cmd.dma);
4557 return status;
4558 }
4559
4560 /* Mark all fields invalid */
4561 static void be_reset_nic_desc(struct be_nic_res_desc *nic)
4562 {
4563 memset(nic, 0, sizeof(*nic));
4564 nic->unicast_mac_count = 0xFFFF;
4565 nic->mcc_count = 0xFFFF;
4566 nic->vlan_count = 0xFFFF;
4567 nic->mcast_mac_count = 0xFFFF;
4568 nic->txq_count = 0xFFFF;
4569 nic->rq_count = 0xFFFF;
4570 nic->rssq_count = 0xFFFF;
4571 nic->lro_count = 0xFFFF;
4572 nic->cq_count = 0xFFFF;
4573 nic->toe_conn_count = 0xFFFF;
4574 nic->eq_count = 0xFFFF;
4575 nic->iface_count = 0xFFFF;
4576 nic->link_param = 0xFF;
4577 nic->channel_id_param = cpu_to_le16(0xF000);
4578 nic->acpi_params = 0xFF;
4579 nic->wol_param = 0x0F;
4580 nic->tunnel_iface_count = 0xFFFF;
4581 nic->direct_tenant_iface_count = 0xFFFF;
4582 nic->bw_min = 0xFFFFFFFF;
4583 nic->bw_max = 0xFFFFFFFF;
4584 }
4585
4586 /* Mark all fields invalid */
4587 static void be_reset_pcie_desc(struct be_pcie_res_desc *pcie)
4588 {
4589 memset(pcie, 0, sizeof(*pcie));
4590 pcie->sriov_state = 0xFF;
4591 pcie->pf_state = 0xFF;
4592 pcie->pf_type = 0xFF;
4593 pcie->num_vfs = 0xFFFF;
4594 }
4595
4596 int be_cmd_config_qos(struct be_adapter *adapter, u32 max_rate, u16 link_speed,
4597 u8 domain)
4598 {
4599 struct be_nic_res_desc nic_desc;
4600 u32 bw_percent;
4601 u16 version = 0;
4602
4603 if (BE3_chip(adapter))
4604 return be_cmd_set_qos(adapter, max_rate / 10, domain);
4605
4606 be_reset_nic_desc(&nic_desc);
4607 nic_desc.pf_num = adapter->pf_num;
4608 nic_desc.vf_num = domain;
4609 nic_desc.bw_min = 0;
4610 if (lancer_chip(adapter)) {
4611 nic_desc.hdr.desc_type = NIC_RESOURCE_DESC_TYPE_V0;
4612 nic_desc.hdr.desc_len = RESOURCE_DESC_SIZE_V0;
4613 nic_desc.flags = (1 << QUN_SHIFT) | (1 << IMM_SHIFT) |
4614 (1 << NOSV_SHIFT);
4615 nic_desc.bw_max = cpu_to_le32(max_rate / 10);
4616 } else {
4617 version = 1;
4618 nic_desc.hdr.desc_type = NIC_RESOURCE_DESC_TYPE_V1;
4619 nic_desc.hdr.desc_len = RESOURCE_DESC_SIZE_V1;
4620 nic_desc.flags = (1 << IMM_SHIFT) | (1 << NOSV_SHIFT);
4621 bw_percent = max_rate ? (max_rate * 100) / link_speed : 100;
4622 nic_desc.bw_max = cpu_to_le32(bw_percent);
4623 }
4624
4625 return be_cmd_set_profile_config(adapter, &nic_desc,
4626 nic_desc.hdr.desc_len,
4627 1, version, domain);
4628 }
4629
4630 int be_cmd_set_sriov_config(struct be_adapter *adapter,
4631 struct be_resources pool_res, u16 num_vfs,
4632 struct be_resources *vft_res)
4633 {
4634 struct {
4635 struct be_pcie_res_desc pcie;
4636 struct be_nic_res_desc nic_vft;
4637 } __packed desc;
4638
4639 /* PF PCIE descriptor */
4640 be_reset_pcie_desc(&desc.pcie);
4641 desc.pcie.hdr.desc_type = PCIE_RESOURCE_DESC_TYPE_V1;
4642 desc.pcie.hdr.desc_len = RESOURCE_DESC_SIZE_V1;
4643 desc.pcie.flags = BIT(IMM_SHIFT) | BIT(NOSV_SHIFT);
4644 desc.pcie.pf_num = adapter->pdev->devfn;
4645 desc.pcie.sriov_state = num_vfs ? 1 : 0;
4646 desc.pcie.num_vfs = cpu_to_le16(num_vfs);
4647
4648 /* VF NIC Template descriptor */
4649 be_reset_nic_desc(&desc.nic_vft);
4650 desc.nic_vft.hdr.desc_type = NIC_RESOURCE_DESC_TYPE_V1;
4651 desc.nic_vft.hdr.desc_len = RESOURCE_DESC_SIZE_V1;
4652 desc.nic_vft.flags = vft_res->flags | BIT(VFT_SHIFT) |
4653 BIT(IMM_SHIFT) | BIT(NOSV_SHIFT);
4654 desc.nic_vft.pf_num = adapter->pdev->devfn;
4655 desc.nic_vft.vf_num = 0;
4656 desc.nic_vft.cap_flags = cpu_to_le32(vft_res->vf_if_cap_flags);
4657 desc.nic_vft.rq_count = cpu_to_le16(vft_res->max_rx_qs);
4658 desc.nic_vft.txq_count = cpu_to_le16(vft_res->max_tx_qs);
4659 desc.nic_vft.rssq_count = cpu_to_le16(vft_res->max_rss_qs);
4660 desc.nic_vft.cq_count = cpu_to_le16(vft_res->max_cq_count);
4661
4662 if (vft_res->max_uc_mac)
4663 desc.nic_vft.unicast_mac_count =
4664 cpu_to_le16(vft_res->max_uc_mac);
4665 if (vft_res->max_vlans)
4666 desc.nic_vft.vlan_count = cpu_to_le16(vft_res->max_vlans);
4667 if (vft_res->max_iface_count)
4668 desc.nic_vft.iface_count =
4669 cpu_to_le16(vft_res->max_iface_count);
4670 if (vft_res->max_mcc_count)
4671 desc.nic_vft.mcc_count = cpu_to_le16(vft_res->max_mcc_count);
4672
4673 return be_cmd_set_profile_config(adapter, &desc,
4674 2 * RESOURCE_DESC_SIZE_V1, 2, 1, 0);
4675 }
4676
4677 int be_cmd_manage_iface(struct be_adapter *adapter, u32 iface, u8 op)
4678 {
4679 struct be_mcc_wrb *wrb;
4680 struct be_cmd_req_manage_iface_filters *req;
4681 int status;
4682
4683 if (iface == 0xFFFFFFFF)
4684 return -1;
4685
4686 mutex_lock(&adapter->mcc_lock);
4687
4688 wrb = wrb_from_mccq(adapter);
4689 if (!wrb) {
4690 status = -EBUSY;
4691 goto err;
4692 }
4693 req = embedded_payload(wrb);
4694
4695 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4696 OPCODE_COMMON_MANAGE_IFACE_FILTERS, sizeof(*req),
4697 wrb, NULL);
4698 req->op = op;
4699 req->target_iface_id = cpu_to_le32(iface);
4700
4701 status = be_mcc_notify_wait(adapter);
4702 err:
4703 mutex_unlock(&adapter->mcc_lock);
4704 return status;
4705 }
4706
4707 int be_cmd_set_vxlan_port(struct be_adapter *adapter, __be16 port)
4708 {
4709 struct be_port_res_desc port_desc;
4710
4711 memset(&port_desc, 0, sizeof(port_desc));
4712 port_desc.hdr.desc_type = PORT_RESOURCE_DESC_TYPE_V1;
4713 port_desc.hdr.desc_len = RESOURCE_DESC_SIZE_V1;
4714 port_desc.flags = (1 << IMM_SHIFT) | (1 << NOSV_SHIFT);
4715 port_desc.link_num = adapter->hba_port_num;
4716 if (port) {
4717 port_desc.nv_flags = NV_TYPE_VXLAN | (1 << SOCVID_SHIFT) |
4718 (1 << RCVID_SHIFT);
4719 port_desc.nv_port = swab16(port);
4720 } else {
4721 port_desc.nv_flags = NV_TYPE_DISABLED;
4722 port_desc.nv_port = 0;
4723 }
4724
4725 return be_cmd_set_profile_config(adapter, &port_desc,
4726 RESOURCE_DESC_SIZE_V1, 1, 1, 0);
4727 }
4728
4729 int be_cmd_get_if_id(struct be_adapter *adapter, struct be_vf_cfg *vf_cfg,
4730 int vf_num)
4731 {
4732 struct be_mcc_wrb *wrb;
4733 struct be_cmd_req_get_iface_list *req;
4734 struct be_cmd_resp_get_iface_list *resp;
4735 int status;
4736
4737 mutex_lock(&adapter->mcc_lock);
4738
4739 wrb = wrb_from_mccq(adapter);
4740 if (!wrb) {
4741 status = -EBUSY;
4742 goto err;
4743 }
4744 req = embedded_payload(wrb);
4745
4746 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4747 OPCODE_COMMON_GET_IFACE_LIST, sizeof(*resp),
4748 wrb, NULL);
4749 req->hdr.domain = vf_num + 1;
4750
4751 status = be_mcc_notify_wait(adapter);
4752 if (!status) {
4753 resp = (struct be_cmd_resp_get_iface_list *)req;
4754 vf_cfg->if_handle = le32_to_cpu(resp->if_desc.if_id);
4755 }
4756
4757 err:
4758 mutex_unlock(&adapter->mcc_lock);
4759 return status;
4760 }
4761
4762 static int lancer_wait_idle(struct be_adapter *adapter)
4763 {
4764 #define SLIPORT_IDLE_TIMEOUT 30
4765 u32 reg_val;
4766 int status = 0, i;
4767
4768 for (i = 0; i < SLIPORT_IDLE_TIMEOUT; i++) {
4769 reg_val = ioread32(adapter->db + PHYSDEV_CONTROL_OFFSET);
4770 if ((reg_val & PHYSDEV_CONTROL_INP_MASK) == 0)
4771 break;
4772
4773 ssleep(1);
4774 }
4775
4776 if (i == SLIPORT_IDLE_TIMEOUT)
4777 status = -1;
4778
4779 return status;
4780 }
4781
4782 int lancer_physdev_ctrl(struct be_adapter *adapter, u32 mask)
4783 {
4784 int status = 0;
4785
4786 status = lancer_wait_idle(adapter);
4787 if (status)
4788 return status;
4789
4790 iowrite32(mask, adapter->db + PHYSDEV_CONTROL_OFFSET);
4791
4792 return status;
4793 }
4794
4795 /* Routine to check whether dump image is present or not */
4796 bool dump_present(struct be_adapter *adapter)
4797 {
4798 u32 sliport_status = 0;
4799
4800 sliport_status = ioread32(adapter->db + SLIPORT_STATUS_OFFSET);
4801 return !!(sliport_status & SLIPORT_STATUS_DIP_MASK);
4802 }
4803
4804 int lancer_initiate_dump(struct be_adapter *adapter)
4805 {
4806 struct device *dev = &adapter->pdev->dev;
4807 int status;
4808
4809 if (dump_present(adapter)) {
4810 dev_info(dev, "Previous dump not cleared, not forcing dump\n");
4811 return -EEXIST;
4812 }
4813
4814 /* give firmware reset and diagnostic dump */
4815 status = lancer_physdev_ctrl(adapter, PHYSDEV_CONTROL_FW_RESET_MASK |
4816 PHYSDEV_CONTROL_DD_MASK);
4817 if (status < 0) {
4818 dev_err(dev, "FW reset failed\n");
4819 return status;
4820 }
4821
4822 status = lancer_wait_idle(adapter);
4823 if (status)
4824 return status;
4825
4826 if (!dump_present(adapter)) {
4827 dev_err(dev, "FW dump not generated\n");
4828 return -EIO;
4829 }
4830
4831 return 0;
4832 }
4833
4834 int lancer_delete_dump(struct be_adapter *adapter)
4835 {
4836 int status;
4837
4838 status = lancer_cmd_delete_object(adapter, LANCER_FW_DUMP_FILE);
4839 return be_cmd_status(status);
4840 }
4841
4842 /* Uses sync mcc */
4843 int be_cmd_enable_vf(struct be_adapter *adapter, u8 domain)
4844 {
4845 struct be_mcc_wrb *wrb;
4846 struct be_cmd_enable_disable_vf *req;
4847 int status;
4848
4849 if (BEx_chip(adapter))
4850 return 0;
4851
4852 mutex_lock(&adapter->mcc_lock);
4853
4854 wrb = wrb_from_mccq(adapter);
4855 if (!wrb) {
4856 status = -EBUSY;
4857 goto err;
4858 }
4859
4860 req = embedded_payload(wrb);
4861
4862 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4863 OPCODE_COMMON_ENABLE_DISABLE_VF, sizeof(*req),
4864 wrb, NULL);
4865
4866 req->hdr.domain = domain;
4867 req->enable = 1;
4868 status = be_mcc_notify_wait(adapter);
4869 err:
4870 mutex_unlock(&adapter->mcc_lock);
4871 return status;
4872 }
4873
4874 int be_cmd_intr_set(struct be_adapter *adapter, bool intr_enable)
4875 {
4876 struct be_mcc_wrb *wrb;
4877 struct be_cmd_req_intr_set *req;
4878 int status;
4879
4880 if (mutex_lock_interruptible(&adapter->mbox_lock))
4881 return -1;
4882
4883 wrb = wrb_from_mbox(adapter);
4884
4885 req = embedded_payload(wrb);
4886
4887 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4888 OPCODE_COMMON_SET_INTERRUPT_ENABLE, sizeof(*req),
4889 wrb, NULL);
4890
4891 req->intr_enabled = intr_enable;
4892
4893 status = be_mbox_notify_wait(adapter);
4894
4895 mutex_unlock(&adapter->mbox_lock);
4896 return status;
4897 }
4898
4899 /* Uses MBOX */
4900 int be_cmd_get_active_profile(struct be_adapter *adapter, u16 *profile_id)
4901 {
4902 struct be_cmd_req_get_active_profile *req;
4903 struct be_mcc_wrb *wrb;
4904 int status;
4905
4906 if (mutex_lock_interruptible(&adapter->mbox_lock))
4907 return -1;
4908
4909 wrb = wrb_from_mbox(adapter);
4910 if (!wrb) {
4911 status = -EBUSY;
4912 goto err;
4913 }
4914
4915 req = embedded_payload(wrb);
4916
4917 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4918 OPCODE_COMMON_GET_ACTIVE_PROFILE, sizeof(*req),
4919 wrb, NULL);
4920
4921 status = be_mbox_notify_wait(adapter);
4922 if (!status) {
4923 struct be_cmd_resp_get_active_profile *resp =
4924 embedded_payload(wrb);
4925
4926 *profile_id = le16_to_cpu(resp->active_profile_id);
4927 }
4928
4929 err:
4930 mutex_unlock(&adapter->mbox_lock);
4931 return status;
4932 }
4933
4934 static int
4935 __be_cmd_set_logical_link_config(struct be_adapter *adapter,
4936 int link_state, int version, u8 domain)
4937 {
4938 struct be_cmd_req_set_ll_link *req;
4939 struct be_mcc_wrb *wrb;
4940 u32 link_config = 0;
4941 int status;
4942
4943 mutex_lock(&adapter->mcc_lock);
4944
4945 wrb = wrb_from_mccq(adapter);
4946 if (!wrb) {
4947 status = -EBUSY;
4948 goto err;
4949 }
4950
4951 req = embedded_payload(wrb);
4952
4953 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
4954 OPCODE_COMMON_SET_LOGICAL_LINK_CONFIG,
4955 sizeof(*req), wrb, NULL);
4956
4957 req->hdr.version = version;
4958 req->hdr.domain = domain;
4959
4960 if (link_state == IFLA_VF_LINK_STATE_ENABLE ||
4961 link_state == IFLA_VF_LINK_STATE_AUTO)
4962 link_config |= PLINK_ENABLE;
4963
4964 if (link_state == IFLA_VF_LINK_STATE_AUTO)
4965 link_config |= PLINK_TRACK;
4966
4967 req->link_config = cpu_to_le32(link_config);
4968
4969 status = be_mcc_notify_wait(adapter);
4970 err:
4971 mutex_unlock(&adapter->mcc_lock);
4972 return status;
4973 }
4974
4975 int be_cmd_set_logical_link_config(struct be_adapter *adapter,
4976 int link_state, u8 domain)
4977 {
4978 int status;
4979
4980 if (BE2_chip(adapter))
4981 return -EOPNOTSUPP;
4982
4983 status = __be_cmd_set_logical_link_config(adapter, link_state,
4984 2, domain);
4985
4986 /* Version 2 of the command will not be recognized by older FW.
4987 * On such a failure issue version 1 of the command.
4988 */
4989 if (base_status(status) == MCC_STATUS_ILLEGAL_REQUEST)
4990 status = __be_cmd_set_logical_link_config(adapter, link_state,
4991 1, domain);
4992 return status;
4993 }
4994
4995 int be_cmd_set_features(struct be_adapter *adapter)
4996 {
4997 struct be_cmd_resp_set_features *resp;
4998 struct be_cmd_req_set_features *req;
4999 struct be_mcc_wrb *wrb;
5000 int status;
5001
5002 if (mutex_lock_interruptible(&adapter->mcc_lock))
5003 return -1;
5004
5005 wrb = wrb_from_mccq(adapter);
5006 if (!wrb) {
5007 status = -EBUSY;
5008 goto err;
5009 }
5010
5011 req = embedded_payload(wrb);
5012
5013 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
5014 OPCODE_COMMON_SET_FEATURES,
5015 sizeof(*req), wrb, NULL);
5016
5017 req->features = cpu_to_le32(BE_FEATURE_UE_RECOVERY);
5018 req->parameter_len = cpu_to_le32(sizeof(struct be_req_ue_recovery));
5019 req->parameter.req.uer = cpu_to_le32(BE_UE_RECOVERY_UER_MASK);
5020
5021 status = be_mcc_notify_wait(adapter);
5022 if (status)
5023 goto err;
5024
5025 resp = embedded_payload(wrb);
5026
5027 adapter->error_recovery.ue_to_poll_time =
5028 le16_to_cpu(resp->parameter.resp.ue2rp);
5029 adapter->error_recovery.ue_to_reset_time =
5030 le16_to_cpu(resp->parameter.resp.ue2sr);
5031 adapter->error_recovery.recovery_supported = true;
5032 err:
5033 /* Checking "MCC_STATUS_INVALID_LENGTH" for SKH as FW
5034 * returns this error in older firmware versions
5035 */
5036 if (base_status(status) == MCC_STATUS_ILLEGAL_REQUEST ||
5037 base_status(status) == MCC_STATUS_INVALID_LENGTH)
5038 dev_info(&adapter->pdev->dev,
5039 "Adapter does not support HW error recovery\n");
5040
5041 mutex_unlock(&adapter->mcc_lock);
5042 return status;
5043 }
5044
5045 int be_roce_mcc_cmd(void *netdev_handle, void *wrb_payload,
5046 int wrb_payload_size, u16 *cmd_status, u16 *ext_status)
5047 {
5048 struct be_adapter *adapter = netdev_priv(netdev_handle);
5049 struct be_mcc_wrb *wrb;
5050 struct be_cmd_req_hdr *hdr = (struct be_cmd_req_hdr *)wrb_payload;
5051 struct be_cmd_req_hdr *req;
5052 struct be_cmd_resp_hdr *resp;
5053 int status;
5054
5055 mutex_lock(&adapter->mcc_lock);
5056
5057 wrb = wrb_from_mccq(adapter);
5058 if (!wrb) {
5059 status = -EBUSY;
5060 goto err;
5061 }
5062 req = embedded_payload(wrb);
5063 resp = embedded_payload(wrb);
5064
5065 be_wrb_cmd_hdr_prepare(req, hdr->subsystem,
5066 hdr->opcode, wrb_payload_size, wrb, NULL);
5067 memcpy(req, wrb_payload, wrb_payload_size);
5068 be_dws_cpu_to_le(req, wrb_payload_size);
5069
5070 status = be_mcc_notify_wait(adapter);
5071 if (cmd_status)
5072 *cmd_status = (status & 0xffff);
5073 if (ext_status)
5074 *ext_status = 0;
5075 memcpy(wrb_payload, resp, sizeof(*resp) + resp->response_length);
5076 be_dws_le_to_cpu(wrb_payload, sizeof(*resp) + resp->response_length);
5077 err:
5078 mutex_unlock(&adapter->mcc_lock);
5079 return status;
5080 }
5081 EXPORT_SYMBOL(be_roce_mcc_cmd);