]>
Commit | Line | Data |
---|---|---|
c0c050c5 MC |
1 | /* Broadcom NetXtreme-C/E network driver. |
2 | * | |
3 | * Copyright (c) 2014-2015 Broadcom Corporation | |
4 | * | |
5 | * This program is free software; you can redistribute it and/or modify | |
6 | * it under the terms of the GNU General Public License as published by | |
7 | * the Free Software Foundation. | |
8 | */ | |
9 | ||
10 | #include <linux/module.h> | |
11 | #include <linux/pci.h> | |
12 | #include <linux/netdevice.h> | |
13 | #include <linux/if_vlan.h> | |
14 | #include <linux/interrupt.h> | |
15 | #include <linux/etherdevice.h> | |
16 | #include "bnxt_hsi.h" | |
17 | #include "bnxt.h" | |
18 | #include "bnxt_sriov.h" | |
19 | #include "bnxt_ethtool.h" | |
20 | ||
21 | #ifdef CONFIG_BNXT_SRIOV | |
22 | static int bnxt_vf_ndo_prep(struct bnxt *bp, int vf_id) | |
23 | { | |
caefe526 | 24 | if (!test_bit(BNXT_STATE_OPEN, &bp->state)) { |
c0c050c5 MC |
25 | netdev_err(bp->dev, "vf ndo called though PF is down\n"); |
26 | return -EINVAL; | |
27 | } | |
28 | if (!bp->pf.active_vfs) { | |
29 | netdev_err(bp->dev, "vf ndo called though sriov is disabled\n"); | |
30 | return -EINVAL; | |
31 | } | |
32 | if (vf_id >= bp->pf.max_vfs) { | |
33 | netdev_err(bp->dev, "Invalid VF id %d\n", vf_id); | |
34 | return -EINVAL; | |
35 | } | |
36 | return 0; | |
37 | } | |
38 | ||
39 | int bnxt_set_vf_spoofchk(struct net_device *dev, int vf_id, bool setting) | |
40 | { | |
41 | struct hwrm_func_cfg_input req = {0}; | |
42 | struct bnxt *bp = netdev_priv(dev); | |
43 | struct bnxt_vf_info *vf; | |
44 | bool old_setting = false; | |
45 | u32 func_flags; | |
46 | int rc; | |
47 | ||
48 | rc = bnxt_vf_ndo_prep(bp, vf_id); | |
49 | if (rc) | |
50 | return rc; | |
51 | ||
52 | vf = &bp->pf.vf[vf_id]; | |
53 | if (vf->flags & BNXT_VF_SPOOFCHK) | |
54 | old_setting = true; | |
55 | if (old_setting == setting) | |
56 | return 0; | |
57 | ||
58 | func_flags = vf->func_flags; | |
59 | if (setting) | |
60 | func_flags |= FUNC_CFG_REQ_FLAGS_SRC_MAC_ADDR_CHECK; | |
61 | else | |
62 | func_flags &= ~FUNC_CFG_REQ_FLAGS_SRC_MAC_ADDR_CHECK; | |
63 | /*TODO: if the driver supports VLAN filter on guest VLAN, | |
64 | * the spoof check should also include vlan anti-spoofing | |
65 | */ | |
66 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_CFG, -1, -1); | |
67 | req.vf_id = cpu_to_le16(vf->fw_fid); | |
68 | req.flags = cpu_to_le32(func_flags); | |
69 | rc = hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
70 | if (!rc) { | |
71 | vf->func_flags = func_flags; | |
72 | if (setting) | |
73 | vf->flags |= BNXT_VF_SPOOFCHK; | |
74 | else | |
75 | vf->flags &= ~BNXT_VF_SPOOFCHK; | |
76 | } | |
77 | return rc; | |
78 | } | |
79 | ||
80 | int bnxt_get_vf_config(struct net_device *dev, int vf_id, | |
81 | struct ifla_vf_info *ivi) | |
82 | { | |
83 | struct bnxt *bp = netdev_priv(dev); | |
84 | struct bnxt_vf_info *vf; | |
85 | int rc; | |
86 | ||
87 | rc = bnxt_vf_ndo_prep(bp, vf_id); | |
88 | if (rc) | |
89 | return rc; | |
90 | ||
91 | ivi->vf = vf_id; | |
92 | vf = &bp->pf.vf[vf_id]; | |
93 | ||
94 | memcpy(&ivi->mac, vf->mac_addr, ETH_ALEN); | |
95 | ivi->max_tx_rate = vf->max_tx_rate; | |
96 | ivi->min_tx_rate = vf->min_tx_rate; | |
97 | ivi->vlan = vf->vlan; | |
98 | ivi->qos = vf->flags & BNXT_VF_QOS; | |
99 | ivi->spoofchk = vf->flags & BNXT_VF_SPOOFCHK; | |
100 | if (!(vf->flags & BNXT_VF_LINK_FORCED)) | |
101 | ivi->linkstate = IFLA_VF_LINK_STATE_AUTO; | |
102 | else if (vf->flags & BNXT_VF_LINK_UP) | |
103 | ivi->linkstate = IFLA_VF_LINK_STATE_ENABLE; | |
104 | else | |
105 | ivi->linkstate = IFLA_VF_LINK_STATE_DISABLE; | |
106 | ||
107 | return 0; | |
108 | } | |
109 | ||
110 | int bnxt_set_vf_mac(struct net_device *dev, int vf_id, u8 *mac) | |
111 | { | |
112 | struct hwrm_func_cfg_input req = {0}; | |
113 | struct bnxt *bp = netdev_priv(dev); | |
114 | struct bnxt_vf_info *vf; | |
115 | int rc; | |
116 | ||
117 | rc = bnxt_vf_ndo_prep(bp, vf_id); | |
118 | if (rc) | |
119 | return rc; | |
120 | /* reject bc or mc mac addr, zero mac addr means allow | |
121 | * VF to use its own mac addr | |
122 | */ | |
123 | if (is_multicast_ether_addr(mac)) { | |
124 | netdev_err(dev, "Invalid VF ethernet address\n"); | |
125 | return -EINVAL; | |
126 | } | |
127 | vf = &bp->pf.vf[vf_id]; | |
128 | ||
129 | memcpy(vf->mac_addr, mac, ETH_ALEN); | |
130 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_CFG, -1, -1); | |
131 | req.vf_id = cpu_to_le16(vf->fw_fid); | |
132 | req.flags = cpu_to_le32(vf->func_flags); | |
133 | req.enables = cpu_to_le32(FUNC_CFG_REQ_ENABLES_DFLT_MAC_ADDR); | |
134 | memcpy(req.dflt_mac_addr, mac, ETH_ALEN); | |
135 | return hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
136 | } | |
137 | ||
138 | int bnxt_set_vf_vlan(struct net_device *dev, int vf_id, u16 vlan_id, u8 qos) | |
139 | { | |
140 | struct hwrm_func_cfg_input req = {0}; | |
141 | struct bnxt *bp = netdev_priv(dev); | |
142 | struct bnxt_vf_info *vf; | |
143 | u16 vlan_tag; | |
144 | int rc; | |
145 | ||
146 | rc = bnxt_vf_ndo_prep(bp, vf_id); | |
147 | if (rc) | |
148 | return rc; | |
149 | ||
150 | /* TODO: needed to implement proper handling of user priority, | |
151 | * currently fail the command if there is valid priority | |
152 | */ | |
153 | if (vlan_id > 4095 || qos) | |
154 | return -EINVAL; | |
155 | ||
156 | vf = &bp->pf.vf[vf_id]; | |
157 | vlan_tag = vlan_id; | |
158 | if (vlan_tag == vf->vlan) | |
159 | return 0; | |
160 | ||
161 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_CFG, -1, -1); | |
162 | req.vf_id = cpu_to_le16(vf->fw_fid); | |
163 | req.flags = cpu_to_le32(vf->func_flags); | |
164 | req.dflt_vlan = cpu_to_le16(vlan_tag); | |
165 | req.enables = cpu_to_le32(FUNC_CFG_REQ_ENABLES_DFLT_VLAN); | |
166 | rc = hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
167 | if (!rc) | |
168 | vf->vlan = vlan_tag; | |
169 | return rc; | |
170 | } | |
171 | ||
172 | int bnxt_set_vf_bw(struct net_device *dev, int vf_id, int min_tx_rate, | |
173 | int max_tx_rate) | |
174 | { | |
175 | struct hwrm_func_cfg_input req = {0}; | |
176 | struct bnxt *bp = netdev_priv(dev); | |
177 | struct bnxt_vf_info *vf; | |
178 | u32 pf_link_speed; | |
179 | int rc; | |
180 | ||
181 | rc = bnxt_vf_ndo_prep(bp, vf_id); | |
182 | if (rc) | |
183 | return rc; | |
184 | ||
185 | vf = &bp->pf.vf[vf_id]; | |
186 | pf_link_speed = bnxt_fw_to_ethtool_speed(bp->link_info.link_speed); | |
187 | if (max_tx_rate > pf_link_speed) { | |
188 | netdev_info(bp->dev, "max tx rate %d exceed PF link speed for VF %d\n", | |
189 | max_tx_rate, vf_id); | |
190 | return -EINVAL; | |
191 | } | |
192 | ||
193 | if (min_tx_rate > pf_link_speed || min_tx_rate > max_tx_rate) { | |
194 | netdev_info(bp->dev, "min tx rate %d is invalid for VF %d\n", | |
195 | min_tx_rate, vf_id); | |
196 | return -EINVAL; | |
197 | } | |
198 | if (min_tx_rate == vf->min_tx_rate && max_tx_rate == vf->max_tx_rate) | |
199 | return 0; | |
200 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_CFG, -1, -1); | |
201 | req.vf_id = cpu_to_le16(vf->fw_fid); | |
202 | req.flags = cpu_to_le32(vf->func_flags); | |
203 | req.enables = cpu_to_le32(FUNC_CFG_REQ_ENABLES_MAX_BW); | |
204 | req.max_bw = cpu_to_le32(max_tx_rate); | |
205 | req.enables |= cpu_to_le32(FUNC_CFG_REQ_ENABLES_MIN_BW); | |
206 | req.min_bw = cpu_to_le32(min_tx_rate); | |
207 | rc = hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
208 | if (!rc) { | |
209 | vf->min_tx_rate = min_tx_rate; | |
210 | vf->max_tx_rate = max_tx_rate; | |
211 | } | |
212 | return rc; | |
213 | } | |
214 | ||
215 | int bnxt_set_vf_link_state(struct net_device *dev, int vf_id, int link) | |
216 | { | |
217 | struct bnxt *bp = netdev_priv(dev); | |
218 | struct bnxt_vf_info *vf; | |
219 | int rc; | |
220 | ||
221 | rc = bnxt_vf_ndo_prep(bp, vf_id); | |
222 | if (rc) | |
223 | return rc; | |
224 | ||
225 | vf = &bp->pf.vf[vf_id]; | |
226 | ||
227 | vf->flags &= ~(BNXT_VF_LINK_UP | BNXT_VF_LINK_FORCED); | |
228 | switch (link) { | |
229 | case IFLA_VF_LINK_STATE_AUTO: | |
230 | vf->flags |= BNXT_VF_LINK_UP; | |
231 | break; | |
232 | case IFLA_VF_LINK_STATE_DISABLE: | |
233 | vf->flags |= BNXT_VF_LINK_FORCED; | |
234 | break; | |
235 | case IFLA_VF_LINK_STATE_ENABLE: | |
236 | vf->flags |= BNXT_VF_LINK_UP | BNXT_VF_LINK_FORCED; | |
237 | break; | |
238 | default: | |
239 | netdev_err(bp->dev, "Invalid link option\n"); | |
240 | rc = -EINVAL; | |
241 | break; | |
242 | } | |
243 | /* CHIMP TODO: send msg to VF to update new link state */ | |
244 | ||
245 | return rc; | |
246 | } | |
247 | ||
248 | static int bnxt_set_vf_attr(struct bnxt *bp, int num_vfs) | |
249 | { | |
250 | int i; | |
251 | struct bnxt_vf_info *vf; | |
252 | ||
253 | for (i = 0; i < num_vfs; i++) { | |
254 | vf = &bp->pf.vf[i]; | |
255 | memset(vf, 0, sizeof(*vf)); | |
256 | vf->flags = BNXT_VF_QOS | BNXT_VF_LINK_UP; | |
257 | } | |
258 | return 0; | |
259 | } | |
260 | ||
4bb6cdce | 261 | static int bnxt_hwrm_func_vf_resource_free(struct bnxt *bp, int num_vfs) |
c0c050c5 MC |
262 | { |
263 | int i, rc = 0; | |
264 | struct bnxt_pf_info *pf = &bp->pf; | |
265 | struct hwrm_func_vf_resc_free_input req = {0}; | |
266 | ||
267 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_VF_RESC_FREE, -1, -1); | |
268 | ||
269 | mutex_lock(&bp->hwrm_cmd_lock); | |
4bb6cdce | 270 | for (i = pf->first_vf_id; i < pf->first_vf_id + num_vfs; i++) { |
c0c050c5 MC |
271 | req.vf_id = cpu_to_le16(i); |
272 | rc = _hwrm_send_message(bp, &req, sizeof(req), | |
273 | HWRM_CMD_TIMEOUT); | |
274 | if (rc) | |
275 | break; | |
276 | } | |
277 | mutex_unlock(&bp->hwrm_cmd_lock); | |
278 | return rc; | |
279 | } | |
280 | ||
281 | static void bnxt_free_vf_resources(struct bnxt *bp) | |
282 | { | |
283 | struct pci_dev *pdev = bp->pdev; | |
284 | int i; | |
285 | ||
286 | kfree(bp->pf.vf_event_bmap); | |
287 | bp->pf.vf_event_bmap = NULL; | |
288 | ||
289 | for (i = 0; i < 4; i++) { | |
290 | if (bp->pf.hwrm_cmd_req_addr[i]) { | |
291 | dma_free_coherent(&pdev->dev, BNXT_PAGE_SIZE, | |
292 | bp->pf.hwrm_cmd_req_addr[i], | |
293 | bp->pf.hwrm_cmd_req_dma_addr[i]); | |
294 | bp->pf.hwrm_cmd_req_addr[i] = NULL; | |
295 | } | |
296 | } | |
297 | ||
298 | kfree(bp->pf.vf); | |
299 | bp->pf.vf = NULL; | |
300 | } | |
301 | ||
302 | static int bnxt_alloc_vf_resources(struct bnxt *bp, int num_vfs) | |
303 | { | |
304 | struct pci_dev *pdev = bp->pdev; | |
305 | u32 nr_pages, size, i, j, k = 0; | |
306 | ||
307 | bp->pf.vf = kcalloc(num_vfs, sizeof(struct bnxt_vf_info), GFP_KERNEL); | |
308 | if (!bp->pf.vf) | |
309 | return -ENOMEM; | |
310 | ||
311 | bnxt_set_vf_attr(bp, num_vfs); | |
312 | ||
313 | size = num_vfs * BNXT_HWRM_REQ_MAX_SIZE; | |
314 | nr_pages = size / BNXT_PAGE_SIZE; | |
315 | if (size & (BNXT_PAGE_SIZE - 1)) | |
316 | nr_pages++; | |
317 | ||
318 | for (i = 0; i < nr_pages; i++) { | |
319 | bp->pf.hwrm_cmd_req_addr[i] = | |
320 | dma_alloc_coherent(&pdev->dev, BNXT_PAGE_SIZE, | |
321 | &bp->pf.hwrm_cmd_req_dma_addr[i], | |
322 | GFP_KERNEL); | |
323 | ||
324 | if (!bp->pf.hwrm_cmd_req_addr[i]) | |
325 | return -ENOMEM; | |
326 | ||
327 | for (j = 0; j < BNXT_HWRM_REQS_PER_PAGE && k < num_vfs; j++) { | |
328 | struct bnxt_vf_info *vf = &bp->pf.vf[k]; | |
329 | ||
330 | vf->hwrm_cmd_req_addr = bp->pf.hwrm_cmd_req_addr[i] + | |
331 | j * BNXT_HWRM_REQ_MAX_SIZE; | |
332 | vf->hwrm_cmd_req_dma_addr = | |
333 | bp->pf.hwrm_cmd_req_dma_addr[i] + j * | |
334 | BNXT_HWRM_REQ_MAX_SIZE; | |
335 | k++; | |
336 | } | |
337 | } | |
338 | ||
339 | /* Max 128 VF's */ | |
340 | bp->pf.vf_event_bmap = kzalloc(16, GFP_KERNEL); | |
341 | if (!bp->pf.vf_event_bmap) | |
342 | return -ENOMEM; | |
343 | ||
344 | bp->pf.hwrm_cmd_req_pages = nr_pages; | |
345 | return 0; | |
346 | } | |
347 | ||
348 | static int bnxt_hwrm_func_buf_rgtr(struct bnxt *bp) | |
349 | { | |
350 | struct hwrm_func_buf_rgtr_input req = {0}; | |
351 | ||
352 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_BUF_RGTR, -1, -1); | |
353 | ||
354 | req.req_buf_num_pages = cpu_to_le16(bp->pf.hwrm_cmd_req_pages); | |
355 | req.req_buf_page_size = cpu_to_le16(BNXT_PAGE_SHIFT); | |
356 | req.req_buf_len = cpu_to_le16(BNXT_HWRM_REQ_MAX_SIZE); | |
357 | req.req_buf_page_addr0 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[0]); | |
358 | req.req_buf_page_addr1 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[1]); | |
359 | req.req_buf_page_addr2 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[2]); | |
360 | req.req_buf_page_addr3 = cpu_to_le64(bp->pf.hwrm_cmd_req_dma_addr[3]); | |
361 | ||
362 | return hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
363 | } | |
364 | ||
365 | /* only call by PF to reserve resources for VF */ | |
92268c32 | 366 | static int bnxt_hwrm_func_cfg(struct bnxt *bp, int num_vfs) |
c0c050c5 MC |
367 | { |
368 | u32 rc = 0, mtu, i; | |
369 | u16 vf_tx_rings, vf_rx_rings, vf_cp_rings, vf_stat_ctx, vf_vnics; | |
370 | struct hwrm_func_cfg_input req = {0}; | |
371 | struct bnxt_pf_info *pf = &bp->pf; | |
372 | ||
373 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_CFG, -1, -1); | |
374 | ||
375 | /* Remaining rings are distributed equally amongs VF's for now */ | |
376 | /* TODO: the following workaroud is needed to restrict total number | |
377 | * of vf_cp_rings not exceed number of HW ring groups. This WA should | |
378 | * be removed once new HWRM provides HW ring groups capability in | |
379 | * hwrm_func_qcap. | |
380 | */ | |
92268c32 MC |
381 | vf_cp_rings = min_t(u16, pf->max_cp_rings, pf->max_stat_ctxs); |
382 | vf_cp_rings = (vf_cp_rings - bp->cp_nr_rings) / num_vfs; | |
c0c050c5 | 383 | /* TODO: restore this logic below once the WA above is removed */ |
92268c32 MC |
384 | /* vf_cp_rings = (pf->max_cp_rings - bp->cp_nr_rings) / num_vfs; */ |
385 | vf_stat_ctx = (pf->max_stat_ctxs - bp->num_stat_ctxs) / num_vfs; | |
c0c050c5 | 386 | if (bp->flags & BNXT_FLAG_AGG_RINGS) |
92268c32 MC |
387 | vf_rx_rings = (pf->max_rx_rings - bp->rx_nr_rings * 2) / |
388 | num_vfs; | |
c0c050c5 | 389 | else |
92268c32 MC |
390 | vf_rx_rings = (pf->max_rx_rings - bp->rx_nr_rings) / num_vfs; |
391 | vf_tx_rings = (pf->max_tx_rings - bp->tx_nr_rings) / num_vfs; | |
c0c050c5 MC |
392 | |
393 | req.enables = cpu_to_le32(FUNC_CFG_REQ_ENABLES_MTU | | |
394 | FUNC_CFG_REQ_ENABLES_MRU | | |
395 | FUNC_CFG_REQ_ENABLES_NUM_RSSCOS_CTXS | | |
396 | FUNC_CFG_REQ_ENABLES_NUM_STAT_CTXS | | |
397 | FUNC_CFG_REQ_ENABLES_NUM_CMPL_RINGS | | |
398 | FUNC_CFG_REQ_ENABLES_NUM_TX_RINGS | | |
399 | FUNC_CFG_REQ_ENABLES_NUM_RX_RINGS | | |
400 | FUNC_CFG_REQ_ENABLES_NUM_L2_CTXS | | |
401 | FUNC_CFG_REQ_ENABLES_NUM_VNICS); | |
402 | ||
403 | mtu = bp->dev->mtu + ETH_HLEN + ETH_FCS_LEN + VLAN_HLEN; | |
404 | req.mru = cpu_to_le16(mtu); | |
405 | req.mtu = cpu_to_le16(mtu); | |
406 | ||
407 | req.num_rsscos_ctxs = cpu_to_le16(1); | |
408 | req.num_cmpl_rings = cpu_to_le16(vf_cp_rings); | |
409 | req.num_tx_rings = cpu_to_le16(vf_tx_rings); | |
410 | req.num_rx_rings = cpu_to_le16(vf_rx_rings); | |
411 | req.num_l2_ctxs = cpu_to_le16(4); | |
412 | vf_vnics = 1; | |
413 | ||
414 | req.num_vnics = cpu_to_le16(vf_vnics); | |
415 | /* FIXME spec currently uses 1 bit for stats ctx */ | |
416 | req.num_stat_ctxs = cpu_to_le16(vf_stat_ctx); | |
417 | ||
418 | mutex_lock(&bp->hwrm_cmd_lock); | |
92268c32 | 419 | for (i = 0; i < num_vfs; i++) { |
c0c050c5 MC |
420 | req.vf_id = cpu_to_le16(pf->first_vf_id + i); |
421 | rc = _hwrm_send_message(bp, &req, sizeof(req), | |
422 | HWRM_CMD_TIMEOUT); | |
423 | if (rc) | |
424 | break; | |
92268c32 MC |
425 | pf->active_vfs = i + 1; |
426 | pf->vf[i].fw_fid = le16_to_cpu(req.vf_id); | |
c0c050c5 MC |
427 | } |
428 | mutex_unlock(&bp->hwrm_cmd_lock); | |
429 | if (!rc) { | |
92268c32 | 430 | pf->max_pf_tx_rings = bp->tx_nr_rings; |
c0c050c5 | 431 | if (bp->flags & BNXT_FLAG_AGG_RINGS) |
92268c32 | 432 | pf->max_pf_rx_rings = bp->rx_nr_rings * 2; |
c0c050c5 | 433 | else |
92268c32 | 434 | pf->max_pf_rx_rings = bp->rx_nr_rings; |
c0c050c5 MC |
435 | } |
436 | return rc; | |
437 | } | |
438 | ||
439 | static int bnxt_sriov_enable(struct bnxt *bp, int *num_vfs) | |
440 | { | |
441 | int rc = 0, vfs_supported; | |
442 | int min_rx_rings, min_tx_rings, min_rss_ctxs; | |
443 | int tx_ok = 0, rx_ok = 0, rss_ok = 0; | |
444 | ||
445 | /* Check if we can enable requested num of vf's. At a mininum | |
446 | * we require 1 RX 1 TX rings for each VF. In this minimum conf | |
447 | * features like TPA will not be available. | |
448 | */ | |
449 | vfs_supported = *num_vfs; | |
450 | ||
451 | while (vfs_supported) { | |
452 | min_rx_rings = vfs_supported; | |
453 | min_tx_rings = vfs_supported; | |
454 | min_rss_ctxs = vfs_supported; | |
455 | ||
456 | if (bp->flags & BNXT_FLAG_AGG_RINGS) { | |
457 | if (bp->pf.max_rx_rings - bp->rx_nr_rings * 2 >= | |
458 | min_rx_rings) | |
459 | rx_ok = 1; | |
460 | } else { | |
461 | if (bp->pf.max_rx_rings - bp->rx_nr_rings >= | |
462 | min_rx_rings) | |
463 | rx_ok = 1; | |
464 | } | |
465 | ||
466 | if (bp->pf.max_tx_rings - bp->tx_nr_rings >= min_tx_rings) | |
467 | tx_ok = 1; | |
468 | ||
469 | if (bp->pf.max_rsscos_ctxs - bp->rsscos_nr_ctxs >= min_rss_ctxs) | |
470 | rss_ok = 1; | |
471 | ||
472 | if (tx_ok && rx_ok && rss_ok) | |
473 | break; | |
474 | ||
475 | vfs_supported--; | |
476 | } | |
477 | ||
478 | if (!vfs_supported) { | |
479 | netdev_err(bp->dev, "Cannot enable VF's as all resources are used by PF\n"); | |
480 | return -EINVAL; | |
481 | } | |
482 | ||
483 | if (vfs_supported != *num_vfs) { | |
484 | netdev_info(bp->dev, "Requested VFs %d, can enable %d\n", | |
485 | *num_vfs, vfs_supported); | |
486 | *num_vfs = vfs_supported; | |
487 | } | |
488 | ||
489 | rc = bnxt_alloc_vf_resources(bp, *num_vfs); | |
490 | if (rc) | |
491 | goto err_out1; | |
492 | ||
493 | /* Reserve resources for VFs */ | |
92268c32 | 494 | rc = bnxt_hwrm_func_cfg(bp, *num_vfs); |
c0c050c5 MC |
495 | if (rc) |
496 | goto err_out2; | |
497 | ||
498 | /* Register buffers for VFs */ | |
499 | rc = bnxt_hwrm_func_buf_rgtr(bp); | |
500 | if (rc) | |
501 | goto err_out2; | |
502 | ||
503 | rc = pci_enable_sriov(bp->pdev, *num_vfs); | |
504 | if (rc) | |
505 | goto err_out2; | |
506 | ||
507 | return 0; | |
508 | ||
509 | err_out2: | |
510 | /* Free the resources reserved for various VF's */ | |
4bb6cdce | 511 | bnxt_hwrm_func_vf_resource_free(bp, *num_vfs); |
c0c050c5 MC |
512 | |
513 | err_out1: | |
514 | bnxt_free_vf_resources(bp); | |
515 | ||
516 | return rc; | |
517 | } | |
518 | ||
519 | void bnxt_sriov_disable(struct bnxt *bp) | |
520 | { | |
4bb6cdce | 521 | u16 num_vfs = pci_num_vf(bp->pdev); |
c0c050c5 | 522 | |
4bb6cdce JH |
523 | if (!num_vfs) |
524 | return; | |
c0c050c5 | 525 | |
4bb6cdce JH |
526 | if (pci_vfs_assigned(bp->pdev)) { |
527 | netdev_warn(bp->dev, "Unable to free %d VFs because some are assigned to VMs.\n", | |
528 | num_vfs); | |
529 | } else { | |
530 | pci_disable_sriov(bp->pdev); | |
531 | /* Free the HW resources reserved for various VF's */ | |
532 | bnxt_hwrm_func_vf_resource_free(bp, num_vfs); | |
533 | } | |
c0c050c5 MC |
534 | |
535 | bnxt_free_vf_resources(bp); | |
536 | ||
537 | bp->pf.active_vfs = 0; | |
538 | bp->pf.max_pf_rx_rings = bp->pf.max_rx_rings; | |
539 | bp->pf.max_pf_tx_rings = bp->pf.max_tx_rings; | |
540 | } | |
541 | ||
542 | int bnxt_sriov_configure(struct pci_dev *pdev, int num_vfs) | |
543 | { | |
544 | struct net_device *dev = pci_get_drvdata(pdev); | |
545 | struct bnxt *bp = netdev_priv(dev); | |
546 | ||
547 | if (!(bp->flags & BNXT_FLAG_USING_MSIX)) { | |
548 | netdev_warn(dev, "Not allow SRIOV if the irq mode is not MSIX\n"); | |
549 | return 0; | |
550 | } | |
551 | ||
552 | rtnl_lock(); | |
553 | if (!netif_running(dev)) { | |
554 | netdev_warn(dev, "Reject SRIOV config request since if is down!\n"); | |
555 | rtnl_unlock(); | |
556 | return 0; | |
557 | } | |
558 | bp->sriov_cfg = true; | |
559 | rtnl_unlock(); | |
4bb6cdce JH |
560 | |
561 | if (pci_vfs_assigned(bp->pdev)) { | |
562 | netdev_warn(dev, "Unable to configure SRIOV since some VFs are assigned to VMs.\n"); | |
563 | num_vfs = 0; | |
564 | goto sriov_cfg_exit; | |
c0c050c5 MC |
565 | } |
566 | ||
567 | /* Check if enabled VFs is same as requested */ | |
4bb6cdce JH |
568 | if (num_vfs && num_vfs == bp->pf.active_vfs) |
569 | goto sriov_cfg_exit; | |
570 | ||
571 | /* if there are previous existing VFs, clean them up */ | |
572 | bnxt_sriov_disable(bp); | |
573 | if (!num_vfs) | |
574 | goto sriov_cfg_exit; | |
c0c050c5 MC |
575 | |
576 | bnxt_sriov_enable(bp, &num_vfs); | |
577 | ||
4bb6cdce | 578 | sriov_cfg_exit: |
c0c050c5 MC |
579 | bp->sriov_cfg = false; |
580 | wake_up(&bp->sriov_cfg_wait); | |
581 | ||
582 | return num_vfs; | |
583 | } | |
584 | ||
585 | static int bnxt_hwrm_fwd_resp(struct bnxt *bp, struct bnxt_vf_info *vf, | |
586 | void *encap_resp, __le64 encap_resp_addr, | |
587 | __le16 encap_resp_cpr, u32 msg_size) | |
588 | { | |
589 | int rc = 0; | |
590 | struct hwrm_fwd_resp_input req = {0}; | |
591 | struct hwrm_fwd_resp_output *resp = bp->hwrm_cmd_resp_addr; | |
592 | ||
593 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FWD_RESP, -1, -1); | |
594 | ||
595 | /* Set the new target id */ | |
596 | req.target_id = cpu_to_le16(vf->fw_fid); | |
597 | req.encap_resp_len = cpu_to_le16(msg_size); | |
598 | req.encap_resp_addr = encap_resp_addr; | |
599 | req.encap_resp_cmpl_ring = encap_resp_cpr; | |
600 | memcpy(req.encap_resp, encap_resp, msg_size); | |
601 | ||
602 | mutex_lock(&bp->hwrm_cmd_lock); | |
603 | rc = _hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
604 | ||
605 | if (rc) { | |
606 | netdev_err(bp->dev, "hwrm_fwd_resp failed. rc:%d\n", rc); | |
607 | goto fwd_resp_exit; | |
608 | } | |
609 | ||
610 | if (resp->error_code) { | |
611 | netdev_err(bp->dev, "hwrm_fwd_resp error %d\n", | |
612 | resp->error_code); | |
613 | rc = -1; | |
614 | } | |
615 | ||
616 | fwd_resp_exit: | |
617 | mutex_unlock(&bp->hwrm_cmd_lock); | |
618 | return rc; | |
619 | } | |
620 | ||
621 | static int bnxt_hwrm_fwd_err_resp(struct bnxt *bp, struct bnxt_vf_info *vf, | |
622 | u32 msg_size) | |
623 | { | |
624 | int rc = 0; | |
625 | struct hwrm_reject_fwd_resp_input req = {0}; | |
626 | struct hwrm_reject_fwd_resp_output *resp = bp->hwrm_cmd_resp_addr; | |
627 | ||
628 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_REJECT_FWD_RESP, -1, -1); | |
629 | /* Set the new target id */ | |
630 | req.target_id = cpu_to_le16(vf->fw_fid); | |
631 | memcpy(req.encap_request, vf->hwrm_cmd_req_addr, msg_size); | |
632 | ||
633 | mutex_lock(&bp->hwrm_cmd_lock); | |
634 | rc = _hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
635 | ||
636 | if (rc) { | |
637 | netdev_err(bp->dev, "hwrm_fwd_err_resp failed. rc:%d\n", rc); | |
638 | goto fwd_err_resp_exit; | |
639 | } | |
640 | ||
641 | if (resp->error_code) { | |
642 | netdev_err(bp->dev, "hwrm_fwd_err_resp error %d\n", | |
643 | resp->error_code); | |
644 | rc = -1; | |
645 | } | |
646 | ||
647 | fwd_err_resp_exit: | |
648 | mutex_unlock(&bp->hwrm_cmd_lock); | |
649 | return rc; | |
650 | } | |
651 | ||
652 | static int bnxt_hwrm_exec_fwd_resp(struct bnxt *bp, struct bnxt_vf_info *vf, | |
653 | u32 msg_size) | |
654 | { | |
655 | int rc = 0; | |
656 | struct hwrm_exec_fwd_resp_input req = {0}; | |
657 | struct hwrm_exec_fwd_resp_output *resp = bp->hwrm_cmd_resp_addr; | |
658 | ||
659 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_EXEC_FWD_RESP, -1, -1); | |
660 | /* Set the new target id */ | |
661 | req.target_id = cpu_to_le16(vf->fw_fid); | |
662 | memcpy(req.encap_request, vf->hwrm_cmd_req_addr, msg_size); | |
663 | ||
664 | mutex_lock(&bp->hwrm_cmd_lock); | |
665 | rc = _hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT); | |
666 | ||
667 | if (rc) { | |
668 | netdev_err(bp->dev, "hwrm_exec_fw_resp failed. rc:%d\n", rc); | |
669 | goto exec_fwd_resp_exit; | |
670 | } | |
671 | ||
672 | if (resp->error_code) { | |
673 | netdev_err(bp->dev, "hwrm_exec_fw_resp error %d\n", | |
674 | resp->error_code); | |
675 | rc = -1; | |
676 | } | |
677 | ||
678 | exec_fwd_resp_exit: | |
679 | mutex_unlock(&bp->hwrm_cmd_lock); | |
680 | return rc; | |
681 | } | |
682 | ||
683 | static int bnxt_vf_validate_set_mac(struct bnxt *bp, struct bnxt_vf_info *vf) | |
684 | { | |
685 | u32 msg_size = sizeof(struct hwrm_cfa_l2_filter_alloc_input); | |
686 | struct hwrm_cfa_l2_filter_alloc_input *req = | |
687 | (struct hwrm_cfa_l2_filter_alloc_input *)vf->hwrm_cmd_req_addr; | |
688 | ||
689 | if (!is_valid_ether_addr(vf->mac_addr) || | |
690 | ether_addr_equal((const u8 *)req->l2_addr, vf->mac_addr)) | |
691 | return bnxt_hwrm_exec_fwd_resp(bp, vf, msg_size); | |
692 | else | |
693 | return bnxt_hwrm_fwd_err_resp(bp, vf, msg_size); | |
694 | } | |
695 | ||
696 | static int bnxt_vf_set_link(struct bnxt *bp, struct bnxt_vf_info *vf) | |
697 | { | |
698 | int rc = 0; | |
699 | ||
700 | if (!(vf->flags & BNXT_VF_LINK_FORCED)) { | |
701 | /* real link */ | |
702 | rc = bnxt_hwrm_exec_fwd_resp( | |
703 | bp, vf, sizeof(struct hwrm_port_phy_qcfg_input)); | |
704 | } else { | |
705 | struct hwrm_port_phy_qcfg_output phy_qcfg_resp; | |
706 | struct hwrm_port_phy_qcfg_input *phy_qcfg_req; | |
707 | ||
708 | phy_qcfg_req = | |
709 | (struct hwrm_port_phy_qcfg_input *)vf->hwrm_cmd_req_addr; | |
710 | mutex_lock(&bp->hwrm_cmd_lock); | |
711 | memcpy(&phy_qcfg_resp, &bp->link_info.phy_qcfg_resp, | |
712 | sizeof(phy_qcfg_resp)); | |
713 | mutex_unlock(&bp->hwrm_cmd_lock); | |
714 | phy_qcfg_resp.seq_id = phy_qcfg_req->seq_id; | |
715 | ||
716 | if (vf->flags & BNXT_VF_LINK_UP) { | |
717 | /* if physical link is down, force link up on VF */ | |
718 | if (phy_qcfg_resp.link == | |
719 | PORT_PHY_QCFG_RESP_LINK_NO_LINK) { | |
720 | phy_qcfg_resp.link = | |
721 | PORT_PHY_QCFG_RESP_LINK_LINK; | |
722 | if (phy_qcfg_resp.auto_link_speed) | |
723 | phy_qcfg_resp.link_speed = | |
724 | phy_qcfg_resp.auto_link_speed; | |
725 | else | |
726 | phy_qcfg_resp.link_speed = | |
727 | phy_qcfg_resp.force_link_speed; | |
728 | phy_qcfg_resp.duplex = | |
729 | PORT_PHY_QCFG_RESP_DUPLEX_FULL; | |
730 | phy_qcfg_resp.pause = | |
731 | (PORT_PHY_QCFG_RESP_PAUSE_TX | | |
732 | PORT_PHY_QCFG_RESP_PAUSE_RX); | |
733 | } | |
734 | } else { | |
735 | /* force link down */ | |
736 | phy_qcfg_resp.link = PORT_PHY_QCFG_RESP_LINK_NO_LINK; | |
737 | phy_qcfg_resp.link_speed = 0; | |
738 | phy_qcfg_resp.duplex = PORT_PHY_QCFG_RESP_DUPLEX_HALF; | |
739 | phy_qcfg_resp.pause = 0; | |
740 | } | |
741 | rc = bnxt_hwrm_fwd_resp(bp, vf, &phy_qcfg_resp, | |
742 | phy_qcfg_req->resp_addr, | |
743 | phy_qcfg_req->cmpl_ring, | |
744 | sizeof(phy_qcfg_resp)); | |
745 | } | |
746 | return rc; | |
747 | } | |
748 | ||
749 | static int bnxt_vf_req_validate_snd(struct bnxt *bp, struct bnxt_vf_info *vf) | |
750 | { | |
751 | int rc = 0; | |
752 | struct hwrm_cmd_req_hdr *encap_req = vf->hwrm_cmd_req_addr; | |
753 | u32 req_type = le32_to_cpu(encap_req->cmpl_ring_req_type) & 0xffff; | |
754 | ||
755 | switch (req_type) { | |
756 | case HWRM_CFA_L2_FILTER_ALLOC: | |
757 | rc = bnxt_vf_validate_set_mac(bp, vf); | |
758 | break; | |
759 | case HWRM_FUNC_CFG: | |
760 | /* TODO Validate if VF is allowed to change mac address, | |
761 | * mtu, num of rings etc | |
762 | */ | |
763 | rc = bnxt_hwrm_exec_fwd_resp( | |
764 | bp, vf, sizeof(struct hwrm_func_cfg_input)); | |
765 | break; | |
766 | case HWRM_PORT_PHY_QCFG: | |
767 | rc = bnxt_vf_set_link(bp, vf); | |
768 | break; | |
769 | default: | |
770 | break; | |
771 | } | |
772 | return rc; | |
773 | } | |
774 | ||
775 | void bnxt_hwrm_exec_fwd_req(struct bnxt *bp) | |
776 | { | |
777 | u32 i = 0, active_vfs = bp->pf.active_vfs, vf_id; | |
778 | ||
779 | /* Scan through VF's and process commands */ | |
780 | while (1) { | |
781 | vf_id = find_next_bit(bp->pf.vf_event_bmap, active_vfs, i); | |
782 | if (vf_id >= active_vfs) | |
783 | break; | |
784 | ||
785 | clear_bit(vf_id, bp->pf.vf_event_bmap); | |
786 | bnxt_vf_req_validate_snd(bp, &bp->pf.vf[vf_id]); | |
787 | i = vf_id + 1; | |
788 | } | |
789 | } | |
379a80a1 MC |
790 | |
791 | void bnxt_update_vf_mac(struct bnxt *bp) | |
792 | { | |
793 | struct hwrm_func_qcaps_input req = {0}; | |
794 | struct hwrm_func_qcaps_output *resp = bp->hwrm_cmd_resp_addr; | |
795 | ||
796 | bnxt_hwrm_cmd_hdr_init(bp, &req, HWRM_FUNC_QCAPS, -1, -1); | |
797 | req.fid = cpu_to_le16(0xffff); | |
798 | ||
799 | mutex_lock(&bp->hwrm_cmd_lock); | |
800 | if (_hwrm_send_message(bp, &req, sizeof(req), HWRM_CMD_TIMEOUT)) | |
801 | goto update_vf_mac_exit; | |
802 | ||
803 | if (!is_valid_ether_addr(resp->perm_mac_address)) | |
804 | goto update_vf_mac_exit; | |
805 | ||
bdd4347b JH |
806 | if (!ether_addr_equal(resp->perm_mac_address, bp->vf.mac_addr)) |
807 | memcpy(bp->vf.mac_addr, resp->perm_mac_address, ETH_ALEN); | |
808 | /* overwrite netdev dev_adr with admin VF MAC */ | |
379a80a1 MC |
809 | memcpy(bp->dev->dev_addr, bp->vf.mac_addr, ETH_ALEN); |
810 | update_vf_mac_exit: | |
811 | mutex_unlock(&bp->hwrm_cmd_lock); | |
812 | } | |
813 | ||
c0c050c5 MC |
814 | #else |
815 | ||
816 | void bnxt_sriov_disable(struct bnxt *bp) | |
817 | { | |
818 | } | |
819 | ||
820 | void bnxt_hwrm_exec_fwd_req(struct bnxt *bp) | |
821 | { | |
379a80a1 MC |
822 | netdev_err(bp->dev, "Invalid VF message received when SRIOV is not enable\n"); |
823 | } | |
824 | ||
825 | void bnxt_update_vf_mac(struct bnxt *bp) | |
826 | { | |
c0c050c5 MC |
827 | } |
828 | #endif |