]>
Commit | Line | Data |
---|---|---|
1 | /* | |
2 | * QEMU e1000 emulation | |
3 | * | |
4 | * Software developer's manual: | |
5 | * http://download.intel.com/design/network/manuals/8254x_GBe_SDM.pdf | |
6 | * | |
7 | * Nir Peleg, Tutis Systems Ltd. for Qumranet Inc. | |
8 | * Copyright (c) 2008 Qumranet | |
9 | * Based on work done by: | |
10 | * Copyright (c) 2007 Dan Aloni | |
11 | * Copyright (c) 2004 Antony T Curtis | |
12 | * | |
13 | * This library is free software; you can redistribute it and/or | |
14 | * modify it under the terms of the GNU Lesser General Public | |
15 | * License as published by the Free Software Foundation; either | |
16 | * version 2 of the License, or (at your option) any later version. | |
17 | * | |
18 | * This library is distributed in the hope that it will be useful, | |
19 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
20 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |
21 | * Lesser General Public License for more details. | |
22 | * | |
23 | * You should have received a copy of the GNU Lesser General Public | |
24 | * License along with this library; if not, see <http://www.gnu.org/licenses/>. | |
25 | */ | |
26 | ||
27 | ||
28 | #include "qemu/osdep.h" | |
29 | #include "hw/pci/pci.h" | |
30 | #include "hw/qdev-properties.h" | |
31 | #include "migration/vmstate.h" | |
32 | #include "net/net.h" | |
33 | #include "net/checksum.h" | |
34 | #include "sysemu/sysemu.h" | |
35 | #include "sysemu/dma.h" | |
36 | #include "qemu/iov.h" | |
37 | #include "qemu/module.h" | |
38 | #include "qemu/range.h" | |
39 | ||
40 | #include "e1000x_common.h" | |
41 | #include "trace.h" | |
42 | ||
43 | static const uint8_t bcast[] = {0xff, 0xff, 0xff, 0xff, 0xff, 0xff}; | |
44 | ||
45 | /* #define E1000_DEBUG */ | |
46 | ||
47 | #ifdef E1000_DEBUG | |
48 | enum { | |
49 | DEBUG_GENERAL, DEBUG_IO, DEBUG_MMIO, DEBUG_INTERRUPT, | |
50 | DEBUG_RX, DEBUG_TX, DEBUG_MDIC, DEBUG_EEPROM, | |
51 | DEBUG_UNKNOWN, DEBUG_TXSUM, DEBUG_TXERR, DEBUG_RXERR, | |
52 | DEBUG_RXFILTER, DEBUG_PHY, DEBUG_NOTYET, | |
53 | }; | |
54 | #define DBGBIT(x) (1<<DEBUG_##x) | |
55 | static int debugflags = DBGBIT(TXERR) | DBGBIT(GENERAL); | |
56 | ||
57 | #define DBGOUT(what, fmt, ...) do { \ | |
58 | if (debugflags & DBGBIT(what)) \ | |
59 | fprintf(stderr, "e1000: " fmt, ## __VA_ARGS__); \ | |
60 | } while (0) | |
61 | #else | |
62 | #define DBGOUT(what, fmt, ...) do {} while (0) | |
63 | #endif | |
64 | ||
65 | #define IOPORT_SIZE 0x40 | |
66 | #define PNPMMIO_SIZE 0x20000 | |
67 | #define MIN_BUF_SIZE 60 /* Min. octets in an ethernet frame sans FCS */ | |
68 | ||
69 | #define MAXIMUM_ETHERNET_HDR_LEN (14+4) | |
70 | ||
71 | /* | |
72 | * HW models: | |
73 | * E1000_DEV_ID_82540EM works with Windows, Linux, and OS X <= 10.8 | |
74 | * E1000_DEV_ID_82544GC_COPPER appears to work; not well tested | |
75 | * E1000_DEV_ID_82545EM_COPPER works with Linux and OS X >= 10.6 | |
76 | * Others never tested | |
77 | */ | |
78 | ||
79 | typedef struct E1000State_st { | |
80 | /*< private >*/ | |
81 | PCIDevice parent_obj; | |
82 | /*< public >*/ | |
83 | ||
84 | NICState *nic; | |
85 | NICConf conf; | |
86 | MemoryRegion mmio; | |
87 | MemoryRegion io; | |
88 | ||
89 | uint32_t mac_reg[0x8000]; | |
90 | uint16_t phy_reg[0x20]; | |
91 | uint16_t eeprom_data[64]; | |
92 | ||
93 | uint32_t rxbuf_size; | |
94 | uint32_t rxbuf_min_shift; | |
95 | struct e1000_tx { | |
96 | unsigned char header[256]; | |
97 | unsigned char vlan_header[4]; | |
98 | /* Fields vlan and data must not be reordered or separated. */ | |
99 | unsigned char vlan[4]; | |
100 | unsigned char data[0x10000]; | |
101 | uint16_t size; | |
102 | unsigned char vlan_needed; | |
103 | unsigned char sum_needed; | |
104 | bool cptse; | |
105 | e1000x_txd_props props; | |
106 | e1000x_txd_props tso_props; | |
107 | uint16_t tso_frames; | |
108 | } tx; | |
109 | ||
110 | struct { | |
111 | uint32_t val_in; /* shifted in from guest driver */ | |
112 | uint16_t bitnum_in; | |
113 | uint16_t bitnum_out; | |
114 | uint16_t reading; | |
115 | uint32_t old_eecd; | |
116 | } eecd_state; | |
117 | ||
118 | QEMUTimer *autoneg_timer; | |
119 | ||
120 | QEMUTimer *mit_timer; /* Mitigation timer. */ | |
121 | bool mit_timer_on; /* Mitigation timer is running. */ | |
122 | bool mit_irq_level; /* Tracks interrupt pin level. */ | |
123 | uint32_t mit_ide; /* Tracks E1000_TXD_CMD_IDE bit. */ | |
124 | ||
125 | QEMUTimer *flush_queue_timer; | |
126 | ||
127 | /* Compatibility flags for migration to/from qemu 1.3.0 and older */ | |
128 | #define E1000_FLAG_AUTONEG_BIT 0 | |
129 | #define E1000_FLAG_MIT_BIT 1 | |
130 | #define E1000_FLAG_MAC_BIT 2 | |
131 | #define E1000_FLAG_TSO_BIT 3 | |
132 | #define E1000_FLAG_AUTONEG (1 << E1000_FLAG_AUTONEG_BIT) | |
133 | #define E1000_FLAG_MIT (1 << E1000_FLAG_MIT_BIT) | |
134 | #define E1000_FLAG_MAC (1 << E1000_FLAG_MAC_BIT) | |
135 | #define E1000_FLAG_TSO (1 << E1000_FLAG_TSO_BIT) | |
136 | uint32_t compat_flags; | |
137 | bool received_tx_tso; | |
138 | bool use_tso_for_migration; | |
139 | e1000x_txd_props mig_props; | |
140 | } E1000State; | |
141 | ||
142 | #define chkflag(x) (s->compat_flags & E1000_FLAG_##x) | |
143 | ||
144 | typedef struct E1000BaseClass { | |
145 | PCIDeviceClass parent_class; | |
146 | uint16_t phy_id2; | |
147 | } E1000BaseClass; | |
148 | ||
149 | #define TYPE_E1000_BASE "e1000-base" | |
150 | ||
151 | #define E1000(obj) \ | |
152 | OBJECT_CHECK(E1000State, (obj), TYPE_E1000_BASE) | |
153 | ||
154 | #define E1000_DEVICE_CLASS(klass) \ | |
155 | OBJECT_CLASS_CHECK(E1000BaseClass, (klass), TYPE_E1000_BASE) | |
156 | #define E1000_DEVICE_GET_CLASS(obj) \ | |
157 | OBJECT_GET_CLASS(E1000BaseClass, (obj), TYPE_E1000_BASE) | |
158 | ||
159 | static void | |
160 | e1000_link_up(E1000State *s) | |
161 | { | |
162 | e1000x_update_regs_on_link_up(s->mac_reg, s->phy_reg); | |
163 | ||
164 | /* E1000_STATUS_LU is tested by e1000_can_receive() */ | |
165 | qemu_flush_queued_packets(qemu_get_queue(s->nic)); | |
166 | } | |
167 | ||
168 | static void | |
169 | e1000_autoneg_done(E1000State *s) | |
170 | { | |
171 | e1000x_update_regs_on_autoneg_done(s->mac_reg, s->phy_reg); | |
172 | ||
173 | /* E1000_STATUS_LU is tested by e1000_can_receive() */ | |
174 | qemu_flush_queued_packets(qemu_get_queue(s->nic)); | |
175 | } | |
176 | ||
177 | static bool | |
178 | have_autoneg(E1000State *s) | |
179 | { | |
180 | return chkflag(AUTONEG) && (s->phy_reg[PHY_CTRL] & MII_CR_AUTO_NEG_EN); | |
181 | } | |
182 | ||
183 | static void | |
184 | set_phy_ctrl(E1000State *s, int index, uint16_t val) | |
185 | { | |
186 | /* bits 0-5 reserved; MII_CR_[RESTART_AUTO_NEG,RESET] are self clearing */ | |
187 | s->phy_reg[PHY_CTRL] = val & ~(0x3f | | |
188 | MII_CR_RESET | | |
189 | MII_CR_RESTART_AUTO_NEG); | |
190 | ||
191 | /* | |
192 | * QEMU 1.3 does not support link auto-negotiation emulation, so if we | |
193 | * migrate during auto negotiation, after migration the link will be | |
194 | * down. | |
195 | */ | |
196 | if (have_autoneg(s) && (val & MII_CR_RESTART_AUTO_NEG)) { | |
197 | e1000x_restart_autoneg(s->mac_reg, s->phy_reg, s->autoneg_timer); | |
198 | } | |
199 | } | |
200 | ||
201 | static void (*phyreg_writeops[])(E1000State *, int, uint16_t) = { | |
202 | [PHY_CTRL] = set_phy_ctrl, | |
203 | }; | |
204 | ||
205 | enum { NPHYWRITEOPS = ARRAY_SIZE(phyreg_writeops) }; | |
206 | ||
207 | enum { PHY_R = 1, PHY_W = 2, PHY_RW = PHY_R | PHY_W }; | |
208 | static const char phy_regcap[0x20] = { | |
209 | [PHY_STATUS] = PHY_R, [M88E1000_EXT_PHY_SPEC_CTRL] = PHY_RW, | |
210 | [PHY_ID1] = PHY_R, [M88E1000_PHY_SPEC_CTRL] = PHY_RW, | |
211 | [PHY_CTRL] = PHY_RW, [PHY_1000T_CTRL] = PHY_RW, | |
212 | [PHY_LP_ABILITY] = PHY_R, [PHY_1000T_STATUS] = PHY_R, | |
213 | [PHY_AUTONEG_ADV] = PHY_RW, [M88E1000_RX_ERR_CNTR] = PHY_R, | |
214 | [PHY_ID2] = PHY_R, [M88E1000_PHY_SPEC_STATUS] = PHY_R, | |
215 | [PHY_AUTONEG_EXP] = PHY_R, | |
216 | }; | |
217 | ||
218 | /* PHY_ID2 documented in 8254x_GBe_SDM.pdf, pp. 250 */ | |
219 | static const uint16_t phy_reg_init[] = { | |
220 | [PHY_CTRL] = MII_CR_SPEED_SELECT_MSB | | |
221 | MII_CR_FULL_DUPLEX | | |
222 | MII_CR_AUTO_NEG_EN, | |
223 | ||
224 | [PHY_STATUS] = MII_SR_EXTENDED_CAPS | | |
225 | MII_SR_LINK_STATUS | /* link initially up */ | |
226 | MII_SR_AUTONEG_CAPS | | |
227 | /* MII_SR_AUTONEG_COMPLETE: initially NOT completed */ | |
228 | MII_SR_PREAMBLE_SUPPRESS | | |
229 | MII_SR_EXTENDED_STATUS | | |
230 | MII_SR_10T_HD_CAPS | | |
231 | MII_SR_10T_FD_CAPS | | |
232 | MII_SR_100X_HD_CAPS | | |
233 | MII_SR_100X_FD_CAPS, | |
234 | ||
235 | [PHY_ID1] = 0x141, | |
236 | /* [PHY_ID2] configured per DevId, from e1000_reset() */ | |
237 | [PHY_AUTONEG_ADV] = 0xde1, | |
238 | [PHY_LP_ABILITY] = 0x1e0, | |
239 | [PHY_1000T_CTRL] = 0x0e00, | |
240 | [PHY_1000T_STATUS] = 0x3c00, | |
241 | [M88E1000_PHY_SPEC_CTRL] = 0x360, | |
242 | [M88E1000_PHY_SPEC_STATUS] = 0xac00, | |
243 | [M88E1000_EXT_PHY_SPEC_CTRL] = 0x0d60, | |
244 | }; | |
245 | ||
246 | static const uint32_t mac_reg_init[] = { | |
247 | [PBA] = 0x00100030, | |
248 | [LEDCTL] = 0x602, | |
249 | [CTRL] = E1000_CTRL_SWDPIN2 | E1000_CTRL_SWDPIN0 | | |
250 | E1000_CTRL_SPD_1000 | E1000_CTRL_SLU, | |
251 | [STATUS] = 0x80000000 | E1000_STATUS_GIO_MASTER_ENABLE | | |
252 | E1000_STATUS_ASDV | E1000_STATUS_MTXCKOK | | |
253 | E1000_STATUS_SPEED_1000 | E1000_STATUS_FD | | |
254 | E1000_STATUS_LU, | |
255 | [MANC] = E1000_MANC_EN_MNG2HOST | E1000_MANC_RCV_TCO_EN | | |
256 | E1000_MANC_ARP_EN | E1000_MANC_0298_EN | | |
257 | E1000_MANC_RMCP_EN, | |
258 | }; | |
259 | ||
260 | /* Helper function, *curr == 0 means the value is not set */ | |
261 | static inline void | |
262 | mit_update_delay(uint32_t *curr, uint32_t value) | |
263 | { | |
264 | if (value && (*curr == 0 || value < *curr)) { | |
265 | *curr = value; | |
266 | } | |
267 | } | |
268 | ||
269 | static void | |
270 | set_interrupt_cause(E1000State *s, int index, uint32_t val) | |
271 | { | |
272 | PCIDevice *d = PCI_DEVICE(s); | |
273 | uint32_t pending_ints; | |
274 | uint32_t mit_delay; | |
275 | ||
276 | s->mac_reg[ICR] = val; | |
277 | ||
278 | /* | |
279 | * Make sure ICR and ICS registers have the same value. | |
280 | * The spec says that the ICS register is write-only. However in practice, | |
281 | * on real hardware ICS is readable, and for reads it has the same value as | |
282 | * ICR (except that ICS does not have the clear on read behaviour of ICR). | |
283 | * | |
284 | * The VxWorks PRO/1000 driver uses this behaviour. | |
285 | */ | |
286 | s->mac_reg[ICS] = val; | |
287 | ||
288 | pending_ints = (s->mac_reg[IMS] & s->mac_reg[ICR]); | |
289 | if (!s->mit_irq_level && pending_ints) { | |
290 | /* | |
291 | * Here we detect a potential raising edge. We postpone raising the | |
292 | * interrupt line if we are inside the mitigation delay window | |
293 | * (s->mit_timer_on == 1). | |
294 | * We provide a partial implementation of interrupt mitigation, | |
295 | * emulating only RADV, TADV and ITR (lower 16 bits, 1024ns units for | |
296 | * RADV and TADV, 256ns units for ITR). RDTR is only used to enable | |
297 | * RADV; relative timers based on TIDV and RDTR are not implemented. | |
298 | */ | |
299 | if (s->mit_timer_on) { | |
300 | return; | |
301 | } | |
302 | if (chkflag(MIT)) { | |
303 | /* Compute the next mitigation delay according to pending | |
304 | * interrupts and the current values of RADV (provided | |
305 | * RDTR!=0), TADV and ITR. | |
306 | * Then rearm the timer. | |
307 | */ | |
308 | mit_delay = 0; | |
309 | if (s->mit_ide && | |
310 | (pending_ints & (E1000_ICR_TXQE | E1000_ICR_TXDW))) { | |
311 | mit_update_delay(&mit_delay, s->mac_reg[TADV] * 4); | |
312 | } | |
313 | if (s->mac_reg[RDTR] && (pending_ints & E1000_ICS_RXT0)) { | |
314 | mit_update_delay(&mit_delay, s->mac_reg[RADV] * 4); | |
315 | } | |
316 | mit_update_delay(&mit_delay, s->mac_reg[ITR]); | |
317 | ||
318 | /* | |
319 | * According to e1000 SPEC, the Ethernet controller guarantees | |
320 | * a maximum observable interrupt rate of 7813 interrupts/sec. | |
321 | * Thus if mit_delay < 500 then the delay should be set to the | |
322 | * minimum delay possible which is 500. | |
323 | */ | |
324 | mit_delay = (mit_delay < 500) ? 500 : mit_delay; | |
325 | ||
326 | s->mit_timer_on = 1; | |
327 | timer_mod(s->mit_timer, qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + | |
328 | mit_delay * 256); | |
329 | s->mit_ide = 0; | |
330 | } | |
331 | } | |
332 | ||
333 | s->mit_irq_level = (pending_ints != 0); | |
334 | pci_set_irq(d, s->mit_irq_level); | |
335 | } | |
336 | ||
337 | static void | |
338 | e1000_mit_timer(void *opaque) | |
339 | { | |
340 | E1000State *s = opaque; | |
341 | ||
342 | s->mit_timer_on = 0; | |
343 | /* Call set_interrupt_cause to update the irq level (if necessary). */ | |
344 | set_interrupt_cause(s, 0, s->mac_reg[ICR]); | |
345 | } | |
346 | ||
347 | static void | |
348 | set_ics(E1000State *s, int index, uint32_t val) | |
349 | { | |
350 | DBGOUT(INTERRUPT, "set_ics %x, ICR %x, IMR %x\n", val, s->mac_reg[ICR], | |
351 | s->mac_reg[IMS]); | |
352 | set_interrupt_cause(s, 0, val | s->mac_reg[ICR]); | |
353 | } | |
354 | ||
355 | static void | |
356 | e1000_autoneg_timer(void *opaque) | |
357 | { | |
358 | E1000State *s = opaque; | |
359 | if (!qemu_get_queue(s->nic)->link_down) { | |
360 | e1000_autoneg_done(s); | |
361 | set_ics(s, 0, E1000_ICS_LSC); /* signal link status change to guest */ | |
362 | } | |
363 | } | |
364 | ||
365 | static void e1000_reset(void *opaque) | |
366 | { | |
367 | E1000State *d = opaque; | |
368 | E1000BaseClass *edc = E1000_DEVICE_GET_CLASS(d); | |
369 | uint8_t *macaddr = d->conf.macaddr.a; | |
370 | ||
371 | timer_del(d->autoneg_timer); | |
372 | timer_del(d->mit_timer); | |
373 | timer_del(d->flush_queue_timer); | |
374 | d->mit_timer_on = 0; | |
375 | d->mit_irq_level = 0; | |
376 | d->mit_ide = 0; | |
377 | memset(d->phy_reg, 0, sizeof d->phy_reg); | |
378 | memmove(d->phy_reg, phy_reg_init, sizeof phy_reg_init); | |
379 | d->phy_reg[PHY_ID2] = edc->phy_id2; | |
380 | memset(d->mac_reg, 0, sizeof d->mac_reg); | |
381 | memmove(d->mac_reg, mac_reg_init, sizeof mac_reg_init); | |
382 | d->rxbuf_min_shift = 1; | |
383 | memset(&d->tx, 0, sizeof d->tx); | |
384 | ||
385 | if (qemu_get_queue(d->nic)->link_down) { | |
386 | e1000x_update_regs_on_link_down(d->mac_reg, d->phy_reg); | |
387 | } | |
388 | ||
389 | e1000x_reset_mac_addr(d->nic, d->mac_reg, macaddr); | |
390 | } | |
391 | ||
392 | static void | |
393 | set_ctrl(E1000State *s, int index, uint32_t val) | |
394 | { | |
395 | /* RST is self clearing */ | |
396 | s->mac_reg[CTRL] = val & ~E1000_CTRL_RST; | |
397 | } | |
398 | ||
399 | static void | |
400 | e1000_flush_queue_timer(void *opaque) | |
401 | { | |
402 | E1000State *s = opaque; | |
403 | ||
404 | qemu_flush_queued_packets(qemu_get_queue(s->nic)); | |
405 | } | |
406 | ||
407 | static void | |
408 | set_rx_control(E1000State *s, int index, uint32_t val) | |
409 | { | |
410 | s->mac_reg[RCTL] = val; | |
411 | s->rxbuf_size = e1000x_rxbufsize(val); | |
412 | s->rxbuf_min_shift = ((val / E1000_RCTL_RDMTS_QUAT) & 3) + 1; | |
413 | DBGOUT(RX, "RCTL: %d, mac_reg[RCTL] = 0x%x\n", s->mac_reg[RDT], | |
414 | s->mac_reg[RCTL]); | |
415 | timer_mod(s->flush_queue_timer, | |
416 | qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL) + 1000); | |
417 | } | |
418 | ||
419 | static void | |
420 | set_mdic(E1000State *s, int index, uint32_t val) | |
421 | { | |
422 | uint32_t data = val & E1000_MDIC_DATA_MASK; | |
423 | uint32_t addr = ((val & E1000_MDIC_REG_MASK) >> E1000_MDIC_REG_SHIFT); | |
424 | ||
425 | if ((val & E1000_MDIC_PHY_MASK) >> E1000_MDIC_PHY_SHIFT != 1) // phy # | |
426 | val = s->mac_reg[MDIC] | E1000_MDIC_ERROR; | |
427 | else if (val & E1000_MDIC_OP_READ) { | |
428 | DBGOUT(MDIC, "MDIC read reg 0x%x\n", addr); | |
429 | if (!(phy_regcap[addr] & PHY_R)) { | |
430 | DBGOUT(MDIC, "MDIC read reg %x unhandled\n", addr); | |
431 | val |= E1000_MDIC_ERROR; | |
432 | } else | |
433 | val = (val ^ data) | s->phy_reg[addr]; | |
434 | } else if (val & E1000_MDIC_OP_WRITE) { | |
435 | DBGOUT(MDIC, "MDIC write reg 0x%x, value 0x%x\n", addr, data); | |
436 | if (!(phy_regcap[addr] & PHY_W)) { | |
437 | DBGOUT(MDIC, "MDIC write reg %x unhandled\n", addr); | |
438 | val |= E1000_MDIC_ERROR; | |
439 | } else { | |
440 | if (addr < NPHYWRITEOPS && phyreg_writeops[addr]) { | |
441 | phyreg_writeops[addr](s, index, data); | |
442 | } else { | |
443 | s->phy_reg[addr] = data; | |
444 | } | |
445 | } | |
446 | } | |
447 | s->mac_reg[MDIC] = val | E1000_MDIC_READY; | |
448 | ||
449 | if (val & E1000_MDIC_INT_EN) { | |
450 | set_ics(s, 0, E1000_ICR_MDAC); | |
451 | } | |
452 | } | |
453 | ||
454 | static uint32_t | |
455 | get_eecd(E1000State *s, int index) | |
456 | { | |
457 | uint32_t ret = E1000_EECD_PRES|E1000_EECD_GNT | s->eecd_state.old_eecd; | |
458 | ||
459 | DBGOUT(EEPROM, "reading eeprom bit %d (reading %d)\n", | |
460 | s->eecd_state.bitnum_out, s->eecd_state.reading); | |
461 | if (!s->eecd_state.reading || | |
462 | ((s->eeprom_data[(s->eecd_state.bitnum_out >> 4) & 0x3f] >> | |
463 | ((s->eecd_state.bitnum_out & 0xf) ^ 0xf))) & 1) | |
464 | ret |= E1000_EECD_DO; | |
465 | return ret; | |
466 | } | |
467 | ||
468 | static void | |
469 | set_eecd(E1000State *s, int index, uint32_t val) | |
470 | { | |
471 | uint32_t oldval = s->eecd_state.old_eecd; | |
472 | ||
473 | s->eecd_state.old_eecd = val & (E1000_EECD_SK | E1000_EECD_CS | | |
474 | E1000_EECD_DI|E1000_EECD_FWE_MASK|E1000_EECD_REQ); | |
475 | if (!(E1000_EECD_CS & val)) { /* CS inactive; nothing to do */ | |
476 | return; | |
477 | } | |
478 | if (E1000_EECD_CS & (val ^ oldval)) { /* CS rise edge; reset state */ | |
479 | s->eecd_state.val_in = 0; | |
480 | s->eecd_state.bitnum_in = 0; | |
481 | s->eecd_state.bitnum_out = 0; | |
482 | s->eecd_state.reading = 0; | |
483 | } | |
484 | if (!(E1000_EECD_SK & (val ^ oldval))) { /* no clock edge */ | |
485 | return; | |
486 | } | |
487 | if (!(E1000_EECD_SK & val)) { /* falling edge */ | |
488 | s->eecd_state.bitnum_out++; | |
489 | return; | |
490 | } | |
491 | s->eecd_state.val_in <<= 1; | |
492 | if (val & E1000_EECD_DI) | |
493 | s->eecd_state.val_in |= 1; | |
494 | if (++s->eecd_state.bitnum_in == 9 && !s->eecd_state.reading) { | |
495 | s->eecd_state.bitnum_out = ((s->eecd_state.val_in & 0x3f)<<4)-1; | |
496 | s->eecd_state.reading = (((s->eecd_state.val_in >> 6) & 7) == | |
497 | EEPROM_READ_OPCODE_MICROWIRE); | |
498 | } | |
499 | DBGOUT(EEPROM, "eeprom bitnum in %d out %d, reading %d\n", | |
500 | s->eecd_state.bitnum_in, s->eecd_state.bitnum_out, | |
501 | s->eecd_state.reading); | |
502 | } | |
503 | ||
504 | static uint32_t | |
505 | flash_eerd_read(E1000State *s, int x) | |
506 | { | |
507 | unsigned int index, r = s->mac_reg[EERD] & ~E1000_EEPROM_RW_REG_START; | |
508 | ||
509 | if ((s->mac_reg[EERD] & E1000_EEPROM_RW_REG_START) == 0) | |
510 | return (s->mac_reg[EERD]); | |
511 | ||
512 | if ((index = r >> E1000_EEPROM_RW_ADDR_SHIFT) > EEPROM_CHECKSUM_REG) | |
513 | return (E1000_EEPROM_RW_REG_DONE | r); | |
514 | ||
515 | return ((s->eeprom_data[index] << E1000_EEPROM_RW_REG_DATA) | | |
516 | E1000_EEPROM_RW_REG_DONE | r); | |
517 | } | |
518 | ||
519 | static void | |
520 | putsum(uint8_t *data, uint32_t n, uint32_t sloc, uint32_t css, uint32_t cse) | |
521 | { | |
522 | uint32_t sum; | |
523 | ||
524 | if (cse && cse < n) | |
525 | n = cse + 1; | |
526 | if (sloc < n-1) { | |
527 | sum = net_checksum_add(n-css, data+css); | |
528 | stw_be_p(data + sloc, net_checksum_finish_nozero(sum)); | |
529 | } | |
530 | } | |
531 | ||
532 | static inline void | |
533 | inc_tx_bcast_or_mcast_count(E1000State *s, const unsigned char *arr) | |
534 | { | |
535 | if (!memcmp(arr, bcast, sizeof bcast)) { | |
536 | e1000x_inc_reg_if_not_full(s->mac_reg, BPTC); | |
537 | } else if (arr[0] & 1) { | |
538 | e1000x_inc_reg_if_not_full(s->mac_reg, MPTC); | |
539 | } | |
540 | } | |
541 | ||
542 | static void | |
543 | e1000_send_packet(E1000State *s, const uint8_t *buf, int size) | |
544 | { | |
545 | static const int PTCregs[6] = { PTC64, PTC127, PTC255, PTC511, | |
546 | PTC1023, PTC1522 }; | |
547 | ||
548 | NetClientState *nc = qemu_get_queue(s->nic); | |
549 | if (s->phy_reg[PHY_CTRL] & MII_CR_LOOPBACK) { | |
550 | nc->info->receive(nc, buf, size); | |
551 | } else { | |
552 | qemu_send_packet(nc, buf, size); | |
553 | } | |
554 | inc_tx_bcast_or_mcast_count(s, buf); | |
555 | e1000x_increase_size_stats(s->mac_reg, PTCregs, size); | |
556 | } | |
557 | ||
558 | static void | |
559 | xmit_seg(E1000State *s) | |
560 | { | |
561 | uint16_t len; | |
562 | unsigned int frames = s->tx.tso_frames, css, sofar; | |
563 | struct e1000_tx *tp = &s->tx; | |
564 | struct e1000x_txd_props *props = tp->cptse ? &tp->tso_props : &tp->props; | |
565 | ||
566 | if (tp->cptse) { | |
567 | css = props->ipcss; | |
568 | DBGOUT(TXSUM, "frames %d size %d ipcss %d\n", | |
569 | frames, tp->size, css); | |
570 | if (props->ip) { /* IPv4 */ | |
571 | stw_be_p(tp->data+css+2, tp->size - css); | |
572 | stw_be_p(tp->data+css+4, | |
573 | lduw_be_p(tp->data + css + 4) + frames); | |
574 | } else { /* IPv6 */ | |
575 | stw_be_p(tp->data+css+4, tp->size - css); | |
576 | } | |
577 | css = props->tucss; | |
578 | len = tp->size - css; | |
579 | DBGOUT(TXSUM, "tcp %d tucss %d len %d\n", props->tcp, css, len); | |
580 | if (props->tcp) { | |
581 | sofar = frames * props->mss; | |
582 | stl_be_p(tp->data+css+4, ldl_be_p(tp->data+css+4)+sofar); /* seq */ | |
583 | if (props->paylen - sofar > props->mss) { | |
584 | tp->data[css + 13] &= ~9; /* PSH, FIN */ | |
585 | } else if (frames) { | |
586 | e1000x_inc_reg_if_not_full(s->mac_reg, TSCTC); | |
587 | } | |
588 | } else { /* UDP */ | |
589 | stw_be_p(tp->data+css+4, len); | |
590 | } | |
591 | if (tp->sum_needed & E1000_TXD_POPTS_TXSM) { | |
592 | unsigned int phsum; | |
593 | // add pseudo-header length before checksum calculation | |
594 | void *sp = tp->data + props->tucso; | |
595 | ||
596 | phsum = lduw_be_p(sp) + len; | |
597 | phsum = (phsum >> 16) + (phsum & 0xffff); | |
598 | stw_be_p(sp, phsum); | |
599 | } | |
600 | tp->tso_frames++; | |
601 | } | |
602 | ||
603 | if (tp->sum_needed & E1000_TXD_POPTS_TXSM) { | |
604 | putsum(tp->data, tp->size, props->tucso, props->tucss, props->tucse); | |
605 | } | |
606 | if (tp->sum_needed & E1000_TXD_POPTS_IXSM) { | |
607 | putsum(tp->data, tp->size, props->ipcso, props->ipcss, props->ipcse); | |
608 | } | |
609 | if (tp->vlan_needed) { | |
610 | memmove(tp->vlan, tp->data, 4); | |
611 | memmove(tp->data, tp->data + 4, 8); | |
612 | memcpy(tp->data + 8, tp->vlan_header, 4); | |
613 | e1000_send_packet(s, tp->vlan, tp->size + 4); | |
614 | } else { | |
615 | e1000_send_packet(s, tp->data, tp->size); | |
616 | } | |
617 | ||
618 | e1000x_inc_reg_if_not_full(s->mac_reg, TPT); | |
619 | e1000x_grow_8reg_if_not_full(s->mac_reg, TOTL, s->tx.size); | |
620 | s->mac_reg[GPTC] = s->mac_reg[TPT]; | |
621 | s->mac_reg[GOTCL] = s->mac_reg[TOTL]; | |
622 | s->mac_reg[GOTCH] = s->mac_reg[TOTH]; | |
623 | } | |
624 | ||
625 | static void | |
626 | process_tx_desc(E1000State *s, struct e1000_tx_desc *dp) | |
627 | { | |
628 | PCIDevice *d = PCI_DEVICE(s); | |
629 | uint32_t txd_lower = le32_to_cpu(dp->lower.data); | |
630 | uint32_t dtype = txd_lower & (E1000_TXD_CMD_DEXT | E1000_TXD_DTYP_D); | |
631 | unsigned int split_size = txd_lower & 0xffff, bytes, sz; | |
632 | unsigned int msh = 0xfffff; | |
633 | uint64_t addr; | |
634 | struct e1000_context_desc *xp = (struct e1000_context_desc *)dp; | |
635 | struct e1000_tx *tp = &s->tx; | |
636 | ||
637 | s->mit_ide |= (txd_lower & E1000_TXD_CMD_IDE); | |
638 | if (dtype == E1000_TXD_CMD_DEXT) { /* context descriptor */ | |
639 | if (le32_to_cpu(xp->cmd_and_length) & E1000_TXD_CMD_TSE) { | |
640 | e1000x_read_tx_ctx_descr(xp, &tp->tso_props); | |
641 | s->use_tso_for_migration = 1; | |
642 | tp->tso_frames = 0; | |
643 | } else { | |
644 | e1000x_read_tx_ctx_descr(xp, &tp->props); | |
645 | s->use_tso_for_migration = 0; | |
646 | } | |
647 | return; | |
648 | } else if (dtype == (E1000_TXD_CMD_DEXT | E1000_TXD_DTYP_D)) { | |
649 | // data descriptor | |
650 | if (tp->size == 0) { | |
651 | tp->sum_needed = le32_to_cpu(dp->upper.data) >> 8; | |
652 | } | |
653 | tp->cptse = (txd_lower & E1000_TXD_CMD_TSE) ? 1 : 0; | |
654 | } else { | |
655 | // legacy descriptor | |
656 | tp->cptse = 0; | |
657 | } | |
658 | ||
659 | if (e1000x_vlan_enabled(s->mac_reg) && | |
660 | e1000x_is_vlan_txd(txd_lower) && | |
661 | (tp->cptse || txd_lower & E1000_TXD_CMD_EOP)) { | |
662 | tp->vlan_needed = 1; | |
663 | stw_be_p(tp->vlan_header, | |
664 | le16_to_cpu(s->mac_reg[VET])); | |
665 | stw_be_p(tp->vlan_header + 2, | |
666 | le16_to_cpu(dp->upper.fields.special)); | |
667 | } | |
668 | ||
669 | addr = le64_to_cpu(dp->buffer_addr); | |
670 | if (tp->cptse) { | |
671 | msh = tp->tso_props.hdr_len + tp->tso_props.mss; | |
672 | do { | |
673 | bytes = split_size; | |
674 | if (tp->size + bytes > msh) | |
675 | bytes = msh - tp->size; | |
676 | ||
677 | bytes = MIN(sizeof(tp->data) - tp->size, bytes); | |
678 | pci_dma_read(d, addr, tp->data + tp->size, bytes); | |
679 | sz = tp->size + bytes; | |
680 | if (sz >= tp->tso_props.hdr_len | |
681 | && tp->size < tp->tso_props.hdr_len) { | |
682 | memmove(tp->header, tp->data, tp->tso_props.hdr_len); | |
683 | } | |
684 | tp->size = sz; | |
685 | addr += bytes; | |
686 | if (sz == msh) { | |
687 | xmit_seg(s); | |
688 | memmove(tp->data, tp->header, tp->tso_props.hdr_len); | |
689 | tp->size = tp->tso_props.hdr_len; | |
690 | } | |
691 | split_size -= bytes; | |
692 | } while (bytes && split_size); | |
693 | } else { | |
694 | split_size = MIN(sizeof(tp->data) - tp->size, split_size); | |
695 | pci_dma_read(d, addr, tp->data + tp->size, split_size); | |
696 | tp->size += split_size; | |
697 | } | |
698 | ||
699 | if (!(txd_lower & E1000_TXD_CMD_EOP)) | |
700 | return; | |
701 | if (!(tp->cptse && tp->size < tp->tso_props.hdr_len)) { | |
702 | xmit_seg(s); | |
703 | } | |
704 | tp->tso_frames = 0; | |
705 | tp->sum_needed = 0; | |
706 | tp->vlan_needed = 0; | |
707 | tp->size = 0; | |
708 | tp->cptse = 0; | |
709 | } | |
710 | ||
711 | static uint32_t | |
712 | txdesc_writeback(E1000State *s, dma_addr_t base, struct e1000_tx_desc *dp) | |
713 | { | |
714 | PCIDevice *d = PCI_DEVICE(s); | |
715 | uint32_t txd_upper, txd_lower = le32_to_cpu(dp->lower.data); | |
716 | ||
717 | if (!(txd_lower & (E1000_TXD_CMD_RS|E1000_TXD_CMD_RPS))) | |
718 | return 0; | |
719 | txd_upper = (le32_to_cpu(dp->upper.data) | E1000_TXD_STAT_DD) & | |
720 | ~(E1000_TXD_STAT_EC | E1000_TXD_STAT_LC | E1000_TXD_STAT_TU); | |
721 | dp->upper.data = cpu_to_le32(txd_upper); | |
722 | pci_dma_write(d, base + ((char *)&dp->upper - (char *)dp), | |
723 | &dp->upper, sizeof(dp->upper)); | |
724 | return E1000_ICR_TXDW; | |
725 | } | |
726 | ||
727 | static uint64_t tx_desc_base(E1000State *s) | |
728 | { | |
729 | uint64_t bah = s->mac_reg[TDBAH]; | |
730 | uint64_t bal = s->mac_reg[TDBAL] & ~0xf; | |
731 | ||
732 | return (bah << 32) + bal; | |
733 | } | |
734 | ||
735 | static void | |
736 | start_xmit(E1000State *s) | |
737 | { | |
738 | PCIDevice *d = PCI_DEVICE(s); | |
739 | dma_addr_t base; | |
740 | struct e1000_tx_desc desc; | |
741 | uint32_t tdh_start = s->mac_reg[TDH], cause = E1000_ICS_TXQE; | |
742 | ||
743 | if (!(s->mac_reg[TCTL] & E1000_TCTL_EN)) { | |
744 | DBGOUT(TX, "tx disabled\n"); | |
745 | return; | |
746 | } | |
747 | ||
748 | while (s->mac_reg[TDH] != s->mac_reg[TDT]) { | |
749 | base = tx_desc_base(s) + | |
750 | sizeof(struct e1000_tx_desc) * s->mac_reg[TDH]; | |
751 | pci_dma_read(d, base, &desc, sizeof(desc)); | |
752 | ||
753 | DBGOUT(TX, "index %d: %p : %x %x\n", s->mac_reg[TDH], | |
754 | (void *)(intptr_t)desc.buffer_addr, desc.lower.data, | |
755 | desc.upper.data); | |
756 | ||
757 | process_tx_desc(s, &desc); | |
758 | cause |= txdesc_writeback(s, base, &desc); | |
759 | ||
760 | if (++s->mac_reg[TDH] * sizeof(desc) >= s->mac_reg[TDLEN]) | |
761 | s->mac_reg[TDH] = 0; | |
762 | /* | |
763 | * the following could happen only if guest sw assigns | |
764 | * bogus values to TDT/TDLEN. | |
765 | * there's nothing too intelligent we could do about this. | |
766 | */ | |
767 | if (s->mac_reg[TDH] == tdh_start || | |
768 | tdh_start >= s->mac_reg[TDLEN] / sizeof(desc)) { | |
769 | DBGOUT(TXERR, "TDH wraparound @%x, TDT %x, TDLEN %x\n", | |
770 | tdh_start, s->mac_reg[TDT], s->mac_reg[TDLEN]); | |
771 | break; | |
772 | } | |
773 | } | |
774 | set_ics(s, 0, cause); | |
775 | } | |
776 | ||
777 | static int | |
778 | receive_filter(E1000State *s, const uint8_t *buf, int size) | |
779 | { | |
780 | uint32_t rctl = s->mac_reg[RCTL]; | |
781 | int isbcast = !memcmp(buf, bcast, sizeof bcast), ismcast = (buf[0] & 1); | |
782 | ||
783 | if (e1000x_is_vlan_packet(buf, le16_to_cpu(s->mac_reg[VET])) && | |
784 | e1000x_vlan_rx_filter_enabled(s->mac_reg)) { | |
785 | uint16_t vid = lduw_be_p(buf + 14); | |
786 | uint32_t vfta = ldl_le_p((uint32_t*)(s->mac_reg + VFTA) + | |
787 | ((vid >> 5) & 0x7f)); | |
788 | if ((vfta & (1 << (vid & 0x1f))) == 0) | |
789 | return 0; | |
790 | } | |
791 | ||
792 | if (!isbcast && !ismcast && (rctl & E1000_RCTL_UPE)) { /* promiscuous ucast */ | |
793 | return 1; | |
794 | } | |
795 | ||
796 | if (ismcast && (rctl & E1000_RCTL_MPE)) { /* promiscuous mcast */ | |
797 | e1000x_inc_reg_if_not_full(s->mac_reg, MPRC); | |
798 | return 1; | |
799 | } | |
800 | ||
801 | if (isbcast && (rctl & E1000_RCTL_BAM)) { /* broadcast enabled */ | |
802 | e1000x_inc_reg_if_not_full(s->mac_reg, BPRC); | |
803 | return 1; | |
804 | } | |
805 | ||
806 | return e1000x_rx_group_filter(s->mac_reg, buf); | |
807 | } | |
808 | ||
809 | static void | |
810 | e1000_set_link_status(NetClientState *nc) | |
811 | { | |
812 | E1000State *s = qemu_get_nic_opaque(nc); | |
813 | uint32_t old_status = s->mac_reg[STATUS]; | |
814 | ||
815 | if (nc->link_down) { | |
816 | e1000x_update_regs_on_link_down(s->mac_reg, s->phy_reg); | |
817 | } else { | |
818 | if (have_autoneg(s) && | |
819 | !(s->phy_reg[PHY_STATUS] & MII_SR_AUTONEG_COMPLETE)) { | |
820 | e1000x_restart_autoneg(s->mac_reg, s->phy_reg, s->autoneg_timer); | |
821 | } else { | |
822 | e1000_link_up(s); | |
823 | } | |
824 | } | |
825 | ||
826 | if (s->mac_reg[STATUS] != old_status) | |
827 | set_ics(s, 0, E1000_ICR_LSC); | |
828 | } | |
829 | ||
830 | static bool e1000_has_rxbufs(E1000State *s, size_t total_size) | |
831 | { | |
832 | int bufs; | |
833 | /* Fast-path short packets */ | |
834 | if (total_size <= s->rxbuf_size) { | |
835 | return s->mac_reg[RDH] != s->mac_reg[RDT]; | |
836 | } | |
837 | if (s->mac_reg[RDH] < s->mac_reg[RDT]) { | |
838 | bufs = s->mac_reg[RDT] - s->mac_reg[RDH]; | |
839 | } else if (s->mac_reg[RDH] > s->mac_reg[RDT]) { | |
840 | bufs = s->mac_reg[RDLEN] / sizeof(struct e1000_rx_desc) + | |
841 | s->mac_reg[RDT] - s->mac_reg[RDH]; | |
842 | } else { | |
843 | return false; | |
844 | } | |
845 | return total_size <= bufs * s->rxbuf_size; | |
846 | } | |
847 | ||
848 | static int | |
849 | e1000_can_receive(NetClientState *nc) | |
850 | { | |
851 | E1000State *s = qemu_get_nic_opaque(nc); | |
852 | ||
853 | return e1000x_rx_ready(&s->parent_obj, s->mac_reg) && | |
854 | e1000_has_rxbufs(s, 1) && !timer_pending(s->flush_queue_timer); | |
855 | } | |
856 | ||
857 | static uint64_t rx_desc_base(E1000State *s) | |
858 | { | |
859 | uint64_t bah = s->mac_reg[RDBAH]; | |
860 | uint64_t bal = s->mac_reg[RDBAL] & ~0xf; | |
861 | ||
862 | return (bah << 32) + bal; | |
863 | } | |
864 | ||
865 | static void | |
866 | e1000_receiver_overrun(E1000State *s, size_t size) | |
867 | { | |
868 | trace_e1000_receiver_overrun(size, s->mac_reg[RDH], s->mac_reg[RDT]); | |
869 | e1000x_inc_reg_if_not_full(s->mac_reg, RNBC); | |
870 | e1000x_inc_reg_if_not_full(s->mac_reg, MPC); | |
871 | set_ics(s, 0, E1000_ICS_RXO); | |
872 | } | |
873 | ||
874 | static ssize_t | |
875 | e1000_receive_iov(NetClientState *nc, const struct iovec *iov, int iovcnt) | |
876 | { | |
877 | E1000State *s = qemu_get_nic_opaque(nc); | |
878 | PCIDevice *d = PCI_DEVICE(s); | |
879 | struct e1000_rx_desc desc; | |
880 | dma_addr_t base; | |
881 | unsigned int n, rdt; | |
882 | uint32_t rdh_start; | |
883 | uint16_t vlan_special = 0; | |
884 | uint8_t vlan_status = 0; | |
885 | uint8_t min_buf[MIN_BUF_SIZE]; | |
886 | struct iovec min_iov; | |
887 | uint8_t *filter_buf = iov->iov_base; | |
888 | size_t size = iov_size(iov, iovcnt); | |
889 | size_t iov_ofs = 0; | |
890 | size_t desc_offset; | |
891 | size_t desc_size; | |
892 | size_t total_size; | |
893 | ||
894 | if (!e1000x_hw_rx_enabled(s->mac_reg)) { | |
895 | return -1; | |
896 | } | |
897 | ||
898 | if (timer_pending(s->flush_queue_timer)) { | |
899 | return 0; | |
900 | } | |
901 | ||
902 | /* Pad to minimum Ethernet frame length */ | |
903 | if (size < sizeof(min_buf)) { | |
904 | iov_to_buf(iov, iovcnt, 0, min_buf, size); | |
905 | memset(&min_buf[size], 0, sizeof(min_buf) - size); | |
906 | min_iov.iov_base = filter_buf = min_buf; | |
907 | min_iov.iov_len = size = sizeof(min_buf); | |
908 | iovcnt = 1; | |
909 | iov = &min_iov; | |
910 | } else if (iov->iov_len < MAXIMUM_ETHERNET_HDR_LEN) { | |
911 | /* This is very unlikely, but may happen. */ | |
912 | iov_to_buf(iov, iovcnt, 0, min_buf, MAXIMUM_ETHERNET_HDR_LEN); | |
913 | filter_buf = min_buf; | |
914 | } | |
915 | ||
916 | /* Discard oversized packets if !LPE and !SBP. */ | |
917 | if (e1000x_is_oversized(s->mac_reg, size)) { | |
918 | return size; | |
919 | } | |
920 | ||
921 | if (!receive_filter(s, filter_buf, size)) { | |
922 | return size; | |
923 | } | |
924 | ||
925 | if (e1000x_vlan_enabled(s->mac_reg) && | |
926 | e1000x_is_vlan_packet(filter_buf, le16_to_cpu(s->mac_reg[VET]))) { | |
927 | vlan_special = cpu_to_le16(lduw_be_p(filter_buf + 14)); | |
928 | iov_ofs = 4; | |
929 | if (filter_buf == iov->iov_base) { | |
930 | memmove(filter_buf + 4, filter_buf, 12); | |
931 | } else { | |
932 | iov_from_buf(iov, iovcnt, 4, filter_buf, 12); | |
933 | while (iov->iov_len <= iov_ofs) { | |
934 | iov_ofs -= iov->iov_len; | |
935 | iov++; | |
936 | } | |
937 | } | |
938 | vlan_status = E1000_RXD_STAT_VP; | |
939 | size -= 4; | |
940 | } | |
941 | ||
942 | rdh_start = s->mac_reg[RDH]; | |
943 | desc_offset = 0; | |
944 | total_size = size + e1000x_fcs_len(s->mac_reg); | |
945 | if (!e1000_has_rxbufs(s, total_size)) { | |
946 | e1000_receiver_overrun(s, total_size); | |
947 | return -1; | |
948 | } | |
949 | do { | |
950 | desc_size = total_size - desc_offset; | |
951 | if (desc_size > s->rxbuf_size) { | |
952 | desc_size = s->rxbuf_size; | |
953 | } | |
954 | base = rx_desc_base(s) + sizeof(desc) * s->mac_reg[RDH]; | |
955 | pci_dma_read(d, base, &desc, sizeof(desc)); | |
956 | desc.special = vlan_special; | |
957 | desc.status |= (vlan_status | E1000_RXD_STAT_DD); | |
958 | if (desc.buffer_addr) { | |
959 | if (desc_offset < size) { | |
960 | size_t iov_copy; | |
961 | hwaddr ba = le64_to_cpu(desc.buffer_addr); | |
962 | size_t copy_size = size - desc_offset; | |
963 | if (copy_size > s->rxbuf_size) { | |
964 | copy_size = s->rxbuf_size; | |
965 | } | |
966 | do { | |
967 | iov_copy = MIN(copy_size, iov->iov_len - iov_ofs); | |
968 | pci_dma_write(d, ba, iov->iov_base + iov_ofs, iov_copy); | |
969 | copy_size -= iov_copy; | |
970 | ba += iov_copy; | |
971 | iov_ofs += iov_copy; | |
972 | if (iov_ofs == iov->iov_len) { | |
973 | iov++; | |
974 | iov_ofs = 0; | |
975 | } | |
976 | } while (copy_size); | |
977 | } | |
978 | desc_offset += desc_size; | |
979 | desc.length = cpu_to_le16(desc_size); | |
980 | if (desc_offset >= total_size) { | |
981 | desc.status |= E1000_RXD_STAT_EOP | E1000_RXD_STAT_IXSM; | |
982 | } else { | |
983 | /* Guest zeroing out status is not a hardware requirement. | |
984 | Clear EOP in case guest didn't do it. */ | |
985 | desc.status &= ~E1000_RXD_STAT_EOP; | |
986 | } | |
987 | } else { // as per intel docs; skip descriptors with null buf addr | |
988 | DBGOUT(RX, "Null RX descriptor!!\n"); | |
989 | } | |
990 | pci_dma_write(d, base, &desc, sizeof(desc)); | |
991 | ||
992 | if (++s->mac_reg[RDH] * sizeof(desc) >= s->mac_reg[RDLEN]) | |
993 | s->mac_reg[RDH] = 0; | |
994 | /* see comment in start_xmit; same here */ | |
995 | if (s->mac_reg[RDH] == rdh_start || | |
996 | rdh_start >= s->mac_reg[RDLEN] / sizeof(desc)) { | |
997 | DBGOUT(RXERR, "RDH wraparound @%x, RDT %x, RDLEN %x\n", | |
998 | rdh_start, s->mac_reg[RDT], s->mac_reg[RDLEN]); | |
999 | e1000_receiver_overrun(s, total_size); | |
1000 | return -1; | |
1001 | } | |
1002 | } while (desc_offset < total_size); | |
1003 | ||
1004 | e1000x_update_rx_total_stats(s->mac_reg, size, total_size); | |
1005 | ||
1006 | n = E1000_ICS_RXT0; | |
1007 | if ((rdt = s->mac_reg[RDT]) < s->mac_reg[RDH]) | |
1008 | rdt += s->mac_reg[RDLEN] / sizeof(desc); | |
1009 | if (((rdt - s->mac_reg[RDH]) * sizeof(desc)) <= s->mac_reg[RDLEN] >> | |
1010 | s->rxbuf_min_shift) | |
1011 | n |= E1000_ICS_RXDMT0; | |
1012 | ||
1013 | set_ics(s, 0, n); | |
1014 | ||
1015 | return size; | |
1016 | } | |
1017 | ||
1018 | static ssize_t | |
1019 | e1000_receive(NetClientState *nc, const uint8_t *buf, size_t size) | |
1020 | { | |
1021 | const struct iovec iov = { | |
1022 | .iov_base = (uint8_t *)buf, | |
1023 | .iov_len = size | |
1024 | }; | |
1025 | ||
1026 | return e1000_receive_iov(nc, &iov, 1); | |
1027 | } | |
1028 | ||
1029 | static uint32_t | |
1030 | mac_readreg(E1000State *s, int index) | |
1031 | { | |
1032 | return s->mac_reg[index]; | |
1033 | } | |
1034 | ||
1035 | static uint32_t | |
1036 | mac_low4_read(E1000State *s, int index) | |
1037 | { | |
1038 | return s->mac_reg[index] & 0xf; | |
1039 | } | |
1040 | ||
1041 | static uint32_t | |
1042 | mac_low11_read(E1000State *s, int index) | |
1043 | { | |
1044 | return s->mac_reg[index] & 0x7ff; | |
1045 | } | |
1046 | ||
1047 | static uint32_t | |
1048 | mac_low13_read(E1000State *s, int index) | |
1049 | { | |
1050 | return s->mac_reg[index] & 0x1fff; | |
1051 | } | |
1052 | ||
1053 | static uint32_t | |
1054 | mac_low16_read(E1000State *s, int index) | |
1055 | { | |
1056 | return s->mac_reg[index] & 0xffff; | |
1057 | } | |
1058 | ||
1059 | static uint32_t | |
1060 | mac_icr_read(E1000State *s, int index) | |
1061 | { | |
1062 | uint32_t ret = s->mac_reg[ICR]; | |
1063 | ||
1064 | DBGOUT(INTERRUPT, "ICR read: %x\n", ret); | |
1065 | set_interrupt_cause(s, 0, 0); | |
1066 | return ret; | |
1067 | } | |
1068 | ||
1069 | static uint32_t | |
1070 | mac_read_clr4(E1000State *s, int index) | |
1071 | { | |
1072 | uint32_t ret = s->mac_reg[index]; | |
1073 | ||
1074 | s->mac_reg[index] = 0; | |
1075 | return ret; | |
1076 | } | |
1077 | ||
1078 | static uint32_t | |
1079 | mac_read_clr8(E1000State *s, int index) | |
1080 | { | |
1081 | uint32_t ret = s->mac_reg[index]; | |
1082 | ||
1083 | s->mac_reg[index] = 0; | |
1084 | s->mac_reg[index-1] = 0; | |
1085 | return ret; | |
1086 | } | |
1087 | ||
1088 | static void | |
1089 | mac_writereg(E1000State *s, int index, uint32_t val) | |
1090 | { | |
1091 | uint32_t macaddr[2]; | |
1092 | ||
1093 | s->mac_reg[index] = val; | |
1094 | ||
1095 | if (index == RA + 1) { | |
1096 | macaddr[0] = cpu_to_le32(s->mac_reg[RA]); | |
1097 | macaddr[1] = cpu_to_le32(s->mac_reg[RA + 1]); | |
1098 | qemu_format_nic_info_str(qemu_get_queue(s->nic), (uint8_t *)macaddr); | |
1099 | } | |
1100 | } | |
1101 | ||
1102 | static void | |
1103 | set_rdt(E1000State *s, int index, uint32_t val) | |
1104 | { | |
1105 | s->mac_reg[index] = val & 0xffff; | |
1106 | if (e1000_has_rxbufs(s, 1)) { | |
1107 | qemu_flush_queued_packets(qemu_get_queue(s->nic)); | |
1108 | } | |
1109 | } | |
1110 | ||
1111 | static void | |
1112 | set_16bit(E1000State *s, int index, uint32_t val) | |
1113 | { | |
1114 | s->mac_reg[index] = val & 0xffff; | |
1115 | } | |
1116 | ||
1117 | static void | |
1118 | set_dlen(E1000State *s, int index, uint32_t val) | |
1119 | { | |
1120 | s->mac_reg[index] = val & 0xfff80; | |
1121 | } | |
1122 | ||
1123 | static void | |
1124 | set_tctl(E1000State *s, int index, uint32_t val) | |
1125 | { | |
1126 | s->mac_reg[index] = val; | |
1127 | s->mac_reg[TDT] &= 0xffff; | |
1128 | start_xmit(s); | |
1129 | } | |
1130 | ||
1131 | static void | |
1132 | set_icr(E1000State *s, int index, uint32_t val) | |
1133 | { | |
1134 | DBGOUT(INTERRUPT, "set_icr %x\n", val); | |
1135 | set_interrupt_cause(s, 0, s->mac_reg[ICR] & ~val); | |
1136 | } | |
1137 | ||
1138 | static void | |
1139 | set_imc(E1000State *s, int index, uint32_t val) | |
1140 | { | |
1141 | s->mac_reg[IMS] &= ~val; | |
1142 | set_ics(s, 0, 0); | |
1143 | } | |
1144 | ||
1145 | static void | |
1146 | set_ims(E1000State *s, int index, uint32_t val) | |
1147 | { | |
1148 | s->mac_reg[IMS] |= val; | |
1149 | set_ics(s, 0, 0); | |
1150 | } | |
1151 | ||
1152 | #define getreg(x) [x] = mac_readreg | |
1153 | static uint32_t (*macreg_readops[])(E1000State *, int) = { | |
1154 | getreg(PBA), getreg(RCTL), getreg(TDH), getreg(TXDCTL), | |
1155 | getreg(WUFC), getreg(TDT), getreg(CTRL), getreg(LEDCTL), | |
1156 | getreg(MANC), getreg(MDIC), getreg(SWSM), getreg(STATUS), | |
1157 | getreg(TORL), getreg(TOTL), getreg(IMS), getreg(TCTL), | |
1158 | getreg(RDH), getreg(RDT), getreg(VET), getreg(ICS), | |
1159 | getreg(TDBAL), getreg(TDBAH), getreg(RDBAH), getreg(RDBAL), | |
1160 | getreg(TDLEN), getreg(RDLEN), getreg(RDTR), getreg(RADV), | |
1161 | getreg(TADV), getreg(ITR), getreg(FCRUC), getreg(IPAV), | |
1162 | getreg(WUC), getreg(WUS), getreg(SCC), getreg(ECOL), | |
1163 | getreg(MCC), getreg(LATECOL), getreg(COLC), getreg(DC), | |
1164 | getreg(TNCRS), getreg(SEQEC), getreg(CEXTERR), getreg(RLEC), | |
1165 | getreg(XONRXC), getreg(XONTXC), getreg(XOFFRXC), getreg(XOFFTXC), | |
1166 | getreg(RFC), getreg(RJC), getreg(RNBC), getreg(TSCTFC), | |
1167 | getreg(MGTPRC), getreg(MGTPDC), getreg(MGTPTC), getreg(GORCL), | |
1168 | getreg(GOTCL), | |
1169 | ||
1170 | [TOTH] = mac_read_clr8, [TORH] = mac_read_clr8, | |
1171 | [GOTCH] = mac_read_clr8, [GORCH] = mac_read_clr8, | |
1172 | [PRC64] = mac_read_clr4, [PRC127] = mac_read_clr4, | |
1173 | [PRC255] = mac_read_clr4, [PRC511] = mac_read_clr4, | |
1174 | [PRC1023] = mac_read_clr4, [PRC1522] = mac_read_clr4, | |
1175 | [PTC64] = mac_read_clr4, [PTC127] = mac_read_clr4, | |
1176 | [PTC255] = mac_read_clr4, [PTC511] = mac_read_clr4, | |
1177 | [PTC1023] = mac_read_clr4, [PTC1522] = mac_read_clr4, | |
1178 | [GPRC] = mac_read_clr4, [GPTC] = mac_read_clr4, | |
1179 | [TPT] = mac_read_clr4, [TPR] = mac_read_clr4, | |
1180 | [RUC] = mac_read_clr4, [ROC] = mac_read_clr4, | |
1181 | [BPRC] = mac_read_clr4, [MPRC] = mac_read_clr4, | |
1182 | [TSCTC] = mac_read_clr4, [BPTC] = mac_read_clr4, | |
1183 | [MPTC] = mac_read_clr4, | |
1184 | [ICR] = mac_icr_read, [EECD] = get_eecd, | |
1185 | [EERD] = flash_eerd_read, | |
1186 | [RDFH] = mac_low13_read, [RDFT] = mac_low13_read, | |
1187 | [RDFHS] = mac_low13_read, [RDFTS] = mac_low13_read, | |
1188 | [RDFPC] = mac_low13_read, | |
1189 | [TDFH] = mac_low11_read, [TDFT] = mac_low11_read, | |
1190 | [TDFHS] = mac_low13_read, [TDFTS] = mac_low13_read, | |
1191 | [TDFPC] = mac_low13_read, | |
1192 | [AIT] = mac_low16_read, | |
1193 | ||
1194 | [CRCERRS ... MPC] = &mac_readreg, | |
1195 | [IP6AT ... IP6AT+3] = &mac_readreg, [IP4AT ... IP4AT+6] = &mac_readreg, | |
1196 | [FFLT ... FFLT+6] = &mac_low11_read, | |
1197 | [RA ... RA+31] = &mac_readreg, | |
1198 | [WUPM ... WUPM+31] = &mac_readreg, | |
1199 | [MTA ... MTA+127] = &mac_readreg, | |
1200 | [VFTA ... VFTA+127] = &mac_readreg, | |
1201 | [FFMT ... FFMT+254] = &mac_low4_read, | |
1202 | [FFVT ... FFVT+254] = &mac_readreg, | |
1203 | [PBM ... PBM+16383] = &mac_readreg, | |
1204 | }; | |
1205 | enum { NREADOPS = ARRAY_SIZE(macreg_readops) }; | |
1206 | ||
1207 | #define putreg(x) [x] = mac_writereg | |
1208 | static void (*macreg_writeops[])(E1000State *, int, uint32_t) = { | |
1209 | putreg(PBA), putreg(EERD), putreg(SWSM), putreg(WUFC), | |
1210 | putreg(TDBAL), putreg(TDBAH), putreg(TXDCTL), putreg(RDBAH), | |
1211 | putreg(RDBAL), putreg(LEDCTL), putreg(VET), putreg(FCRUC), | |
1212 | putreg(TDFH), putreg(TDFT), putreg(TDFHS), putreg(TDFTS), | |
1213 | putreg(TDFPC), putreg(RDFH), putreg(RDFT), putreg(RDFHS), | |
1214 | putreg(RDFTS), putreg(RDFPC), putreg(IPAV), putreg(WUC), | |
1215 | putreg(WUS), putreg(AIT), | |
1216 | ||
1217 | [TDLEN] = set_dlen, [RDLEN] = set_dlen, [TCTL] = set_tctl, | |
1218 | [TDT] = set_tctl, [MDIC] = set_mdic, [ICS] = set_ics, | |
1219 | [TDH] = set_16bit, [RDH] = set_16bit, [RDT] = set_rdt, | |
1220 | [IMC] = set_imc, [IMS] = set_ims, [ICR] = set_icr, | |
1221 | [EECD] = set_eecd, [RCTL] = set_rx_control, [CTRL] = set_ctrl, | |
1222 | [RDTR] = set_16bit, [RADV] = set_16bit, [TADV] = set_16bit, | |
1223 | [ITR] = set_16bit, | |
1224 | ||
1225 | [IP6AT ... IP6AT+3] = &mac_writereg, [IP4AT ... IP4AT+6] = &mac_writereg, | |
1226 | [FFLT ... FFLT+6] = &mac_writereg, | |
1227 | [RA ... RA+31] = &mac_writereg, | |
1228 | [WUPM ... WUPM+31] = &mac_writereg, | |
1229 | [MTA ... MTA+127] = &mac_writereg, | |
1230 | [VFTA ... VFTA+127] = &mac_writereg, | |
1231 | [FFMT ... FFMT+254] = &mac_writereg, [FFVT ... FFVT+254] = &mac_writereg, | |
1232 | [PBM ... PBM+16383] = &mac_writereg, | |
1233 | }; | |
1234 | ||
1235 | enum { NWRITEOPS = ARRAY_SIZE(macreg_writeops) }; | |
1236 | ||
1237 | enum { MAC_ACCESS_PARTIAL = 1, MAC_ACCESS_FLAG_NEEDED = 2 }; | |
1238 | ||
1239 | #define markflag(x) ((E1000_FLAG_##x << 2) | MAC_ACCESS_FLAG_NEEDED) | |
1240 | /* In the array below the meaning of the bits is: [f|f|f|f|f|f|n|p] | |
1241 | * f - flag bits (up to 6 possible flags) | |
1242 | * n - flag needed | |
1243 | * p - partially implenented */ | |
1244 | static const uint8_t mac_reg_access[0x8000] = { | |
1245 | [RDTR] = markflag(MIT), [TADV] = markflag(MIT), | |
1246 | [RADV] = markflag(MIT), [ITR] = markflag(MIT), | |
1247 | ||
1248 | [IPAV] = markflag(MAC), [WUC] = markflag(MAC), | |
1249 | [IP6AT] = markflag(MAC), [IP4AT] = markflag(MAC), | |
1250 | [FFVT] = markflag(MAC), [WUPM] = markflag(MAC), | |
1251 | [ECOL] = markflag(MAC), [MCC] = markflag(MAC), | |
1252 | [DC] = markflag(MAC), [TNCRS] = markflag(MAC), | |
1253 | [RLEC] = markflag(MAC), [XONRXC] = markflag(MAC), | |
1254 | [XOFFTXC] = markflag(MAC), [RFC] = markflag(MAC), | |
1255 | [TSCTFC] = markflag(MAC), [MGTPRC] = markflag(MAC), | |
1256 | [WUS] = markflag(MAC), [AIT] = markflag(MAC), | |
1257 | [FFLT] = markflag(MAC), [FFMT] = markflag(MAC), | |
1258 | [SCC] = markflag(MAC), [FCRUC] = markflag(MAC), | |
1259 | [LATECOL] = markflag(MAC), [COLC] = markflag(MAC), | |
1260 | [SEQEC] = markflag(MAC), [CEXTERR] = markflag(MAC), | |
1261 | [XONTXC] = markflag(MAC), [XOFFRXC] = markflag(MAC), | |
1262 | [RJC] = markflag(MAC), [RNBC] = markflag(MAC), | |
1263 | [MGTPDC] = markflag(MAC), [MGTPTC] = markflag(MAC), | |
1264 | [RUC] = markflag(MAC), [ROC] = markflag(MAC), | |
1265 | [GORCL] = markflag(MAC), [GORCH] = markflag(MAC), | |
1266 | [GOTCL] = markflag(MAC), [GOTCH] = markflag(MAC), | |
1267 | [BPRC] = markflag(MAC), [MPRC] = markflag(MAC), | |
1268 | [TSCTC] = markflag(MAC), [PRC64] = markflag(MAC), | |
1269 | [PRC127] = markflag(MAC), [PRC255] = markflag(MAC), | |
1270 | [PRC511] = markflag(MAC), [PRC1023] = markflag(MAC), | |
1271 | [PRC1522] = markflag(MAC), [PTC64] = markflag(MAC), | |
1272 | [PTC127] = markflag(MAC), [PTC255] = markflag(MAC), | |
1273 | [PTC511] = markflag(MAC), [PTC1023] = markflag(MAC), | |
1274 | [PTC1522] = markflag(MAC), [MPTC] = markflag(MAC), | |
1275 | [BPTC] = markflag(MAC), | |
1276 | ||
1277 | [TDFH] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1278 | [TDFT] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1279 | [TDFHS] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1280 | [TDFTS] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1281 | [TDFPC] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1282 | [RDFH] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1283 | [RDFT] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1284 | [RDFHS] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1285 | [RDFTS] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1286 | [RDFPC] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1287 | [PBM] = markflag(MAC) | MAC_ACCESS_PARTIAL, | |
1288 | }; | |
1289 | ||
1290 | static void | |
1291 | e1000_mmio_write(void *opaque, hwaddr addr, uint64_t val, | |
1292 | unsigned size) | |
1293 | { | |
1294 | E1000State *s = opaque; | |
1295 | unsigned int index = (addr & 0x1ffff) >> 2; | |
1296 | ||
1297 | if (index < NWRITEOPS && macreg_writeops[index]) { | |
1298 | if (!(mac_reg_access[index] & MAC_ACCESS_FLAG_NEEDED) | |
1299 | || (s->compat_flags & (mac_reg_access[index] >> 2))) { | |
1300 | if (mac_reg_access[index] & MAC_ACCESS_PARTIAL) { | |
1301 | DBGOUT(GENERAL, "Writing to register at offset: 0x%08x. " | |
1302 | "It is not fully implemented.\n", index<<2); | |
1303 | } | |
1304 | macreg_writeops[index](s, index, val); | |
1305 | } else { /* "flag needed" bit is set, but the flag is not active */ | |
1306 | DBGOUT(MMIO, "MMIO write attempt to disabled reg. addr=0x%08x\n", | |
1307 | index<<2); | |
1308 | } | |
1309 | } else if (index < NREADOPS && macreg_readops[index]) { | |
1310 | DBGOUT(MMIO, "e1000_mmio_writel RO %x: 0x%04"PRIx64"\n", | |
1311 | index<<2, val); | |
1312 | } else { | |
1313 | DBGOUT(UNKNOWN, "MMIO unknown write addr=0x%08x,val=0x%08"PRIx64"\n", | |
1314 | index<<2, val); | |
1315 | } | |
1316 | } | |
1317 | ||
1318 | static uint64_t | |
1319 | e1000_mmio_read(void *opaque, hwaddr addr, unsigned size) | |
1320 | { | |
1321 | E1000State *s = opaque; | |
1322 | unsigned int index = (addr & 0x1ffff) >> 2; | |
1323 | ||
1324 | if (index < NREADOPS && macreg_readops[index]) { | |
1325 | if (!(mac_reg_access[index] & MAC_ACCESS_FLAG_NEEDED) | |
1326 | || (s->compat_flags & (mac_reg_access[index] >> 2))) { | |
1327 | if (mac_reg_access[index] & MAC_ACCESS_PARTIAL) { | |
1328 | DBGOUT(GENERAL, "Reading register at offset: 0x%08x. " | |
1329 | "It is not fully implemented.\n", index<<2); | |
1330 | } | |
1331 | return macreg_readops[index](s, index); | |
1332 | } else { /* "flag needed" bit is set, but the flag is not active */ | |
1333 | DBGOUT(MMIO, "MMIO read attempt of disabled reg. addr=0x%08x\n", | |
1334 | index<<2); | |
1335 | } | |
1336 | } else { | |
1337 | DBGOUT(UNKNOWN, "MMIO unknown read addr=0x%08x\n", index<<2); | |
1338 | } | |
1339 | return 0; | |
1340 | } | |
1341 | ||
1342 | static const MemoryRegionOps e1000_mmio_ops = { | |
1343 | .read = e1000_mmio_read, | |
1344 | .write = e1000_mmio_write, | |
1345 | .endianness = DEVICE_LITTLE_ENDIAN, | |
1346 | .impl = { | |
1347 | .min_access_size = 4, | |
1348 | .max_access_size = 4, | |
1349 | }, | |
1350 | }; | |
1351 | ||
1352 | static uint64_t e1000_io_read(void *opaque, hwaddr addr, | |
1353 | unsigned size) | |
1354 | { | |
1355 | E1000State *s = opaque; | |
1356 | ||
1357 | (void)s; | |
1358 | return 0; | |
1359 | } | |
1360 | ||
1361 | static void e1000_io_write(void *opaque, hwaddr addr, | |
1362 | uint64_t val, unsigned size) | |
1363 | { | |
1364 | E1000State *s = opaque; | |
1365 | ||
1366 | (void)s; | |
1367 | } | |
1368 | ||
1369 | static const MemoryRegionOps e1000_io_ops = { | |
1370 | .read = e1000_io_read, | |
1371 | .write = e1000_io_write, | |
1372 | .endianness = DEVICE_LITTLE_ENDIAN, | |
1373 | }; | |
1374 | ||
1375 | static bool is_version_1(void *opaque, int version_id) | |
1376 | { | |
1377 | return version_id == 1; | |
1378 | } | |
1379 | ||
1380 | static int e1000_pre_save(void *opaque) | |
1381 | { | |
1382 | E1000State *s = opaque; | |
1383 | NetClientState *nc = qemu_get_queue(s->nic); | |
1384 | ||
1385 | /* | |
1386 | * If link is down and auto-negotiation is supported and ongoing, | |
1387 | * complete auto-negotiation immediately. This allows us to look | |
1388 | * at MII_SR_AUTONEG_COMPLETE to infer link status on load. | |
1389 | */ | |
1390 | if (nc->link_down && have_autoneg(s)) { | |
1391 | s->phy_reg[PHY_STATUS] |= MII_SR_AUTONEG_COMPLETE; | |
1392 | } | |
1393 | ||
1394 | /* Decide which set of props to migrate in the main structure */ | |
1395 | if (chkflag(TSO) || !s->use_tso_for_migration) { | |
1396 | /* Either we're migrating with the extra subsection, in which | |
1397 | * case the mig_props is always 'props' OR | |
1398 | * we've not got the subsection, but 'props' was the last | |
1399 | * updated. | |
1400 | */ | |
1401 | s->mig_props = s->tx.props; | |
1402 | } else { | |
1403 | /* We're not using the subsection, and 'tso_props' was | |
1404 | * the last updated. | |
1405 | */ | |
1406 | s->mig_props = s->tx.tso_props; | |
1407 | } | |
1408 | return 0; | |
1409 | } | |
1410 | ||
1411 | static int e1000_post_load(void *opaque, int version_id) | |
1412 | { | |
1413 | E1000State *s = opaque; | |
1414 | NetClientState *nc = qemu_get_queue(s->nic); | |
1415 | ||
1416 | if (!chkflag(MIT)) { | |
1417 | s->mac_reg[ITR] = s->mac_reg[RDTR] = s->mac_reg[RADV] = | |
1418 | s->mac_reg[TADV] = 0; | |
1419 | s->mit_irq_level = false; | |
1420 | } | |
1421 | s->mit_ide = 0; | |
1422 | s->mit_timer_on = true; | |
1423 | timer_mod(s->mit_timer, qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + 1); | |
1424 | ||
1425 | /* nc.link_down can't be migrated, so infer link_down according | |
1426 | * to link status bit in mac_reg[STATUS]. | |
1427 | * Alternatively, restart link negotiation if it was in progress. */ | |
1428 | nc->link_down = (s->mac_reg[STATUS] & E1000_STATUS_LU) == 0; | |
1429 | ||
1430 | if (have_autoneg(s) && | |
1431 | !(s->phy_reg[PHY_STATUS] & MII_SR_AUTONEG_COMPLETE)) { | |
1432 | nc->link_down = false; | |
1433 | timer_mod(s->autoneg_timer, | |
1434 | qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL) + 500); | |
1435 | } | |
1436 | ||
1437 | s->tx.props = s->mig_props; | |
1438 | if (!s->received_tx_tso) { | |
1439 | /* We received only one set of offload data (tx.props) | |
1440 | * and haven't got tx.tso_props. The best we can do | |
1441 | * is dupe the data. | |
1442 | */ | |
1443 | s->tx.tso_props = s->mig_props; | |
1444 | } | |
1445 | return 0; | |
1446 | } | |
1447 | ||
1448 | static int e1000_tx_tso_post_load(void *opaque, int version_id) | |
1449 | { | |
1450 | E1000State *s = opaque; | |
1451 | s->received_tx_tso = true; | |
1452 | return 0; | |
1453 | } | |
1454 | ||
1455 | static bool e1000_mit_state_needed(void *opaque) | |
1456 | { | |
1457 | E1000State *s = opaque; | |
1458 | ||
1459 | return chkflag(MIT); | |
1460 | } | |
1461 | ||
1462 | static bool e1000_full_mac_needed(void *opaque) | |
1463 | { | |
1464 | E1000State *s = opaque; | |
1465 | ||
1466 | return chkflag(MAC); | |
1467 | } | |
1468 | ||
1469 | static bool e1000_tso_state_needed(void *opaque) | |
1470 | { | |
1471 | E1000State *s = opaque; | |
1472 | ||
1473 | return chkflag(TSO); | |
1474 | } | |
1475 | ||
1476 | static const VMStateDescription vmstate_e1000_mit_state = { | |
1477 | .name = "e1000/mit_state", | |
1478 | .version_id = 1, | |
1479 | .minimum_version_id = 1, | |
1480 | .needed = e1000_mit_state_needed, | |
1481 | .fields = (VMStateField[]) { | |
1482 | VMSTATE_UINT32(mac_reg[RDTR], E1000State), | |
1483 | VMSTATE_UINT32(mac_reg[RADV], E1000State), | |
1484 | VMSTATE_UINT32(mac_reg[TADV], E1000State), | |
1485 | VMSTATE_UINT32(mac_reg[ITR], E1000State), | |
1486 | VMSTATE_BOOL(mit_irq_level, E1000State), | |
1487 | VMSTATE_END_OF_LIST() | |
1488 | } | |
1489 | }; | |
1490 | ||
1491 | static const VMStateDescription vmstate_e1000_full_mac_state = { | |
1492 | .name = "e1000/full_mac_state", | |
1493 | .version_id = 1, | |
1494 | .minimum_version_id = 1, | |
1495 | .needed = e1000_full_mac_needed, | |
1496 | .fields = (VMStateField[]) { | |
1497 | VMSTATE_UINT32_ARRAY(mac_reg, E1000State, 0x8000), | |
1498 | VMSTATE_END_OF_LIST() | |
1499 | } | |
1500 | }; | |
1501 | ||
1502 | static const VMStateDescription vmstate_e1000_tx_tso_state = { | |
1503 | .name = "e1000/tx_tso_state", | |
1504 | .version_id = 1, | |
1505 | .minimum_version_id = 1, | |
1506 | .needed = e1000_tso_state_needed, | |
1507 | .post_load = e1000_tx_tso_post_load, | |
1508 | .fields = (VMStateField[]) { | |
1509 | VMSTATE_UINT8(tx.tso_props.ipcss, E1000State), | |
1510 | VMSTATE_UINT8(tx.tso_props.ipcso, E1000State), | |
1511 | VMSTATE_UINT16(tx.tso_props.ipcse, E1000State), | |
1512 | VMSTATE_UINT8(tx.tso_props.tucss, E1000State), | |
1513 | VMSTATE_UINT8(tx.tso_props.tucso, E1000State), | |
1514 | VMSTATE_UINT16(tx.tso_props.tucse, E1000State), | |
1515 | VMSTATE_UINT32(tx.tso_props.paylen, E1000State), | |
1516 | VMSTATE_UINT8(tx.tso_props.hdr_len, E1000State), | |
1517 | VMSTATE_UINT16(tx.tso_props.mss, E1000State), | |
1518 | VMSTATE_INT8(tx.tso_props.ip, E1000State), | |
1519 | VMSTATE_INT8(tx.tso_props.tcp, E1000State), | |
1520 | VMSTATE_END_OF_LIST() | |
1521 | } | |
1522 | }; | |
1523 | ||
1524 | static const VMStateDescription vmstate_e1000 = { | |
1525 | .name = "e1000", | |
1526 | .version_id = 2, | |
1527 | .minimum_version_id = 1, | |
1528 | .pre_save = e1000_pre_save, | |
1529 | .post_load = e1000_post_load, | |
1530 | .fields = (VMStateField[]) { | |
1531 | VMSTATE_PCI_DEVICE(parent_obj, E1000State), | |
1532 | VMSTATE_UNUSED_TEST(is_version_1, 4), /* was instance id */ | |
1533 | VMSTATE_UNUSED(4), /* Was mmio_base. */ | |
1534 | VMSTATE_UINT32(rxbuf_size, E1000State), | |
1535 | VMSTATE_UINT32(rxbuf_min_shift, E1000State), | |
1536 | VMSTATE_UINT32(eecd_state.val_in, E1000State), | |
1537 | VMSTATE_UINT16(eecd_state.bitnum_in, E1000State), | |
1538 | VMSTATE_UINT16(eecd_state.bitnum_out, E1000State), | |
1539 | VMSTATE_UINT16(eecd_state.reading, E1000State), | |
1540 | VMSTATE_UINT32(eecd_state.old_eecd, E1000State), | |
1541 | VMSTATE_UINT8(mig_props.ipcss, E1000State), | |
1542 | VMSTATE_UINT8(mig_props.ipcso, E1000State), | |
1543 | VMSTATE_UINT16(mig_props.ipcse, E1000State), | |
1544 | VMSTATE_UINT8(mig_props.tucss, E1000State), | |
1545 | VMSTATE_UINT8(mig_props.tucso, E1000State), | |
1546 | VMSTATE_UINT16(mig_props.tucse, E1000State), | |
1547 | VMSTATE_UINT32(mig_props.paylen, E1000State), | |
1548 | VMSTATE_UINT8(mig_props.hdr_len, E1000State), | |
1549 | VMSTATE_UINT16(mig_props.mss, E1000State), | |
1550 | VMSTATE_UINT16(tx.size, E1000State), | |
1551 | VMSTATE_UINT16(tx.tso_frames, E1000State), | |
1552 | VMSTATE_UINT8(tx.sum_needed, E1000State), | |
1553 | VMSTATE_INT8(mig_props.ip, E1000State), | |
1554 | VMSTATE_INT8(mig_props.tcp, E1000State), | |
1555 | VMSTATE_BUFFER(tx.header, E1000State), | |
1556 | VMSTATE_BUFFER(tx.data, E1000State), | |
1557 | VMSTATE_UINT16_ARRAY(eeprom_data, E1000State, 64), | |
1558 | VMSTATE_UINT16_ARRAY(phy_reg, E1000State, 0x20), | |
1559 | VMSTATE_UINT32(mac_reg[CTRL], E1000State), | |
1560 | VMSTATE_UINT32(mac_reg[EECD], E1000State), | |
1561 | VMSTATE_UINT32(mac_reg[EERD], E1000State), | |
1562 | VMSTATE_UINT32(mac_reg[GPRC], E1000State), | |
1563 | VMSTATE_UINT32(mac_reg[GPTC], E1000State), | |
1564 | VMSTATE_UINT32(mac_reg[ICR], E1000State), | |
1565 | VMSTATE_UINT32(mac_reg[ICS], E1000State), | |
1566 | VMSTATE_UINT32(mac_reg[IMC], E1000State), | |
1567 | VMSTATE_UINT32(mac_reg[IMS], E1000State), | |
1568 | VMSTATE_UINT32(mac_reg[LEDCTL], E1000State), | |
1569 | VMSTATE_UINT32(mac_reg[MANC], E1000State), | |
1570 | VMSTATE_UINT32(mac_reg[MDIC], E1000State), | |
1571 | VMSTATE_UINT32(mac_reg[MPC], E1000State), | |
1572 | VMSTATE_UINT32(mac_reg[PBA], E1000State), | |
1573 | VMSTATE_UINT32(mac_reg[RCTL], E1000State), | |
1574 | VMSTATE_UINT32(mac_reg[RDBAH], E1000State), | |
1575 | VMSTATE_UINT32(mac_reg[RDBAL], E1000State), | |
1576 | VMSTATE_UINT32(mac_reg[RDH], E1000State), | |
1577 | VMSTATE_UINT32(mac_reg[RDLEN], E1000State), | |
1578 | VMSTATE_UINT32(mac_reg[RDT], E1000State), | |
1579 | VMSTATE_UINT32(mac_reg[STATUS], E1000State), | |
1580 | VMSTATE_UINT32(mac_reg[SWSM], E1000State), | |
1581 | VMSTATE_UINT32(mac_reg[TCTL], E1000State), | |
1582 | VMSTATE_UINT32(mac_reg[TDBAH], E1000State), | |
1583 | VMSTATE_UINT32(mac_reg[TDBAL], E1000State), | |
1584 | VMSTATE_UINT32(mac_reg[TDH], E1000State), | |
1585 | VMSTATE_UINT32(mac_reg[TDLEN], E1000State), | |
1586 | VMSTATE_UINT32(mac_reg[TDT], E1000State), | |
1587 | VMSTATE_UINT32(mac_reg[TORH], E1000State), | |
1588 | VMSTATE_UINT32(mac_reg[TORL], E1000State), | |
1589 | VMSTATE_UINT32(mac_reg[TOTH], E1000State), | |
1590 | VMSTATE_UINT32(mac_reg[TOTL], E1000State), | |
1591 | VMSTATE_UINT32(mac_reg[TPR], E1000State), | |
1592 | VMSTATE_UINT32(mac_reg[TPT], E1000State), | |
1593 | VMSTATE_UINT32(mac_reg[TXDCTL], E1000State), | |
1594 | VMSTATE_UINT32(mac_reg[WUFC], E1000State), | |
1595 | VMSTATE_UINT32(mac_reg[VET], E1000State), | |
1596 | VMSTATE_UINT32_SUB_ARRAY(mac_reg, E1000State, RA, 32), | |
1597 | VMSTATE_UINT32_SUB_ARRAY(mac_reg, E1000State, MTA, 128), | |
1598 | VMSTATE_UINT32_SUB_ARRAY(mac_reg, E1000State, VFTA, 128), | |
1599 | VMSTATE_END_OF_LIST() | |
1600 | }, | |
1601 | .subsections = (const VMStateDescription*[]) { | |
1602 | &vmstate_e1000_mit_state, | |
1603 | &vmstate_e1000_full_mac_state, | |
1604 | &vmstate_e1000_tx_tso_state, | |
1605 | NULL | |
1606 | } | |
1607 | }; | |
1608 | ||
1609 | /* | |
1610 | * EEPROM contents documented in Tables 5-2 and 5-3, pp. 98-102. | |
1611 | * Note: A valid DevId will be inserted during pci_e1000_init(). | |
1612 | */ | |
1613 | static const uint16_t e1000_eeprom_template[64] = { | |
1614 | 0x0000, 0x0000, 0x0000, 0x0000, 0xffff, 0x0000, 0x0000, 0x0000, | |
1615 | 0x3000, 0x1000, 0x6403, 0 /*DevId*/, 0x8086, 0 /*DevId*/, 0x8086, 0x3040, | |
1616 | 0x0008, 0x2000, 0x7e14, 0x0048, 0x1000, 0x00d8, 0x0000, 0x2700, | |
1617 | 0x6cc9, 0x3150, 0x0722, 0x040b, 0x0984, 0x0000, 0xc000, 0x0706, | |
1618 | 0x1008, 0x0000, 0x0f04, 0x7fff, 0x4d01, 0xffff, 0xffff, 0xffff, | |
1619 | 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, | |
1620 | 0x0100, 0x4000, 0x121c, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, | |
1621 | 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0xffff, 0x0000, | |
1622 | }; | |
1623 | ||
1624 | /* PCI interface */ | |
1625 | ||
1626 | static void | |
1627 | e1000_mmio_setup(E1000State *d) | |
1628 | { | |
1629 | int i; | |
1630 | const uint32_t excluded_regs[] = { | |
1631 | E1000_MDIC, E1000_ICR, E1000_ICS, E1000_IMS, | |
1632 | E1000_IMC, E1000_TCTL, E1000_TDT, PNPMMIO_SIZE | |
1633 | }; | |
1634 | ||
1635 | memory_region_init_io(&d->mmio, OBJECT(d), &e1000_mmio_ops, d, | |
1636 | "e1000-mmio", PNPMMIO_SIZE); | |
1637 | memory_region_add_coalescing(&d->mmio, 0, excluded_regs[0]); | |
1638 | for (i = 0; excluded_regs[i] != PNPMMIO_SIZE; i++) | |
1639 | memory_region_add_coalescing(&d->mmio, excluded_regs[i] + 4, | |
1640 | excluded_regs[i+1] - excluded_regs[i] - 4); | |
1641 | memory_region_init_io(&d->io, OBJECT(d), &e1000_io_ops, d, "e1000-io", IOPORT_SIZE); | |
1642 | } | |
1643 | ||
1644 | static void | |
1645 | pci_e1000_uninit(PCIDevice *dev) | |
1646 | { | |
1647 | E1000State *d = E1000(dev); | |
1648 | ||
1649 | timer_del(d->autoneg_timer); | |
1650 | timer_free(d->autoneg_timer); | |
1651 | timer_del(d->mit_timer); | |
1652 | timer_free(d->mit_timer); | |
1653 | timer_del(d->flush_queue_timer); | |
1654 | timer_free(d->flush_queue_timer); | |
1655 | qemu_del_nic(d->nic); | |
1656 | } | |
1657 | ||
1658 | static NetClientInfo net_e1000_info = { | |
1659 | .type = NET_CLIENT_DRIVER_NIC, | |
1660 | .size = sizeof(NICState), | |
1661 | .can_receive = e1000_can_receive, | |
1662 | .receive = e1000_receive, | |
1663 | .receive_iov = e1000_receive_iov, | |
1664 | .link_status_changed = e1000_set_link_status, | |
1665 | }; | |
1666 | ||
1667 | static void e1000_write_config(PCIDevice *pci_dev, uint32_t address, | |
1668 | uint32_t val, int len) | |
1669 | { | |
1670 | E1000State *s = E1000(pci_dev); | |
1671 | ||
1672 | pci_default_write_config(pci_dev, address, val, len); | |
1673 | ||
1674 | if (range_covers_byte(address, len, PCI_COMMAND) && | |
1675 | (pci_dev->config[PCI_COMMAND] & PCI_COMMAND_MASTER)) { | |
1676 | qemu_flush_queued_packets(qemu_get_queue(s->nic)); | |
1677 | } | |
1678 | } | |
1679 | ||
1680 | static void pci_e1000_realize(PCIDevice *pci_dev, Error **errp) | |
1681 | { | |
1682 | DeviceState *dev = DEVICE(pci_dev); | |
1683 | E1000State *d = E1000(pci_dev); | |
1684 | uint8_t *pci_conf; | |
1685 | uint8_t *macaddr; | |
1686 | ||
1687 | pci_dev->config_write = e1000_write_config; | |
1688 | ||
1689 | pci_conf = pci_dev->config; | |
1690 | ||
1691 | /* TODO: RST# value should be 0, PCI spec 6.2.4 */ | |
1692 | pci_conf[PCI_CACHE_LINE_SIZE] = 0x10; | |
1693 | ||
1694 | pci_conf[PCI_INTERRUPT_PIN] = 1; /* interrupt pin A */ | |
1695 | ||
1696 | e1000_mmio_setup(d); | |
1697 | ||
1698 | pci_register_bar(pci_dev, 0, PCI_BASE_ADDRESS_SPACE_MEMORY, &d->mmio); | |
1699 | ||
1700 | pci_register_bar(pci_dev, 1, PCI_BASE_ADDRESS_SPACE_IO, &d->io); | |
1701 | ||
1702 | qemu_macaddr_default_if_unset(&d->conf.macaddr); | |
1703 | macaddr = d->conf.macaddr.a; | |
1704 | ||
1705 | e1000x_core_prepare_eeprom(d->eeprom_data, | |
1706 | e1000_eeprom_template, | |
1707 | sizeof(e1000_eeprom_template), | |
1708 | PCI_DEVICE_GET_CLASS(pci_dev)->device_id, | |
1709 | macaddr); | |
1710 | ||
1711 | d->nic = qemu_new_nic(&net_e1000_info, &d->conf, | |
1712 | object_get_typename(OBJECT(d)), dev->id, d); | |
1713 | ||
1714 | qemu_format_nic_info_str(qemu_get_queue(d->nic), macaddr); | |
1715 | ||
1716 | d->autoneg_timer = timer_new_ms(QEMU_CLOCK_VIRTUAL, e1000_autoneg_timer, d); | |
1717 | d->mit_timer = timer_new_ns(QEMU_CLOCK_VIRTUAL, e1000_mit_timer, d); | |
1718 | d->flush_queue_timer = timer_new_ms(QEMU_CLOCK_VIRTUAL, | |
1719 | e1000_flush_queue_timer, d); | |
1720 | } | |
1721 | ||
1722 | static void qdev_e1000_reset(DeviceState *dev) | |
1723 | { | |
1724 | E1000State *d = E1000(dev); | |
1725 | e1000_reset(d); | |
1726 | } | |
1727 | ||
1728 | static Property e1000_properties[] = { | |
1729 | DEFINE_NIC_PROPERTIES(E1000State, conf), | |
1730 | DEFINE_PROP_BIT("autonegotiation", E1000State, | |
1731 | compat_flags, E1000_FLAG_AUTONEG_BIT, true), | |
1732 | DEFINE_PROP_BIT("mitigation", E1000State, | |
1733 | compat_flags, E1000_FLAG_MIT_BIT, true), | |
1734 | DEFINE_PROP_BIT("extra_mac_registers", E1000State, | |
1735 | compat_flags, E1000_FLAG_MAC_BIT, true), | |
1736 | DEFINE_PROP_BIT("migrate_tso_props", E1000State, | |
1737 | compat_flags, E1000_FLAG_TSO_BIT, true), | |
1738 | DEFINE_PROP_END_OF_LIST(), | |
1739 | }; | |
1740 | ||
1741 | typedef struct E1000Info { | |
1742 | const char *name; | |
1743 | uint16_t device_id; | |
1744 | uint8_t revision; | |
1745 | uint16_t phy_id2; | |
1746 | } E1000Info; | |
1747 | ||
1748 | static void e1000_class_init(ObjectClass *klass, void *data) | |
1749 | { | |
1750 | DeviceClass *dc = DEVICE_CLASS(klass); | |
1751 | PCIDeviceClass *k = PCI_DEVICE_CLASS(klass); | |
1752 | E1000BaseClass *e = E1000_DEVICE_CLASS(klass); | |
1753 | const E1000Info *info = data; | |
1754 | ||
1755 | k->realize = pci_e1000_realize; | |
1756 | k->exit = pci_e1000_uninit; | |
1757 | k->romfile = "efi-e1000.rom"; | |
1758 | k->vendor_id = PCI_VENDOR_ID_INTEL; | |
1759 | k->device_id = info->device_id; | |
1760 | k->revision = info->revision; | |
1761 | e->phy_id2 = info->phy_id2; | |
1762 | k->class_id = PCI_CLASS_NETWORK_ETHERNET; | |
1763 | set_bit(DEVICE_CATEGORY_NETWORK, dc->categories); | |
1764 | dc->desc = "Intel Gigabit Ethernet"; | |
1765 | dc->reset = qdev_e1000_reset; | |
1766 | dc->vmsd = &vmstate_e1000; | |
1767 | dc->props = e1000_properties; | |
1768 | } | |
1769 | ||
1770 | static void e1000_instance_init(Object *obj) | |
1771 | { | |
1772 | E1000State *n = E1000(obj); | |
1773 | device_add_bootindex_property(obj, &n->conf.bootindex, | |
1774 | "bootindex", "/ethernet-phy@0", | |
1775 | DEVICE(n), NULL); | |
1776 | } | |
1777 | ||
1778 | static const TypeInfo e1000_base_info = { | |
1779 | .name = TYPE_E1000_BASE, | |
1780 | .parent = TYPE_PCI_DEVICE, | |
1781 | .instance_size = sizeof(E1000State), | |
1782 | .instance_init = e1000_instance_init, | |
1783 | .class_size = sizeof(E1000BaseClass), | |
1784 | .abstract = true, | |
1785 | .interfaces = (InterfaceInfo[]) { | |
1786 | { INTERFACE_CONVENTIONAL_PCI_DEVICE }, | |
1787 | { }, | |
1788 | }, | |
1789 | }; | |
1790 | ||
1791 | static const E1000Info e1000_devices[] = { | |
1792 | { | |
1793 | .name = "e1000", | |
1794 | .device_id = E1000_DEV_ID_82540EM, | |
1795 | .revision = 0x03, | |
1796 | .phy_id2 = E1000_PHY_ID2_8254xx_DEFAULT, | |
1797 | }, | |
1798 | { | |
1799 | .name = "e1000-82544gc", | |
1800 | .device_id = E1000_DEV_ID_82544GC_COPPER, | |
1801 | .revision = 0x03, | |
1802 | .phy_id2 = E1000_PHY_ID2_82544x, | |
1803 | }, | |
1804 | { | |
1805 | .name = "e1000-82545em", | |
1806 | .device_id = E1000_DEV_ID_82545EM_COPPER, | |
1807 | .revision = 0x03, | |
1808 | .phy_id2 = E1000_PHY_ID2_8254xx_DEFAULT, | |
1809 | }, | |
1810 | }; | |
1811 | ||
1812 | static void e1000_register_types(void) | |
1813 | { | |
1814 | int i; | |
1815 | ||
1816 | type_register_static(&e1000_base_info); | |
1817 | for (i = 0; i < ARRAY_SIZE(e1000_devices); i++) { | |
1818 | const E1000Info *info = &e1000_devices[i]; | |
1819 | TypeInfo type_info = {}; | |
1820 | ||
1821 | type_info.name = info->name; | |
1822 | type_info.parent = TYPE_E1000_BASE; | |
1823 | type_info.class_data = (void *)info; | |
1824 | type_info.class_init = e1000_class_init; | |
1825 | type_info.instance_init = e1000_instance_init; | |
1826 | ||
1827 | type_register(&type_info); | |
1828 | } | |
1829 | } | |
1830 | ||
1831 | type_init(e1000_register_types) |