1/*
2 * QEMU Sun Happy Meal Ethernet emulation
3 *
4 * Copyright (c) 2017 Mark Cave-Ayland
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
24
25#include "qemu/osdep.h"
26#include "hw/pci/pci.h"
27#include "hw/qdev-properties.h"
28#include "migration/vmstate.h"
29#include "hw/net/mii.h"
30#include "net/net.h"
31#include "qemu/module.h"
32#include "net/checksum.h"
33#include "net/eth.h"
34#include "sysemu/sysemu.h"
35#include "trace.h"
36
37#define HME_REG_SIZE 0x8000
38
39#define HME_SEB_REG_SIZE 0x2000
40
41#define HME_SEBI_RESET 0x0
42#define HME_SEB_RESET_ETX 0x1
43#define HME_SEB_RESET_ERX 0x2
44
45#define HME_SEBI_STAT 0x100
46#define HME_SEBI_STAT_LINUXBUG 0x108
47#define HME_SEB_STAT_RXTOHOST 0x10000
48#define HME_SEB_STAT_NORXD 0x20000
49#define HME_SEB_STAT_MIFIRQ 0x800000
50#define HME_SEB_STAT_HOSTTOTX 0x1000000
51#define HME_SEB_STAT_TXALL 0x2000000
52
53#define HME_SEBI_IMASK 0x104
54#define HME_SEBI_IMASK_LINUXBUG 0x10c
55
56#define HME_ETX_REG_SIZE 0x2000
57
58#define HME_ETXI_PENDING 0x0
59
60#define HME_ETXI_RING 0x8
61#define HME_ETXI_RING_ADDR 0xffffff00
62#define HME_ETXI_RING_OFFSET 0xff
63
64#define HME_ETXI_RSIZE 0x2c
65
66#define HME_ERX_REG_SIZE 0x2000
67
68#define HME_ERXI_CFG 0x0
69#define HME_ERX_CFG_RINGSIZE 0x600
70#define HME_ERX_CFG_RINGSIZE_SHIFT 9
71#define HME_ERX_CFG_BYTEOFFSET 0x38
72#define HME_ERX_CFG_BYTEOFFSET_SHIFT 3
73#define HME_ERX_CFG_CSUMSTART 0x7f0000
74#define HME_ERX_CFG_CSUMSHIFT 16
75
76#define HME_ERXI_RING 0x4
77#define HME_ERXI_RING_ADDR 0xffffff00
78#define HME_ERXI_RING_OFFSET 0xff
79
80#define HME_MAC_REG_SIZE 0x1000
81
82#define HME_MACI_TXCFG 0x20c
83#define HME_MAC_TXCFG_ENABLE 0x1
84
85#define HME_MACI_RXCFG 0x30c
86#define HME_MAC_RXCFG_ENABLE 0x1
87#define HME_MAC_RXCFG_PMISC 0x40
88#define HME_MAC_RXCFG_HENABLE 0x800
89
90#define HME_MACI_MACADDR2 0x318
91#define HME_MACI_MACADDR1 0x31c
92#define HME_MACI_MACADDR0 0x320
93
94#define HME_MACI_HASHTAB3 0x340
95#define HME_MACI_HASHTAB2 0x344
96#define HME_MACI_HASHTAB1 0x348
97#define HME_MACI_HASHTAB0 0x34c
98
99#define HME_MIF_REG_SIZE 0x20
100
101#define HME_MIFI_FO 0xc
102#define HME_MIF_FO_ST 0xc0000000
103#define HME_MIF_FO_ST_SHIFT 30
104#define HME_MIF_FO_OPC 0x30000000
105#define HME_MIF_FO_OPC_SHIFT 28
106#define HME_MIF_FO_PHYAD 0x0f800000
107#define HME_MIF_FO_PHYAD_SHIFT 23
108#define HME_MIF_FO_REGAD 0x007c0000
109#define HME_MIF_FO_REGAD_SHIFT 18
110#define HME_MIF_FO_TAMSB 0x20000
111#define HME_MIF_FO_TALSB 0x10000
112#define HME_MIF_FO_DATA 0xffff
113
114#define HME_MIFI_CFG 0x10
115#define HME_MIF_CFG_MDI0 0x100
116#define HME_MIF_CFG_MDI1 0x200
117
118#define HME_MIFI_IMASK 0x14
119
120#define HME_MIFI_STAT 0x18
121
122
123/* Wired HME PHY addresses */
124#define HME_PHYAD_INTERNAL 1
125#define HME_PHYAD_EXTERNAL 0
126
127#define MII_COMMAND_START 0x1
128#define MII_COMMAND_READ 0x2
129#define MII_COMMAND_WRITE 0x1
130
131#define TYPE_SUNHME "sunhme"
132#define SUNHME(obj) OBJECT_CHECK(SunHMEState, (obj), TYPE_SUNHME)
133
134/* Maximum size of buffer */
135#define HME_FIFO_SIZE 0x800
136
137/* Size of TX/RX descriptor */
138#define HME_DESC_SIZE 0x8
139
140#define HME_XD_OWN 0x80000000
141#define HME_XD_OFL 0x40000000
142#define HME_XD_SOP 0x40000000
143#define HME_XD_EOP 0x20000000
144#define HME_XD_RXLENMSK 0x3fff0000
145#define HME_XD_RXLENSHIFT 16
146#define HME_XD_RXCKSUM 0xffff
147#define HME_XD_TXLENMSK 0x00001fff
148#define HME_XD_TXCKSUM 0x10000000
149#define HME_XD_TXCSSTUFF 0xff00000
150#define HME_XD_TXCSSTUFFSHIFT 20
151#define HME_XD_TXCSSTART 0xfc000
152#define HME_XD_TXCSSTARTSHIFT 14
153
154#define HME_MII_REGS_SIZE 0x20
155
156typedef struct SunHMEState {
157 /*< private >*/
158 PCIDevice parent_obj;
159
160 NICState *nic;
161 NICConf conf;
162
163 MemoryRegion hme;
164 MemoryRegion sebreg;
165 MemoryRegion etxreg;
166 MemoryRegion erxreg;
167 MemoryRegion macreg;
168 MemoryRegion mifreg;
169
170 uint32_t sebregs[HME_SEB_REG_SIZE >> 2];
171 uint32_t etxregs[HME_ETX_REG_SIZE >> 2];
172 uint32_t erxregs[HME_ERX_REG_SIZE >> 2];
173 uint32_t macregs[HME_MAC_REG_SIZE >> 2];
174 uint32_t mifregs[HME_MIF_REG_SIZE >> 2];
175
176 uint16_t miiregs[HME_MII_REGS_SIZE];
177} SunHMEState;
178
179static Property sunhme_properties[] = {
180 DEFINE_NIC_PROPERTIES(SunHMEState, conf),
181 DEFINE_PROP_END_OF_LIST(),
182};
183
184static void sunhme_reset_tx(SunHMEState *s)
185{
186 /* Indicate TX reset complete */
187 s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ETX;
188}
189
190static void sunhme_reset_rx(SunHMEState *s)
191{
192 /* Indicate RX reset complete */
193 s->sebregs[HME_SEBI_RESET] &= ~HME_SEB_RESET_ERX;
194}
195
196static void sunhme_update_irq(SunHMEState *s)
197{
198 PCIDevice *d = PCI_DEVICE(s);
199 int level;
200
201 /* MIF interrupt mask (16-bit) */
202 uint32_t mifmask = ~(s->mifregs[HME_MIFI_IMASK >> 2]) & 0xffff;
203 uint32_t mif = s->mifregs[HME_MIFI_STAT >> 2] & mifmask;
204
205 /* Main SEB interrupt mask (include MIF status from above) */
206 uint32_t sebmask = ~(s->sebregs[HME_SEBI_IMASK >> 2]) &
207 ~HME_SEB_STAT_MIFIRQ;
208 uint32_t seb = s->sebregs[HME_SEBI_STAT >> 2] & sebmask;
209 if (mif) {
210 seb |= HME_SEB_STAT_MIFIRQ;
211 }
212
213 level = (seb ? 1 : 0);
214 trace_sunhme_update_irq(mifmask, mif, sebmask, seb, level);
215
216 pci_set_irq(d, level);
217}
218
219static void sunhme_seb_write(void *opaque, hwaddr addr,
220 uint64_t val, unsigned size)
221{
222 SunHMEState *s = SUNHME(opaque);
223
224 trace_sunhme_seb_write(addr, val);
225
226 /* Handly buggy Linux drivers before 4.13 which have
227 the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */
228 switch (addr) {
229 case HME_SEBI_STAT_LINUXBUG:
230 addr = HME_SEBI_STAT;
231 break;
232 case HME_SEBI_IMASK_LINUXBUG:
233 addr = HME_SEBI_IMASK;
234 break;
235 default:
236 break;
237 }
238
239 switch (addr) {
240 case HME_SEBI_RESET:
241 if (val & HME_SEB_RESET_ETX) {
242 sunhme_reset_tx(s);
243 }
244 if (val & HME_SEB_RESET_ERX) {
245 sunhme_reset_rx(s);
246 }
247 val = s->sebregs[HME_SEBI_RESET >> 2];
248 break;
249 }
250
251 s->sebregs[addr >> 2] = val;
252}
253
254static uint64_t sunhme_seb_read(void *opaque, hwaddr addr,
255 unsigned size)
256{
257 SunHMEState *s = SUNHME(opaque);
258 uint64_t val;
259
260 /* Handly buggy Linux drivers before 4.13 which have
261 the wrong offsets for HME_SEBI_STAT and HME_SEBI_IMASK */
262 switch (addr) {
263 case HME_SEBI_STAT_LINUXBUG:
264 addr = HME_SEBI_STAT;
265 break;
266 case HME_SEBI_IMASK_LINUXBUG:
267 addr = HME_SEBI_IMASK;
268 break;
269 default:
270 break;
271 }
272
273 val = s->sebregs[addr >> 2];
274
275 switch (addr) {
276 case HME_SEBI_STAT:
277 /* Autoclear status (except MIF) */
278 s->sebregs[HME_SEBI_STAT >> 2] &= HME_SEB_STAT_MIFIRQ;
279 sunhme_update_irq(s);
280 break;
281 }
282
283 trace_sunhme_seb_read(addr, val);
284
285 return val;
286}
287
288static const MemoryRegionOps sunhme_seb_ops = {
289 .read = sunhme_seb_read,
290 .write = sunhme_seb_write,
291 .endianness = DEVICE_LITTLE_ENDIAN,
292 .valid = {
293 .min_access_size = 4,
294 .max_access_size = 4,
295 },
296};
297
298static void sunhme_transmit(SunHMEState *s);
299
300static void sunhme_etx_write(void *opaque, hwaddr addr,
301 uint64_t val, unsigned size)
302{
303 SunHMEState *s = SUNHME(opaque);
304
305 trace_sunhme_etx_write(addr, val);
306
307 switch (addr) {
308 case HME_ETXI_PENDING:
309 if (val) {
310 sunhme_transmit(s);
311 }
312 break;
313 }
314
315 s->etxregs[addr >> 2] = val;
316}
317
318static uint64_t sunhme_etx_read(void *opaque, hwaddr addr,
319 unsigned size)
320{
321 SunHMEState *s = SUNHME(opaque);
322 uint64_t val;
323
324 val = s->etxregs[addr >> 2];
325
326 trace_sunhme_etx_read(addr, val);
327
328 return val;
329}
330
331static const MemoryRegionOps sunhme_etx_ops = {
332 .read = sunhme_etx_read,
333 .write = sunhme_etx_write,
334 .endianness = DEVICE_LITTLE_ENDIAN,
335 .valid = {
336 .min_access_size = 4,
337 .max_access_size = 4,
338 },
339};
340
341static void sunhme_erx_write(void *opaque, hwaddr addr,
342 uint64_t val, unsigned size)
343{
344 SunHMEState *s = SUNHME(opaque);
345
346 trace_sunhme_erx_write(addr, val);
347
348 s->erxregs[addr >> 2] = val;
349}
350
351static uint64_t sunhme_erx_read(void *opaque, hwaddr addr,
352 unsigned size)
353{
354 SunHMEState *s = SUNHME(opaque);
355 uint64_t val;
356
357 val = s->erxregs[addr >> 2];
358
359 trace_sunhme_erx_read(addr, val);
360
361 return val;
362}
363
364static const MemoryRegionOps sunhme_erx_ops = {
365 .read = sunhme_erx_read,
366 .write = sunhme_erx_write,
367 .endianness = DEVICE_LITTLE_ENDIAN,
368 .valid = {
369 .min_access_size = 4,
370 .max_access_size = 4,
371 },
372};
373
374static void sunhme_mac_write(void *opaque, hwaddr addr,
375 uint64_t val, unsigned size)
376{
377 SunHMEState *s = SUNHME(opaque);
378 uint64_t oldval = s->macregs[addr >> 2];
379
380 trace_sunhme_mac_write(addr, val);
381
382 s->macregs[addr >> 2] = val;
383
384 switch (addr) {
385 case HME_MACI_RXCFG:
386 if (!(oldval & HME_MAC_RXCFG_ENABLE) &&
387 (val & HME_MAC_RXCFG_ENABLE)) {
388 qemu_flush_queued_packets(qemu_get_queue(s->nic));
389 }
390 break;
391 }
392}
393
394static uint64_t sunhme_mac_read(void *opaque, hwaddr addr,
395 unsigned size)
396{
397 SunHMEState *s = SUNHME(opaque);
398 uint64_t val;
399
400 val = s->macregs[addr >> 2];
401
402 trace_sunhme_mac_read(addr, val);
403
404 return val;
405}
406
407static const MemoryRegionOps sunhme_mac_ops = {
408 .read = sunhme_mac_read,
409 .write = sunhme_mac_write,
410 .endianness = DEVICE_LITTLE_ENDIAN,
411 .valid = {
412 .min_access_size = 4,
413 .max_access_size = 4,
414 },
415};
416
417static void sunhme_mii_write(SunHMEState *s, uint8_t reg, uint16_t data)
418{
419 trace_sunhme_mii_write(reg, data);
420
421 switch (reg) {
422 case MII_BMCR:
423 if (data & MII_BMCR_RESET) {
424 /* Autoclear reset bit, enable auto negotiation */
425 data &= ~MII_BMCR_RESET;
426 data |= MII_BMCR_AUTOEN;
427 }
428 if (data & MII_BMCR_ANRESTART) {
429 /* Autoclear auto negotiation restart */
430 data &= ~MII_BMCR_ANRESTART;
431
432 /* Indicate negotiation complete */
433 s->miiregs[MII_BMSR] |= MII_BMSR_AN_COMP;
434
435 if (!qemu_get_queue(s->nic)->link_down) {
436 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD;
437 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST;
438 }
439 }
440 break;
441 }
442
443 s->miiregs[reg] = data;
444}
445
446static uint16_t sunhme_mii_read(SunHMEState *s, uint8_t reg)
447{
448 uint16_t data = s->miiregs[reg];
449
450 trace_sunhme_mii_read(reg, data);
451
452 return data;
453}
454
455static void sunhme_mif_write(void *opaque, hwaddr addr,
456 uint64_t val, unsigned size)
457{
458 SunHMEState *s = SUNHME(opaque);
459 uint8_t cmd, reg;
460 uint16_t data;
461
462 trace_sunhme_mif_write(addr, val);
463
464 switch (addr) {
465 case HME_MIFI_CFG:
466 /* Mask the read-only bits */
467 val &= ~(HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1);
468 val |= s->mifregs[HME_MIFI_CFG >> 2] &
469 (HME_MIF_CFG_MDI0 | HME_MIF_CFG_MDI1);
470 break;
471 case HME_MIFI_FO:
472 /* Detect start of MII command */
473 if ((val & HME_MIF_FO_ST) >> HME_MIF_FO_ST_SHIFT
474 != MII_COMMAND_START) {
475 val |= HME_MIF_FO_TALSB;
476 break;
477 }
478
479 /* Internal phy only */
480 if ((val & HME_MIF_FO_PHYAD) >> HME_MIF_FO_PHYAD_SHIFT
481 != HME_PHYAD_INTERNAL) {
482 val |= HME_MIF_FO_TALSB;
483 break;
484 }
485
486 cmd = (val & HME_MIF_FO_OPC) >> HME_MIF_FO_OPC_SHIFT;
487 reg = (val & HME_MIF_FO_REGAD) >> HME_MIF_FO_REGAD_SHIFT;
488 data = (val & HME_MIF_FO_DATA);
489
490 switch (cmd) {
491 case MII_COMMAND_WRITE:
492 sunhme_mii_write(s, reg, data);
493 break;
494
495 case MII_COMMAND_READ:
496 val &= ~HME_MIF_FO_DATA;
497 val |= sunhme_mii_read(s, reg);
498 break;
499 }
500
501 val |= HME_MIF_FO_TALSB;
502 break;
503 }
504
505 s->mifregs[addr >> 2] = val;
506}
507
508static uint64_t sunhme_mif_read(void *opaque, hwaddr addr,
509 unsigned size)
510{
511 SunHMEState *s = SUNHME(opaque);
512 uint64_t val;
513
514 val = s->mifregs[addr >> 2];
515
516 switch (addr) {
517 case HME_MIFI_STAT:
518 /* Autoclear MIF interrupt status */
519 s->mifregs[HME_MIFI_STAT >> 2] = 0;
520 sunhme_update_irq(s);
521 break;
522 }
523
524 trace_sunhme_mif_read(addr, val);
525
526 return val;
527}
528
529static const MemoryRegionOps sunhme_mif_ops = {
530 .read = sunhme_mif_read,
531 .write = sunhme_mif_write,
532 .endianness = DEVICE_LITTLE_ENDIAN,
533 .valid = {
534 .min_access_size = 4,
535 .max_access_size = 4,
536 },
537};
538
539static void sunhme_transmit_frame(SunHMEState *s, uint8_t *buf, int size)
540{
541 qemu_send_packet(qemu_get_queue(s->nic), buf, size);
542}
543
544static inline int sunhme_get_tx_ring_count(SunHMEState *s)
545{
546 return (s->etxregs[HME_ETXI_RSIZE >> 2] + 1) << 4;
547}
548
549static inline int sunhme_get_tx_ring_nr(SunHMEState *s)
550{
551 return s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_OFFSET;
552}
553
554static inline void sunhme_set_tx_ring_nr(SunHMEState *s, int i)
555{
556 uint32_t ring = s->etxregs[HME_ETXI_RING >> 2] & ~HME_ETXI_RING_OFFSET;
557 ring |= i & HME_ETXI_RING_OFFSET;
558
559 s->etxregs[HME_ETXI_RING >> 2] = ring;
560}
561
562static void sunhme_transmit(SunHMEState *s)
563{
564 PCIDevice *d = PCI_DEVICE(s);
565 dma_addr_t tb, addr;
566 uint32_t intstatus, status, buffer, sum = 0;
567 int cr, nr, len, xmit_pos, csum_offset = 0, csum_stuff_offset = 0;
568 uint16_t csum = 0;
569 uint8_t xmit_buffer[HME_FIFO_SIZE];
570
571 tb = s->etxregs[HME_ETXI_RING >> 2] & HME_ETXI_RING_ADDR;
572 nr = sunhme_get_tx_ring_count(s);
573 cr = sunhme_get_tx_ring_nr(s);
574
575 pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4);
576 pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4);
577
578 xmit_pos = 0;
579 while (status & HME_XD_OWN) {
580 trace_sunhme_tx_desc(buffer, status, cr, nr);
581
582 /* Copy data into transmit buffer */
583 addr = buffer;
584 len = status & HME_XD_TXLENMSK;
585
586 if (xmit_pos + len > HME_FIFO_SIZE) {
587 len = HME_FIFO_SIZE - xmit_pos;
588 }
589
590 pci_dma_read(d, addr, &xmit_buffer[xmit_pos], len);
591 xmit_pos += len;
592
593 /* Detect start of packet for TX checksum */
594 if (status & HME_XD_SOP) {
595 sum = 0;
596 csum_offset = (status & HME_XD_TXCSSTART) >> HME_XD_TXCSSTARTSHIFT;
597 csum_stuff_offset = (status & HME_XD_TXCSSTUFF) >>
598 HME_XD_TXCSSTUFFSHIFT;
599 }
600
601 if (status & HME_XD_TXCKSUM) {
602 /* Only start calculation from csum_offset */
603 if (xmit_pos - len <= csum_offset && xmit_pos > csum_offset) {
604 sum += net_checksum_add(xmit_pos - csum_offset,
605 xmit_buffer + csum_offset);
606 trace_sunhme_tx_xsum_add(csum_offset, xmit_pos - csum_offset);
607 } else {
608 sum += net_checksum_add(len, xmit_buffer + xmit_pos - len);
609 trace_sunhme_tx_xsum_add(xmit_pos - len, len);
610 }
611 }
612
613 /* Detect end of packet for TX checksum */
614 if (status & HME_XD_EOP) {
615 /* Stuff the checksum if required */
616 if (status & HME_XD_TXCKSUM) {
617 csum = net_checksum_finish(sum);
618 stw_be_p(xmit_buffer + csum_stuff_offset, csum);
619 trace_sunhme_tx_xsum_stuff(csum, csum_stuff_offset);
620 }
621
622 if (s->macregs[HME_MACI_TXCFG >> 2] & HME_MAC_TXCFG_ENABLE) {
623 sunhme_transmit_frame(s, xmit_buffer, xmit_pos);
624 trace_sunhme_tx_done(xmit_pos);
625 }
626 }
627
628 /* Update status */
629 status &= ~HME_XD_OWN;
630 pci_dma_write(d, tb + cr * HME_DESC_SIZE, &status, 4);
631
632 /* Move onto next descriptor */
633 cr++;
634 if (cr >= nr) {
635 cr = 0;
636 }
637 sunhme_set_tx_ring_nr(s, cr);
638
639 pci_dma_read(d, tb + cr * HME_DESC_SIZE, &status, 4);
640 pci_dma_read(d, tb + cr * HME_DESC_SIZE + 4, &buffer, 4);
641
642 /* Indicate TX complete */
643 intstatus = s->sebregs[HME_SEBI_STAT >> 2];
644 intstatus |= HME_SEB_STAT_HOSTTOTX;
645 s->sebregs[HME_SEBI_STAT >> 2] = intstatus;
646
647 /* Autoclear TX pending */
648 s->etxregs[HME_ETXI_PENDING >> 2] = 0;
649
650 sunhme_update_irq(s);
651 }
652
653 /* TX FIFO now clear */
654 intstatus = s->sebregs[HME_SEBI_STAT >> 2];
655 intstatus |= HME_SEB_STAT_TXALL;
656 s->sebregs[HME_SEBI_STAT >> 2] = intstatus;
657 sunhme_update_irq(s);
658}
659
660static int sunhme_can_receive(NetClientState *nc)
661{
662 SunHMEState *s = qemu_get_nic_opaque(nc);
663
664 return s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE;
665}
666
667static void sunhme_link_status_changed(NetClientState *nc)
668{
669 SunHMEState *s = qemu_get_nic_opaque(nc);
670
671 if (nc->link_down) {
672 s->miiregs[MII_ANLPAR] &= ~MII_ANLPAR_TXFD;
673 s->miiregs[MII_BMSR] &= ~MII_BMSR_LINK_ST;
674 } else {
675 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD;
676 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST;
677 }
678
679 /* Exact bits unknown */
680 s->mifregs[HME_MIFI_STAT >> 2] = 0xffff;
681 sunhme_update_irq(s);
682}
683
684static inline int sunhme_get_rx_ring_count(SunHMEState *s)
685{
686 uint32_t rings = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_RINGSIZE)
687 >> HME_ERX_CFG_RINGSIZE_SHIFT;
688
689 switch (rings) {
690 case 0:
691 return 32;
692 case 1:
693 return 64;
694 case 2:
695 return 128;
696 case 3:
697 return 256;
698 }
699
700 return 0;
701}
702
703static inline int sunhme_get_rx_ring_nr(SunHMEState *s)
704{
705 return s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_OFFSET;
706}
707
708static inline void sunhme_set_rx_ring_nr(SunHMEState *s, int i)
709{
710 uint32_t ring = s->erxregs[HME_ERXI_RING >> 2] & ~HME_ERXI_RING_OFFSET;
711 ring |= i & HME_ERXI_RING_OFFSET;
712
713 s->erxregs[HME_ERXI_RING >> 2] = ring;
714}
715
716#define MIN_BUF_SIZE 60
717
718static ssize_t sunhme_receive(NetClientState *nc, const uint8_t *buf,
719 size_t size)
720{
721 SunHMEState *s = qemu_get_nic_opaque(nc);
722 PCIDevice *d = PCI_DEVICE(s);
723 dma_addr_t rb, addr;
724 uint32_t intstatus, status, buffer, buffersize, sum;
725 uint16_t csum;
726 uint8_t buf1[60];
727 int nr, cr, len, rxoffset, csum_offset;
728
729 trace_sunhme_rx_incoming(size);
730
731 /* Do nothing if MAC RX disabled */
732 if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_ENABLE)) {
733 return 0;
734 }
735
736 trace_sunhme_rx_filter_destmac(buf[0], buf[1], buf[2],
737 buf[3], buf[4], buf[5]);
738
739 /* Check destination MAC address */
740 if (!(s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_PMISC)) {
741 /* Try and match local MAC address */
742 if (((s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff00) >> 8) == buf[0] &&
743 (s->macregs[HME_MACI_MACADDR0 >> 2] & 0xff) == buf[1] &&
744 ((s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff00) >> 8) == buf[2] &&
745 (s->macregs[HME_MACI_MACADDR1 >> 2] & 0xff) == buf[3] &&
746 ((s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff00) >> 8) == buf[4] &&
747 (s->macregs[HME_MACI_MACADDR2 >> 2] & 0xff) == buf[5]) {
748 /* Matched local MAC address */
749 trace_sunhme_rx_filter_local_match();
750 } else if (buf[0] == 0xff && buf[1] == 0xff && buf[2] == 0xff &&
751 buf[3] == 0xff && buf[4] == 0xff && buf[5] == 0xff) {
752 /* Matched broadcast address */
753 trace_sunhme_rx_filter_bcast_match();
754 } else if (s->macregs[HME_MACI_RXCFG >> 2] & HME_MAC_RXCFG_HENABLE) {
755 /* Didn't match local address, check hash filter */
756 int mcast_idx = net_crc32_le(buf, ETH_ALEN) >> 26;
757 if (!(s->macregs[(HME_MACI_HASHTAB0 >> 2) - (mcast_idx >> 4)] &
758 (1 << (mcast_idx & 0xf)))) {
759 /* Didn't match hash filter */
760 trace_sunhme_rx_filter_hash_nomatch();
761 trace_sunhme_rx_filter_reject();
762 return -1;
763 } else {
764 trace_sunhme_rx_filter_hash_match();
765 }
766 } else {
767 /* Not for us */
768 trace_sunhme_rx_filter_reject();
769 return -1;
770 }
771 } else {
772 trace_sunhme_rx_filter_promisc_match();
773 }
774
775 trace_sunhme_rx_filter_accept();
776
777 /* If too small buffer, then expand it */
778 if (size < MIN_BUF_SIZE) {
779 memcpy(buf1, buf, size);
780 memset(buf1 + size, 0, MIN_BUF_SIZE - size);
781 buf = buf1;
782 size = MIN_BUF_SIZE;
783 }
784
785 rb = s->erxregs[HME_ERXI_RING >> 2] & HME_ERXI_RING_ADDR;
786 nr = sunhme_get_rx_ring_count(s);
787 cr = sunhme_get_rx_ring_nr(s);
788
789 pci_dma_read(d, rb + cr * HME_DESC_SIZE, &status, 4);
790 pci_dma_read(d, rb + cr * HME_DESC_SIZE + 4, &buffer, 4);
791
792 /* If we don't own the current descriptor then indicate overflow error */
793 if (!(status & HME_XD_OWN)) {
794 s->sebregs[HME_SEBI_STAT >> 2] |= HME_SEB_STAT_NORXD;
795 sunhme_update_irq(s);
796 trace_sunhme_rx_norxd();
797 return -1;
798 }
799
800 rxoffset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_BYTEOFFSET) >>
801 HME_ERX_CFG_BYTEOFFSET_SHIFT;
802
803 addr = buffer + rxoffset;
804 buffersize = (status & HME_XD_RXLENMSK) >> HME_XD_RXLENSHIFT;
805
806 /* Detect receive overflow */
807 len = size;
808 if (size > buffersize) {
809 status |= HME_XD_OFL;
810 len = buffersize;
811 }
812
813 pci_dma_write(d, addr, buf, len);
814
815 trace_sunhme_rx_desc(buffer, rxoffset, status, len, cr, nr);
816
817 /* Calculate the receive checksum */
818 csum_offset = (s->erxregs[HME_ERXI_CFG >> 2] & HME_ERX_CFG_CSUMSTART) >>
819 HME_ERX_CFG_CSUMSHIFT << 1;
820 sum = 0;
821 sum += net_checksum_add(len - csum_offset, (uint8_t *)buf + csum_offset);
822 csum = net_checksum_finish(sum);
823
824 trace_sunhme_rx_xsum_calc(csum);
825
826 /* Update status */
827 status &= ~HME_XD_OWN;
828 status &= ~HME_XD_RXLENMSK;
829 status |= len << HME_XD_RXLENSHIFT;
830 status &= ~HME_XD_RXCKSUM;
831 status |= csum;
832
833 pci_dma_write(d, rb + cr * HME_DESC_SIZE, &status, 4);
834
835 cr++;
836 if (cr >= nr) {
837 cr = 0;
838 }
839
840 sunhme_set_rx_ring_nr(s, cr);
841
842 /* Indicate RX complete */
843 intstatus = s->sebregs[HME_SEBI_STAT >> 2];
844 intstatus |= HME_SEB_STAT_RXTOHOST;
845 s->sebregs[HME_SEBI_STAT >> 2] = intstatus;
846
847 sunhme_update_irq(s);
848
849 return len;
850}
851
852static NetClientInfo net_sunhme_info = {
853 .type = NET_CLIENT_DRIVER_NIC,
854 .size = sizeof(NICState),
855 .can_receive = sunhme_can_receive,
856 .receive = sunhme_receive,
857 .link_status_changed = sunhme_link_status_changed,
858};
859
860static void sunhme_realize(PCIDevice *pci_dev, Error **errp)
861{
862 SunHMEState *s = SUNHME(pci_dev);
863 DeviceState *d = DEVICE(pci_dev);
864 uint8_t *pci_conf;
865
866 pci_conf = pci_dev->config;
867 pci_conf[PCI_INTERRUPT_PIN] = 1; /* interrupt pin A */
868
869 memory_region_init(&s->hme, OBJECT(pci_dev), "sunhme", HME_REG_SIZE);
870 pci_register_bar(pci_dev, 0, PCI_BASE_ADDRESS_SPACE_MEMORY, &s->hme);
871
872 memory_region_init_io(&s->sebreg, OBJECT(pci_dev), &sunhme_seb_ops, s,
873 "sunhme.seb", HME_SEB_REG_SIZE);
874 memory_region_add_subregion(&s->hme, 0, &s->sebreg);
875
876 memory_region_init_io(&s->etxreg, OBJECT(pci_dev), &sunhme_etx_ops, s,
877 "sunhme.etx", HME_ETX_REG_SIZE);
878 memory_region_add_subregion(&s->hme, 0x2000, &s->etxreg);
879
880 memory_region_init_io(&s->erxreg, OBJECT(pci_dev), &sunhme_erx_ops, s,
881 "sunhme.erx", HME_ERX_REG_SIZE);
882 memory_region_add_subregion(&s->hme, 0x4000, &s->erxreg);
883
884 memory_region_init_io(&s->macreg, OBJECT(pci_dev), &sunhme_mac_ops, s,
885 "sunhme.mac", HME_MAC_REG_SIZE);
886 memory_region_add_subregion(&s->hme, 0x6000, &s->macreg);
887
888 memory_region_init_io(&s->mifreg, OBJECT(pci_dev), &sunhme_mif_ops, s,
889 "sunhme.mif", HME_MIF_REG_SIZE);
890 memory_region_add_subregion(&s->hme, 0x7000, &s->mifreg);
891
892 qemu_macaddr_default_if_unset(&s->conf.macaddr);
893 s->nic = qemu_new_nic(&net_sunhme_info, &s->conf,
894 object_get_typename(OBJECT(d)), d->id, s);
895 qemu_format_nic_info_str(qemu_get_queue(s->nic), s->conf.macaddr.a);
896}
897
898static void sunhme_instance_init(Object *obj)
899{
900 SunHMEState *s = SUNHME(obj);
901
902 device_add_bootindex_property(obj, &s->conf.bootindex,
903 "bootindex", "/ethernet-phy@0",
904 DEVICE(obj), NULL);
905}
906
907static void sunhme_reset(DeviceState *ds)
908{
909 SunHMEState *s = SUNHME(ds);
910
911 /* Configure internal transceiver */
912 s->mifregs[HME_MIFI_CFG >> 2] |= HME_MIF_CFG_MDI0;
913
914 /* Advetise auto, 100Mbps FD */
915 s->miiregs[MII_ANAR] = MII_ANAR_TXFD;
916 s->miiregs[MII_BMSR] = MII_BMSR_AUTONEG | MII_BMSR_100TX_FD |
917 MII_BMSR_AN_COMP;
918
919 if (!qemu_get_queue(s->nic)->link_down) {
920 s->miiregs[MII_ANLPAR] |= MII_ANLPAR_TXFD;
921 s->miiregs[MII_BMSR] |= MII_BMSR_LINK_ST;
922 }
923
924 /* Set manufacturer */
925 s->miiregs[MII_PHYID1] = DP83840_PHYID1;
926 s->miiregs[MII_PHYID2] = DP83840_PHYID2;
927
928 /* Configure default interrupt mask */
929 s->mifregs[HME_MIFI_IMASK >> 2] = 0xffff;
930 s->sebregs[HME_SEBI_IMASK >> 2] = 0xff7fffff;
931}
932
933static const VMStateDescription vmstate_hme = {
934 .name = "sunhme",
935 .version_id = 0,
936 .minimum_version_id = 0,
937 .fields = (VMStateField[]) {
938 VMSTATE_PCI_DEVICE(parent_obj, SunHMEState),
939 VMSTATE_MACADDR(conf.macaddr, SunHMEState),
940 VMSTATE_UINT32_ARRAY(sebregs, SunHMEState, (HME_SEB_REG_SIZE >> 2)),
941 VMSTATE_UINT32_ARRAY(etxregs, SunHMEState, (HME_ETX_REG_SIZE >> 2)),
942 VMSTATE_UINT32_ARRAY(erxregs, SunHMEState, (HME_ERX_REG_SIZE >> 2)),
943 VMSTATE_UINT32_ARRAY(macregs, SunHMEState, (HME_MAC_REG_SIZE >> 2)),
944 VMSTATE_UINT32_ARRAY(mifregs, SunHMEState, (HME_MIF_REG_SIZE >> 2)),
945 VMSTATE_UINT16_ARRAY(miiregs, SunHMEState, HME_MII_REGS_SIZE),
946 VMSTATE_END_OF_LIST()
947 }
948};
949
950static void sunhme_class_init(ObjectClass *klass, void *data)
951{
952 DeviceClass *dc = DEVICE_CLASS(klass);
953 PCIDeviceClass *k = PCI_DEVICE_CLASS(klass);
954
955 k->realize = sunhme_realize;
956 k->vendor_id = PCI_VENDOR_ID_SUN;
957 k->device_id = PCI_DEVICE_ID_SUN_HME;
958 k->class_id = PCI_CLASS_NETWORK_ETHERNET;
959 dc->vmsd = &vmstate_hme;
960 dc->reset = sunhme_reset;
961 dc->props = sunhme_properties;
962 set_bit(DEVICE_CATEGORY_NETWORK, dc->categories);
963}
964
965static const TypeInfo sunhme_info = {
966 .name = TYPE_SUNHME,
967 .parent = TYPE_PCI_DEVICE,
968 .class_init = sunhme_class_init,
969 .instance_size = sizeof(SunHMEState),
970 .instance_init = sunhme_instance_init,
971 .interfaces = (InterfaceInfo[]) {
972 { INTERFACE_CONVENTIONAL_PCI_DEVICE },
973 { }
974 }
975};
976
977static void sunhme_register_types(void)
978{
979 type_register_static(&sunhme_info);
980}
981
982type_init(sunhme_register_types)
983