]> git.proxmox.com Git - mirror_ubuntu-focal-kernel.git/blob - drivers/net/can/ti_hecc.c
ASoC: tlv320aic31xx: Reset registers during power up
[mirror_ubuntu-focal-kernel.git] / drivers / net / can / ti_hecc.c
1 /*
2 * TI HECC (CAN) device driver
3 *
4 * This driver supports TI's HECC (High End CAN Controller module) and the
5 * specs for the same is available at <http://www.ti.com>
6 *
7 * Copyright (C) 2009 Texas Instruments Incorporated - http://www.ti.com/
8 *
9 * This program is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU General Public License as
11 * published by the Free Software Foundation version 2.
12 *
13 * This program is distributed as is WITHOUT ANY WARRANTY of any
14 * kind, whether express or implied; without even the implied warranty
15 * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 * GNU General Public License for more details.
17 *
18 */
19
20 #include <linux/module.h>
21 #include <linux/kernel.h>
22 #include <linux/types.h>
23 #include <linux/interrupt.h>
24 #include <linux/errno.h>
25 #include <linux/netdevice.h>
26 #include <linux/skbuff.h>
27 #include <linux/platform_device.h>
28 #include <linux/clk.h>
29 #include <linux/io.h>
30 #include <linux/of.h>
31 #include <linux/of_device.h>
32 #include <linux/regulator/consumer.h>
33
34 #include <linux/can/dev.h>
35 #include <linux/can/error.h>
36 #include <linux/can/led.h>
37
38 #define DRV_NAME "ti_hecc"
39 #define HECC_MODULE_VERSION "0.7"
40 MODULE_VERSION(HECC_MODULE_VERSION);
41 #define DRV_DESC "TI High End CAN Controller Driver " HECC_MODULE_VERSION
42
43 /* TX / RX Mailbox Configuration */
44 #define HECC_MAX_MAILBOXES 32 /* hardware mailboxes - do not change */
45 #define MAX_TX_PRIO 0x3F /* hardware value - do not change */
46
47 /*
48 * Important Note: TX mailbox configuration
49 * TX mailboxes should be restricted to the number of SKB buffers to avoid
50 * maintaining SKB buffers separately. TX mailboxes should be a power of 2
51 * for the mailbox logic to work. Top mailbox numbers are reserved for RX
52 * and lower mailboxes for TX.
53 *
54 * HECC_MAX_TX_MBOX HECC_MB_TX_SHIFT
55 * 4 (default) 2
56 * 8 3
57 * 16 4
58 */
59 #define HECC_MB_TX_SHIFT 2 /* as per table above */
60 #define HECC_MAX_TX_MBOX BIT(HECC_MB_TX_SHIFT)
61
62 #define HECC_TX_PRIO_SHIFT (HECC_MB_TX_SHIFT)
63 #define HECC_TX_PRIO_MASK (MAX_TX_PRIO << HECC_MB_TX_SHIFT)
64 #define HECC_TX_MB_MASK (HECC_MAX_TX_MBOX - 1)
65 #define HECC_TX_MASK ((HECC_MAX_TX_MBOX - 1) | HECC_TX_PRIO_MASK)
66 #define HECC_TX_MBOX_MASK (~(BIT(HECC_MAX_TX_MBOX) - 1))
67 #define HECC_DEF_NAPI_WEIGHT HECC_MAX_RX_MBOX
68
69 /*
70 * Important Note: RX mailbox configuration
71 * RX mailboxes are further logically split into two - main and buffer
72 * mailboxes. The goal is to get all packets into main mailboxes as
73 * driven by mailbox number and receive priority (higher to lower) and
74 * buffer mailboxes are used to receive pkts while main mailboxes are being
75 * processed. This ensures in-order packet reception.
76 *
77 * Here are the recommended values for buffer mailbox. Note that RX mailboxes
78 * start after TX mailboxes:
79 *
80 * HECC_MAX_RX_MBOX HECC_RX_BUFFER_MBOX No of buffer mailboxes
81 * 28 12 8
82 * 16 20 4
83 */
84
85 #define HECC_MAX_RX_MBOX (HECC_MAX_MAILBOXES - HECC_MAX_TX_MBOX)
86 #define HECC_RX_BUFFER_MBOX 12 /* as per table above */
87 #define HECC_RX_FIRST_MBOX (HECC_MAX_MAILBOXES - 1)
88 #define HECC_RX_HIGH_MBOX_MASK (~(BIT(HECC_RX_BUFFER_MBOX) - 1))
89
90 /* TI HECC module registers */
91 #define HECC_CANME 0x0 /* Mailbox enable */
92 #define HECC_CANMD 0x4 /* Mailbox direction */
93 #define HECC_CANTRS 0x8 /* Transmit request set */
94 #define HECC_CANTRR 0xC /* Transmit request */
95 #define HECC_CANTA 0x10 /* Transmission acknowledge */
96 #define HECC_CANAA 0x14 /* Abort acknowledge */
97 #define HECC_CANRMP 0x18 /* Receive message pending */
98 #define HECC_CANRML 0x1C /* Remote message lost */
99 #define HECC_CANRFP 0x20 /* Remote frame pending */
100 #define HECC_CANGAM 0x24 /* SECC only:Global acceptance mask */
101 #define HECC_CANMC 0x28 /* Master control */
102 #define HECC_CANBTC 0x2C /* Bit timing configuration */
103 #define HECC_CANES 0x30 /* Error and status */
104 #define HECC_CANTEC 0x34 /* Transmit error counter */
105 #define HECC_CANREC 0x38 /* Receive error counter */
106 #define HECC_CANGIF0 0x3C /* Global interrupt flag 0 */
107 #define HECC_CANGIM 0x40 /* Global interrupt mask */
108 #define HECC_CANGIF1 0x44 /* Global interrupt flag 1 */
109 #define HECC_CANMIM 0x48 /* Mailbox interrupt mask */
110 #define HECC_CANMIL 0x4C /* Mailbox interrupt level */
111 #define HECC_CANOPC 0x50 /* Overwrite protection control */
112 #define HECC_CANTIOC 0x54 /* Transmit I/O control */
113 #define HECC_CANRIOC 0x58 /* Receive I/O control */
114 #define HECC_CANLNT 0x5C /* HECC only: Local network time */
115 #define HECC_CANTOC 0x60 /* HECC only: Time-out control */
116 #define HECC_CANTOS 0x64 /* HECC only: Time-out status */
117 #define HECC_CANTIOCE 0x68 /* SCC only:Enhanced TX I/O control */
118 #define HECC_CANRIOCE 0x6C /* SCC only:Enhanced RX I/O control */
119
120 /* Mailbox registers */
121 #define HECC_CANMID 0x0
122 #define HECC_CANMCF 0x4
123 #define HECC_CANMDL 0x8
124 #define HECC_CANMDH 0xC
125
126 #define HECC_SET_REG 0xFFFFFFFF
127 #define HECC_CANID_MASK 0x3FF /* 18 bits mask for extended id's */
128 #define HECC_CCE_WAIT_COUNT 100 /* Wait for ~1 sec for CCE bit */
129
130 #define HECC_CANMC_SCM BIT(13) /* SCC compat mode */
131 #define HECC_CANMC_CCR BIT(12) /* Change config request */
132 #define HECC_CANMC_PDR BIT(11) /* Local Power down - for sleep mode */
133 #define HECC_CANMC_ABO BIT(7) /* Auto Bus On */
134 #define HECC_CANMC_STM BIT(6) /* Self test mode - loopback */
135 #define HECC_CANMC_SRES BIT(5) /* Software reset */
136
137 #define HECC_CANTIOC_EN BIT(3) /* Enable CAN TX I/O pin */
138 #define HECC_CANRIOC_EN BIT(3) /* Enable CAN RX I/O pin */
139
140 #define HECC_CANMID_IDE BIT(31) /* Extended frame format */
141 #define HECC_CANMID_AME BIT(30) /* Acceptance mask enable */
142 #define HECC_CANMID_AAM BIT(29) /* Auto answer mode */
143
144 #define HECC_CANES_FE BIT(24) /* form error */
145 #define HECC_CANES_BE BIT(23) /* bit error */
146 #define HECC_CANES_SA1 BIT(22) /* stuck at dominant error */
147 #define HECC_CANES_CRCE BIT(21) /* CRC error */
148 #define HECC_CANES_SE BIT(20) /* stuff bit error */
149 #define HECC_CANES_ACKE BIT(19) /* ack error */
150 #define HECC_CANES_BO BIT(18) /* Bus off status */
151 #define HECC_CANES_EP BIT(17) /* Error passive status */
152 #define HECC_CANES_EW BIT(16) /* Error warning status */
153 #define HECC_CANES_SMA BIT(5) /* suspend mode ack */
154 #define HECC_CANES_CCE BIT(4) /* Change config enabled */
155 #define HECC_CANES_PDA BIT(3) /* Power down mode ack */
156
157 #define HECC_CANBTC_SAM BIT(7) /* sample points */
158
159 #define HECC_BUS_ERROR (HECC_CANES_FE | HECC_CANES_BE |\
160 HECC_CANES_CRCE | HECC_CANES_SE |\
161 HECC_CANES_ACKE)
162
163 #define HECC_CANMCF_RTR BIT(4) /* Remote transmit request */
164
165 #define HECC_CANGIF_MAIF BIT(17) /* Message alarm interrupt */
166 #define HECC_CANGIF_TCOIF BIT(16) /* Timer counter overflow int */
167 #define HECC_CANGIF_GMIF BIT(15) /* Global mailbox interrupt */
168 #define HECC_CANGIF_AAIF BIT(14) /* Abort ack interrupt */
169 #define HECC_CANGIF_WDIF BIT(13) /* Write denied interrupt */
170 #define HECC_CANGIF_WUIF BIT(12) /* Wake up interrupt */
171 #define HECC_CANGIF_RMLIF BIT(11) /* Receive message lost interrupt */
172 #define HECC_CANGIF_BOIF BIT(10) /* Bus off interrupt */
173 #define HECC_CANGIF_EPIF BIT(9) /* Error passive interrupt */
174 #define HECC_CANGIF_WLIF BIT(8) /* Warning level interrupt */
175 #define HECC_CANGIF_MBOX_MASK 0x1F /* Mailbox number mask */
176 #define HECC_CANGIM_I1EN BIT(1) /* Int line 1 enable */
177 #define HECC_CANGIM_I0EN BIT(0) /* Int line 0 enable */
178 #define HECC_CANGIM_DEF_MASK 0x700 /* only busoff/warning/passive */
179 #define HECC_CANGIM_SIL BIT(2) /* system interrupts to int line 1 */
180
181 /* CAN Bittiming constants as per HECC specs */
182 static const struct can_bittiming_const ti_hecc_bittiming_const = {
183 .name = DRV_NAME,
184 .tseg1_min = 1,
185 .tseg1_max = 16,
186 .tseg2_min = 1,
187 .tseg2_max = 8,
188 .sjw_max = 4,
189 .brp_min = 1,
190 .brp_max = 256,
191 .brp_inc = 1,
192 };
193
194 struct ti_hecc_priv {
195 struct can_priv can; /* MUST be first member/field */
196 struct napi_struct napi;
197 struct net_device *ndev;
198 struct clk *clk;
199 void __iomem *base;
200 void __iomem *hecc_ram;
201 void __iomem *mbx;
202 bool use_hecc1int;
203 spinlock_t mbx_lock; /* CANME register needs protection */
204 u32 tx_head;
205 u32 tx_tail;
206 u32 rx_next;
207 struct regulator *reg_xceiver;
208 };
209
210 static inline int get_tx_head_mb(struct ti_hecc_priv *priv)
211 {
212 return priv->tx_head & HECC_TX_MB_MASK;
213 }
214
215 static inline int get_tx_tail_mb(struct ti_hecc_priv *priv)
216 {
217 return priv->tx_tail & HECC_TX_MB_MASK;
218 }
219
220 static inline int get_tx_head_prio(struct ti_hecc_priv *priv)
221 {
222 return (priv->tx_head >> HECC_TX_PRIO_SHIFT) & MAX_TX_PRIO;
223 }
224
225 static inline void hecc_write_lam(struct ti_hecc_priv *priv, u32 mbxno, u32 val)
226 {
227 __raw_writel(val, priv->hecc_ram + mbxno * 4);
228 }
229
230 static inline void hecc_write_mbx(struct ti_hecc_priv *priv, u32 mbxno,
231 u32 reg, u32 val)
232 {
233 __raw_writel(val, priv->mbx + mbxno * 0x10 + reg);
234 }
235
236 static inline u32 hecc_read_mbx(struct ti_hecc_priv *priv, u32 mbxno, u32 reg)
237 {
238 return __raw_readl(priv->mbx + mbxno * 0x10 + reg);
239 }
240
241 static inline void hecc_write(struct ti_hecc_priv *priv, u32 reg, u32 val)
242 {
243 __raw_writel(val, priv->base + reg);
244 }
245
246 static inline u32 hecc_read(struct ti_hecc_priv *priv, int reg)
247 {
248 return __raw_readl(priv->base + reg);
249 }
250
251 static inline void hecc_set_bit(struct ti_hecc_priv *priv, int reg,
252 u32 bit_mask)
253 {
254 hecc_write(priv, reg, hecc_read(priv, reg) | bit_mask);
255 }
256
257 static inline void hecc_clear_bit(struct ti_hecc_priv *priv, int reg,
258 u32 bit_mask)
259 {
260 hecc_write(priv, reg, hecc_read(priv, reg) & ~bit_mask);
261 }
262
263 static inline u32 hecc_get_bit(struct ti_hecc_priv *priv, int reg, u32 bit_mask)
264 {
265 return (hecc_read(priv, reg) & bit_mask) ? 1 : 0;
266 }
267
268 static int ti_hecc_set_btc(struct ti_hecc_priv *priv)
269 {
270 struct can_bittiming *bit_timing = &priv->can.bittiming;
271 u32 can_btc;
272
273 can_btc = (bit_timing->phase_seg2 - 1) & 0x7;
274 can_btc |= ((bit_timing->phase_seg1 + bit_timing->prop_seg - 1)
275 & 0xF) << 3;
276 if (priv->can.ctrlmode & CAN_CTRLMODE_3_SAMPLES) {
277 if (bit_timing->brp > 4)
278 can_btc |= HECC_CANBTC_SAM;
279 else
280 netdev_warn(priv->ndev, "WARN: Triple"
281 "sampling not set due to h/w limitations");
282 }
283 can_btc |= ((bit_timing->sjw - 1) & 0x3) << 8;
284 can_btc |= ((bit_timing->brp - 1) & 0xFF) << 16;
285
286 /* ERM being set to 0 by default meaning resync at falling edge */
287
288 hecc_write(priv, HECC_CANBTC, can_btc);
289 netdev_info(priv->ndev, "setting CANBTC=%#x\n", can_btc);
290
291 return 0;
292 }
293
294 static int ti_hecc_transceiver_switch(const struct ti_hecc_priv *priv,
295 int on)
296 {
297 if (!priv->reg_xceiver)
298 return 0;
299
300 if (on)
301 return regulator_enable(priv->reg_xceiver);
302 else
303 return regulator_disable(priv->reg_xceiver);
304 }
305
306 static void ti_hecc_reset(struct net_device *ndev)
307 {
308 u32 cnt;
309 struct ti_hecc_priv *priv = netdev_priv(ndev);
310
311 netdev_dbg(ndev, "resetting hecc ...\n");
312 hecc_set_bit(priv, HECC_CANMC, HECC_CANMC_SRES);
313
314 /* Set change control request and wait till enabled */
315 hecc_set_bit(priv, HECC_CANMC, HECC_CANMC_CCR);
316
317 /*
318 * INFO: It has been observed that at times CCE bit may not be
319 * set and hw seems to be ok even if this bit is not set so
320 * timing out with a timing of 1ms to respect the specs
321 */
322 cnt = HECC_CCE_WAIT_COUNT;
323 while (!hecc_get_bit(priv, HECC_CANES, HECC_CANES_CCE) && cnt != 0) {
324 --cnt;
325 udelay(10);
326 }
327
328 /*
329 * Note: On HECC, BTC can be programmed only in initialization mode, so
330 * it is expected that the can bittiming parameters are set via ip
331 * utility before the device is opened
332 */
333 ti_hecc_set_btc(priv);
334
335 /* Clear CCR (and CANMC register) and wait for CCE = 0 enable */
336 hecc_write(priv, HECC_CANMC, 0);
337
338 /*
339 * INFO: CAN net stack handles bus off and hence disabling auto-bus-on
340 * hecc_set_bit(priv, HECC_CANMC, HECC_CANMC_ABO);
341 */
342
343 /*
344 * INFO: It has been observed that at times CCE bit may not be
345 * set and hw seems to be ok even if this bit is not set so
346 */
347 cnt = HECC_CCE_WAIT_COUNT;
348 while (hecc_get_bit(priv, HECC_CANES, HECC_CANES_CCE) && cnt != 0) {
349 --cnt;
350 udelay(10);
351 }
352
353 /* Enable TX and RX I/O Control pins */
354 hecc_write(priv, HECC_CANTIOC, HECC_CANTIOC_EN);
355 hecc_write(priv, HECC_CANRIOC, HECC_CANRIOC_EN);
356
357 /* Clear registers for clean operation */
358 hecc_write(priv, HECC_CANTA, HECC_SET_REG);
359 hecc_write(priv, HECC_CANRMP, HECC_SET_REG);
360 hecc_write(priv, HECC_CANGIF0, HECC_SET_REG);
361 hecc_write(priv, HECC_CANGIF1, HECC_SET_REG);
362 hecc_write(priv, HECC_CANME, 0);
363 hecc_write(priv, HECC_CANMD, 0);
364
365 /* SCC compat mode NOT supported (and not needed too) */
366 hecc_set_bit(priv, HECC_CANMC, HECC_CANMC_SCM);
367 }
368
369 static void ti_hecc_start(struct net_device *ndev)
370 {
371 struct ti_hecc_priv *priv = netdev_priv(ndev);
372 u32 cnt, mbxno, mbx_mask;
373
374 /* put HECC in initialization mode and set btc */
375 ti_hecc_reset(ndev);
376
377 priv->tx_head = priv->tx_tail = HECC_TX_MASK;
378 priv->rx_next = HECC_RX_FIRST_MBOX;
379
380 /* Enable local and global acceptance mask registers */
381 hecc_write(priv, HECC_CANGAM, HECC_SET_REG);
382
383 /* Prepare configured mailboxes to receive messages */
384 for (cnt = 0; cnt < HECC_MAX_RX_MBOX; cnt++) {
385 mbxno = HECC_MAX_MAILBOXES - 1 - cnt;
386 mbx_mask = BIT(mbxno);
387 hecc_clear_bit(priv, HECC_CANME, mbx_mask);
388 hecc_write_mbx(priv, mbxno, HECC_CANMID, HECC_CANMID_AME);
389 hecc_write_lam(priv, mbxno, HECC_SET_REG);
390 hecc_set_bit(priv, HECC_CANMD, mbx_mask);
391 hecc_set_bit(priv, HECC_CANME, mbx_mask);
392 hecc_set_bit(priv, HECC_CANMIM, mbx_mask);
393 }
394
395 /* Prevent message over-write & Enable interrupts */
396 hecc_write(priv, HECC_CANOPC, HECC_SET_REG);
397 if (priv->use_hecc1int) {
398 hecc_write(priv, HECC_CANMIL, HECC_SET_REG);
399 hecc_write(priv, HECC_CANGIM, HECC_CANGIM_DEF_MASK |
400 HECC_CANGIM_I1EN | HECC_CANGIM_SIL);
401 } else {
402 hecc_write(priv, HECC_CANMIL, 0);
403 hecc_write(priv, HECC_CANGIM,
404 HECC_CANGIM_DEF_MASK | HECC_CANGIM_I0EN);
405 }
406 priv->can.state = CAN_STATE_ERROR_ACTIVE;
407 }
408
409 static void ti_hecc_stop(struct net_device *ndev)
410 {
411 struct ti_hecc_priv *priv = netdev_priv(ndev);
412
413 /* Disable interrupts and disable mailboxes */
414 hecc_write(priv, HECC_CANGIM, 0);
415 hecc_write(priv, HECC_CANMIM, 0);
416 hecc_write(priv, HECC_CANME, 0);
417 priv->can.state = CAN_STATE_STOPPED;
418 }
419
420 static int ti_hecc_do_set_mode(struct net_device *ndev, enum can_mode mode)
421 {
422 int ret = 0;
423
424 switch (mode) {
425 case CAN_MODE_START:
426 ti_hecc_start(ndev);
427 netif_wake_queue(ndev);
428 break;
429 default:
430 ret = -EOPNOTSUPP;
431 break;
432 }
433
434 return ret;
435 }
436
437 static int ti_hecc_get_berr_counter(const struct net_device *ndev,
438 struct can_berr_counter *bec)
439 {
440 struct ti_hecc_priv *priv = netdev_priv(ndev);
441
442 bec->txerr = hecc_read(priv, HECC_CANTEC);
443 bec->rxerr = hecc_read(priv, HECC_CANREC);
444
445 return 0;
446 }
447
448 /*
449 * ti_hecc_xmit: HECC Transmit
450 *
451 * The transmit mailboxes start from 0 to HECC_MAX_TX_MBOX. In HECC the
452 * priority of the mailbox for tranmission is dependent upon priority setting
453 * field in mailbox registers. The mailbox with highest value in priority field
454 * is transmitted first. Only when two mailboxes have the same value in
455 * priority field the highest numbered mailbox is transmitted first.
456 *
457 * To utilize the HECC priority feature as described above we start with the
458 * highest numbered mailbox with highest priority level and move on to the next
459 * mailbox with the same priority level and so on. Once we loop through all the
460 * transmit mailboxes we choose the next priority level (lower) and so on
461 * until we reach the lowest priority level on the lowest numbered mailbox
462 * when we stop transmission until all mailboxes are transmitted and then
463 * restart at highest numbered mailbox with highest priority.
464 *
465 * Two counters (head and tail) are used to track the next mailbox to transmit
466 * and to track the echo buffer for already transmitted mailbox. The queue
467 * is stopped when all the mailboxes are busy or when there is a priority
468 * value roll-over happens.
469 */
470 static netdev_tx_t ti_hecc_xmit(struct sk_buff *skb, struct net_device *ndev)
471 {
472 struct ti_hecc_priv *priv = netdev_priv(ndev);
473 struct can_frame *cf = (struct can_frame *)skb->data;
474 u32 mbxno, mbx_mask, data;
475 unsigned long flags;
476
477 if (can_dropped_invalid_skb(ndev, skb))
478 return NETDEV_TX_OK;
479
480 mbxno = get_tx_head_mb(priv);
481 mbx_mask = BIT(mbxno);
482 spin_lock_irqsave(&priv->mbx_lock, flags);
483 if (unlikely(hecc_read(priv, HECC_CANME) & mbx_mask)) {
484 spin_unlock_irqrestore(&priv->mbx_lock, flags);
485 netif_stop_queue(ndev);
486 netdev_err(priv->ndev,
487 "BUG: TX mbx not ready tx_head=%08X, tx_tail=%08X\n",
488 priv->tx_head, priv->tx_tail);
489 return NETDEV_TX_BUSY;
490 }
491 spin_unlock_irqrestore(&priv->mbx_lock, flags);
492
493 /* Prepare mailbox for transmission */
494 data = cf->can_dlc | (get_tx_head_prio(priv) << 8);
495 if (cf->can_id & CAN_RTR_FLAG) /* Remote transmission request */
496 data |= HECC_CANMCF_RTR;
497 hecc_write_mbx(priv, mbxno, HECC_CANMCF, data);
498
499 if (cf->can_id & CAN_EFF_FLAG) /* Extended frame format */
500 data = (cf->can_id & CAN_EFF_MASK) | HECC_CANMID_IDE;
501 else /* Standard frame format */
502 data = (cf->can_id & CAN_SFF_MASK) << 18;
503 hecc_write_mbx(priv, mbxno, HECC_CANMID, data);
504 hecc_write_mbx(priv, mbxno, HECC_CANMDL,
505 be32_to_cpu(*(__be32 *)(cf->data)));
506 if (cf->can_dlc > 4)
507 hecc_write_mbx(priv, mbxno, HECC_CANMDH,
508 be32_to_cpu(*(__be32 *)(cf->data + 4)));
509 else
510 *(u32 *)(cf->data + 4) = 0;
511 can_put_echo_skb(skb, ndev, mbxno);
512
513 spin_lock_irqsave(&priv->mbx_lock, flags);
514 --priv->tx_head;
515 if ((hecc_read(priv, HECC_CANME) & BIT(get_tx_head_mb(priv))) ||
516 (priv->tx_head & HECC_TX_MASK) == HECC_TX_MASK) {
517 netif_stop_queue(ndev);
518 }
519 hecc_set_bit(priv, HECC_CANME, mbx_mask);
520 spin_unlock_irqrestore(&priv->mbx_lock, flags);
521
522 hecc_clear_bit(priv, HECC_CANMD, mbx_mask);
523 hecc_set_bit(priv, HECC_CANMIM, mbx_mask);
524 hecc_write(priv, HECC_CANTRS, mbx_mask);
525
526 return NETDEV_TX_OK;
527 }
528
529 static int ti_hecc_rx_pkt(struct ti_hecc_priv *priv, int mbxno)
530 {
531 struct net_device_stats *stats = &priv->ndev->stats;
532 struct can_frame *cf;
533 struct sk_buff *skb;
534 u32 data, mbx_mask;
535 unsigned long flags;
536
537 skb = alloc_can_skb(priv->ndev, &cf);
538 if (!skb) {
539 if (printk_ratelimit())
540 netdev_err(priv->ndev,
541 "ti_hecc_rx_pkt: alloc_can_skb() failed\n");
542 return -ENOMEM;
543 }
544
545 mbx_mask = BIT(mbxno);
546 data = hecc_read_mbx(priv, mbxno, HECC_CANMID);
547 if (data & HECC_CANMID_IDE)
548 cf->can_id = (data & CAN_EFF_MASK) | CAN_EFF_FLAG;
549 else
550 cf->can_id = (data >> 18) & CAN_SFF_MASK;
551 data = hecc_read_mbx(priv, mbxno, HECC_CANMCF);
552 if (data & HECC_CANMCF_RTR)
553 cf->can_id |= CAN_RTR_FLAG;
554 cf->can_dlc = get_can_dlc(data & 0xF);
555 data = hecc_read_mbx(priv, mbxno, HECC_CANMDL);
556 *(__be32 *)(cf->data) = cpu_to_be32(data);
557 if (cf->can_dlc > 4) {
558 data = hecc_read_mbx(priv, mbxno, HECC_CANMDH);
559 *(__be32 *)(cf->data + 4) = cpu_to_be32(data);
560 }
561 spin_lock_irqsave(&priv->mbx_lock, flags);
562 hecc_clear_bit(priv, HECC_CANME, mbx_mask);
563 hecc_write(priv, HECC_CANRMP, mbx_mask);
564 /* enable mailbox only if it is part of rx buffer mailboxes */
565 if (priv->rx_next < HECC_RX_BUFFER_MBOX)
566 hecc_set_bit(priv, HECC_CANME, mbx_mask);
567 spin_unlock_irqrestore(&priv->mbx_lock, flags);
568
569 stats->rx_bytes += cf->can_dlc;
570 can_led_event(priv->ndev, CAN_LED_EVENT_RX);
571 netif_receive_skb(skb);
572 stats->rx_packets++;
573
574 return 0;
575 }
576
577 /*
578 * ti_hecc_rx_poll - HECC receive pkts
579 *
580 * The receive mailboxes start from highest numbered mailbox till last xmit
581 * mailbox. On CAN frame reception the hardware places the data into highest
582 * numbered mailbox that matches the CAN ID filter. Since all receive mailboxes
583 * have same filtering (ALL CAN frames) packets will arrive in the highest
584 * available RX mailbox and we need to ensure in-order packet reception.
585 *
586 * To ensure the packets are received in the right order we logically divide
587 * the RX mailboxes into main and buffer mailboxes. Packets are received as per
588 * mailbox priotity (higher to lower) in the main bank and once it is full we
589 * disable further reception into main mailboxes. While the main mailboxes are
590 * processed in NAPI, further packets are received in buffer mailboxes.
591 *
592 * We maintain a RX next mailbox counter to process packets and once all main
593 * mailboxe packets are passed to the upper stack we enable all of them but
594 * continue to process packets received in buffer mailboxes. With each packet
595 * received from buffer mailbox we enable it immediately so as to handle the
596 * overflow from higher mailboxes.
597 */
598 static int ti_hecc_rx_poll(struct napi_struct *napi, int quota)
599 {
600 struct net_device *ndev = napi->dev;
601 struct ti_hecc_priv *priv = netdev_priv(ndev);
602 u32 num_pkts = 0;
603 u32 mbx_mask;
604 unsigned long pending_pkts, flags;
605
606 if (!netif_running(ndev))
607 return 0;
608
609 while ((pending_pkts = hecc_read(priv, HECC_CANRMP)) &&
610 num_pkts < quota) {
611 mbx_mask = BIT(priv->rx_next); /* next rx mailbox to process */
612 if (mbx_mask & pending_pkts) {
613 if (ti_hecc_rx_pkt(priv, priv->rx_next) < 0)
614 return num_pkts;
615 ++num_pkts;
616 } else if (priv->rx_next > HECC_RX_BUFFER_MBOX) {
617 break; /* pkt not received yet */
618 }
619 --priv->rx_next;
620 if (priv->rx_next == HECC_RX_BUFFER_MBOX) {
621 /* enable high bank mailboxes */
622 spin_lock_irqsave(&priv->mbx_lock, flags);
623 mbx_mask = hecc_read(priv, HECC_CANME);
624 mbx_mask |= HECC_RX_HIGH_MBOX_MASK;
625 hecc_write(priv, HECC_CANME, mbx_mask);
626 spin_unlock_irqrestore(&priv->mbx_lock, flags);
627 } else if (priv->rx_next == HECC_MAX_TX_MBOX - 1) {
628 priv->rx_next = HECC_RX_FIRST_MBOX;
629 break;
630 }
631 }
632
633 /* Enable packet interrupt if all pkts are handled */
634 if (hecc_read(priv, HECC_CANRMP) == 0) {
635 napi_complete(napi);
636 /* Re-enable RX mailbox interrupts */
637 mbx_mask = hecc_read(priv, HECC_CANMIM);
638 mbx_mask |= HECC_TX_MBOX_MASK;
639 hecc_write(priv, HECC_CANMIM, mbx_mask);
640 }
641
642 return num_pkts;
643 }
644
645 static int ti_hecc_error(struct net_device *ndev, int int_status,
646 int err_status)
647 {
648 struct ti_hecc_priv *priv = netdev_priv(ndev);
649 struct net_device_stats *stats = &ndev->stats;
650 struct can_frame *cf;
651 struct sk_buff *skb;
652
653 /* propagate the error condition to the can stack */
654 skb = alloc_can_err_skb(ndev, &cf);
655 if (!skb) {
656 if (printk_ratelimit())
657 netdev_err(priv->ndev,
658 "ti_hecc_error: alloc_can_err_skb() failed\n");
659 return -ENOMEM;
660 }
661
662 if (int_status & HECC_CANGIF_WLIF) { /* warning level int */
663 if ((int_status & HECC_CANGIF_BOIF) == 0) {
664 priv->can.state = CAN_STATE_ERROR_WARNING;
665 ++priv->can.can_stats.error_warning;
666 cf->can_id |= CAN_ERR_CRTL;
667 if (hecc_read(priv, HECC_CANTEC) > 96)
668 cf->data[1] |= CAN_ERR_CRTL_TX_WARNING;
669 if (hecc_read(priv, HECC_CANREC) > 96)
670 cf->data[1] |= CAN_ERR_CRTL_RX_WARNING;
671 }
672 hecc_set_bit(priv, HECC_CANES, HECC_CANES_EW);
673 netdev_dbg(priv->ndev, "Error Warning interrupt\n");
674 hecc_clear_bit(priv, HECC_CANMC, HECC_CANMC_CCR);
675 }
676
677 if (int_status & HECC_CANGIF_EPIF) { /* error passive int */
678 if ((int_status & HECC_CANGIF_BOIF) == 0) {
679 priv->can.state = CAN_STATE_ERROR_PASSIVE;
680 ++priv->can.can_stats.error_passive;
681 cf->can_id |= CAN_ERR_CRTL;
682 if (hecc_read(priv, HECC_CANTEC) > 127)
683 cf->data[1] |= CAN_ERR_CRTL_TX_PASSIVE;
684 if (hecc_read(priv, HECC_CANREC) > 127)
685 cf->data[1] |= CAN_ERR_CRTL_RX_PASSIVE;
686 }
687 hecc_set_bit(priv, HECC_CANES, HECC_CANES_EP);
688 netdev_dbg(priv->ndev, "Error passive interrupt\n");
689 hecc_clear_bit(priv, HECC_CANMC, HECC_CANMC_CCR);
690 }
691
692 /*
693 * Need to check busoff condition in error status register too to
694 * ensure warning interrupts don't hog the system
695 */
696 if ((int_status & HECC_CANGIF_BOIF) || (err_status & HECC_CANES_BO)) {
697 priv->can.state = CAN_STATE_BUS_OFF;
698 cf->can_id |= CAN_ERR_BUSOFF;
699 hecc_set_bit(priv, HECC_CANES, HECC_CANES_BO);
700 hecc_clear_bit(priv, HECC_CANMC, HECC_CANMC_CCR);
701 /* Disable all interrupts in bus-off to avoid int hog */
702 hecc_write(priv, HECC_CANGIM, 0);
703 ++priv->can.can_stats.bus_off;
704 can_bus_off(ndev);
705 }
706
707 if (err_status & HECC_BUS_ERROR) {
708 ++priv->can.can_stats.bus_error;
709 cf->can_id |= CAN_ERR_BUSERROR | CAN_ERR_PROT;
710 if (err_status & HECC_CANES_FE) {
711 hecc_set_bit(priv, HECC_CANES, HECC_CANES_FE);
712 cf->data[2] |= CAN_ERR_PROT_FORM;
713 }
714 if (err_status & HECC_CANES_BE) {
715 hecc_set_bit(priv, HECC_CANES, HECC_CANES_BE);
716 cf->data[2] |= CAN_ERR_PROT_BIT;
717 }
718 if (err_status & HECC_CANES_SE) {
719 hecc_set_bit(priv, HECC_CANES, HECC_CANES_SE);
720 cf->data[2] |= CAN_ERR_PROT_STUFF;
721 }
722 if (err_status & HECC_CANES_CRCE) {
723 hecc_set_bit(priv, HECC_CANES, HECC_CANES_CRCE);
724 cf->data[3] = CAN_ERR_PROT_LOC_CRC_SEQ;
725 }
726 if (err_status & HECC_CANES_ACKE) {
727 hecc_set_bit(priv, HECC_CANES, HECC_CANES_ACKE);
728 cf->data[3] = CAN_ERR_PROT_LOC_ACK;
729 }
730 }
731
732 stats->rx_packets++;
733 stats->rx_bytes += cf->can_dlc;
734 netif_rx(skb);
735
736 return 0;
737 }
738
739 static irqreturn_t ti_hecc_interrupt(int irq, void *dev_id)
740 {
741 struct net_device *ndev = (struct net_device *)dev_id;
742 struct ti_hecc_priv *priv = netdev_priv(ndev);
743 struct net_device_stats *stats = &ndev->stats;
744 u32 mbxno, mbx_mask, int_status, err_status;
745 unsigned long ack, flags;
746
747 int_status = hecc_read(priv,
748 (priv->use_hecc1int) ? HECC_CANGIF1 : HECC_CANGIF0);
749
750 if (!int_status)
751 return IRQ_NONE;
752
753 err_status = hecc_read(priv, HECC_CANES);
754 if (err_status & (HECC_BUS_ERROR | HECC_CANES_BO |
755 HECC_CANES_EP | HECC_CANES_EW))
756 ti_hecc_error(ndev, int_status, err_status);
757
758 if (int_status & HECC_CANGIF_GMIF) {
759 while (priv->tx_tail - priv->tx_head > 0) {
760 mbxno = get_tx_tail_mb(priv);
761 mbx_mask = BIT(mbxno);
762 if (!(mbx_mask & hecc_read(priv, HECC_CANTA)))
763 break;
764 hecc_clear_bit(priv, HECC_CANMIM, mbx_mask);
765 hecc_write(priv, HECC_CANTA, mbx_mask);
766 spin_lock_irqsave(&priv->mbx_lock, flags);
767 hecc_clear_bit(priv, HECC_CANME, mbx_mask);
768 spin_unlock_irqrestore(&priv->mbx_lock, flags);
769 stats->tx_bytes += hecc_read_mbx(priv, mbxno,
770 HECC_CANMCF) & 0xF;
771 stats->tx_packets++;
772 can_led_event(ndev, CAN_LED_EVENT_TX);
773 can_get_echo_skb(ndev, mbxno);
774 --priv->tx_tail;
775 }
776
777 /* restart queue if wrap-up or if queue stalled on last pkt */
778 if (((priv->tx_head == priv->tx_tail) &&
779 ((priv->tx_head & HECC_TX_MASK) != HECC_TX_MASK)) ||
780 (((priv->tx_tail & HECC_TX_MASK) == HECC_TX_MASK) &&
781 ((priv->tx_head & HECC_TX_MASK) == HECC_TX_MASK)))
782 netif_wake_queue(ndev);
783
784 /* Disable RX mailbox interrupts and let NAPI reenable them */
785 if (hecc_read(priv, HECC_CANRMP)) {
786 ack = hecc_read(priv, HECC_CANMIM);
787 ack &= BIT(HECC_MAX_TX_MBOX) - 1;
788 hecc_write(priv, HECC_CANMIM, ack);
789 napi_schedule(&priv->napi);
790 }
791 }
792
793 /* clear all interrupt conditions - read back to avoid spurious ints */
794 if (priv->use_hecc1int) {
795 hecc_write(priv, HECC_CANGIF1, HECC_SET_REG);
796 int_status = hecc_read(priv, HECC_CANGIF1);
797 } else {
798 hecc_write(priv, HECC_CANGIF0, HECC_SET_REG);
799 int_status = hecc_read(priv, HECC_CANGIF0);
800 }
801
802 return IRQ_HANDLED;
803 }
804
805 static int ti_hecc_open(struct net_device *ndev)
806 {
807 struct ti_hecc_priv *priv = netdev_priv(ndev);
808 int err;
809
810 err = request_irq(ndev->irq, ti_hecc_interrupt, IRQF_SHARED,
811 ndev->name, ndev);
812 if (err) {
813 netdev_err(ndev, "error requesting interrupt\n");
814 return err;
815 }
816
817 ti_hecc_transceiver_switch(priv, 1);
818
819 /* Open common can device */
820 err = open_candev(ndev);
821 if (err) {
822 netdev_err(ndev, "open_candev() failed %d\n", err);
823 ti_hecc_transceiver_switch(priv, 0);
824 free_irq(ndev->irq, ndev);
825 return err;
826 }
827
828 can_led_event(ndev, CAN_LED_EVENT_OPEN);
829
830 ti_hecc_start(ndev);
831 napi_enable(&priv->napi);
832 netif_start_queue(ndev);
833
834 return 0;
835 }
836
837 static int ti_hecc_close(struct net_device *ndev)
838 {
839 struct ti_hecc_priv *priv = netdev_priv(ndev);
840
841 netif_stop_queue(ndev);
842 napi_disable(&priv->napi);
843 ti_hecc_stop(ndev);
844 free_irq(ndev->irq, ndev);
845 close_candev(ndev);
846 ti_hecc_transceiver_switch(priv, 0);
847
848 can_led_event(ndev, CAN_LED_EVENT_STOP);
849
850 return 0;
851 }
852
853 static const struct net_device_ops ti_hecc_netdev_ops = {
854 .ndo_open = ti_hecc_open,
855 .ndo_stop = ti_hecc_close,
856 .ndo_start_xmit = ti_hecc_xmit,
857 .ndo_change_mtu = can_change_mtu,
858 };
859
860 static const struct of_device_id ti_hecc_dt_ids[] = {
861 {
862 .compatible = "ti,am3517-hecc",
863 },
864 { }
865 };
866 MODULE_DEVICE_TABLE(of, ti_hecc_dt_ids);
867
868 static int ti_hecc_probe(struct platform_device *pdev)
869 {
870 struct net_device *ndev = (struct net_device *)0;
871 struct ti_hecc_priv *priv;
872 struct device_node *np = pdev->dev.of_node;
873 struct resource *res, *irq;
874 struct regulator *reg_xceiver;
875 int err = -ENODEV;
876
877 if (!IS_ENABLED(CONFIG_OF) || !np)
878 return -EINVAL;
879
880 reg_xceiver = devm_regulator_get(&pdev->dev, "xceiver");
881 if (PTR_ERR(reg_xceiver) == -EPROBE_DEFER)
882 return -EPROBE_DEFER;
883 else if (IS_ERR(reg_xceiver))
884 reg_xceiver = NULL;
885
886 ndev = alloc_candev(sizeof(struct ti_hecc_priv), HECC_MAX_TX_MBOX);
887 if (!ndev) {
888 dev_err(&pdev->dev, "alloc_candev failed\n");
889 return -ENOMEM;
890 }
891 priv = netdev_priv(ndev);
892
893 /* handle hecc memory */
894 res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "hecc");
895 if (!res) {
896 dev_err(&pdev->dev, "can't get IORESOURCE_MEM hecc\n");
897 return -EINVAL;
898 }
899
900 priv->base = devm_ioremap_resource(&pdev->dev, res);
901 if (IS_ERR(priv->base)) {
902 dev_err(&pdev->dev, "hecc ioremap failed\n");
903 return PTR_ERR(priv->base);
904 }
905
906 /* handle hecc-ram memory */
907 res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "hecc-ram");
908 if (!res) {
909 dev_err(&pdev->dev, "can't get IORESOURCE_MEM hecc-ram\n");
910 return -EINVAL;
911 }
912
913 priv->hecc_ram = devm_ioremap_resource(&pdev->dev, res);
914 if (IS_ERR(priv->hecc_ram)) {
915 dev_err(&pdev->dev, "hecc-ram ioremap failed\n");
916 return PTR_ERR(priv->hecc_ram);
917 }
918
919 /* handle mbx memory */
920 res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "mbx");
921 if (!res) {
922 dev_err(&pdev->dev, "can't get IORESOURCE_MEM mbx\n");
923 return -EINVAL;
924 }
925
926 priv->mbx = devm_ioremap_resource(&pdev->dev, res);
927 if (IS_ERR(priv->mbx)) {
928 dev_err(&pdev->dev, "mbx ioremap failed\n");
929 return PTR_ERR(priv->mbx);
930 }
931
932 irq = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
933 if (!irq) {
934 dev_err(&pdev->dev, "No irq resource\n");
935 goto probe_exit;
936 }
937
938 priv->ndev = ndev;
939 priv->reg_xceiver = reg_xceiver;
940 priv->use_hecc1int = of_property_read_bool(np, "ti,use-hecc1int");
941
942 priv->can.bittiming_const = &ti_hecc_bittiming_const;
943 priv->can.do_set_mode = ti_hecc_do_set_mode;
944 priv->can.do_get_berr_counter = ti_hecc_get_berr_counter;
945 priv->can.ctrlmode_supported = CAN_CTRLMODE_3_SAMPLES;
946
947 spin_lock_init(&priv->mbx_lock);
948 ndev->irq = irq->start;
949 ndev->flags |= IFF_ECHO;
950 platform_set_drvdata(pdev, ndev);
951 SET_NETDEV_DEV(ndev, &pdev->dev);
952 ndev->netdev_ops = &ti_hecc_netdev_ops;
953
954 priv->clk = clk_get(&pdev->dev, "hecc_ck");
955 if (IS_ERR(priv->clk)) {
956 dev_err(&pdev->dev, "No clock available\n");
957 err = PTR_ERR(priv->clk);
958 priv->clk = NULL;
959 goto probe_exit_candev;
960 }
961 priv->can.clock.freq = clk_get_rate(priv->clk);
962 netif_napi_add(ndev, &priv->napi, ti_hecc_rx_poll,
963 HECC_DEF_NAPI_WEIGHT);
964
965 err = clk_prepare_enable(priv->clk);
966 if (err) {
967 dev_err(&pdev->dev, "clk_prepare_enable() failed\n");
968 goto probe_exit_clk;
969 }
970
971 err = register_candev(ndev);
972 if (err) {
973 dev_err(&pdev->dev, "register_candev() failed\n");
974 goto probe_exit_clk;
975 }
976
977 devm_can_led_init(ndev);
978
979 dev_info(&pdev->dev, "device registered (reg_base=%p, irq=%u)\n",
980 priv->base, (u32) ndev->irq);
981
982 return 0;
983
984 probe_exit_clk:
985 clk_put(priv->clk);
986 probe_exit_candev:
987 free_candev(ndev);
988 probe_exit:
989 return err;
990 }
991
992 static int ti_hecc_remove(struct platform_device *pdev)
993 {
994 struct net_device *ndev = platform_get_drvdata(pdev);
995 struct ti_hecc_priv *priv = netdev_priv(ndev);
996
997 unregister_candev(ndev);
998 clk_disable_unprepare(priv->clk);
999 clk_put(priv->clk);
1000 free_candev(ndev);
1001
1002 return 0;
1003 }
1004
1005 #ifdef CONFIG_PM
1006 static int ti_hecc_suspend(struct platform_device *pdev, pm_message_t state)
1007 {
1008 struct net_device *dev = platform_get_drvdata(pdev);
1009 struct ti_hecc_priv *priv = netdev_priv(dev);
1010
1011 if (netif_running(dev)) {
1012 netif_stop_queue(dev);
1013 netif_device_detach(dev);
1014 }
1015
1016 hecc_set_bit(priv, HECC_CANMC, HECC_CANMC_PDR);
1017 priv->can.state = CAN_STATE_SLEEPING;
1018
1019 clk_disable_unprepare(priv->clk);
1020
1021 return 0;
1022 }
1023
1024 static int ti_hecc_resume(struct platform_device *pdev)
1025 {
1026 struct net_device *dev = platform_get_drvdata(pdev);
1027 struct ti_hecc_priv *priv = netdev_priv(dev);
1028 int err;
1029
1030 err = clk_prepare_enable(priv->clk);
1031 if (err)
1032 return err;
1033
1034 hecc_clear_bit(priv, HECC_CANMC, HECC_CANMC_PDR);
1035 priv->can.state = CAN_STATE_ERROR_ACTIVE;
1036
1037 if (netif_running(dev)) {
1038 netif_device_attach(dev);
1039 netif_start_queue(dev);
1040 }
1041
1042 return 0;
1043 }
1044 #else
1045 #define ti_hecc_suspend NULL
1046 #define ti_hecc_resume NULL
1047 #endif
1048
1049 /* TI HECC netdevice driver: platform driver structure */
1050 static struct platform_driver ti_hecc_driver = {
1051 .driver = {
1052 .name = DRV_NAME,
1053 .of_match_table = ti_hecc_dt_ids,
1054 },
1055 .probe = ti_hecc_probe,
1056 .remove = ti_hecc_remove,
1057 .suspend = ti_hecc_suspend,
1058 .resume = ti_hecc_resume,
1059 };
1060
1061 module_platform_driver(ti_hecc_driver);
1062
1063 MODULE_AUTHOR("Anant Gole <anantgole@ti.com>");
1064 MODULE_LICENSE("GPL v2");
1065 MODULE_DESCRIPTION(DRV_DESC);
1066 MODULE_ALIAS("platform:" DRV_NAME);