]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blob - arch/cris/arch-v32/drivers/sync_serial.c
sched/headers: Prepare to move signal wakeup & sigpending methods from <linux/sched...
[mirror_ubuntu-artful-kernel.git] / arch / cris / arch-v32 / drivers / sync_serial.c
1 /*
2 * Simple synchronous serial port driver for ETRAX FS and ARTPEC-3.
3 *
4 * Copyright (c) 2005, 2008 Axis Communications AB
5 * Author: Mikael Starvik
6 *
7 */
8
9 #include <linux/module.h>
10 #include <linux/kernel.h>
11 #include <linux/types.h>
12 #include <linux/errno.h>
13 #include <linux/major.h>
14 #include <linux/sched/signal.h>
15 #include <linux/mutex.h>
16 #include <linux/interrupt.h>
17 #include <linux/poll.h>
18 #include <linux/fs.h>
19 #include <linux/cdev.h>
20 #include <linux/device.h>
21 #include <linux/wait.h>
22
23 #include <asm/io.h>
24 #include <mach/dma.h>
25 #include <pinmux.h>
26 #include <hwregs/reg_rdwr.h>
27 #include <hwregs/sser_defs.h>
28 #include <hwregs/timer_defs.h>
29 #include <hwregs/dma_defs.h>
30 #include <hwregs/dma.h>
31 #include <hwregs/intr_vect_defs.h>
32 #include <hwregs/intr_vect.h>
33 #include <hwregs/reg_map.h>
34 #include <asm/sync_serial.h>
35
36
37 /* The receiver is a bit tricky because of the continuous stream of data.*/
38 /* */
39 /* Three DMA descriptors are linked together. Each DMA descriptor is */
40 /* responsible for port->bufchunk of a common buffer. */
41 /* */
42 /* +---------------------------------------------+ */
43 /* | +----------+ +----------+ +----------+ | */
44 /* +-> | Descr[0] |-->| Descr[1] |-->| Descr[2] |-+ */
45 /* +----------+ +----------+ +----------+ */
46 /* | | | */
47 /* v v v */
48 /* +-------------------------------------+ */
49 /* | BUFFER | */
50 /* +-------------------------------------+ */
51 /* |<- data_avail ->| */
52 /* readp writep */
53 /* */
54 /* If the application keeps up the pace readp will be right after writep.*/
55 /* If the application can't keep the pace we have to throw away data. */
56 /* The idea is that readp should be ready with the data pointed out by */
57 /* Descr[i] when the DMA has filled in Descr[i+1]. */
58 /* Otherwise we will discard */
59 /* the rest of the data pointed out by Descr1 and set readp to the start */
60 /* of Descr2 */
61
62 /* IN_BUFFER_SIZE should be a multiple of 6 to make sure that 24 bit */
63 /* words can be handled */
64 #define IN_DESCR_SIZE SSP_INPUT_CHUNK_SIZE
65 #define NBR_IN_DESCR (8*6)
66 #define IN_BUFFER_SIZE (IN_DESCR_SIZE * NBR_IN_DESCR)
67
68 #define NBR_OUT_DESCR 8
69 #define OUT_BUFFER_SIZE (1024 * NBR_OUT_DESCR)
70
71 #define DEFAULT_FRAME_RATE 0
72 #define DEFAULT_WORD_RATE 7
73
74 /* To be removed when we move to pure udev. */
75 #define SYNC_SERIAL_MAJOR 125
76
77 /* NOTE: Enabling some debug will likely cause overrun or underrun,
78 * especially if manual mode is used.
79 */
80 #define DEBUG(x)
81 #define DEBUGREAD(x)
82 #define DEBUGWRITE(x)
83 #define DEBUGPOLL(x)
84 #define DEBUGRXINT(x)
85 #define DEBUGTXINT(x)
86 #define DEBUGTRDMA(x)
87 #define DEBUGOUTBUF(x)
88
89 enum syncser_irq_setup {
90 no_irq_setup = 0,
91 dma_irq_setup = 1,
92 manual_irq_setup = 2,
93 };
94
95 struct sync_port {
96 unsigned long regi_sser;
97 unsigned long regi_dmain;
98 unsigned long regi_dmaout;
99
100 /* Interrupt vectors. */
101 unsigned long dma_in_intr_vect; /* Used for DMA in. */
102 unsigned long dma_out_intr_vect; /* Used for DMA out. */
103 unsigned long syncser_intr_vect; /* Used when no DMA. */
104
105 /* DMA number for in and out. */
106 unsigned int dma_in_nbr;
107 unsigned int dma_out_nbr;
108
109 /* DMA owner. */
110 enum dma_owner req_dma;
111
112 char started; /* 1 if port has been started */
113 char port_nbr; /* Port 0 or 1 */
114 char busy; /* 1 if port is busy */
115
116 char enabled; /* 1 if port is enabled */
117 char use_dma; /* 1 if port uses dma */
118 char tr_running;
119
120 enum syncser_irq_setup init_irqs;
121 int output;
122 int input;
123
124 /* Next byte to be read by application */
125 unsigned char *readp;
126 /* Next byte to be written by etrax */
127 unsigned char *writep;
128
129 unsigned int in_buffer_size;
130 unsigned int in_buffer_len;
131 unsigned int inbufchunk;
132 /* Data buffers for in and output. */
133 unsigned char out_buffer[OUT_BUFFER_SIZE] __aligned(32);
134 unsigned char in_buffer[IN_BUFFER_SIZE] __aligned(32);
135 unsigned char flip[IN_BUFFER_SIZE] __aligned(32);
136 struct timespec timestamp[NBR_IN_DESCR];
137 struct dma_descr_data *next_rx_desc;
138 struct dma_descr_data *prev_rx_desc;
139
140 struct timeval last_timestamp;
141 int read_ts_idx;
142 int write_ts_idx;
143
144 /* Pointer to the first available descriptor in the ring,
145 * unless active_tr_descr == catch_tr_descr and a dma
146 * transfer is active */
147 struct dma_descr_data *active_tr_descr;
148
149 /* Pointer to the first allocated descriptor in the ring */
150 struct dma_descr_data *catch_tr_descr;
151
152 /* Pointer to the descriptor with the current end-of-list */
153 struct dma_descr_data *prev_tr_descr;
154 int full;
155
156 /* Pointer to the first byte being read by DMA
157 * or current position in out_buffer if not using DMA. */
158 unsigned char *out_rd_ptr;
159
160 /* Number of bytes currently locked for being read by DMA */
161 int out_buf_count;
162
163 dma_descr_context in_context __aligned(32);
164 dma_descr_context out_context __aligned(32);
165 dma_descr_data in_descr[NBR_IN_DESCR] __aligned(16);
166 dma_descr_data out_descr[NBR_OUT_DESCR] __aligned(16);
167
168 wait_queue_head_t out_wait_q;
169 wait_queue_head_t in_wait_q;
170
171 spinlock_t lock;
172 };
173
174 static DEFINE_MUTEX(sync_serial_mutex);
175 static int etrax_sync_serial_init(void);
176 static void initialize_port(int portnbr);
177 static inline int sync_data_avail(struct sync_port *port);
178
179 static int sync_serial_open(struct inode *, struct file *);
180 static int sync_serial_release(struct inode *, struct file *);
181 static unsigned int sync_serial_poll(struct file *filp, poll_table *wait);
182
183 static long sync_serial_ioctl(struct file *file,
184 unsigned int cmd, unsigned long arg);
185 static int sync_serial_ioctl_unlocked(struct file *file,
186 unsigned int cmd, unsigned long arg);
187 static ssize_t sync_serial_write(struct file *file, const char __user *buf,
188 size_t count, loff_t *ppos);
189 static ssize_t sync_serial_read(struct file *file, char __user *buf,
190 size_t count, loff_t *ppos);
191
192 #if ((defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT0) && \
193 defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL0_DMA)) || \
194 (defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT1) && \
195 defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL1_DMA)))
196 #define SYNC_SER_DMA
197 #else
198 #define SYNC_SER_MANUAL
199 #endif
200
201 #ifdef SYNC_SER_DMA
202 static void start_dma_out(struct sync_port *port, const char *data, int count);
203 static void start_dma_in(struct sync_port *port);
204 static irqreturn_t tr_interrupt(int irq, void *dev_id);
205 static irqreturn_t rx_interrupt(int irq, void *dev_id);
206 #endif
207 #ifdef SYNC_SER_MANUAL
208 static void send_word(struct sync_port *port);
209 static irqreturn_t manual_interrupt(int irq, void *dev_id);
210 #endif
211
212 #define artpec_pinmux_alloc_fixed crisv32_pinmux_alloc_fixed
213 #define artpec_request_dma crisv32_request_dma
214 #define artpec_free_dma crisv32_free_dma
215
216 #ifdef CONFIG_ETRAXFS
217 /* ETRAX FS */
218 #define DMA_OUT_NBR0 SYNC_SER0_TX_DMA_NBR
219 #define DMA_IN_NBR0 SYNC_SER0_RX_DMA_NBR
220 #define DMA_OUT_NBR1 SYNC_SER1_TX_DMA_NBR
221 #define DMA_IN_NBR1 SYNC_SER1_RX_DMA_NBR
222 #define PINMUX_SSER0 pinmux_sser0
223 #define PINMUX_SSER1 pinmux_sser1
224 #define SYNCSER_INST0 regi_sser0
225 #define SYNCSER_INST1 regi_sser1
226 #define SYNCSER_INTR_VECT0 SSER0_INTR_VECT
227 #define SYNCSER_INTR_VECT1 SSER1_INTR_VECT
228 #define OUT_DMA_INST0 regi_dma4
229 #define IN_DMA_INST0 regi_dma5
230 #define DMA_OUT_INTR_VECT0 DMA4_INTR_VECT
231 #define DMA_OUT_INTR_VECT1 DMA7_INTR_VECT
232 #define DMA_IN_INTR_VECT0 DMA5_INTR_VECT
233 #define DMA_IN_INTR_VECT1 DMA6_INTR_VECT
234 #define REQ_DMA_SYNCSER0 dma_sser0
235 #define REQ_DMA_SYNCSER1 dma_sser1
236 #if defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL1_DMA)
237 #define PORT1_DMA 1
238 #else
239 #define PORT1_DMA 0
240 #endif
241 #elif defined(CONFIG_CRIS_MACH_ARTPEC3)
242 /* ARTPEC-3 */
243 #define DMA_OUT_NBR0 SYNC_SER_TX_DMA_NBR
244 #define DMA_IN_NBR0 SYNC_SER_RX_DMA_NBR
245 #define PINMUX_SSER0 pinmux_sser
246 #define SYNCSER_INST0 regi_sser
247 #define SYNCSER_INTR_VECT0 SSER_INTR_VECT
248 #define OUT_DMA_INST0 regi_dma6
249 #define IN_DMA_INST0 regi_dma7
250 #define DMA_OUT_INTR_VECT0 DMA6_INTR_VECT
251 #define DMA_IN_INTR_VECT0 DMA7_INTR_VECT
252 #define REQ_DMA_SYNCSER0 dma_sser
253 #define REQ_DMA_SYNCSER1 dma_sser
254 #endif
255
256 #if defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL0_DMA)
257 #define PORT0_DMA 1
258 #else
259 #define PORT0_DMA 0
260 #endif
261
262 /* The ports */
263 static struct sync_port ports[] = {
264 {
265 .regi_sser = SYNCSER_INST0,
266 .regi_dmaout = OUT_DMA_INST0,
267 .regi_dmain = IN_DMA_INST0,
268 .use_dma = PORT0_DMA,
269 .dma_in_intr_vect = DMA_IN_INTR_VECT0,
270 .dma_out_intr_vect = DMA_OUT_INTR_VECT0,
271 .dma_in_nbr = DMA_IN_NBR0,
272 .dma_out_nbr = DMA_OUT_NBR0,
273 .req_dma = REQ_DMA_SYNCSER0,
274 .syncser_intr_vect = SYNCSER_INTR_VECT0,
275 },
276 #ifdef CONFIG_ETRAXFS
277 {
278 .regi_sser = SYNCSER_INST1,
279 .regi_dmaout = regi_dma6,
280 .regi_dmain = regi_dma7,
281 .use_dma = PORT1_DMA,
282 .dma_in_intr_vect = DMA_IN_INTR_VECT1,
283 .dma_out_intr_vect = DMA_OUT_INTR_VECT1,
284 .dma_in_nbr = DMA_IN_NBR1,
285 .dma_out_nbr = DMA_OUT_NBR1,
286 .req_dma = REQ_DMA_SYNCSER1,
287 .syncser_intr_vect = SYNCSER_INTR_VECT1,
288 },
289 #endif
290 };
291
292 #define NBR_PORTS ARRAY_SIZE(ports)
293
294 static const struct file_operations syncser_fops = {
295 .owner = THIS_MODULE,
296 .write = sync_serial_write,
297 .read = sync_serial_read,
298 .poll = sync_serial_poll,
299 .unlocked_ioctl = sync_serial_ioctl,
300 .open = sync_serial_open,
301 .release = sync_serial_release,
302 .llseek = noop_llseek,
303 };
304
305 static dev_t syncser_first;
306 static int minor_count = NBR_PORTS;
307 #define SYNCSER_NAME "syncser"
308 static struct cdev *syncser_cdev;
309 static struct class *syncser_class;
310
311 static void sync_serial_start_port(struct sync_port *port)
312 {
313 reg_sser_rw_cfg cfg = REG_RD(sser, port->regi_sser, rw_cfg);
314 reg_sser_rw_tr_cfg tr_cfg =
315 REG_RD(sser, port->regi_sser, rw_tr_cfg);
316 reg_sser_rw_rec_cfg rec_cfg =
317 REG_RD(sser, port->regi_sser, rw_rec_cfg);
318 cfg.en = regk_sser_yes;
319 tr_cfg.tr_en = regk_sser_yes;
320 rec_cfg.rec_en = regk_sser_yes;
321 REG_WR(sser, port->regi_sser, rw_cfg, cfg);
322 REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
323 REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
324 port->started = 1;
325 }
326
327 static void __init initialize_port(int portnbr)
328 {
329 struct sync_port *port = &ports[portnbr];
330 reg_sser_rw_cfg cfg = { 0 };
331 reg_sser_rw_frm_cfg frm_cfg = { 0 };
332 reg_sser_rw_tr_cfg tr_cfg = { 0 };
333 reg_sser_rw_rec_cfg rec_cfg = { 0 };
334
335 DEBUG(pr_info("Init sync serial port %d\n", portnbr));
336
337 port->port_nbr = portnbr;
338 port->init_irqs = no_irq_setup;
339
340 port->out_rd_ptr = port->out_buffer;
341 port->out_buf_count = 0;
342
343 port->output = 1;
344 port->input = 0;
345
346 port->readp = port->flip;
347 port->writep = port->flip;
348 port->in_buffer_size = IN_BUFFER_SIZE;
349 port->in_buffer_len = 0;
350 port->inbufchunk = IN_DESCR_SIZE;
351
352 port->read_ts_idx = 0;
353 port->write_ts_idx = 0;
354
355 init_waitqueue_head(&port->out_wait_q);
356 init_waitqueue_head(&port->in_wait_q);
357
358 spin_lock_init(&port->lock);
359
360 cfg.out_clk_src = regk_sser_intern_clk;
361 cfg.out_clk_pol = regk_sser_pos;
362 cfg.clk_od_mode = regk_sser_no;
363 cfg.clk_dir = regk_sser_out;
364 cfg.gate_clk = regk_sser_no;
365 cfg.base_freq = regk_sser_f29_493;
366 cfg.clk_div = 256;
367 REG_WR(sser, port->regi_sser, rw_cfg, cfg);
368
369 frm_cfg.wordrate = DEFAULT_WORD_RATE;
370 frm_cfg.type = regk_sser_edge;
371 frm_cfg.frame_pin_dir = regk_sser_out;
372 frm_cfg.frame_pin_use = regk_sser_frm;
373 frm_cfg.status_pin_dir = regk_sser_in;
374 frm_cfg.status_pin_use = regk_sser_hold;
375 frm_cfg.out_on = regk_sser_tr;
376 frm_cfg.tr_delay = 1;
377 REG_WR(sser, port->regi_sser, rw_frm_cfg, frm_cfg);
378
379 tr_cfg.urun_stop = regk_sser_no;
380 tr_cfg.sample_size = 7;
381 tr_cfg.sh_dir = regk_sser_msbfirst;
382 tr_cfg.use_dma = port->use_dma ? regk_sser_yes : regk_sser_no;
383 #if 0
384 tr_cfg.rate_ctrl = regk_sser_bulk;
385 tr_cfg.data_pin_use = regk_sser_dout;
386 #else
387 tr_cfg.rate_ctrl = regk_sser_iso;
388 tr_cfg.data_pin_use = regk_sser_dout;
389 #endif
390 tr_cfg.bulk_wspace = 1;
391 REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
392
393 rec_cfg.sample_size = 7;
394 rec_cfg.sh_dir = regk_sser_msbfirst;
395 rec_cfg.use_dma = port->use_dma ? regk_sser_yes : regk_sser_no;
396 rec_cfg.fifo_thr = regk_sser_inf;
397 REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
398
399 #ifdef SYNC_SER_DMA
400 {
401 int i;
402 /* Setup the descriptor ring for dma out/transmit. */
403 for (i = 0; i < NBR_OUT_DESCR; i++) {
404 dma_descr_data *descr = &port->out_descr[i];
405 descr->wait = 0;
406 descr->intr = 1;
407 descr->eol = 0;
408 descr->out_eop = 0;
409 descr->next =
410 (dma_descr_data *)virt_to_phys(&descr[i+1]);
411 }
412 }
413
414 /* Create a ring from the list. */
415 port->out_descr[NBR_OUT_DESCR-1].next =
416 (dma_descr_data *)virt_to_phys(&port->out_descr[0]);
417
418 /* Setup context for traversing the ring. */
419 port->active_tr_descr = &port->out_descr[0];
420 port->prev_tr_descr = &port->out_descr[NBR_OUT_DESCR-1];
421 port->catch_tr_descr = &port->out_descr[0];
422 #endif
423 }
424
425 static inline int sync_data_avail(struct sync_port *port)
426 {
427 return port->in_buffer_len;
428 }
429
430 static int sync_serial_open(struct inode *inode, struct file *file)
431 {
432 int ret = 0;
433 int dev = iminor(inode);
434 struct sync_port *port;
435 #ifdef SYNC_SER_DMA
436 reg_dma_rw_cfg cfg = { .en = regk_dma_yes };
437 reg_dma_rw_intr_mask intr_mask = { .data = regk_dma_yes };
438 #endif
439
440 DEBUG(pr_debug("Open sync serial port %d\n", dev));
441
442 if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled) {
443 DEBUG(pr_info("Invalid minor %d\n", dev));
444 return -ENODEV;
445 }
446 port = &ports[dev];
447 /* Allow open this device twice (assuming one reader and one writer) */
448 if (port->busy == 2) {
449 DEBUG(pr_info("syncser%d is busy\n", dev));
450 return -EBUSY;
451 }
452
453 mutex_lock(&sync_serial_mutex);
454
455 /* Clear any stale date left in the flip buffer */
456 port->readp = port->writep = port->flip;
457 port->in_buffer_len = 0;
458 port->read_ts_idx = 0;
459 port->write_ts_idx = 0;
460
461 if (port->init_irqs != no_irq_setup) {
462 /* Init only on first call. */
463 port->busy++;
464 mutex_unlock(&sync_serial_mutex);
465 return 0;
466 }
467 if (port->use_dma) {
468 #ifdef SYNC_SER_DMA
469 const char *tmp;
470 DEBUG(pr_info("Using DMA for syncser%d\n", dev));
471
472 tmp = dev == 0 ? "syncser0 tx" : "syncser1 tx";
473 if (request_irq(port->dma_out_intr_vect, tr_interrupt, 0,
474 tmp, port)) {
475 pr_err("Can't alloc syncser%d TX IRQ", dev);
476 ret = -EBUSY;
477 goto unlock_and_exit;
478 }
479 if (artpec_request_dma(port->dma_out_nbr, tmp,
480 DMA_VERBOSE_ON_ERROR, 0, port->req_dma)) {
481 free_irq(port->dma_out_intr_vect, port);
482 pr_err("Can't alloc syncser%d TX DMA", dev);
483 ret = -EBUSY;
484 goto unlock_and_exit;
485 }
486 tmp = dev == 0 ? "syncser0 rx" : "syncser1 rx";
487 if (request_irq(port->dma_in_intr_vect, rx_interrupt, 0,
488 tmp, port)) {
489 artpec_free_dma(port->dma_out_nbr);
490 free_irq(port->dma_out_intr_vect, port);
491 pr_err("Can't alloc syncser%d RX IRQ", dev);
492 ret = -EBUSY;
493 goto unlock_and_exit;
494 }
495 if (artpec_request_dma(port->dma_in_nbr, tmp,
496 DMA_VERBOSE_ON_ERROR, 0, port->req_dma)) {
497 artpec_free_dma(port->dma_out_nbr);
498 free_irq(port->dma_out_intr_vect, port);
499 free_irq(port->dma_in_intr_vect, port);
500 pr_err("Can't alloc syncser%d RX DMA", dev);
501 ret = -EBUSY;
502 goto unlock_and_exit;
503 }
504 /* Enable DMAs */
505 REG_WR(dma, port->regi_dmain, rw_cfg, cfg);
506 REG_WR(dma, port->regi_dmaout, rw_cfg, cfg);
507 /* Enable DMA IRQs */
508 REG_WR(dma, port->regi_dmain, rw_intr_mask, intr_mask);
509 REG_WR(dma, port->regi_dmaout, rw_intr_mask, intr_mask);
510 /* Set up wordsize = 1 for DMAs. */
511 DMA_WR_CMD(port->regi_dmain, regk_dma_set_w_size1);
512 DMA_WR_CMD(port->regi_dmaout, regk_dma_set_w_size1);
513
514 start_dma_in(port);
515 port->init_irqs = dma_irq_setup;
516 #endif
517 } else { /* !port->use_dma */
518 #ifdef SYNC_SER_MANUAL
519 const char *tmp = dev == 0 ? "syncser0 manual irq" :
520 "syncser1 manual irq";
521 if (request_irq(port->syncser_intr_vect, manual_interrupt,
522 0, tmp, port)) {
523 pr_err("Can't alloc syncser%d manual irq",
524 dev);
525 ret = -EBUSY;
526 goto unlock_and_exit;
527 }
528 port->init_irqs = manual_irq_setup;
529 #else
530 panic("sync_serial: Manual mode not supported\n");
531 #endif /* SYNC_SER_MANUAL */
532 }
533 port->busy++;
534 ret = 0;
535
536 unlock_and_exit:
537 mutex_unlock(&sync_serial_mutex);
538 return ret;
539 }
540
541 static int sync_serial_release(struct inode *inode, struct file *file)
542 {
543 int dev = iminor(inode);
544 struct sync_port *port;
545
546 if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled) {
547 DEBUG(pr_info("Invalid minor %d\n", dev));
548 return -ENODEV;
549 }
550 port = &ports[dev];
551 if (port->busy)
552 port->busy--;
553 if (!port->busy)
554 /* XXX */;
555 return 0;
556 }
557
558 static unsigned int sync_serial_poll(struct file *file, poll_table *wait)
559 {
560 int dev = iminor(file_inode(file));
561 unsigned int mask = 0;
562 struct sync_port *port;
563 DEBUGPOLL(
564 static unsigned int prev_mask;
565 );
566
567 port = &ports[dev];
568
569 if (!port->started)
570 sync_serial_start_port(port);
571
572 poll_wait(file, &port->out_wait_q, wait);
573 poll_wait(file, &port->in_wait_q, wait);
574
575 /* No active transfer, descriptors are available */
576 if (port->output && !port->tr_running)
577 mask |= POLLOUT | POLLWRNORM;
578
579 /* Descriptor and buffer space available. */
580 if (port->output &&
581 port->active_tr_descr != port->catch_tr_descr &&
582 port->out_buf_count < OUT_BUFFER_SIZE)
583 mask |= POLLOUT | POLLWRNORM;
584
585 /* At least an inbufchunk of data */
586 if (port->input && sync_data_avail(port) >= port->inbufchunk)
587 mask |= POLLIN | POLLRDNORM;
588
589 DEBUGPOLL(
590 if (mask != prev_mask)
591 pr_info("sync_serial_poll: mask 0x%08X %s %s\n",
592 mask,
593 mask & POLLOUT ? "POLLOUT" : "",
594 mask & POLLIN ? "POLLIN" : "");
595 prev_mask = mask;
596 );
597 return mask;
598 }
599
600 static ssize_t __sync_serial_read(struct file *file,
601 char __user *buf,
602 size_t count,
603 loff_t *ppos,
604 struct timespec *ts)
605 {
606 unsigned long flags;
607 int dev = MINOR(file_inode(file)->i_rdev);
608 int avail;
609 struct sync_port *port;
610 unsigned char *start;
611 unsigned char *end;
612
613 if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled) {
614 DEBUG(pr_info("Invalid minor %d\n", dev));
615 return -ENODEV;
616 }
617 port = &ports[dev];
618
619 if (!port->started)
620 sync_serial_start_port(port);
621
622 /* Calculate number of available bytes */
623 /* Save pointers to avoid that they are modified by interrupt */
624 spin_lock_irqsave(&port->lock, flags);
625 start = port->readp;
626 end = port->writep;
627 spin_unlock_irqrestore(&port->lock, flags);
628
629 while ((start == end) && !port->in_buffer_len) {
630 if (file->f_flags & O_NONBLOCK)
631 return -EAGAIN;
632
633 wait_event_interruptible(port->in_wait_q,
634 !(start == end && !port->full));
635
636 if (signal_pending(current))
637 return -EINTR;
638
639 spin_lock_irqsave(&port->lock, flags);
640 start = port->readp;
641 end = port->writep;
642 spin_unlock_irqrestore(&port->lock, flags);
643 }
644
645 DEBUGREAD(pr_info("R%d c %d ri %u wi %u /%u\n",
646 dev, count,
647 start - port->flip, end - port->flip,
648 port->in_buffer_size));
649
650 /* Lazy read, never return wrapped data. */
651 if (end > start)
652 avail = end - start;
653 else
654 avail = port->flip + port->in_buffer_size - start;
655
656 count = count > avail ? avail : count;
657 if (copy_to_user(buf, start, count))
658 return -EFAULT;
659
660 /* If timestamp requested, find timestamp of first returned byte
661 * and copy it.
662 * N.B: Applications that request timstamps MUST read data in
663 * chunks that are multiples of IN_DESCR_SIZE.
664 * Otherwise the timestamps will not be aligned to the data read.
665 */
666 if (ts != NULL) {
667 int idx = port->read_ts_idx;
668 memcpy(ts, &port->timestamp[idx], sizeof(struct timespec));
669 port->read_ts_idx += count / IN_DESCR_SIZE;
670 if (port->read_ts_idx >= NBR_IN_DESCR)
671 port->read_ts_idx = 0;
672 }
673
674 spin_lock_irqsave(&port->lock, flags);
675 port->readp += count;
676 /* Check for wrap */
677 if (port->readp >= port->flip + port->in_buffer_size)
678 port->readp = port->flip;
679 port->in_buffer_len -= count;
680 port->full = 0;
681 spin_unlock_irqrestore(&port->lock, flags);
682
683 DEBUGREAD(pr_info("r %d\n", count));
684
685 return count;
686 }
687
688 static ssize_t sync_serial_input(struct file *file, unsigned long arg)
689 {
690 struct ssp_request req;
691 int count;
692 int ret;
693
694 /* Copy the request structure from user-mode. */
695 ret = copy_from_user(&req, (struct ssp_request __user *)arg,
696 sizeof(struct ssp_request));
697
698 if (ret) {
699 DEBUG(pr_info("sync_serial_input copy from user failed\n"));
700 return -EFAULT;
701 }
702
703 /* To get the timestamps aligned, make sure that 'len'
704 * is a multiple of IN_DESCR_SIZE.
705 */
706 if ((req.len % IN_DESCR_SIZE) != 0) {
707 DEBUG(pr_info("sync_serial: req.len %x, IN_DESCR_SIZE %x\n",
708 req.len, IN_DESCR_SIZE));
709 return -EFAULT;
710 }
711
712 /* Do the actual read. */
713 /* Note that req.buf is actually a pointer to user space. */
714 count = __sync_serial_read(file, req.buf, req.len,
715 NULL, &req.ts);
716
717 if (count < 0) {
718 DEBUG(pr_info("sync_serial_input read failed\n"));
719 return count;
720 }
721
722 /* Copy the request back to user-mode. */
723 ret = copy_to_user((struct ssp_request __user *)arg, &req,
724 sizeof(struct ssp_request));
725
726 if (ret) {
727 DEBUG(pr_info("syncser input copy2user failed\n"));
728 return -EFAULT;
729 }
730
731 /* Return the number of bytes read. */
732 return count;
733 }
734
735
736 static int sync_serial_ioctl_unlocked(struct file *file,
737 unsigned int cmd, unsigned long arg)
738 {
739 int return_val = 0;
740 int dma_w_size = regk_dma_set_w_size1;
741 int dev = iminor(file_inode(file));
742 struct sync_port *port;
743 reg_sser_rw_tr_cfg tr_cfg;
744 reg_sser_rw_rec_cfg rec_cfg;
745 reg_sser_rw_frm_cfg frm_cfg;
746 reg_sser_rw_cfg gen_cfg;
747 reg_sser_rw_intr_mask intr_mask;
748
749 if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled) {
750 DEBUG(pr_info("Invalid minor %d\n", dev));
751 return -1;
752 }
753
754 if (cmd == SSP_INPUT)
755 return sync_serial_input(file, arg);
756
757 port = &ports[dev];
758 spin_lock_irq(&port->lock);
759
760 tr_cfg = REG_RD(sser, port->regi_sser, rw_tr_cfg);
761 rec_cfg = REG_RD(sser, port->regi_sser, rw_rec_cfg);
762 frm_cfg = REG_RD(sser, port->regi_sser, rw_frm_cfg);
763 gen_cfg = REG_RD(sser, port->regi_sser, rw_cfg);
764 intr_mask = REG_RD(sser, port->regi_sser, rw_intr_mask);
765
766 switch (cmd) {
767 case SSP_SPEED:
768 if (GET_SPEED(arg) == CODEC) {
769 unsigned int freq;
770
771 gen_cfg.base_freq = regk_sser_f32;
772
773 /* Clock divider will internally be
774 * gen_cfg.clk_div + 1.
775 */
776
777 freq = GET_FREQ(arg);
778 switch (freq) {
779 case FREQ_32kHz:
780 case FREQ_64kHz:
781 case FREQ_128kHz:
782 case FREQ_256kHz:
783 gen_cfg.clk_div = 125 *
784 (1 << (freq - FREQ_256kHz)) - 1;
785 break;
786 case FREQ_512kHz:
787 gen_cfg.clk_div = 62;
788 break;
789 case FREQ_1MHz:
790 case FREQ_2MHz:
791 case FREQ_4MHz:
792 gen_cfg.clk_div = 8 * (1 << freq) - 1;
793 break;
794 }
795 } else if (GET_SPEED(arg) == CODEC_f32768) {
796 gen_cfg.base_freq = regk_sser_f32_768;
797 switch (GET_FREQ(arg)) {
798 case FREQ_4096kHz:
799 gen_cfg.clk_div = 7;
800 break;
801 default:
802 spin_unlock_irq(&port->lock);
803 return -EINVAL;
804 }
805 } else {
806 gen_cfg.base_freq = regk_sser_f29_493;
807 switch (GET_SPEED(arg)) {
808 case SSP150:
809 gen_cfg.clk_div = 29493000 / (150 * 8) - 1;
810 break;
811 case SSP300:
812 gen_cfg.clk_div = 29493000 / (300 * 8) - 1;
813 break;
814 case SSP600:
815 gen_cfg.clk_div = 29493000 / (600 * 8) - 1;
816 break;
817 case SSP1200:
818 gen_cfg.clk_div = 29493000 / (1200 * 8) - 1;
819 break;
820 case SSP2400:
821 gen_cfg.clk_div = 29493000 / (2400 * 8) - 1;
822 break;
823 case SSP4800:
824 gen_cfg.clk_div = 29493000 / (4800 * 8) - 1;
825 break;
826 case SSP9600:
827 gen_cfg.clk_div = 29493000 / (9600 * 8) - 1;
828 break;
829 case SSP19200:
830 gen_cfg.clk_div = 29493000 / (19200 * 8) - 1;
831 break;
832 case SSP28800:
833 gen_cfg.clk_div = 29493000 / (28800 * 8) - 1;
834 break;
835 case SSP57600:
836 gen_cfg.clk_div = 29493000 / (57600 * 8) - 1;
837 break;
838 case SSP115200:
839 gen_cfg.clk_div = 29493000 / (115200 * 8) - 1;
840 break;
841 case SSP230400:
842 gen_cfg.clk_div = 29493000 / (230400 * 8) - 1;
843 break;
844 case SSP460800:
845 gen_cfg.clk_div = 29493000 / (460800 * 8) - 1;
846 break;
847 case SSP921600:
848 gen_cfg.clk_div = 29493000 / (921600 * 8) - 1;
849 break;
850 case SSP3125000:
851 gen_cfg.base_freq = regk_sser_f100;
852 gen_cfg.clk_div = 100000000 / (3125000 * 8) - 1;
853 break;
854
855 }
856 }
857 frm_cfg.wordrate = GET_WORD_RATE(arg);
858
859 break;
860 case SSP_MODE:
861 switch (arg) {
862 case MASTER_OUTPUT:
863 port->output = 1;
864 port->input = 0;
865 frm_cfg.out_on = regk_sser_tr;
866 frm_cfg.frame_pin_dir = regk_sser_out;
867 gen_cfg.clk_dir = regk_sser_out;
868 break;
869 case SLAVE_OUTPUT:
870 port->output = 1;
871 port->input = 0;
872 frm_cfg.frame_pin_dir = regk_sser_in;
873 gen_cfg.clk_dir = regk_sser_in;
874 break;
875 case MASTER_INPUT:
876 port->output = 0;
877 port->input = 1;
878 frm_cfg.frame_pin_dir = regk_sser_out;
879 frm_cfg.out_on = regk_sser_intern_tb;
880 gen_cfg.clk_dir = regk_sser_out;
881 break;
882 case SLAVE_INPUT:
883 port->output = 0;
884 port->input = 1;
885 frm_cfg.frame_pin_dir = regk_sser_in;
886 gen_cfg.clk_dir = regk_sser_in;
887 break;
888 case MASTER_BIDIR:
889 port->output = 1;
890 port->input = 1;
891 frm_cfg.frame_pin_dir = regk_sser_out;
892 frm_cfg.out_on = regk_sser_intern_tb;
893 gen_cfg.clk_dir = regk_sser_out;
894 break;
895 case SLAVE_BIDIR:
896 port->output = 1;
897 port->input = 1;
898 frm_cfg.frame_pin_dir = regk_sser_in;
899 gen_cfg.clk_dir = regk_sser_in;
900 break;
901 default:
902 spin_unlock_irq(&port->lock);
903 return -EINVAL;
904 }
905 if (!port->use_dma || arg == MASTER_OUTPUT ||
906 arg == SLAVE_OUTPUT)
907 intr_mask.rdav = regk_sser_yes;
908 break;
909 case SSP_FRAME_SYNC:
910 if (arg & NORMAL_SYNC) {
911 frm_cfg.rec_delay = 1;
912 frm_cfg.tr_delay = 1;
913 } else if (arg & EARLY_SYNC)
914 frm_cfg.rec_delay = frm_cfg.tr_delay = 0;
915 else if (arg & LATE_SYNC) {
916 frm_cfg.tr_delay = 2;
917 frm_cfg.rec_delay = 2;
918 } else if (arg & SECOND_WORD_SYNC) {
919 frm_cfg.rec_delay = 7;
920 frm_cfg.tr_delay = 1;
921 }
922
923 tr_cfg.bulk_wspace = frm_cfg.tr_delay;
924 frm_cfg.early_wend = regk_sser_yes;
925 if (arg & BIT_SYNC)
926 frm_cfg.type = regk_sser_edge;
927 else if (arg & WORD_SYNC)
928 frm_cfg.type = regk_sser_level;
929 else if (arg & EXTENDED_SYNC)
930 frm_cfg.early_wend = regk_sser_no;
931
932 if (arg & SYNC_ON)
933 frm_cfg.frame_pin_use = regk_sser_frm;
934 else if (arg & SYNC_OFF)
935 frm_cfg.frame_pin_use = regk_sser_gio0;
936
937 dma_w_size = regk_dma_set_w_size2;
938 if (arg & WORD_SIZE_8) {
939 rec_cfg.sample_size = tr_cfg.sample_size = 7;
940 dma_w_size = regk_dma_set_w_size1;
941 } else if (arg & WORD_SIZE_12)
942 rec_cfg.sample_size = tr_cfg.sample_size = 11;
943 else if (arg & WORD_SIZE_16)
944 rec_cfg.sample_size = tr_cfg.sample_size = 15;
945 else if (arg & WORD_SIZE_24)
946 rec_cfg.sample_size = tr_cfg.sample_size = 23;
947 else if (arg & WORD_SIZE_32)
948 rec_cfg.sample_size = tr_cfg.sample_size = 31;
949
950 if (arg & BIT_ORDER_MSB)
951 rec_cfg.sh_dir = tr_cfg.sh_dir = regk_sser_msbfirst;
952 else if (arg & BIT_ORDER_LSB)
953 rec_cfg.sh_dir = tr_cfg.sh_dir = regk_sser_lsbfirst;
954
955 if (arg & FLOW_CONTROL_ENABLE) {
956 frm_cfg.status_pin_use = regk_sser_frm;
957 rec_cfg.fifo_thr = regk_sser_thr16;
958 } else if (arg & FLOW_CONTROL_DISABLE) {
959 frm_cfg.status_pin_use = regk_sser_gio0;
960 rec_cfg.fifo_thr = regk_sser_inf;
961 }
962
963 if (arg & CLOCK_NOT_GATED)
964 gen_cfg.gate_clk = regk_sser_no;
965 else if (arg & CLOCK_GATED)
966 gen_cfg.gate_clk = regk_sser_yes;
967
968 break;
969 case SSP_IPOLARITY:
970 /* NOTE!! negedge is considered NORMAL */
971 if (arg & CLOCK_NORMAL)
972 rec_cfg.clk_pol = regk_sser_neg;
973 else if (arg & CLOCK_INVERT)
974 rec_cfg.clk_pol = regk_sser_pos;
975
976 if (arg & FRAME_NORMAL)
977 frm_cfg.level = regk_sser_pos_hi;
978 else if (arg & FRAME_INVERT)
979 frm_cfg.level = regk_sser_neg_lo;
980
981 if (arg & STATUS_NORMAL)
982 gen_cfg.hold_pol = regk_sser_pos;
983 else if (arg & STATUS_INVERT)
984 gen_cfg.hold_pol = regk_sser_neg;
985 break;
986 case SSP_OPOLARITY:
987 if (arg & CLOCK_NORMAL)
988 gen_cfg.out_clk_pol = regk_sser_pos;
989 else if (arg & CLOCK_INVERT)
990 gen_cfg.out_clk_pol = regk_sser_neg;
991
992 if (arg & FRAME_NORMAL)
993 frm_cfg.level = regk_sser_pos_hi;
994 else if (arg & FRAME_INVERT)
995 frm_cfg.level = regk_sser_neg_lo;
996
997 if (arg & STATUS_NORMAL)
998 gen_cfg.hold_pol = regk_sser_pos;
999 else if (arg & STATUS_INVERT)
1000 gen_cfg.hold_pol = regk_sser_neg;
1001 break;
1002 case SSP_SPI:
1003 rec_cfg.fifo_thr = regk_sser_inf;
1004 rec_cfg.sh_dir = tr_cfg.sh_dir = regk_sser_msbfirst;
1005 rec_cfg.sample_size = tr_cfg.sample_size = 7;
1006 frm_cfg.frame_pin_use = regk_sser_frm;
1007 frm_cfg.type = regk_sser_level;
1008 frm_cfg.tr_delay = 1;
1009 frm_cfg.level = regk_sser_neg_lo;
1010 if (arg & SPI_SLAVE) {
1011 rec_cfg.clk_pol = regk_sser_neg;
1012 gen_cfg.clk_dir = regk_sser_in;
1013 port->input = 1;
1014 port->output = 0;
1015 } else {
1016 gen_cfg.out_clk_pol = regk_sser_pos;
1017 port->input = 0;
1018 port->output = 1;
1019 gen_cfg.clk_dir = regk_sser_out;
1020 }
1021 break;
1022 case SSP_INBUFCHUNK:
1023 break;
1024 default:
1025 return_val = -1;
1026 }
1027
1028
1029 if (port->started) {
1030 rec_cfg.rec_en = port->input;
1031 gen_cfg.en = (port->output | port->input);
1032 }
1033
1034 REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
1035 REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
1036 REG_WR(sser, port->regi_sser, rw_frm_cfg, frm_cfg);
1037 REG_WR(sser, port->regi_sser, rw_intr_mask, intr_mask);
1038 REG_WR(sser, port->regi_sser, rw_cfg, gen_cfg);
1039
1040
1041 if (cmd == SSP_FRAME_SYNC && (arg & (WORD_SIZE_8 | WORD_SIZE_12 |
1042 WORD_SIZE_16 | WORD_SIZE_24 | WORD_SIZE_32))) {
1043 int en = gen_cfg.en;
1044 gen_cfg.en = 0;
1045 REG_WR(sser, port->regi_sser, rw_cfg, gen_cfg);
1046 /* ##### Should DMA be stoped before we change dma size? */
1047 DMA_WR_CMD(port->regi_dmain, dma_w_size);
1048 DMA_WR_CMD(port->regi_dmaout, dma_w_size);
1049 gen_cfg.en = en;
1050 REG_WR(sser, port->regi_sser, rw_cfg, gen_cfg);
1051 }
1052
1053 spin_unlock_irq(&port->lock);
1054 return return_val;
1055 }
1056
1057 static long sync_serial_ioctl(struct file *file,
1058 unsigned int cmd, unsigned long arg)
1059 {
1060 long ret;
1061
1062 mutex_lock(&sync_serial_mutex);
1063 ret = sync_serial_ioctl_unlocked(file, cmd, arg);
1064 mutex_unlock(&sync_serial_mutex);
1065
1066 return ret;
1067 }
1068
1069 /* NOTE: sync_serial_write does not support concurrency */
1070 static ssize_t sync_serial_write(struct file *file, const char __user *buf,
1071 size_t count, loff_t *ppos)
1072 {
1073 int dev = iminor(file_inode(file));
1074 DECLARE_WAITQUEUE(wait, current);
1075 struct sync_port *port;
1076 int trunc_count;
1077 unsigned long flags;
1078 int bytes_free;
1079 int out_buf_count;
1080
1081 unsigned char *rd_ptr; /* First allocated byte in the buffer */
1082 unsigned char *wr_ptr; /* First free byte in the buffer */
1083 unsigned char *buf_stop_ptr; /* Last byte + 1 */
1084
1085 if (dev < 0 || dev >= NBR_PORTS || !ports[dev].enabled) {
1086 DEBUG(pr_info("Invalid minor %d\n", dev));
1087 return -ENODEV;
1088 }
1089 port = &ports[dev];
1090
1091 /* |<- OUT_BUFFER_SIZE ->|
1092 * |<- out_buf_count ->|
1093 * |<- trunc_count ->| ...->|
1094 * ______________________________________________________
1095 * | free | data | free |
1096 * |_________|___________________|________________________|
1097 * ^ rd_ptr ^ wr_ptr
1098 */
1099 DEBUGWRITE(pr_info("W d%d c %u a: %p c: %p\n",
1100 port->port_nbr, count, port->active_tr_descr,
1101 port->catch_tr_descr));
1102
1103 /* Read variables that may be updated by interrupts */
1104 spin_lock_irqsave(&port->lock, flags);
1105 rd_ptr = port->out_rd_ptr;
1106 out_buf_count = port->out_buf_count;
1107 spin_unlock_irqrestore(&port->lock, flags);
1108
1109 /* Check if resources are available */
1110 if (port->tr_running &&
1111 ((port->use_dma && port->active_tr_descr == port->catch_tr_descr) ||
1112 out_buf_count >= OUT_BUFFER_SIZE)) {
1113 DEBUGWRITE(pr_info("sser%d full\n", dev));
1114 return -EAGAIN;
1115 }
1116
1117 buf_stop_ptr = port->out_buffer + OUT_BUFFER_SIZE;
1118
1119 /* Determine pointer to the first free byte, before copying. */
1120 wr_ptr = rd_ptr + out_buf_count;
1121 if (wr_ptr >= buf_stop_ptr)
1122 wr_ptr -= OUT_BUFFER_SIZE;
1123
1124 /* If we wrap the ring buffer, let the user space program handle it by
1125 * truncating the data. This could be more elegant, small buffer
1126 * fragments may occur.
1127 */
1128 bytes_free = OUT_BUFFER_SIZE - out_buf_count;
1129 if (wr_ptr + bytes_free > buf_stop_ptr)
1130 bytes_free = buf_stop_ptr - wr_ptr;
1131 trunc_count = (count < bytes_free) ? count : bytes_free;
1132
1133 if (copy_from_user(wr_ptr, buf, trunc_count))
1134 return -EFAULT;
1135
1136 DEBUGOUTBUF(pr_info("%-4d + %-4d = %-4d %p %p %p\n",
1137 out_buf_count, trunc_count,
1138 port->out_buf_count, port->out_buffer,
1139 wr_ptr, buf_stop_ptr));
1140
1141 /* Make sure transmitter/receiver is running */
1142 if (!port->started) {
1143 reg_sser_rw_cfg cfg = REG_RD(sser, port->regi_sser, rw_cfg);
1144 reg_sser_rw_rec_cfg rec_cfg =
1145 REG_RD(sser, port->regi_sser, rw_rec_cfg);
1146 cfg.en = regk_sser_yes;
1147 rec_cfg.rec_en = port->input;
1148 REG_WR(sser, port->regi_sser, rw_cfg, cfg);
1149 REG_WR(sser, port->regi_sser, rw_rec_cfg, rec_cfg);
1150 port->started = 1;
1151 }
1152
1153 /* Setup wait if blocking */
1154 if (!(file->f_flags & O_NONBLOCK)) {
1155 add_wait_queue(&port->out_wait_q, &wait);
1156 set_current_state(TASK_INTERRUPTIBLE);
1157 }
1158
1159 spin_lock_irqsave(&port->lock, flags);
1160 port->out_buf_count += trunc_count;
1161 if (port->use_dma) {
1162 #ifdef SYNC_SER_DMA
1163 start_dma_out(port, wr_ptr, trunc_count);
1164 #endif
1165 } else if (!port->tr_running) {
1166 #ifdef SYNC_SER_MANUAL
1167 reg_sser_rw_intr_mask intr_mask;
1168 intr_mask = REG_RD(sser, port->regi_sser, rw_intr_mask);
1169 /* Start sender by writing data */
1170 send_word(port);
1171 /* and enable transmitter ready IRQ */
1172 intr_mask.trdy = 1;
1173 REG_WR(sser, port->regi_sser, rw_intr_mask, intr_mask);
1174 #endif
1175 }
1176 spin_unlock_irqrestore(&port->lock, flags);
1177
1178 /* Exit if non blocking */
1179 if (file->f_flags & O_NONBLOCK) {
1180 DEBUGWRITE(pr_info("w d%d c %u %08x\n",
1181 port->port_nbr, trunc_count,
1182 REG_RD_INT(dma, port->regi_dmaout, r_intr)));
1183 return trunc_count;
1184 }
1185
1186 schedule();
1187 remove_wait_queue(&port->out_wait_q, &wait);
1188
1189 if (signal_pending(current))
1190 return -EINTR;
1191
1192 DEBUGWRITE(pr_info("w d%d c %u\n", port->port_nbr, trunc_count));
1193 return trunc_count;
1194 }
1195
1196 static ssize_t sync_serial_read(struct file *file, char __user *buf,
1197 size_t count, loff_t *ppos)
1198 {
1199 return __sync_serial_read(file, buf, count, ppos, NULL);
1200 }
1201
1202 #ifdef SYNC_SER_MANUAL
1203 static void send_word(struct sync_port *port)
1204 {
1205 reg_sser_rw_tr_cfg tr_cfg = REG_RD(sser, port->regi_sser, rw_tr_cfg);
1206 reg_sser_rw_tr_data tr_data = {0};
1207
1208 switch (tr_cfg.sample_size) {
1209 case 8:
1210 port->out_buf_count--;
1211 tr_data.data = *port->out_rd_ptr++;
1212 REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
1213 if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
1214 port->out_rd_ptr = port->out_buffer;
1215 break;
1216 case 12:
1217 {
1218 int data = (*port->out_rd_ptr++) << 8;
1219 data |= *port->out_rd_ptr++;
1220 port->out_buf_count -= 2;
1221 tr_data.data = data;
1222 REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
1223 if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
1224 port->out_rd_ptr = port->out_buffer;
1225 break;
1226 }
1227 case 16:
1228 port->out_buf_count -= 2;
1229 tr_data.data = *(unsigned short *)port->out_rd_ptr;
1230 REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
1231 port->out_rd_ptr += 2;
1232 if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
1233 port->out_rd_ptr = port->out_buffer;
1234 break;
1235 case 24:
1236 port->out_buf_count -= 3;
1237 tr_data.data = *(unsigned short *)port->out_rd_ptr;
1238 REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
1239 port->out_rd_ptr += 2;
1240 tr_data.data = *port->out_rd_ptr++;
1241 REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
1242 if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
1243 port->out_rd_ptr = port->out_buffer;
1244 break;
1245 case 32:
1246 port->out_buf_count -= 4;
1247 tr_data.data = *(unsigned short *)port->out_rd_ptr;
1248 REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
1249 port->out_rd_ptr += 2;
1250 tr_data.data = *(unsigned short *)port->out_rd_ptr;
1251 REG_WR(sser, port->regi_sser, rw_tr_data, tr_data);
1252 port->out_rd_ptr += 2;
1253 if (port->out_rd_ptr >= port->out_buffer + OUT_BUFFER_SIZE)
1254 port->out_rd_ptr = port->out_buffer;
1255 break;
1256 }
1257 }
1258 #endif
1259
1260 #ifdef SYNC_SER_DMA
1261 static void start_dma_out(struct sync_port *port, const char *data, int count)
1262 {
1263 port->active_tr_descr->buf = (char *)virt_to_phys((char *)data);
1264 port->active_tr_descr->after = port->active_tr_descr->buf + count;
1265 port->active_tr_descr->intr = 1;
1266
1267 port->active_tr_descr->eol = 1;
1268 port->prev_tr_descr->eol = 0;
1269
1270 DEBUGTRDMA(pr_info("Inserting eolr:%p eol@:%p\n",
1271 port->prev_tr_descr, port->active_tr_descr));
1272 port->prev_tr_descr = port->active_tr_descr;
1273 port->active_tr_descr = phys_to_virt((int)port->active_tr_descr->next);
1274
1275 if (!port->tr_running) {
1276 reg_sser_rw_tr_cfg tr_cfg = REG_RD(sser, port->regi_sser,
1277 rw_tr_cfg);
1278
1279 port->out_context.next = NULL;
1280 port->out_context.saved_data =
1281 (dma_descr_data *)virt_to_phys(port->prev_tr_descr);
1282 port->out_context.saved_data_buf = port->prev_tr_descr->buf;
1283
1284 DMA_START_CONTEXT(port->regi_dmaout,
1285 virt_to_phys((char *)&port->out_context));
1286
1287 tr_cfg.tr_en = regk_sser_yes;
1288 REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
1289 DEBUGTRDMA(pr_info("dma s\n"););
1290 } else {
1291 DMA_CONTINUE_DATA(port->regi_dmaout);
1292 DEBUGTRDMA(pr_info("dma c\n"););
1293 }
1294
1295 port->tr_running = 1;
1296 }
1297
1298 static void start_dma_in(struct sync_port *port)
1299 {
1300 int i;
1301 char *buf;
1302 unsigned long flags;
1303 spin_lock_irqsave(&port->lock, flags);
1304 port->writep = port->flip;
1305 spin_unlock_irqrestore(&port->lock, flags);
1306
1307 buf = (char *)virt_to_phys(port->in_buffer);
1308 for (i = 0; i < NBR_IN_DESCR; i++) {
1309 port->in_descr[i].buf = buf;
1310 port->in_descr[i].after = buf + port->inbufchunk;
1311 port->in_descr[i].intr = 1;
1312 port->in_descr[i].next =
1313 (dma_descr_data *)virt_to_phys(&port->in_descr[i+1]);
1314 port->in_descr[i].buf = buf;
1315 buf += port->inbufchunk;
1316 }
1317 /* Link the last descriptor to the first */
1318 port->in_descr[i-1].next =
1319 (dma_descr_data *)virt_to_phys(&port->in_descr[0]);
1320 port->in_descr[i-1].eol = regk_sser_yes;
1321 port->next_rx_desc = &port->in_descr[0];
1322 port->prev_rx_desc = &port->in_descr[NBR_IN_DESCR - 1];
1323 port->in_context.saved_data =
1324 (dma_descr_data *)virt_to_phys(&port->in_descr[0]);
1325 port->in_context.saved_data_buf = port->in_descr[0].buf;
1326 DMA_START_CONTEXT(port->regi_dmain, virt_to_phys(&port->in_context));
1327 }
1328
1329 static irqreturn_t tr_interrupt(int irq, void *dev_id)
1330 {
1331 reg_dma_r_masked_intr masked;
1332 reg_dma_rw_ack_intr ack_intr = { .data = regk_dma_yes };
1333 reg_dma_rw_stat stat;
1334 int i;
1335 int found = 0;
1336 int stop_sser = 0;
1337
1338 for (i = 0; i < NBR_PORTS; i++) {
1339 struct sync_port *port = &ports[i];
1340 if (!port->enabled || !port->use_dma)
1341 continue;
1342
1343 /* IRQ active for the port? */
1344 masked = REG_RD(dma, port->regi_dmaout, r_masked_intr);
1345 if (!masked.data)
1346 continue;
1347
1348 found = 1;
1349
1350 /* Check if we should stop the DMA transfer */
1351 stat = REG_RD(dma, port->regi_dmaout, rw_stat);
1352 if (stat.list_state == regk_dma_data_at_eol)
1353 stop_sser = 1;
1354
1355 /* Clear IRQ */
1356 REG_WR(dma, port->regi_dmaout, rw_ack_intr, ack_intr);
1357
1358 if (!stop_sser) {
1359 /* The DMA has completed a descriptor, EOL was not
1360 * encountered, so step relevant descriptor and
1361 * datapointers forward. */
1362 int sent;
1363 sent = port->catch_tr_descr->after -
1364 port->catch_tr_descr->buf;
1365 DEBUGTXINT(pr_info("%-4d - %-4d = %-4d\t"
1366 "in descr %p (ac: %p)\n",
1367 port->out_buf_count, sent,
1368 port->out_buf_count - sent,
1369 port->catch_tr_descr,
1370 port->active_tr_descr););
1371 port->out_buf_count -= sent;
1372 port->catch_tr_descr =
1373 phys_to_virt((int) port->catch_tr_descr->next);
1374 port->out_rd_ptr =
1375 phys_to_virt((int) port->catch_tr_descr->buf);
1376 } else {
1377 reg_sser_rw_tr_cfg tr_cfg;
1378 int j, sent;
1379 /* EOL handler.
1380 * Note that if an EOL was encountered during the irq
1381 * locked section of sync_ser_write the DMA will be
1382 * restarted and the eol flag will be cleared.
1383 * The remaining descriptors will be traversed by
1384 * the descriptor interrupts as usual.
1385 */
1386 j = 0;
1387 while (!port->catch_tr_descr->eol) {
1388 sent = port->catch_tr_descr->after -
1389 port->catch_tr_descr->buf;
1390 DEBUGOUTBUF(pr_info(
1391 "traversing descr %p -%d (%d)\n",
1392 port->catch_tr_descr,
1393 sent,
1394 port->out_buf_count));
1395 port->out_buf_count -= sent;
1396 port->catch_tr_descr = phys_to_virt(
1397 (int)port->catch_tr_descr->next);
1398 j++;
1399 if (j >= NBR_OUT_DESCR) {
1400 /* TODO: Reset and recover */
1401 panic("sync_serial: missing eol");
1402 }
1403 }
1404 sent = port->catch_tr_descr->after -
1405 port->catch_tr_descr->buf;
1406 DEBUGOUTBUF(pr_info("eol at descr %p -%d (%d)\n",
1407 port->catch_tr_descr,
1408 sent,
1409 port->out_buf_count));
1410
1411 port->out_buf_count -= sent;
1412
1413 /* Update read pointer to first free byte, we
1414 * may already be writing data there. */
1415 port->out_rd_ptr =
1416 phys_to_virt((int) port->catch_tr_descr->after);
1417 if (port->out_rd_ptr > port->out_buffer +
1418 OUT_BUFFER_SIZE)
1419 port->out_rd_ptr = port->out_buffer;
1420
1421 tr_cfg = REG_RD(sser, port->regi_sser, rw_tr_cfg);
1422 DEBUGTXINT(pr_info(
1423 "tr_int DMA stop %d, set catch @ %p\n",
1424 port->out_buf_count,
1425 port->active_tr_descr));
1426 if (port->out_buf_count != 0)
1427 pr_err("sync_ser: buf not empty after eol\n");
1428 port->catch_tr_descr = port->active_tr_descr;
1429 port->tr_running = 0;
1430 tr_cfg.tr_en = regk_sser_no;
1431 REG_WR(sser, port->regi_sser, rw_tr_cfg, tr_cfg);
1432 }
1433 /* wake up the waiting process */
1434 wake_up_interruptible(&port->out_wait_q);
1435 }
1436 return IRQ_RETVAL(found);
1437 } /* tr_interrupt */
1438
1439
1440 static inline void handle_rx_packet(struct sync_port *port)
1441 {
1442 int idx;
1443 reg_dma_rw_ack_intr ack_intr = { .data = regk_dma_yes };
1444 unsigned long flags;
1445
1446 DEBUGRXINT(pr_info("!"));
1447 spin_lock_irqsave(&port->lock, flags);
1448
1449 /* If we overrun the user experience is crap regardless if we
1450 * drop new or old data. Its much easier to get it right when
1451 * dropping new data so lets do that.
1452 */
1453 if ((port->writep + port->inbufchunk <=
1454 port->flip + port->in_buffer_size) &&
1455 (port->in_buffer_len + port->inbufchunk < IN_BUFFER_SIZE)) {
1456 memcpy(port->writep,
1457 phys_to_virt((unsigned)port->next_rx_desc->buf),
1458 port->inbufchunk);
1459 port->writep += port->inbufchunk;
1460 if (port->writep >= port->flip + port->in_buffer_size)
1461 port->writep = port->flip;
1462
1463 /* Timestamp the new data chunk. */
1464 if (port->write_ts_idx == NBR_IN_DESCR)
1465 port->write_ts_idx = 0;
1466 idx = port->write_ts_idx++;
1467 ktime_get_ts(&port->timestamp[idx]);
1468 port->in_buffer_len += port->inbufchunk;
1469 }
1470 spin_unlock_irqrestore(&port->lock, flags);
1471
1472 port->next_rx_desc->eol = 1;
1473 port->prev_rx_desc->eol = 0;
1474 /* Cache bug workaround */
1475 flush_dma_descr(port->prev_rx_desc, 0);
1476 port->prev_rx_desc = port->next_rx_desc;
1477 port->next_rx_desc = phys_to_virt((unsigned)port->next_rx_desc->next);
1478 /* Cache bug workaround */
1479 flush_dma_descr(port->prev_rx_desc, 1);
1480 /* wake up the waiting process */
1481 wake_up_interruptible(&port->in_wait_q);
1482 DMA_CONTINUE(port->regi_dmain);
1483 REG_WR(dma, port->regi_dmain, rw_ack_intr, ack_intr);
1484
1485 }
1486
1487 static irqreturn_t rx_interrupt(int irq, void *dev_id)
1488 {
1489 reg_dma_r_masked_intr masked;
1490
1491 int i;
1492 int found = 0;
1493
1494 DEBUG(pr_info("rx_interrupt\n"));
1495
1496 for (i = 0; i < NBR_PORTS; i++) {
1497 struct sync_port *port = &ports[i];
1498
1499 if (!port->enabled || !port->use_dma)
1500 continue;
1501
1502 masked = REG_RD(dma, port->regi_dmain, r_masked_intr);
1503
1504 if (!masked.data)
1505 continue;
1506
1507 /* Descriptor interrupt */
1508 found = 1;
1509 while (REG_RD(dma, port->regi_dmain, rw_data) !=
1510 virt_to_phys(port->next_rx_desc))
1511 handle_rx_packet(port);
1512 }
1513 return IRQ_RETVAL(found);
1514 } /* rx_interrupt */
1515 #endif /* SYNC_SER_DMA */
1516
1517 #ifdef SYNC_SER_MANUAL
1518 static irqreturn_t manual_interrupt(int irq, void *dev_id)
1519 {
1520 unsigned long flags;
1521 int i;
1522 int found = 0;
1523 reg_sser_r_masked_intr masked;
1524
1525 for (i = 0; i < NBR_PORTS; i++) {
1526 struct sync_port *port = &ports[i];
1527
1528 if (!port->enabled || port->use_dma)
1529 continue;
1530
1531 masked = REG_RD(sser, port->regi_sser, r_masked_intr);
1532 /* Data received? */
1533 if (masked.rdav) {
1534 reg_sser_rw_rec_cfg rec_cfg =
1535 REG_RD(sser, port->regi_sser, rw_rec_cfg);
1536 reg_sser_r_rec_data data = REG_RD(sser,
1537 port->regi_sser, r_rec_data);
1538 found = 1;
1539 /* Read data */
1540 spin_lock_irqsave(&port->lock, flags);
1541 switch (rec_cfg.sample_size) {
1542 case 8:
1543 *port->writep++ = data.data & 0xff;
1544 break;
1545 case 12:
1546 *port->writep = (data.data & 0x0ff0) >> 4;
1547 *(port->writep + 1) = data.data & 0x0f;
1548 port->writep += 2;
1549 break;
1550 case 16:
1551 *(unsigned short *)port->writep = data.data;
1552 port->writep += 2;
1553 break;
1554 case 24:
1555 *(unsigned int *)port->writep = data.data;
1556 port->writep += 3;
1557 break;
1558 case 32:
1559 *(unsigned int *)port->writep = data.data;
1560 port->writep += 4;
1561 break;
1562 }
1563
1564 /* Wrap? */
1565 if (port->writep >= port->flip + port->in_buffer_size)
1566 port->writep = port->flip;
1567 if (port->writep == port->readp) {
1568 /* Receive buf overrun, discard oldest data */
1569 port->readp++;
1570 /* Wrap? */
1571 if (port->readp >= port->flip +
1572 port->in_buffer_size)
1573 port->readp = port->flip;
1574 }
1575 spin_unlock_irqrestore(&port->lock, flags);
1576 if (sync_data_avail(port) >= port->inbufchunk)
1577 /* Wake up application */
1578 wake_up_interruptible(&port->in_wait_q);
1579 }
1580
1581 /* Transmitter ready? */
1582 if (masked.trdy) {
1583 found = 1;
1584 /* More data to send */
1585 if (port->out_buf_count > 0)
1586 send_word(port);
1587 else {
1588 /* Transmission finished */
1589 reg_sser_rw_intr_mask intr_mask;
1590 intr_mask = REG_RD(sser, port->regi_sser,
1591 rw_intr_mask);
1592 intr_mask.trdy = 0;
1593 REG_WR(sser, port->regi_sser,
1594 rw_intr_mask, intr_mask);
1595 /* Wake up application */
1596 wake_up_interruptible(&port->out_wait_q);
1597 }
1598 }
1599 }
1600 return IRQ_RETVAL(found);
1601 }
1602 #endif
1603
1604 static int __init etrax_sync_serial_init(void)
1605 {
1606 #if 1
1607 /* This code will be removed when we move to udev for all devices. */
1608 syncser_first = MKDEV(SYNC_SERIAL_MAJOR, 0);
1609 if (register_chrdev_region(syncser_first, minor_count, SYNCSER_NAME)) {
1610 pr_err("Failed to register major %d\n", SYNC_SERIAL_MAJOR);
1611 return -1;
1612 }
1613 #else
1614 /* Allocate dynamic major number. */
1615 if (alloc_chrdev_region(&syncser_first, 0, minor_count, SYNCSER_NAME)) {
1616 pr_err("Failed to allocate character device region\n");
1617 return -1;
1618 }
1619 #endif
1620 syncser_cdev = cdev_alloc();
1621 if (!syncser_cdev) {
1622 pr_err("Failed to allocate cdev for syncser\n");
1623 unregister_chrdev_region(syncser_first, minor_count);
1624 return -1;
1625 }
1626 cdev_init(syncser_cdev, &syncser_fops);
1627
1628 /* Create a sysfs class for syncser */
1629 syncser_class = class_create(THIS_MODULE, "syncser_class");
1630 if (IS_ERR(syncser_class)) {
1631 pr_err("Failed to create a sysfs class for syncser\n");
1632 unregister_chrdev_region(syncser_first, minor_count);
1633 cdev_del(syncser_cdev);
1634 return -1;
1635 }
1636
1637 /* Initialize Ports */
1638 #if defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT0)
1639 if (artpec_pinmux_alloc_fixed(PINMUX_SSER0)) {
1640 pr_warn("Unable to alloc pins for synchronous serial port 0\n");
1641 unregister_chrdev_region(syncser_first, minor_count);
1642 return -EIO;
1643 }
1644 initialize_port(0);
1645 ports[0].enabled = 1;
1646 /* Register with sysfs so udev can pick it up. */
1647 device_create(syncser_class, NULL, syncser_first, NULL,
1648 "%s%d", SYNCSER_NAME, 0);
1649 #endif
1650
1651 #if defined(CONFIG_ETRAXFS) && defined(CONFIG_ETRAX_SYNCHRONOUS_SERIAL_PORT1)
1652 if (artpec_pinmux_alloc_fixed(PINMUX_SSER1)) {
1653 pr_warn("Unable to alloc pins for synchronous serial port 1\n");
1654 unregister_chrdev_region(syncser_first, minor_count);
1655 class_destroy(syncser_class);
1656 return -EIO;
1657 }
1658 initialize_port(1);
1659 ports[1].enabled = 1;
1660 /* Register with sysfs so udev can pick it up. */
1661 device_create(syncser_class, NULL, syncser_first, NULL,
1662 "%s%d", SYNCSER_NAME, 0);
1663 #endif
1664
1665 /* Add it to system */
1666 if (cdev_add(syncser_cdev, syncser_first, minor_count) < 0) {
1667 pr_err("Failed to add syncser as char device\n");
1668 device_destroy(syncser_class, syncser_first);
1669 class_destroy(syncser_class);
1670 cdev_del(syncser_cdev);
1671 unregister_chrdev_region(syncser_first, minor_count);
1672 return -1;
1673 }
1674
1675
1676 pr_info("ARTPEC synchronous serial port (%s: %d, %d)\n",
1677 SYNCSER_NAME, MAJOR(syncser_first), MINOR(syncser_first));
1678
1679 return 0;
1680 }
1681
1682 static void __exit etrax_sync_serial_exit(void)
1683 {
1684 int i;
1685 device_destroy(syncser_class, syncser_first);
1686 class_destroy(syncser_class);
1687
1688 if (syncser_cdev) {
1689 cdev_del(syncser_cdev);
1690 unregister_chrdev_region(syncser_first, minor_count);
1691 }
1692 for (i = 0; i < NBR_PORTS; i++) {
1693 struct sync_port *port = &ports[i];
1694 if (port->init_irqs == dma_irq_setup) {
1695 /* Free dma irqs and dma channels. */
1696 #ifdef SYNC_SER_DMA
1697 artpec_free_dma(port->dma_in_nbr);
1698 artpec_free_dma(port->dma_out_nbr);
1699 free_irq(port->dma_out_intr_vect, port);
1700 free_irq(port->dma_in_intr_vect, port);
1701 #endif
1702 } else if (port->init_irqs == manual_irq_setup) {
1703 /* Free manual irq. */
1704 free_irq(port->syncser_intr_vect, port);
1705 }
1706 }
1707
1708 pr_info("ARTPEC synchronous serial port unregistered\n");
1709 }
1710
1711 module_init(etrax_sync_serial_init);
1712 module_exit(etrax_sync_serial_exit);
1713
1714 MODULE_LICENSE("GPL");
1715