1 /*
2  * SuperH IrDA Driver
3  *
4  * Copyright (C) 2010 Renesas Solutions Corp.
5  * Kuninori Morimoto <kuninori.morimoto.gx@renesas.com>
6  *
7  * Based on sh_sir.c
8  * Copyright (C) 2009 Renesas Solutions Corp.
9  * Copyright 2006-2009 Analog Devices Inc.
10  *
11  * This program is free software; you can redistribute it and/or modify
12  * it under the terms of the GNU General Public License version 2 as
13  * published by the Free Software Foundation.
14  */
15 
16 /*
17  * CAUTION
18  *
19  * This driver is very simple.
20  * So, it doesn't have below support now
21  *  - MIR/FIR support
22  *  - DMA transfer support
23  *  - FIFO mode support
24  */
25 #include <linux/io.h>
26 #include <linux/interrupt.h>
27 #include <linux/module.h>
28 #include <linux/platform_device.h>
29 #include <linux/pm_runtime.h>
30 #include <linux/clk.h>
31 #include <net/irda/wrapper.h>
32 #include <net/irda/irda_device.h>
33 
34 #define DRIVER_NAME "sh_irda"
35 
36 #if defined(CONFIG_ARCH_SH7367) || defined(CONFIG_ARCH_SH7377)
37 #define __IRDARAM_LEN	0x13FF
38 #else
39 #define __IRDARAM_LEN	0x1039
40 #endif
41 
42 #define IRTMR		0x1F00 /* Transfer mode */
43 #define IRCFR		0x1F02 /* Configuration */
44 #define IRCTR		0x1F04 /* IR control */
45 #define IRTFLR		0x1F20 /* Transmit frame length */
46 #define IRTCTR		0x1F22 /* Transmit control */
47 #define IRRFLR		0x1F40 /* Receive frame length */
48 #define IRRCTR		0x1F42 /* Receive control */
49 #define SIRISR		0x1F60 /* SIR-UART mode interrupt source */
50 #define SIRIMR		0x1F62 /* SIR-UART mode interrupt mask */
51 #define SIRICR		0x1F64 /* SIR-UART mode interrupt clear */
52 #define SIRBCR		0x1F68 /* SIR-UART mode baud rate count */
53 #define MFIRISR		0x1F70 /* MIR/FIR mode interrupt source */
54 #define MFIRIMR		0x1F72 /* MIR/FIR mode interrupt mask */
55 #define MFIRICR		0x1F74 /* MIR/FIR mode interrupt clear */
56 #define CRCCTR		0x1F80 /* CRC engine control */
57 #define CRCIR		0x1F86 /* CRC engine input data */
58 #define CRCCR		0x1F8A /* CRC engine calculation */
59 #define CRCOR		0x1F8E /* CRC engine output data */
60 #define FIFOCP		0x1FC0 /* FIFO current pointer */
61 #define FIFOFP		0x1FC2 /* FIFO follow pointer */
62 #define FIFORSMSK	0x1FC4 /* FIFO receive status mask */
63 #define FIFORSOR	0x1FC6 /* FIFO receive status OR */
64 #define FIFOSEL		0x1FC8 /* FIFO select */
65 #define FIFORS		0x1FCA /* FIFO receive status */
66 #define FIFORFL		0x1FCC /* FIFO receive frame length */
67 #define FIFORAMCP	0x1FCE /* FIFO RAM current pointer */
68 #define FIFORAMFP	0x1FD0 /* FIFO RAM follow pointer */
69 #define BIFCTL		0x1FD2 /* BUS interface control */
70 #define IRDARAM		0x0000 /* IrDA buffer RAM */
71 #define IRDARAM_LEN	__IRDARAM_LEN /* - 8/16/32 (read-only for 32) */
72 
73 /* IRTMR */
74 #define TMD_MASK	(0x3 << 14) /* Transfer Mode */
75 #define TMD_SIR		(0x0 << 14)
76 #define TMD_MIR		(0x3 << 14)
77 #define TMD_FIR		(0x2 << 14)
78 
79 #define FIFORIM		(1 << 8) /* FIFO receive interrupt mask */
80 #define MIM		(1 << 4) /* MIR/FIR Interrupt Mask */
81 #define SIM		(1 << 0) /* SIR Interrupt Mask */
82 #define xIM_MASK	(FIFORIM | MIM | SIM)
83 
84 /* IRCFR */
85 #define RTO_SHIFT	8 /* shift for Receive Timeout */
86 #define RTO		(0x3 << RTO_SHIFT)
87 
88 /* IRTCTR */
89 #define ARMOD		(1 << 15) /* Auto-Receive Mode */
90 #define TE		(1 <<  0) /* Transmit Enable */
91 
92 /* IRRFLR */
93 #define RFL_MASK	(0x1FFF) /* mask for Receive Frame Length */
94 
95 /* IRRCTR */
96 #define RE		(1 <<  0) /* Receive Enable */
97 
98 /*
99  * SIRISR,  SIRIMR,  SIRICR,
100  * MFIRISR, MFIRIMR, MFIRICR
101  */
102 #define FRE		(1 << 15) /* Frame Receive End */
103 #define TROV		(1 << 11) /* Transfer Area Overflow */
104 #define xIR_9		(1 << 9)
105 #define TOT		xIR_9     /* for SIR     Timeout */
106 #define ABTD		xIR_9     /* for MIR/FIR Abort Detection */
107 #define xIR_8		(1 << 8)
108 #define FER		xIR_8     /* for SIR     Framing Error */
109 #define CRCER		xIR_8     /* for MIR/FIR CRC error */
110 #define FTE		(1 << 7)  /* Frame Transmit End */
111 #define xIR_MASK	(FRE | TROV | xIR_9 | xIR_8 | FTE)
112 
113 /* SIRBCR */
114 #define BRC_MASK	(0x3F) /* mask for Baud Rate Count */
115 
116 /* CRCCTR */
117 #define CRC_RST		(1 << 15) /* CRC Engine Reset */
118 #define CRC_CT_MASK	0x0FFF    /* mask for CRC Engine Input Data Count */
119 
120 /* CRCIR */
121 #define CRC_IN_MASK	0x0FFF    /* mask for CRC Engine Input Data */
122 
123 /************************************************************************
124 
125 
126 			enum / structure
127 
128 
129 ************************************************************************/
130 enum sh_irda_mode {
131 	SH_IRDA_NONE = 0,
132 	SH_IRDA_SIR,
133 	SH_IRDA_MIR,
134 	SH_IRDA_FIR,
135 };
136 
137 struct sh_irda_self;
138 struct sh_irda_xir_func {
139 	int (*xir_fre)	(struct sh_irda_self *self);
140 	int (*xir_trov)	(struct sh_irda_self *self);
141 	int (*xir_9)	(struct sh_irda_self *self);
142 	int (*xir_8)	(struct sh_irda_self *self);
143 	int (*xir_fte)	(struct sh_irda_self *self);
144 };
145 
146 struct sh_irda_self {
147 	void __iomem		*membase;
148 	unsigned int		irq;
149 	struct platform_device	*pdev;
150 
151 	struct net_device	*ndev;
152 
153 	struct irlap_cb		*irlap;
154 	struct qos_info		qos;
155 
156 	iobuff_t		tx_buff;
157 	iobuff_t		rx_buff;
158 
159 	enum sh_irda_mode	mode;
160 	spinlock_t		lock;
161 
162 	struct sh_irda_xir_func	*xir_func;
163 };
164 
165 /************************************************************************
166 
167 
168 			common function
169 
170 
171 ************************************************************************/
sh_irda_write(struct sh_irda_self * self,u32 offset,u16 data)172 static void sh_irda_write(struct sh_irda_self *self, u32 offset, u16 data)
173 {
174 	unsigned long flags;
175 
176 	spin_lock_irqsave(&self->lock, flags);
177 	iowrite16(data, self->membase + offset);
178 	spin_unlock_irqrestore(&self->lock, flags);
179 }
180 
sh_irda_read(struct sh_irda_self * self,u32 offset)181 static u16 sh_irda_read(struct sh_irda_self *self, u32 offset)
182 {
183 	unsigned long flags;
184 	u16 ret;
185 
186 	spin_lock_irqsave(&self->lock, flags);
187 	ret = ioread16(self->membase + offset);
188 	spin_unlock_irqrestore(&self->lock, flags);
189 
190 	return ret;
191 }
192 
sh_irda_update_bits(struct sh_irda_self * self,u32 offset,u16 mask,u16 data)193 static void sh_irda_update_bits(struct sh_irda_self *self, u32 offset,
194 			       u16 mask, u16 data)
195 {
196 	unsigned long flags;
197 	u16 old, new;
198 
199 	spin_lock_irqsave(&self->lock, flags);
200 	old = ioread16(self->membase + offset);
201 	new = (old & ~mask) | data;
202 	if (old != new)
203 		iowrite16(data, self->membase + offset);
204 	spin_unlock_irqrestore(&self->lock, flags);
205 }
206 
207 /************************************************************************
208 
209 
210 			mode function
211 
212 
213 ************************************************************************/
214 /*=====================================
215  *
216  *		common
217  *
218  *=====================================*/
sh_irda_rcv_ctrl(struct sh_irda_self * self,int enable)219 static void sh_irda_rcv_ctrl(struct sh_irda_self *self, int enable)
220 {
221 	struct device *dev = &self->ndev->dev;
222 
223 	sh_irda_update_bits(self, IRRCTR, RE, enable ? RE : 0);
224 	dev_dbg(dev, "recv %s\n", enable ? "enable" : "disable");
225 }
226 
sh_irda_set_timeout(struct sh_irda_self * self,int interval)227 static int sh_irda_set_timeout(struct sh_irda_self *self, int interval)
228 {
229 	struct device *dev = &self->ndev->dev;
230 
231 	if (SH_IRDA_SIR != self->mode)
232 		interval = 0;
233 
234 	if (interval < 0 || interval > 2) {
235 		dev_err(dev, "unsupported timeout interval\n");
236 		return -EINVAL;
237 	}
238 
239 	sh_irda_update_bits(self, IRCFR, RTO, interval << RTO_SHIFT);
240 	return 0;
241 }
242 
sh_irda_set_baudrate(struct sh_irda_self * self,int baudrate)243 static int sh_irda_set_baudrate(struct sh_irda_self *self, int baudrate)
244 {
245 	struct device *dev = &self->ndev->dev;
246 	u16 val;
247 
248 	if (baudrate < 0)
249 		return 0;
250 
251 	if (SH_IRDA_SIR != self->mode) {
252 		dev_err(dev, "it is not SIR mode\n");
253 		return -EINVAL;
254 	}
255 
256 	/*
257 	 * Baud rate (bits/s) =
258 	 *   (48 MHz / 26) / (baud rate counter value + 1) x 16
259 	 */
260 	val = (48000000 / 26 / 16 / baudrate) - 1;
261 	dev_dbg(dev, "baudrate = %d,  val = 0x%02x\n", baudrate, val);
262 
263 	sh_irda_update_bits(self, SIRBCR, BRC_MASK, val);
264 
265 	return 0;
266 }
267 
sh_irda_get_rcv_length(struct sh_irda_self * self)268 static int sh_irda_get_rcv_length(struct sh_irda_self *self)
269 {
270 	return RFL_MASK & sh_irda_read(self, IRRFLR);
271 }
272 
273 /*=====================================
274  *
275  *		NONE MODE
276  *
277  *=====================================*/
sh_irda_xir_fre(struct sh_irda_self * self)278 static int sh_irda_xir_fre(struct sh_irda_self *self)
279 {
280 	struct device *dev = &self->ndev->dev;
281 	dev_err(dev, "none mode: frame recv\n");
282 	return 0;
283 }
284 
sh_irda_xir_trov(struct sh_irda_self * self)285 static int sh_irda_xir_trov(struct sh_irda_self *self)
286 {
287 	struct device *dev = &self->ndev->dev;
288 	dev_err(dev, "none mode: buffer ram over\n");
289 	return 0;
290 }
291 
sh_irda_xir_9(struct sh_irda_self * self)292 static int sh_irda_xir_9(struct sh_irda_self *self)
293 {
294 	struct device *dev = &self->ndev->dev;
295 	dev_err(dev, "none mode: time over\n");
296 	return 0;
297 }
298 
sh_irda_xir_8(struct sh_irda_self * self)299 static int sh_irda_xir_8(struct sh_irda_self *self)
300 {
301 	struct device *dev = &self->ndev->dev;
302 	dev_err(dev, "none mode: framing error\n");
303 	return 0;
304 }
305 
sh_irda_xir_fte(struct sh_irda_self * self)306 static int sh_irda_xir_fte(struct sh_irda_self *self)
307 {
308 	struct device *dev = &self->ndev->dev;
309 	dev_err(dev, "none mode: frame transmit end\n");
310 	return 0;
311 }
312 
313 static struct sh_irda_xir_func sh_irda_xir_func = {
314 	.xir_fre	= sh_irda_xir_fre,
315 	.xir_trov	= sh_irda_xir_trov,
316 	.xir_9		= sh_irda_xir_9,
317 	.xir_8		= sh_irda_xir_8,
318 	.xir_fte	= sh_irda_xir_fte,
319 };
320 
321 /*=====================================
322  *
323  *		MIR/FIR MODE
324  *
325  * MIR/FIR are not supported now
326  *=====================================*/
327 static struct sh_irda_xir_func sh_irda_mfir_func = {
328 	.xir_fre	= sh_irda_xir_fre,
329 	.xir_trov	= sh_irda_xir_trov,
330 	.xir_9		= sh_irda_xir_9,
331 	.xir_8		= sh_irda_xir_8,
332 	.xir_fte	= sh_irda_xir_fte,
333 };
334 
335 /*=====================================
336  *
337  *		SIR MODE
338  *
339  *=====================================*/
sh_irda_sir_fre(struct sh_irda_self * self)340 static int sh_irda_sir_fre(struct sh_irda_self *self)
341 {
342 	struct device *dev = &self->ndev->dev;
343 	u16 data16;
344 	u8  *data = (u8 *)&data16;
345 	int len = sh_irda_get_rcv_length(self);
346 	int i, j;
347 
348 	if (len > IRDARAM_LEN)
349 		len = IRDARAM_LEN;
350 
351 	dev_dbg(dev, "frame recv length = %d\n", len);
352 
353 	for (i = 0; i < len; i++) {
354 		j = i % 2;
355 		if (!j)
356 			data16 = sh_irda_read(self, IRDARAM + i);
357 
358 		async_unwrap_char(self->ndev, &self->ndev->stats,
359 				  &self->rx_buff, data[j]);
360 	}
361 	self->ndev->last_rx = jiffies;
362 
363 	sh_irda_rcv_ctrl(self, 1);
364 
365 	return 0;
366 }
367 
sh_irda_sir_trov(struct sh_irda_self * self)368 static int sh_irda_sir_trov(struct sh_irda_self *self)
369 {
370 	struct device *dev = &self->ndev->dev;
371 
372 	dev_err(dev, "buffer ram over\n");
373 	sh_irda_rcv_ctrl(self, 1);
374 	return 0;
375 }
376 
sh_irda_sir_tot(struct sh_irda_self * self)377 static int sh_irda_sir_tot(struct sh_irda_self *self)
378 {
379 	struct device *dev = &self->ndev->dev;
380 
381 	dev_err(dev, "time over\n");
382 	sh_irda_set_baudrate(self, 9600);
383 	sh_irda_rcv_ctrl(self, 1);
384 	return 0;
385 }
386 
sh_irda_sir_fer(struct sh_irda_self * self)387 static int sh_irda_sir_fer(struct sh_irda_self *self)
388 {
389 	struct device *dev = &self->ndev->dev;
390 
391 	dev_err(dev, "framing error\n");
392 	sh_irda_rcv_ctrl(self, 1);
393 	return 0;
394 }
395 
sh_irda_sir_fte(struct sh_irda_self * self)396 static int sh_irda_sir_fte(struct sh_irda_self *self)
397 {
398 	struct device *dev = &self->ndev->dev;
399 
400 	dev_dbg(dev, "frame transmit end\n");
401 	netif_wake_queue(self->ndev);
402 
403 	return 0;
404 }
405 
406 static struct sh_irda_xir_func sh_irda_sir_func = {
407 	.xir_fre	= sh_irda_sir_fre,
408 	.xir_trov	= sh_irda_sir_trov,
409 	.xir_9		= sh_irda_sir_tot,
410 	.xir_8		= sh_irda_sir_fer,
411 	.xir_fte	= sh_irda_sir_fte,
412 };
413 
sh_irda_set_mode(struct sh_irda_self * self,enum sh_irda_mode mode)414 static void sh_irda_set_mode(struct sh_irda_self *self, enum sh_irda_mode mode)
415 {
416 	struct device *dev = &self->ndev->dev;
417 	struct sh_irda_xir_func	*func;
418 	const char *name;
419 	u16 data;
420 
421 	switch (mode) {
422 	case SH_IRDA_SIR:
423 		name	= "SIR";
424 		data	= TMD_SIR;
425 		func	= &sh_irda_sir_func;
426 		break;
427 	case SH_IRDA_MIR:
428 		name	= "MIR";
429 		data	= TMD_MIR;
430 		func	= &sh_irda_mfir_func;
431 		break;
432 	case SH_IRDA_FIR:
433 		name	= "FIR";
434 		data	= TMD_FIR;
435 		func	= &sh_irda_mfir_func;
436 		break;
437 	default:
438 		name	= "NONE";
439 		data	= 0;
440 		func	= &sh_irda_xir_func;
441 		break;
442 	}
443 
444 	self->mode = mode;
445 	self->xir_func = func;
446 	sh_irda_update_bits(self, IRTMR, TMD_MASK, data);
447 
448 	dev_dbg(dev, "switch to %s mode", name);
449 }
450 
451 /************************************************************************
452 
453 
454 			irq function
455 
456 
457 ************************************************************************/
sh_irda_set_irq_mask(struct sh_irda_self * self)458 static void sh_irda_set_irq_mask(struct sh_irda_self *self)
459 {
460 	u16 tmr_hole;
461 	u16 xir_reg;
462 
463 	/* set all mask */
464 	sh_irda_update_bits(self, IRTMR,   xIM_MASK, xIM_MASK);
465 	sh_irda_update_bits(self, SIRIMR,  xIR_MASK, xIR_MASK);
466 	sh_irda_update_bits(self, MFIRIMR, xIR_MASK, xIR_MASK);
467 
468 	/* clear irq */
469 	sh_irda_update_bits(self, SIRICR,  xIR_MASK, xIR_MASK);
470 	sh_irda_update_bits(self, MFIRICR, xIR_MASK, xIR_MASK);
471 
472 	switch (self->mode) {
473 	case SH_IRDA_SIR:
474 		tmr_hole	= SIM;
475 		xir_reg		= SIRIMR;
476 		break;
477 	case SH_IRDA_MIR:
478 	case SH_IRDA_FIR:
479 		tmr_hole	= MIM;
480 		xir_reg		= MFIRIMR;
481 		break;
482 	default:
483 		tmr_hole	= 0;
484 		xir_reg		= 0;
485 		break;
486 	}
487 
488 	/* open mask */
489 	if (xir_reg) {
490 		sh_irda_update_bits(self, IRTMR, tmr_hole, 0);
491 		sh_irda_update_bits(self, xir_reg, xIR_MASK, 0);
492 	}
493 }
494 
sh_irda_irq(int irq,void * dev_id)495 static irqreturn_t sh_irda_irq(int irq, void *dev_id)
496 {
497 	struct sh_irda_self *self = dev_id;
498 	struct sh_irda_xir_func	*func = self->xir_func;
499 	u16 isr = sh_irda_read(self, SIRISR);
500 
501 	/* clear irq */
502 	sh_irda_write(self, SIRICR, isr);
503 
504 	if (isr & FRE)
505 		func->xir_fre(self);
506 	if (isr & TROV)
507 		func->xir_trov(self);
508 	if (isr & xIR_9)
509 		func->xir_9(self);
510 	if (isr & xIR_8)
511 		func->xir_8(self);
512 	if (isr & FTE)
513 		func->xir_fte(self);
514 
515 	return IRQ_HANDLED;
516 }
517 
518 /************************************************************************
519 
520 
521 			CRC function
522 
523 
524 ************************************************************************/
sh_irda_crc_reset(struct sh_irda_self * self)525 static void sh_irda_crc_reset(struct sh_irda_self *self)
526 {
527 	sh_irda_write(self, CRCCTR, CRC_RST);
528 }
529 
sh_irda_crc_add(struct sh_irda_self * self,u16 data)530 static void sh_irda_crc_add(struct sh_irda_self *self, u16 data)
531 {
532 	sh_irda_write(self, CRCIR, data & CRC_IN_MASK);
533 }
534 
sh_irda_crc_cnt(struct sh_irda_self * self)535 static u16 sh_irda_crc_cnt(struct sh_irda_self *self)
536 {
537 	return CRC_CT_MASK & sh_irda_read(self, CRCCTR);
538 }
539 
sh_irda_crc_out(struct sh_irda_self * self)540 static u16 sh_irda_crc_out(struct sh_irda_self *self)
541 {
542 	return sh_irda_read(self, CRCOR);
543 }
544 
sh_irda_crc_init(struct sh_irda_self * self)545 static int sh_irda_crc_init(struct sh_irda_self *self)
546 {
547 	struct device *dev = &self->ndev->dev;
548 	int ret = -EIO;
549 	u16 val;
550 
551 	sh_irda_crc_reset(self);
552 
553 	sh_irda_crc_add(self, 0xCC);
554 	sh_irda_crc_add(self, 0xF5);
555 	sh_irda_crc_add(self, 0xF1);
556 	sh_irda_crc_add(self, 0xA7);
557 
558 	val = sh_irda_crc_cnt(self);
559 	if (4 != val) {
560 		dev_err(dev, "CRC count error %x\n", val);
561 		goto crc_init_out;
562 	}
563 
564 	val = sh_irda_crc_out(self);
565 	if (0x51DF != val) {
566 		dev_err(dev, "CRC result error%x\n", val);
567 		goto crc_init_out;
568 	}
569 
570 	ret = 0;
571 
572 crc_init_out:
573 
574 	sh_irda_crc_reset(self);
575 	return ret;
576 }
577 
578 /************************************************************************
579 
580 
581 			iobuf function
582 
583 
584 ************************************************************************/
sh_irda_remove_iobuf(struct sh_irda_self * self)585 static void sh_irda_remove_iobuf(struct sh_irda_self *self)
586 {
587 	kfree(self->rx_buff.head);
588 
589 	self->tx_buff.head = NULL;
590 	self->tx_buff.data = NULL;
591 	self->rx_buff.head = NULL;
592 	self->rx_buff.data = NULL;
593 }
594 
sh_irda_init_iobuf(struct sh_irda_self * self,int rxsize,int txsize)595 static int sh_irda_init_iobuf(struct sh_irda_self *self, int rxsize, int txsize)
596 {
597 	if (self->rx_buff.head ||
598 	    self->tx_buff.head) {
599 		dev_err(&self->ndev->dev, "iobuff has already existed.");
600 		return -EINVAL;
601 	}
602 
603 	/* rx_buff */
604 	self->rx_buff.head = kmalloc(rxsize, GFP_KERNEL);
605 	if (!self->rx_buff.head)
606 		return -ENOMEM;
607 
608 	self->rx_buff.truesize	= rxsize;
609 	self->rx_buff.in_frame	= FALSE;
610 	self->rx_buff.state	= OUTSIDE_FRAME;
611 	self->rx_buff.data	= self->rx_buff.head;
612 
613 	/* tx_buff */
614 	self->tx_buff.head	= self->membase + IRDARAM;
615 	self->tx_buff.truesize	= IRDARAM_LEN;
616 
617 	return 0;
618 }
619 
620 /************************************************************************
621 
622 
623 			net_device_ops function
624 
625 
626 ************************************************************************/
sh_irda_hard_xmit(struct sk_buff * skb,struct net_device * ndev)627 static int sh_irda_hard_xmit(struct sk_buff *skb, struct net_device *ndev)
628 {
629 	struct sh_irda_self *self = netdev_priv(ndev);
630 	struct device *dev = &self->ndev->dev;
631 	int speed = irda_get_next_speed(skb);
632 	int ret;
633 
634 	dev_dbg(dev, "hard xmit\n");
635 
636 	netif_stop_queue(ndev);
637 	sh_irda_rcv_ctrl(self, 0);
638 
639 	ret = sh_irda_set_baudrate(self, speed);
640 	if (ret < 0)
641 		goto sh_irda_hard_xmit_end;
642 
643 	self->tx_buff.len = 0;
644 	if (skb->len) {
645 		unsigned long flags;
646 
647 		spin_lock_irqsave(&self->lock, flags);
648 		self->tx_buff.len = async_wrap_skb(skb,
649 						   self->tx_buff.head,
650 						   self->tx_buff.truesize);
651 		spin_unlock_irqrestore(&self->lock, flags);
652 
653 		if (self->tx_buff.len > self->tx_buff.truesize)
654 			self->tx_buff.len = self->tx_buff.truesize;
655 
656 		sh_irda_write(self, IRTFLR, self->tx_buff.len);
657 		sh_irda_write(self, IRTCTR, ARMOD | TE);
658 	} else
659 		goto sh_irda_hard_xmit_end;
660 
661 	dev_kfree_skb(skb);
662 
663 	return 0;
664 
665 sh_irda_hard_xmit_end:
666 	sh_irda_set_baudrate(self, 9600);
667 	netif_wake_queue(self->ndev);
668 	sh_irda_rcv_ctrl(self, 1);
669 	dev_kfree_skb(skb);
670 
671 	return ret;
672 
673 }
674 
sh_irda_ioctl(struct net_device * ndev,struct ifreq * ifreq,int cmd)675 static int sh_irda_ioctl(struct net_device *ndev, struct ifreq *ifreq, int cmd)
676 {
677 	/*
678 	 * FIXME
679 	 *
680 	 * This function is needed for irda framework.
681 	 * But nothing to do now
682 	 */
683 	return 0;
684 }
685 
sh_irda_stats(struct net_device * ndev)686 static struct net_device_stats *sh_irda_stats(struct net_device *ndev)
687 {
688 	struct sh_irda_self *self = netdev_priv(ndev);
689 
690 	return &self->ndev->stats;
691 }
692 
sh_irda_open(struct net_device * ndev)693 static int sh_irda_open(struct net_device *ndev)
694 {
695 	struct sh_irda_self *self = netdev_priv(ndev);
696 	int err;
697 
698 	pm_runtime_get_sync(&self->pdev->dev);
699 	err = sh_irda_crc_init(self);
700 	if (err)
701 		goto open_err;
702 
703 	sh_irda_set_mode(self, SH_IRDA_SIR);
704 	sh_irda_set_timeout(self, 2);
705 	sh_irda_set_baudrate(self, 9600);
706 
707 	self->irlap = irlap_open(ndev, &self->qos, DRIVER_NAME);
708 	if (!self->irlap) {
709 		err = -ENODEV;
710 		goto open_err;
711 	}
712 
713 	netif_start_queue(ndev);
714 	sh_irda_rcv_ctrl(self, 1);
715 	sh_irda_set_irq_mask(self);
716 
717 	dev_info(&ndev->dev, "opened\n");
718 
719 	return 0;
720 
721 open_err:
722 	pm_runtime_put_sync(&self->pdev->dev);
723 
724 	return err;
725 }
726 
sh_irda_stop(struct net_device * ndev)727 static int sh_irda_stop(struct net_device *ndev)
728 {
729 	struct sh_irda_self *self = netdev_priv(ndev);
730 
731 	/* Stop IrLAP */
732 	if (self->irlap) {
733 		irlap_close(self->irlap);
734 		self->irlap = NULL;
735 	}
736 
737 	netif_stop_queue(ndev);
738 	pm_runtime_put_sync(&self->pdev->dev);
739 
740 	dev_info(&ndev->dev, "stoped\n");
741 
742 	return 0;
743 }
744 
745 static const struct net_device_ops sh_irda_ndo = {
746 	.ndo_open		= sh_irda_open,
747 	.ndo_stop		= sh_irda_stop,
748 	.ndo_start_xmit		= sh_irda_hard_xmit,
749 	.ndo_do_ioctl		= sh_irda_ioctl,
750 	.ndo_get_stats		= sh_irda_stats,
751 };
752 
753 /************************************************************************
754 
755 
756 			platform_driver function
757 
758 
759 ************************************************************************/
sh_irda_probe(struct platform_device * pdev)760 static int __devinit sh_irda_probe(struct platform_device *pdev)
761 {
762 	struct net_device *ndev;
763 	struct sh_irda_self *self;
764 	struct resource *res;
765 	int irq;
766 	int err = -ENOMEM;
767 
768 	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
769 	irq = platform_get_irq(pdev, 0);
770 	if (!res || irq < 0) {
771 		dev_err(&pdev->dev, "Not enough platform resources.\n");
772 		goto exit;
773 	}
774 
775 	ndev = alloc_irdadev(sizeof(*self));
776 	if (!ndev)
777 		goto exit;
778 
779 	self = netdev_priv(ndev);
780 	self->membase = ioremap_nocache(res->start, resource_size(res));
781 	if (!self->membase) {
782 		err = -ENXIO;
783 		dev_err(&pdev->dev, "Unable to ioremap.\n");
784 		goto err_mem_1;
785 	}
786 
787 	err = sh_irda_init_iobuf(self, IRDA_SKB_MAX_MTU, IRDA_SIR_MAX_FRAME);
788 	if (err)
789 		goto err_mem_2;
790 
791 	self->pdev = pdev;
792 	pm_runtime_enable(&pdev->dev);
793 
794 	irda_init_max_qos_capabilies(&self->qos);
795 
796 	ndev->netdev_ops	= &sh_irda_ndo;
797 	ndev->irq		= irq;
798 
799 	self->ndev			= ndev;
800 	self->qos.baud_rate.bits	&= IR_9600; /* FIXME */
801 	self->qos.min_turn_time.bits	= 1; /* 10 ms or more */
802 	spin_lock_init(&self->lock);
803 
804 	irda_qos_bits_to_value(&self->qos);
805 
806 	err = register_netdev(ndev);
807 	if (err)
808 		goto err_mem_4;
809 
810 	platform_set_drvdata(pdev, ndev);
811 
812 	if (request_irq(irq, sh_irda_irq, IRQF_DISABLED, "sh_irda", self)) {
813 		dev_warn(&pdev->dev, "Unable to attach sh_irda interrupt\n");
814 		goto err_mem_4;
815 	}
816 
817 	dev_info(&pdev->dev, "SuperH IrDA probed\n");
818 
819 	goto exit;
820 
821 err_mem_4:
822 	pm_runtime_disable(&pdev->dev);
823 	sh_irda_remove_iobuf(self);
824 err_mem_2:
825 	iounmap(self->membase);
826 err_mem_1:
827 	free_netdev(ndev);
828 exit:
829 	return err;
830 }
831 
sh_irda_remove(struct platform_device * pdev)832 static int __devexit sh_irda_remove(struct platform_device *pdev)
833 {
834 	struct net_device *ndev = platform_get_drvdata(pdev);
835 	struct sh_irda_self *self = netdev_priv(ndev);
836 
837 	if (!self)
838 		return 0;
839 
840 	unregister_netdev(ndev);
841 	pm_runtime_disable(&pdev->dev);
842 	sh_irda_remove_iobuf(self);
843 	iounmap(self->membase);
844 	free_netdev(ndev);
845 	platform_set_drvdata(pdev, NULL);
846 
847 	return 0;
848 }
849 
sh_irda_runtime_nop(struct device * dev)850 static int sh_irda_runtime_nop(struct device *dev)
851 {
852 	/* Runtime PM callback shared between ->runtime_suspend()
853 	 * and ->runtime_resume(). Simply returns success.
854 	 *
855 	 * This driver re-initializes all registers after
856 	 * pm_runtime_get_sync() anyway so there is no need
857 	 * to save and restore registers here.
858 	 */
859 	return 0;
860 }
861 
862 static const struct dev_pm_ops sh_irda_pm_ops = {
863 	.runtime_suspend	= sh_irda_runtime_nop,
864 	.runtime_resume		= sh_irda_runtime_nop,
865 };
866 
867 static struct platform_driver sh_irda_driver = {
868 	.probe	= sh_irda_probe,
869 	.remove	= __devexit_p(sh_irda_remove),
870 	.driver	= {
871 		.name	= DRIVER_NAME,
872 		.pm	= &sh_irda_pm_ops,
873 	},
874 };
875 
876 module_platform_driver(sh_irda_driver);
877 
878 MODULE_AUTHOR("Kuninori Morimoto <kuninori.morimoto.gx@renesas.com>");
879 MODULE_DESCRIPTION("SuperH IrDA driver");
880 MODULE_LICENSE("GPL");
881