1 /*
2  * NETJet mISDN driver
3  *
4  * Author       Karsten Keil <keil@isdn4linux.de>
5  *
6  * Copyright 2009  by Karsten Keil <keil@isdn4linux.de>
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License version 2 as
10  * published by the Free Software Foundation.
11  *
12  * This program is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
15  * GNU General Public License for more details.
16  *
17  * You should have received a copy of the GNU General Public License
18  * along with this program; if not, write to the Free Software
19  * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
20  *
21  */
22 
23 #include <linux/interrupt.h>
24 #include <linux/module.h>
25 #include <linux/pci.h>
26 #include <linux/delay.h>
27 #include <linux/mISDNhw.h>
28 #include <linux/slab.h>
29 #include "ipac.h"
30 #include "iohelper.h"
31 #include "netjet.h"
32 #include <linux/isdn/hdlc.h>
33 
34 #define NETJET_REV	"2.0"
35 
36 enum nj_types {
37 	NETJET_S_TJ300,
38 	NETJET_S_TJ320,
39 	ENTERNOW__TJ320,
40 };
41 
42 struct tiger_dma {
43 	size_t		size;
44 	u32		*start;
45 	int		idx;
46 	u32		dmastart;
47 	u32		dmairq;
48 	u32		dmaend;
49 	u32		dmacur;
50 };
51 
52 struct tiger_hw;
53 
54 struct tiger_ch {
55 	struct bchannel		bch;
56 	struct tiger_hw		*nj;
57 	int			idx;
58 	int			free;
59 	int			lastrx;
60 	u16			rxstate;
61 	u16			txstate;
62 	struct isdnhdlc_vars	hsend;
63 	struct isdnhdlc_vars	hrecv;
64 	u8			*hsbuf;
65 	u8			*hrbuf;
66 };
67 
68 #define TX_INIT		0x0001
69 #define TX_IDLE		0x0002
70 #define TX_RUN		0x0004
71 #define TX_UNDERRUN	0x0100
72 #define RX_OVERRUN	0x0100
73 
74 #define LOG_SIZE	64
75 
76 struct tiger_hw {
77 	struct list_head	list;
78 	struct pci_dev		*pdev;
79 	char			name[MISDN_MAX_IDLEN];
80 	enum nj_types		typ;
81 	int			irq;
82 	u32			irqcnt;
83 	u32			base;
84 	size_t			base_s;
85 	dma_addr_t		dma;
86 	void			*dma_p;
87 	spinlock_t		lock;	/* lock HW */
88 	struct isac_hw		isac;
89 	struct tiger_dma	send;
90 	struct tiger_dma	recv;
91 	struct tiger_ch		bc[2];
92 	u8			ctrlreg;
93 	u8			dmactrl;
94 	u8			auxd;
95 	u8			last_is0;
96 	u8			irqmask0;
97 	char			log[LOG_SIZE];
98 };
99 
100 static LIST_HEAD(Cards);
101 static DEFINE_RWLOCK(card_lock); /* protect Cards */
102 static u32 debug;
103 static int nj_cnt;
104 
105 static void
_set_debug(struct tiger_hw * card)106 _set_debug(struct tiger_hw *card)
107 {
108 	card->isac.dch.debug = debug;
109 	card->bc[0].bch.debug = debug;
110 	card->bc[1].bch.debug = debug;
111 }
112 
113 static int
set_debug(const char * val,struct kernel_param * kp)114 set_debug(const char *val, struct kernel_param *kp)
115 {
116 	int ret;
117 	struct tiger_hw *card;
118 
119 	ret = param_set_uint(val, kp);
120 	if (!ret) {
121 		read_lock(&card_lock);
122 		list_for_each_entry(card, &Cards, list)
123 			_set_debug(card);
124 		read_unlock(&card_lock);
125 	}
126 	return ret;
127 }
128 
129 MODULE_AUTHOR("Karsten Keil");
130 MODULE_LICENSE("GPL v2");
131 MODULE_VERSION(NETJET_REV);
132 module_param_call(debug, set_debug, param_get_uint, &debug, S_IRUGO | S_IWUSR);
133 MODULE_PARM_DESC(debug, "Netjet debug mask");
134 
135 static void
nj_disable_hwirq(struct tiger_hw * card)136 nj_disable_hwirq(struct tiger_hw *card)
137 {
138 	outb(0, card->base + NJ_IRQMASK0);
139 	outb(0, card->base + NJ_IRQMASK1);
140 }
141 
142 
143 static u8
ReadISAC_nj(void * p,u8 offset)144 ReadISAC_nj(void *p, u8 offset)
145 {
146 	struct tiger_hw *card = p;
147 	u8 ret;
148 
149 	card->auxd &= 0xfc;
150 	card->auxd |= (offset >> 4) & 3;
151 	outb(card->auxd, card->base + NJ_AUXDATA);
152 	ret = inb(card->base + NJ_ISAC_OFF + ((offset & 0x0f) << 2));
153 	return ret;
154 }
155 
156 static void
WriteISAC_nj(void * p,u8 offset,u8 value)157 WriteISAC_nj(void *p, u8 offset, u8 value)
158 {
159 	struct tiger_hw *card = p;
160 
161 	card->auxd &= 0xfc;
162 	card->auxd |= (offset >> 4) & 3;
163 	outb(card->auxd, card->base + NJ_AUXDATA);
164 	outb(value, card->base + NJ_ISAC_OFF + ((offset & 0x0f) << 2));
165 }
166 
167 static void
ReadFiFoISAC_nj(void * p,u8 offset,u8 * data,int size)168 ReadFiFoISAC_nj(void *p, u8 offset, u8 *data, int size)
169 {
170 	struct tiger_hw *card = p;
171 
172 	card->auxd &= 0xfc;
173 	outb(card->auxd, card->base + NJ_AUXDATA);
174 	insb(card->base + NJ_ISAC_OFF, data, size);
175 }
176 
177 static void
WriteFiFoISAC_nj(void * p,u8 offset,u8 * data,int size)178 WriteFiFoISAC_nj(void *p, u8 offset, u8 *data, int size)
179 {
180 	struct tiger_hw *card = p;
181 
182 	card->auxd &= 0xfc;
183 	outb(card->auxd, card->base + NJ_AUXDATA);
184 	outsb(card->base + NJ_ISAC_OFF, data, size);
185 }
186 
187 static void
fill_mem(struct tiger_ch * bc,u32 idx,u32 cnt,u32 fill)188 fill_mem(struct tiger_ch *bc, u32 idx, u32 cnt, u32 fill)
189 {
190 	struct tiger_hw *card = bc->bch.hw;
191 	u32 mask = 0xff, val;
192 
193 	pr_debug("%s: B%1d fill %02x len %d idx %d/%d\n", card->name,
194 		 bc->bch.nr, fill, cnt, idx, card->send.idx);
195 	if (bc->bch.nr & 2) {
196 		fill  <<= 8;
197 		mask <<= 8;
198 	}
199 	mask ^= 0xffffffff;
200 	while (cnt--) {
201 		val = card->send.start[idx];
202 		val &= mask;
203 		val |= fill;
204 		card->send.start[idx++] = val;
205 		if (idx >= card->send.size)
206 			idx = 0;
207 	}
208 }
209 
210 static int
mode_tiger(struct tiger_ch * bc,u32 protocol)211 mode_tiger(struct tiger_ch *bc, u32 protocol)
212 {
213 	struct tiger_hw *card = bc->bch.hw;
214 
215 	pr_debug("%s: B%1d protocol %x-->%x\n", card->name,
216 		 bc->bch.nr, bc->bch.state, protocol);
217 	switch (protocol) {
218 	case ISDN_P_NONE:
219 		if (bc->bch.state == ISDN_P_NONE)
220 			break;
221 		fill_mem(bc, 0, card->send.size, 0xff);
222 		bc->bch.state = protocol;
223 		/* only stop dma and interrupts if both channels NULL */
224 		if ((card->bc[0].bch.state == ISDN_P_NONE) &&
225 		    (card->bc[1].bch.state == ISDN_P_NONE)) {
226 			card->dmactrl = 0;
227 			outb(card->dmactrl, card->base + NJ_DMACTRL);
228 			outb(0, card->base + NJ_IRQMASK0);
229 		}
230 		test_and_clear_bit(FLG_HDLC, &bc->bch.Flags);
231 		test_and_clear_bit(FLG_TRANSPARENT, &bc->bch.Flags);
232 		bc->txstate = 0;
233 		bc->rxstate = 0;
234 		bc->lastrx = -1;
235 		break;
236 	case ISDN_P_B_RAW:
237 		test_and_set_bit(FLG_TRANSPARENT, &bc->bch.Flags);
238 		bc->bch.state = protocol;
239 		bc->idx = 0;
240 		bc->free = card->send.size / 2;
241 		bc->rxstate = 0;
242 		bc->txstate = TX_INIT | TX_IDLE;
243 		bc->lastrx = -1;
244 		if (!card->dmactrl) {
245 			card->dmactrl = 1;
246 			outb(card->dmactrl, card->base + NJ_DMACTRL);
247 			outb(0x0f, card->base + NJ_IRQMASK0);
248 		}
249 		break;
250 	case ISDN_P_B_HDLC:
251 		test_and_set_bit(FLG_HDLC, &bc->bch.Flags);
252 		bc->bch.state = protocol;
253 		bc->idx = 0;
254 		bc->free = card->send.size / 2;
255 		bc->rxstate = 0;
256 		bc->txstate = TX_INIT | TX_IDLE;
257 		isdnhdlc_rcv_init(&bc->hrecv, 0);
258 		isdnhdlc_out_init(&bc->hsend, 0);
259 		bc->lastrx = -1;
260 		if (!card->dmactrl) {
261 			card->dmactrl = 1;
262 			outb(card->dmactrl, card->base + NJ_DMACTRL);
263 			outb(0x0f, card->base + NJ_IRQMASK0);
264 		}
265 		break;
266 	default:
267 		pr_info("%s: %s protocol %x not handled\n", card->name,
268 			__func__, protocol);
269 		return -ENOPROTOOPT;
270 	}
271 	card->send.dmacur = inl(card->base + NJ_DMA_READ_ADR);
272 	card->recv.dmacur = inl(card->base + NJ_DMA_WRITE_ADR);
273 	card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2;
274 	card->recv.idx = (card->recv.dmacur - card->recv.dmastart) >> 2;
275 	pr_debug("%s: %s ctrl %x irq  %02x/%02x idx %d/%d\n",
276 		 card->name, __func__,
277 		 inb(card->base + NJ_DMACTRL),
278 		 inb(card->base + NJ_IRQMASK0),
279 		 inb(card->base + NJ_IRQSTAT0),
280 		 card->send.idx,
281 		 card->recv.idx);
282 	return 0;
283 }
284 
285 static void
nj_reset(struct tiger_hw * card)286 nj_reset(struct tiger_hw *card)
287 {
288 	outb(0xff, card->base + NJ_CTRL); /* Reset On */
289 	mdelay(1);
290 
291 	/* now edge triggered for TJ320 GE 13/07/00 */
292 	/* see comment in IRQ function */
293 	if (card->typ == NETJET_S_TJ320) /* TJ320 */
294 		card->ctrlreg = 0x40;  /* Reset Off and status read clear */
295 	else
296 		card->ctrlreg = 0x00;  /* Reset Off and status read clear */
297 	outb(card->ctrlreg, card->base + NJ_CTRL);
298 	mdelay(10);
299 
300 	/* configure AUX pins (all output except ISAC IRQ pin) */
301 	card->auxd = 0;
302 	card->dmactrl = 0;
303 	outb(~NJ_ISACIRQ, card->base + NJ_AUXCTRL);
304 	outb(NJ_ISACIRQ,  card->base + NJ_IRQMASK1);
305 	outb(card->auxd, card->base + NJ_AUXDATA);
306 }
307 
308 static int
inittiger(struct tiger_hw * card)309 inittiger(struct tiger_hw *card)
310 {
311 	int i;
312 
313 	card->dma_p = pci_alloc_consistent(card->pdev, NJ_DMA_SIZE,
314 					   &card->dma);
315 	if (!card->dma_p) {
316 		pr_info("%s: No DMA memory\n", card->name);
317 		return -ENOMEM;
318 	}
319 	if ((u64)card->dma > 0xffffffff) {
320 		pr_info("%s: DMA outside 32 bit\n", card->name);
321 		return -ENOMEM;
322 	}
323 	for (i = 0; i < 2; i++) {
324 		card->bc[i].hsbuf = kmalloc(NJ_DMA_TXSIZE, GFP_ATOMIC);
325 		if (!card->bc[i].hsbuf) {
326 			pr_info("%s: no B%d send buffer\n", card->name, i + 1);
327 			return -ENOMEM;
328 		}
329 		card->bc[i].hrbuf = kmalloc(NJ_DMA_RXSIZE, GFP_ATOMIC);
330 		if (!card->bc[i].hrbuf) {
331 			pr_info("%s: no B%d recv buffer\n", card->name, i + 1);
332 			return -ENOMEM;
333 		}
334 	}
335 	memset(card->dma_p, 0xff, NJ_DMA_SIZE);
336 
337 	card->send.start = card->dma_p;
338 	card->send.dmastart = (u32)card->dma;
339 	card->send.dmaend = card->send.dmastart +
340 		(4 * (NJ_DMA_TXSIZE - 1));
341 	card->send.dmairq = card->send.dmastart +
342 		(4 * ((NJ_DMA_TXSIZE / 2) - 1));
343 	card->send.size = NJ_DMA_TXSIZE;
344 
345 	if (debug & DEBUG_HW)
346 		pr_notice("%s: send buffer phy %#x - %#x - %#x  virt %p"
347 			  " size %zu u32\n", card->name,
348 			  card->send.dmastart, card->send.dmairq,
349 			  card->send.dmaend, card->send.start, card->send.size);
350 
351 	outl(card->send.dmastart, card->base + NJ_DMA_READ_START);
352 	outl(card->send.dmairq, card->base + NJ_DMA_READ_IRQ);
353 	outl(card->send.dmaend, card->base + NJ_DMA_READ_END);
354 
355 	card->recv.start = card->dma_p + (NJ_DMA_SIZE / 2);
356 	card->recv.dmastart = (u32)card->dma  + (NJ_DMA_SIZE / 2);
357 	card->recv.dmaend = card->recv.dmastart +
358 		(4 * (NJ_DMA_RXSIZE - 1));
359 	card->recv.dmairq = card->recv.dmastart +
360 		(4 * ((NJ_DMA_RXSIZE / 2) - 1));
361 	card->recv.size = NJ_DMA_RXSIZE;
362 
363 	if (debug & DEBUG_HW)
364 		pr_notice("%s: recv buffer phy %#x - %#x - %#x  virt %p"
365 			  " size %zu u32\n", card->name,
366 			  card->recv.dmastart, card->recv.dmairq,
367 			  card->recv.dmaend, card->recv.start, card->recv.size);
368 
369 	outl(card->recv.dmastart, card->base + NJ_DMA_WRITE_START);
370 	outl(card->recv.dmairq, card->base + NJ_DMA_WRITE_IRQ);
371 	outl(card->recv.dmaend, card->base + NJ_DMA_WRITE_END);
372 	return 0;
373 }
374 
375 static void
read_dma(struct tiger_ch * bc,u32 idx,int cnt)376 read_dma(struct tiger_ch *bc, u32 idx, int cnt)
377 {
378 	struct tiger_hw *card = bc->bch.hw;
379 	int i, stat;
380 	u32 val;
381 	u8 *p, *pn;
382 
383 	if (bc->lastrx == idx) {
384 		bc->rxstate |= RX_OVERRUN;
385 		pr_info("%s: B%1d overrun at idx %d\n", card->name,
386 			bc->bch.nr, idx);
387 	}
388 	bc->lastrx = idx;
389 	if (!bc->bch.rx_skb) {
390 		bc->bch.rx_skb = mI_alloc_skb(bc->bch.maxlen, GFP_ATOMIC);
391 		if (!bc->bch.rx_skb) {
392 			pr_info("%s: B%1d receive out of memory\n",
393 				card->name, bc->bch.nr);
394 			return;
395 		}
396 	}
397 
398 	if (test_bit(FLG_TRANSPARENT, &bc->bch.Flags)) {
399 		if ((bc->bch.rx_skb->len + cnt) > bc->bch.maxlen) {
400 			pr_debug("%s: B%1d overrun %d\n", card->name,
401 				 bc->bch.nr, bc->bch.rx_skb->len + cnt);
402 			skb_trim(bc->bch.rx_skb, 0);
403 			return;
404 		}
405 		p = skb_put(bc->bch.rx_skb, cnt);
406 	} else
407 		p = bc->hrbuf;
408 
409 	for (i = 0; i < cnt; i++) {
410 		val = card->recv.start[idx++];
411 		if (bc->bch.nr & 2)
412 			val >>= 8;
413 		if (idx >= card->recv.size)
414 			idx = 0;
415 		p[i] = val & 0xff;
416 	}
417 	pn = bc->hrbuf;
418 next_frame:
419 	if (test_bit(FLG_HDLC, &bc->bch.Flags)) {
420 		stat = isdnhdlc_decode(&bc->hrecv, pn, cnt, &i,
421 				       bc->bch.rx_skb->data, bc->bch.maxlen);
422 		if (stat > 0) /* valid frame received */
423 			p = skb_put(bc->bch.rx_skb, stat);
424 		else if (stat == -HDLC_CRC_ERROR)
425 			pr_info("%s: B%1d receive frame CRC error\n",
426 				card->name, bc->bch.nr);
427 		else if (stat == -HDLC_FRAMING_ERROR)
428 			pr_info("%s: B%1d receive framing error\n",
429 				card->name, bc->bch.nr);
430 		else if (stat == -HDLC_LENGTH_ERROR)
431 			pr_info("%s: B%1d receive frame too long (> %d)\n",
432 				card->name, bc->bch.nr, bc->bch.maxlen);
433 	} else
434 		stat = cnt;
435 
436 	if (stat > 0) {
437 		if (debug & DEBUG_HW_BFIFO) {
438 			snprintf(card->log, LOG_SIZE, "B%1d-recv %s %d ",
439 				 bc->bch.nr, card->name, stat);
440 			print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET,
441 					     p, stat);
442 		}
443 		recv_Bchannel(&bc->bch, 0);
444 	}
445 	if (test_bit(FLG_HDLC, &bc->bch.Flags)) {
446 		pn += i;
447 		cnt -= i;
448 		if (!bc->bch.rx_skb) {
449 			bc->bch.rx_skb = mI_alloc_skb(bc->bch.maxlen,
450 						      GFP_ATOMIC);
451 			if (!bc->bch.rx_skb) {
452 				pr_info("%s: B%1d receive out of memory\n",
453 					card->name, bc->bch.nr);
454 				return;
455 			}
456 		}
457 		if (cnt > 0)
458 			goto next_frame;
459 	}
460 }
461 
462 static void
recv_tiger(struct tiger_hw * card,u8 irq_stat)463 recv_tiger(struct tiger_hw *card, u8 irq_stat)
464 {
465 	u32 idx;
466 	int cnt = card->recv.size / 2;
467 
468 	/* Note receive is via the WRITE DMA channel */
469 	card->last_is0 &= ~NJ_IRQM0_WR_MASK;
470 	card->last_is0 |= (irq_stat & NJ_IRQM0_WR_MASK);
471 
472 	if (irq_stat & NJ_IRQM0_WR_END)
473 		idx = cnt - 1;
474 	else
475 		idx = card->recv.size - 1;
476 
477 	if (test_bit(FLG_ACTIVE, &card->bc[0].bch.Flags))
478 		read_dma(&card->bc[0], idx, cnt);
479 	if (test_bit(FLG_ACTIVE, &card->bc[1].bch.Flags))
480 		read_dma(&card->bc[1], idx, cnt);
481 }
482 
483 /* sync with current DMA address at start or after exception */
484 static void
resync(struct tiger_ch * bc,struct tiger_hw * card)485 resync(struct tiger_ch *bc, struct tiger_hw *card)
486 {
487 	card->send.dmacur = inl(card->base | NJ_DMA_READ_ADR);
488 	card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2;
489 	if (bc->free > card->send.size / 2)
490 		bc->free = card->send.size / 2;
491 	/* currently we simple sync to the next complete free area
492 	 * this hast the advantage that we have always maximum time to
493 	 * handle TX irq
494 	 */
495 	if (card->send.idx < ((card->send.size / 2) - 1))
496 		bc->idx = (card->recv.size / 2) - 1;
497 	else
498 		bc->idx = card->recv.size - 1;
499 	bc->txstate = TX_RUN;
500 	pr_debug("%s: %s B%1d free %d idx %d/%d\n", card->name,
501 		 __func__, bc->bch.nr, bc->free, bc->idx, card->send.idx);
502 }
503 
504 static int bc_next_frame(struct tiger_ch *);
505 
506 static void
fill_hdlc_flag(struct tiger_ch * bc)507 fill_hdlc_flag(struct tiger_ch *bc)
508 {
509 	struct tiger_hw *card = bc->bch.hw;
510 	int count, i;
511 	u32 m, v;
512 	u8  *p;
513 
514 	if (bc->free == 0)
515 		return;
516 	pr_debug("%s: %s B%1d %d state %x idx %d/%d\n", card->name,
517 		 __func__, bc->bch.nr, bc->free, bc->txstate,
518 		 bc->idx, card->send.idx);
519 	if (bc->txstate & (TX_IDLE | TX_INIT | TX_UNDERRUN))
520 		resync(bc, card);
521 	count = isdnhdlc_encode(&bc->hsend, NULL, 0, &i,
522 				bc->hsbuf, bc->free);
523 	pr_debug("%s: B%1d hdlc encoded %d flags\n", card->name,
524 		 bc->bch.nr, count);
525 	bc->free -= count;
526 	p = bc->hsbuf;
527 	m = (bc->bch.nr & 1) ? 0xffffff00 : 0xffff00ff;
528 	for (i = 0; i < count; i++) {
529 		if (bc->idx >= card->send.size)
530 			bc->idx = 0;
531 		v = card->send.start[bc->idx];
532 		v &= m;
533 		v |= (bc->bch.nr & 1) ? (u32)(p[i]) : ((u32)(p[i])) << 8;
534 		card->send.start[bc->idx++] = v;
535 	}
536 	if (debug & DEBUG_HW_BFIFO) {
537 		snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ",
538 			 bc->bch.nr, card->name, count);
539 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, p, count);
540 	}
541 }
542 
543 static void
fill_dma(struct tiger_ch * bc)544 fill_dma(struct tiger_ch *bc)
545 {
546 	struct tiger_hw *card = bc->bch.hw;
547 	int count, i;
548 	u32 m, v;
549 	u8  *p;
550 
551 	if (bc->free == 0)
552 		return;
553 	count = bc->bch.tx_skb->len - bc->bch.tx_idx;
554 	if (count <= 0)
555 		return;
556 	pr_debug("%s: %s B%1d %d/%d/%d/%d state %x idx %d/%d\n", card->name,
557 		 __func__, bc->bch.nr, count, bc->free, bc->bch.tx_idx,
558 		 bc->bch.tx_skb->len, bc->txstate, bc->idx, card->send.idx);
559 	if (bc->txstate & (TX_IDLE | TX_INIT | TX_UNDERRUN))
560 		resync(bc, card);
561 	p = bc->bch.tx_skb->data + bc->bch.tx_idx;
562 	if (test_bit(FLG_HDLC, &bc->bch.Flags)) {
563 		count = isdnhdlc_encode(&bc->hsend, p, count, &i,
564 					bc->hsbuf, bc->free);
565 		pr_debug("%s: B%1d hdlc encoded %d in %d\n", card->name,
566 			 bc->bch.nr, i, count);
567 		bc->bch.tx_idx += i;
568 		bc->free -= count;
569 		p = bc->hsbuf;
570 	} else {
571 		if (count > bc->free)
572 			count = bc->free;
573 		bc->bch.tx_idx += count;
574 		bc->free -= count;
575 	}
576 	m = (bc->bch.nr & 1) ? 0xffffff00 : 0xffff00ff;
577 	for (i = 0; i < count; i++) {
578 		if (bc->idx >= card->send.size)
579 			bc->idx = 0;
580 		v = card->send.start[bc->idx];
581 		v &= m;
582 		v |= (bc->bch.nr & 1) ? (u32)(p[i]) : ((u32)(p[i])) << 8;
583 		card->send.start[bc->idx++] = v;
584 	}
585 	if (debug & DEBUG_HW_BFIFO) {
586 		snprintf(card->log, LOG_SIZE, "B%1d-send %s %d ",
587 			 bc->bch.nr, card->name, count);
588 		print_hex_dump_bytes(card->log, DUMP_PREFIX_OFFSET, p, count);
589 	}
590 	if (bc->free)
591 		bc_next_frame(bc);
592 }
593 
594 
595 static int
bc_next_frame(struct tiger_ch * bc)596 bc_next_frame(struct tiger_ch *bc)
597 {
598 	if (bc->bch.tx_skb && bc->bch.tx_idx < bc->bch.tx_skb->len)
599 		fill_dma(bc);
600 	else {
601 		if (bc->bch.tx_skb) {
602 			/* send confirm, on trans, free on hdlc. */
603 			if (test_bit(FLG_TRANSPARENT, &bc->bch.Flags))
604 				confirm_Bsend(&bc->bch);
605 			dev_kfree_skb(bc->bch.tx_skb);
606 		}
607 		if (get_next_bframe(&bc->bch))
608 			fill_dma(bc);
609 		else
610 			return 0;
611 	}
612 	return 1;
613 }
614 
615 static void
send_tiger_bc(struct tiger_hw * card,struct tiger_ch * bc)616 send_tiger_bc(struct tiger_hw *card, struct tiger_ch *bc)
617 {
618 	int ret;
619 
620 	bc->free += card->send.size / 2;
621 	if (bc->free >= card->send.size) {
622 		if (!(bc->txstate & (TX_UNDERRUN | TX_INIT))) {
623 			pr_info("%s: B%1d TX underrun state %x\n", card->name,
624 				bc->bch.nr, bc->txstate);
625 			bc->txstate |= TX_UNDERRUN;
626 		}
627 		bc->free = card->send.size;
628 	}
629 	ret = bc_next_frame(bc);
630 	if (!ret) {
631 		if (test_bit(FLG_HDLC, &bc->bch.Flags)) {
632 			fill_hdlc_flag(bc);
633 			return;
634 		}
635 		pr_debug("%s: B%1d TX no data free %d idx %d/%d\n", card->name,
636 			 bc->bch.nr, bc->free, bc->idx, card->send.idx);
637 		if (!(bc->txstate & (TX_IDLE | TX_INIT))) {
638 			fill_mem(bc, bc->idx, bc->free, 0xff);
639 			if (bc->free == card->send.size)
640 				bc->txstate |= TX_IDLE;
641 		}
642 	}
643 }
644 
645 static void
send_tiger(struct tiger_hw * card,u8 irq_stat)646 send_tiger(struct tiger_hw *card, u8 irq_stat)
647 {
648 	int i;
649 
650 	/* Note send is via the READ DMA channel */
651 	if ((irq_stat & card->last_is0) & NJ_IRQM0_RD_MASK) {
652 		pr_info("%s: tiger warn write double dma %x/%x\n",
653 			card->name, irq_stat, card->last_is0);
654 		return;
655 	} else {
656 		card->last_is0 &= ~NJ_IRQM0_RD_MASK;
657 		card->last_is0 |= (irq_stat & NJ_IRQM0_RD_MASK);
658 	}
659 	for (i = 0; i < 2; i++) {
660 		if (test_bit(FLG_ACTIVE, &card->bc[i].bch.Flags))
661 			send_tiger_bc(card, &card->bc[i]);
662 	}
663 }
664 
665 static irqreturn_t
nj_irq(int intno,void * dev_id)666 nj_irq(int intno, void *dev_id)
667 {
668 	struct tiger_hw *card = dev_id;
669 	u8 val, s1val, s0val;
670 
671 	spin_lock(&card->lock);
672 	s0val = inb(card->base | NJ_IRQSTAT0);
673 	s1val = inb(card->base | NJ_IRQSTAT1);
674 	if ((s1val & NJ_ISACIRQ) && (s0val == 0)) {
675 		/* shared IRQ */
676 		spin_unlock(&card->lock);
677 		return IRQ_NONE;
678 	}
679 	pr_debug("%s: IRQSTAT0 %02x IRQSTAT1 %02x\n", card->name, s0val, s1val);
680 	card->irqcnt++;
681 	if (!(s1val & NJ_ISACIRQ)) {
682 		val = ReadISAC_nj(card, ISAC_ISTA);
683 		if (val)
684 			mISDNisac_irq(&card->isac, val);
685 	}
686 
687 	if (s0val)
688 		/* write to clear */
689 		outb(s0val, card->base | NJ_IRQSTAT0);
690 	else
691 		goto end;
692 	s1val = s0val;
693 	/* set bits in sval to indicate which page is free */
694 	card->recv.dmacur = inl(card->base | NJ_DMA_WRITE_ADR);
695 	card->recv.idx = (card->recv.dmacur - card->recv.dmastart) >> 2;
696 	if (card->recv.dmacur < card->recv.dmairq)
697 		s0val = 0x08;	/* the 2nd write area is free */
698 	else
699 		s0val = 0x04;	/* the 1st write area is free */
700 
701 	card->send.dmacur = inl(card->base | NJ_DMA_READ_ADR);
702 	card->send.idx = (card->send.dmacur - card->send.dmastart) >> 2;
703 	if (card->send.dmacur < card->send.dmairq)
704 		s0val |= 0x02;	/* the 2nd read area is free */
705 	else
706 		s0val |= 0x01;	/* the 1st read area is free */
707 
708 	pr_debug("%s: DMA Status %02x/%02x/%02x %d/%d\n", card->name,
709 		 s1val, s0val, card->last_is0,
710 		 card->recv.idx, card->send.idx);
711 	/* test if we have a DMA interrupt */
712 	if (s0val != card->last_is0) {
713 		if ((s0val & NJ_IRQM0_RD_MASK) !=
714 		    (card->last_is0 & NJ_IRQM0_RD_MASK))
715 			/* got a write dma int */
716 			send_tiger(card, s0val);
717 		if ((s0val & NJ_IRQM0_WR_MASK) !=
718 		    (card->last_is0 & NJ_IRQM0_WR_MASK))
719 			/* got a read dma int */
720 			recv_tiger(card, s0val);
721 	}
722 end:
723 	spin_unlock(&card->lock);
724 	return IRQ_HANDLED;
725 }
726 
727 static int
nj_l2l1B(struct mISDNchannel * ch,struct sk_buff * skb)728 nj_l2l1B(struct mISDNchannel *ch, struct sk_buff *skb)
729 {
730 	int ret = -EINVAL;
731 	struct bchannel *bch = container_of(ch, struct bchannel, ch);
732 	struct tiger_ch *bc = container_of(bch, struct tiger_ch, bch);
733 	struct tiger_hw *card = bch->hw;
734 	struct mISDNhead *hh = mISDN_HEAD_P(skb);
735 	u32 id;
736 	u_long flags;
737 
738 	switch (hh->prim) {
739 	case PH_DATA_REQ:
740 		spin_lock_irqsave(&card->lock, flags);
741 		ret = bchannel_senddata(bch, skb);
742 		if (ret > 0) { /* direct TX */
743 			id = hh->id; /* skb can be freed */
744 			fill_dma(bc);
745 			ret = 0;
746 			spin_unlock_irqrestore(&card->lock, flags);
747 			if (!test_bit(FLG_TRANSPARENT, &bch->Flags))
748 				queue_ch_frame(ch, PH_DATA_CNF, id, NULL);
749 		} else
750 			spin_unlock_irqrestore(&card->lock, flags);
751 		return ret;
752 	case PH_ACTIVATE_REQ:
753 		spin_lock_irqsave(&card->lock, flags);
754 		if (!test_and_set_bit(FLG_ACTIVE, &bch->Flags))
755 			ret = mode_tiger(bc, ch->protocol);
756 		else
757 			ret = 0;
758 		spin_unlock_irqrestore(&card->lock, flags);
759 		if (!ret)
760 			_queue_data(ch, PH_ACTIVATE_IND, MISDN_ID_ANY, 0,
761 				    NULL, GFP_KERNEL);
762 		break;
763 	case PH_DEACTIVATE_REQ:
764 		spin_lock_irqsave(&card->lock, flags);
765 		mISDN_clear_bchannel(bch);
766 		mode_tiger(bc, ISDN_P_NONE);
767 		spin_unlock_irqrestore(&card->lock, flags);
768 		_queue_data(ch, PH_DEACTIVATE_IND, MISDN_ID_ANY, 0,
769 			    NULL, GFP_KERNEL);
770 		ret = 0;
771 		break;
772 	}
773 	if (!ret)
774 		dev_kfree_skb(skb);
775 	return ret;
776 }
777 
778 static int
channel_bctrl(struct tiger_ch * bc,struct mISDN_ctrl_req * cq)779 channel_bctrl(struct tiger_ch *bc, struct mISDN_ctrl_req *cq)
780 {
781 	int ret = 0;
782 	struct tiger_hw *card  = bc->bch.hw;
783 
784 	switch (cq->op) {
785 	case MISDN_CTRL_GETOP:
786 		cq->op = 0;
787 		break;
788 		/* Nothing implemented yet */
789 	case MISDN_CTRL_FILL_EMPTY:
790 	default:
791 		pr_info("%s: %s unknown Op %x\n", card->name, __func__, cq->op);
792 		ret = -EINVAL;
793 		break;
794 	}
795 	return ret;
796 }
797 
798 static int
nj_bctrl(struct mISDNchannel * ch,u32 cmd,void * arg)799 nj_bctrl(struct mISDNchannel *ch, u32 cmd, void *arg)
800 {
801 	struct bchannel *bch = container_of(ch, struct bchannel, ch);
802 	struct tiger_ch *bc = container_of(bch, struct tiger_ch, bch);
803 	struct tiger_hw *card  = bch->hw;
804 	int ret = -EINVAL;
805 	u_long flags;
806 
807 	pr_debug("%s: %s cmd:%x %p\n", card->name, __func__, cmd, arg);
808 	switch (cmd) {
809 	case CLOSE_CHANNEL:
810 		test_and_clear_bit(FLG_OPEN, &bch->Flags);
811 		if (test_bit(FLG_ACTIVE, &bch->Flags)) {
812 			spin_lock_irqsave(&card->lock, flags);
813 			mISDN_freebchannel(bch);
814 			test_and_clear_bit(FLG_TX_BUSY, &bch->Flags);
815 			test_and_clear_bit(FLG_ACTIVE, &bch->Flags);
816 			mode_tiger(bc, ISDN_P_NONE);
817 			spin_unlock_irqrestore(&card->lock, flags);
818 		}
819 		ch->protocol = ISDN_P_NONE;
820 		ch->peer = NULL;
821 		module_put(THIS_MODULE);
822 		ret = 0;
823 		break;
824 	case CONTROL_CHANNEL:
825 		ret = channel_bctrl(bc, arg);
826 		break;
827 	default:
828 		pr_info("%s: %s unknown prim(%x)\n", card->name, __func__, cmd);
829 	}
830 	return ret;
831 }
832 
833 static int
channel_ctrl(struct tiger_hw * card,struct mISDN_ctrl_req * cq)834 channel_ctrl(struct tiger_hw *card, struct mISDN_ctrl_req *cq)
835 {
836 	int	ret = 0;
837 
838 	switch (cq->op) {
839 	case MISDN_CTRL_GETOP:
840 		cq->op = MISDN_CTRL_LOOP;
841 		break;
842 	case MISDN_CTRL_LOOP:
843 		/* cq->channel: 0 disable, 1 B1 loop 2 B2 loop, 3 both */
844 		if (cq->channel < 0 || cq->channel > 3) {
845 			ret = -EINVAL;
846 			break;
847 		}
848 		ret = card->isac.ctrl(&card->isac, HW_TESTLOOP, cq->channel);
849 		break;
850 	default:
851 		pr_info("%s: %s unknown Op %x\n", card->name, __func__, cq->op);
852 		ret = -EINVAL;
853 		break;
854 	}
855 	return ret;
856 }
857 
858 static int
open_bchannel(struct tiger_hw * card,struct channel_req * rq)859 open_bchannel(struct tiger_hw *card, struct channel_req *rq)
860 {
861 	struct bchannel *bch;
862 
863 	if (rq->adr.channel == 0 || rq->adr.channel > 2)
864 		return -EINVAL;
865 	if (rq->protocol == ISDN_P_NONE)
866 		return -EINVAL;
867 	bch = &card->bc[rq->adr.channel - 1].bch;
868 	if (test_and_set_bit(FLG_OPEN, &bch->Flags))
869 		return -EBUSY; /* b-channel can be only open once */
870 	test_and_clear_bit(FLG_FILLEMPTY, &bch->Flags);
871 	bch->ch.protocol = rq->protocol;
872 	rq->ch = &bch->ch;
873 	return 0;
874 }
875 
876 /*
877  * device control function
878  */
879 static int
nj_dctrl(struct mISDNchannel * ch,u32 cmd,void * arg)880 nj_dctrl(struct mISDNchannel *ch, u32 cmd, void *arg)
881 {
882 	struct mISDNdevice	*dev = container_of(ch, struct mISDNdevice, D);
883 	struct dchannel		*dch = container_of(dev, struct dchannel, dev);
884 	struct tiger_hw	*card = dch->hw;
885 	struct channel_req	*rq;
886 	int			err = 0;
887 
888 	pr_debug("%s: %s cmd:%x %p\n", card->name, __func__, cmd, arg);
889 	switch (cmd) {
890 	case OPEN_CHANNEL:
891 		rq = arg;
892 		if (rq->protocol == ISDN_P_TE_S0)
893 			err = card->isac.open(&card->isac, rq);
894 		else
895 			err = open_bchannel(card, rq);
896 		if (err)
897 			break;
898 		if (!try_module_get(THIS_MODULE))
899 			pr_info("%s: cannot get module\n", card->name);
900 		break;
901 	case CLOSE_CHANNEL:
902 		pr_debug("%s: dev(%d) close from %p\n", card->name, dch->dev.id,
903 			 __builtin_return_address(0));
904 		module_put(THIS_MODULE);
905 		break;
906 	case CONTROL_CHANNEL:
907 		err = channel_ctrl(card, arg);
908 		break;
909 	default:
910 		pr_debug("%s: %s unknown command %x\n",
911 			 card->name, __func__, cmd);
912 		return -EINVAL;
913 	}
914 	return err;
915 }
916 
917 static int
nj_init_card(struct tiger_hw * card)918 nj_init_card(struct tiger_hw *card)
919 {
920 	u_long flags;
921 	int ret;
922 
923 	spin_lock_irqsave(&card->lock, flags);
924 	nj_disable_hwirq(card);
925 	spin_unlock_irqrestore(&card->lock, flags);
926 
927 	card->irq = card->pdev->irq;
928 	if (request_irq(card->irq, nj_irq, IRQF_SHARED, card->name, card)) {
929 		pr_info("%s: couldn't get interrupt %d\n",
930 			card->name, card->irq);
931 		card->irq = -1;
932 		return -EIO;
933 	}
934 
935 	spin_lock_irqsave(&card->lock, flags);
936 	nj_reset(card);
937 	ret = card->isac.init(&card->isac);
938 	if (ret)
939 		goto error;
940 	ret = inittiger(card);
941 	if (ret)
942 		goto error;
943 	mode_tiger(&card->bc[0], ISDN_P_NONE);
944 	mode_tiger(&card->bc[1], ISDN_P_NONE);
945 error:
946 	spin_unlock_irqrestore(&card->lock, flags);
947 	return ret;
948 }
949 
950 
951 static void
nj_release(struct tiger_hw * card)952 nj_release(struct tiger_hw *card)
953 {
954 	u_long flags;
955 	int i;
956 
957 	if (card->base_s) {
958 		spin_lock_irqsave(&card->lock, flags);
959 		nj_disable_hwirq(card);
960 		mode_tiger(&card->bc[0], ISDN_P_NONE);
961 		mode_tiger(&card->bc[1], ISDN_P_NONE);
962 		card->isac.release(&card->isac);
963 		spin_unlock_irqrestore(&card->lock, flags);
964 		release_region(card->base, card->base_s);
965 		card->base_s = 0;
966 	}
967 	if (card->irq > 0)
968 		free_irq(card->irq, card);
969 	if (card->isac.dch.dev.dev.class)
970 		mISDN_unregister_device(&card->isac.dch.dev);
971 
972 	for (i = 0; i < 2; i++) {
973 		mISDN_freebchannel(&card->bc[i].bch);
974 		kfree(card->bc[i].hsbuf);
975 		kfree(card->bc[i].hrbuf);
976 	}
977 	if (card->dma_p)
978 		pci_free_consistent(card->pdev, NJ_DMA_SIZE,
979 				    card->dma_p, card->dma);
980 	write_lock_irqsave(&card_lock, flags);
981 	list_del(&card->list);
982 	write_unlock_irqrestore(&card_lock, flags);
983 	pci_clear_master(card->pdev);
984 	pci_disable_device(card->pdev);
985 	pci_set_drvdata(card->pdev, NULL);
986 	kfree(card);
987 }
988 
989 
990 static int
nj_setup(struct tiger_hw * card)991 nj_setup(struct tiger_hw *card)
992 {
993 	card->base = pci_resource_start(card->pdev, 0);
994 	card->base_s = pci_resource_len(card->pdev, 0);
995 	if (!request_region(card->base, card->base_s, card->name)) {
996 		pr_info("%s: NETjet config port %#x-%#x already in use\n",
997 			card->name, card->base,
998 			(u32)(card->base + card->base_s - 1));
999 		card->base_s = 0;
1000 		return -EIO;
1001 	}
1002 	ASSIGN_FUNC(nj, ISAC, card->isac);
1003 	return 0;
1004 }
1005 
1006 
1007 static int __devinit
setup_instance(struct tiger_hw * card)1008 setup_instance(struct tiger_hw *card)
1009 {
1010 	int i, err;
1011 	u_long flags;
1012 
1013 	snprintf(card->name, MISDN_MAX_IDLEN - 1, "netjet.%d", nj_cnt + 1);
1014 	write_lock_irqsave(&card_lock, flags);
1015 	list_add_tail(&card->list, &Cards);
1016 	write_unlock_irqrestore(&card_lock, flags);
1017 
1018 	_set_debug(card);
1019 	card->isac.name = card->name;
1020 	spin_lock_init(&card->lock);
1021 	card->isac.hwlock = &card->lock;
1022 	mISDNisac_init(&card->isac, card);
1023 
1024 	card->isac.dch.dev.Bprotocols = (1 << (ISDN_P_B_RAW & ISDN_P_B_MASK)) |
1025 		(1 << (ISDN_P_B_HDLC & ISDN_P_B_MASK));
1026 	card->isac.dch.dev.D.ctrl = nj_dctrl;
1027 	for (i = 0; i < 2; i++) {
1028 		card->bc[i].bch.nr = i + 1;
1029 		set_channelmap(i + 1, card->isac.dch.dev.channelmap);
1030 		mISDN_initbchannel(&card->bc[i].bch, MAX_DATA_MEM);
1031 		card->bc[i].bch.hw = card;
1032 		card->bc[i].bch.ch.send = nj_l2l1B;
1033 		card->bc[i].bch.ch.ctrl = nj_bctrl;
1034 		card->bc[i].bch.ch.nr = i + 1;
1035 		list_add(&card->bc[i].bch.ch.list,
1036 			 &card->isac.dch.dev.bchannels);
1037 		card->bc[i].bch.hw = card;
1038 	}
1039 	err = nj_setup(card);
1040 	if (err)
1041 		goto error;
1042 	err = mISDN_register_device(&card->isac.dch.dev, &card->pdev->dev,
1043 				    card->name);
1044 	if (err)
1045 		goto error;
1046 	err = nj_init_card(card);
1047 	if (!err)  {
1048 		nj_cnt++;
1049 		pr_notice("Netjet %d cards installed\n", nj_cnt);
1050 		return 0;
1051 	}
1052 error:
1053 	nj_release(card);
1054 	return err;
1055 }
1056 
1057 static int __devinit
nj_probe(struct pci_dev * pdev,const struct pci_device_id * ent)1058 nj_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
1059 {
1060 	int err = -ENOMEM;
1061 	int cfg;
1062 	struct tiger_hw *card;
1063 
1064 	if (pdev->subsystem_vendor == 0x8086 &&
1065 	    pdev->subsystem_device == 0x0003) {
1066 		pr_notice("Netjet: Digium X100P/X101P not handled\n");
1067 		return -ENODEV;
1068 	}
1069 
1070 	if (pdev->subsystem_vendor == 0x55 &&
1071 	    pdev->subsystem_device == 0x02) {
1072 		pr_notice("Netjet: Enter!Now not handled yet\n");
1073 		return -ENODEV;
1074 	}
1075 
1076 	if (pdev->subsystem_vendor == 0xb100 &&
1077 	    pdev->subsystem_device == 0x0003) {
1078 		pr_notice("Netjet: Digium TDM400P not handled yet\n");
1079 		return -ENODEV;
1080 	}
1081 
1082 	card = kzalloc(sizeof(struct tiger_hw), GFP_ATOMIC);
1083 	if (!card) {
1084 		pr_info("No kmem for Netjet\n");
1085 		return err;
1086 	}
1087 
1088 	card->pdev = pdev;
1089 
1090 	err = pci_enable_device(pdev);
1091 	if (err) {
1092 		kfree(card);
1093 		return err;
1094 	}
1095 
1096 	printk(KERN_INFO "nj_probe(mISDN): found adapter at %s\n",
1097 	       pci_name(pdev));
1098 
1099 	pci_set_master(pdev);
1100 
1101 	/* the TJ300 and TJ320 must be detected, the IRQ handling is different
1102 	 * unfortunately the chips use the same device ID, but the TJ320 has
1103 	 * the bit20 in status PCI cfg register set
1104 	 */
1105 	pci_read_config_dword(pdev, 0x04, &cfg);
1106 	if (cfg & 0x00100000)
1107 		card->typ = NETJET_S_TJ320;
1108 	else
1109 		card->typ = NETJET_S_TJ300;
1110 
1111 	card->base = pci_resource_start(pdev, 0);
1112 	card->irq = pdev->irq;
1113 	pci_set_drvdata(pdev, card);
1114 	err = setup_instance(card);
1115 	if (err)
1116 		pci_set_drvdata(pdev, NULL);
1117 
1118 	return err;
1119 }
1120 
1121 
nj_remove(struct pci_dev * pdev)1122 static void __devexit nj_remove(struct pci_dev *pdev)
1123 {
1124 	struct tiger_hw *card = pci_get_drvdata(pdev);
1125 
1126 	if (card)
1127 		nj_release(card);
1128 	else
1129 		pr_info("%s drvdata already removed\n", __func__);
1130 }
1131 
1132 /* We cannot select cards with PCI_SUB... IDs, since here are cards with
1133  * SUB IDs set to PCI_ANY_ID, so we need to match all and reject
1134  * known other cards which not work with this driver - see probe function */
1135 static struct pci_device_id nj_pci_ids[] __devinitdata = {
1136 	{ PCI_VENDOR_ID_TIGERJET, PCI_DEVICE_ID_TIGERJET_300,
1137 	  PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0},
1138 	{ }
1139 };
1140 MODULE_DEVICE_TABLE(pci, nj_pci_ids);
1141 
1142 static struct pci_driver nj_driver = {
1143 	.name = "netjet",
1144 	.probe = nj_probe,
1145 	.remove = __devexit_p(nj_remove),
1146 	.id_table = nj_pci_ids,
1147 };
1148 
nj_init(void)1149 static int __init nj_init(void)
1150 {
1151 	int err;
1152 
1153 	pr_notice("Netjet PCI driver Rev. %s\n", NETJET_REV);
1154 	err = pci_register_driver(&nj_driver);
1155 	return err;
1156 }
1157 
nj_cleanup(void)1158 static void __exit nj_cleanup(void)
1159 {
1160 	pci_unregister_driver(&nj_driver);
1161 }
1162 
1163 module_init(nj_init);
1164 module_exit(nj_cleanup);
1165