1/* $Id: hfc_pci.c,v 1.48.2.4 2004/02/11 13:21:33 keil Exp $
2 *
3 * low level driver for CCD's hfc-pci based cards
4 *
5 * Author       Werner Cornelius
6 *              based on existing driver for CCD hfc ISA cards
7 * Copyright    by Werner Cornelius  <werner@isdn4linux.de>
8 *              by Karsten Keil      <keil@isdn4linux.de>
9 *
10 * This software may be used and distributed according to the terms
11 * of the GNU General Public License, incorporated herein by reference.
12 *
13 * For changes and modifications please read
14 * Documentation/isdn/HiSax.cert
15 *
16 */
17
18#include <linux/init.h>
19#include "hisax.h"
20#include "hfc_pci.h"
21#include "isdnl1.h"
22#include <linux/pci.h>
23#include <linux/sched.h>
24#include <linux/interrupt.h>
25
26static const char *hfcpci_revision = "$Revision: 1.48.2.4 $";
27
28/* table entry in the PCI devices list */
29typedef struct {
30	int vendor_id;
31	int device_id;
32	char *vendor_name;
33	char *card_name;
34} PCI_ENTRY;
35
36#define NT_T1_COUNT	20	/* number of 3.125ms interrupts for G2 timeout */
37#define CLKDEL_TE	0x0e	/* CLKDEL in TE mode */
38#define CLKDEL_NT	0x6c	/* CLKDEL in NT mode */
39
40static const PCI_ENTRY id_list[] =
41{
42	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_2BD0, "CCD/Billion/Asuscom", "2BD0"},
43	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B000, "Billion", "B000"},
44	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B006, "Billion", "B006"},
45	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B007, "Billion", "B007"},
46	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B008, "Billion", "B008"},
47	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B009, "Billion", "B009"},
48	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B00A, "Billion", "B00A"},
49	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B00B, "Billion", "B00B"},
50	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B00C, "Billion", "B00C"},
51	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B100, "Seyeon", "B100"},
52	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B700, "Primux II S0", "B700"},
53	{PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B701, "Primux II S0 NT", "B701"},
54	{PCI_VENDOR_ID_ABOCOM, PCI_DEVICE_ID_ABOCOM_2BD1, "Abocom/Magitek", "2BD1"},
55	{PCI_VENDOR_ID_ASUSTEK, PCI_DEVICE_ID_ASUSTEK_0675, "Asuscom/Askey", "675"},
56	{PCI_VENDOR_ID_BERKOM, PCI_DEVICE_ID_BERKOM_T_CONCEPT, "German telekom", "T-Concept"},
57	{PCI_VENDOR_ID_BERKOM, PCI_DEVICE_ID_BERKOM_A1T, "German telekom", "A1T"},
58	{PCI_VENDOR_ID_ANIGMA, PCI_DEVICE_ID_ANIGMA_MC145575, "Motorola MC145575", "MC145575"},
59	{PCI_VENDOR_ID_ZOLTRIX, PCI_DEVICE_ID_ZOLTRIX_2BD0, "Zoltrix", "2BD0"},
60	{PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_IOM2_E, "Digi International", "Digi DataFire Micro V IOM2 (Europe)"},
61	{PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_E, "Digi International", "Digi DataFire Micro V (Europe)"},
62	{PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_IOM2_A, "Digi International", "Digi DataFire Micro V IOM2 (North America)"},
63	{PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_A, "Digi International", "Digi DataFire Micro V (North America)"},
64	{PCI_VENDOR_ID_SITECOM, PCI_DEVICE_ID_SITECOM_DC105V2, "Sitecom Europe", "DC-105 ISDN PCI"},
65	{0, 0, NULL, NULL},
66};
67
68
69/******************************************/
70/* free hardware resources used by driver */
71/******************************************/
72static void
73release_io_hfcpci(struct IsdnCardState *cs)
74{
75	printk(KERN_INFO "HiSax: release hfcpci at %p\n",
76	       cs->hw.hfcpci.pci_io);
77	cs->hw.hfcpci.int_m2 = 0;					/* interrupt output off ! */
78	Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
79	Write_hfc(cs, HFCPCI_CIRM, HFCPCI_RESET);			/* Reset On */
80	mdelay(10);
81	Write_hfc(cs, HFCPCI_CIRM, 0);					/* Reset Off */
82	mdelay(10);
83	Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
84	pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, 0);	/* disable memory mapped ports + busmaster */
85	del_timer(&cs->hw.hfcpci.timer);
86	pci_free_consistent(cs->hw.hfcpci.dev, 0x8000,
87			    cs->hw.hfcpci.fifos, cs->hw.hfcpci.dma);
88	cs->hw.hfcpci.fifos = NULL;
89	iounmap((void *)cs->hw.hfcpci.pci_io);
90}
91
92/********************************************************************************/
93/* function called to reset the HFC PCI chip. A complete software reset of chip */
94/* and fifos is done.                                                           */
95/********************************************************************************/
96static void
97reset_hfcpci(struct IsdnCardState *cs)
98{
99	pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, PCI_ENA_MEMIO);	/* enable memory mapped ports, disable busmaster */
100	cs->hw.hfcpci.int_m2 = 0;	/* interrupt output off ! */
101	Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
102
103	printk(KERN_INFO "HFC_PCI: resetting card\n");
104	pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, PCI_ENA_MEMIO + PCI_ENA_MASTER);	/* enable memory ports + busmaster */
105	Write_hfc(cs, HFCPCI_CIRM, HFCPCI_RESET);	/* Reset On */
106	mdelay(10);
107	Write_hfc(cs, HFCPCI_CIRM, 0);	/* Reset Off */
108	mdelay(10);
109	if (Read_hfc(cs, HFCPCI_STATUS) & 2)
110		printk(KERN_WARNING "HFC-PCI init bit busy\n");
111
112	cs->hw.hfcpci.fifo_en = 0x30;	/* only D fifos enabled */
113	Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
114
115	cs->hw.hfcpci.trm = 0 + HFCPCI_BTRANS_THRESMASK;	/* no echo connect , threshold */
116	Write_hfc(cs, HFCPCI_TRM, cs->hw.hfcpci.trm);
117
118	Write_hfc(cs, HFCPCI_CLKDEL, CLKDEL_TE); /* ST-Bit delay for TE-Mode */
119	cs->hw.hfcpci.sctrl_e = HFCPCI_AUTO_AWAKE;
120	Write_hfc(cs, HFCPCI_SCTRL_E, cs->hw.hfcpci.sctrl_e);	/* S/T Auto awake */
121	cs->hw.hfcpci.bswapped = 0;	/* no exchange */
122	cs->hw.hfcpci.nt_mode = 0;	/* we are in TE mode */
123	cs->hw.hfcpci.ctmt = HFCPCI_TIM3_125 | HFCPCI_AUTO_TIMER;
124	Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt);
125
126	cs->hw.hfcpci.int_m1 = HFCPCI_INTS_DTRANS | HFCPCI_INTS_DREC |
127		HFCPCI_INTS_L1STATE | HFCPCI_INTS_TIMER;
128	Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
129
130	/* Clear already pending ints */
131	if (Read_hfc(cs, HFCPCI_INT_S1));
132
133	Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 2);	/* HFC ST 2 */
134	udelay(10);
135	Write_hfc(cs, HFCPCI_STATES, 2);	/* HFC ST 2 */
136	cs->hw.hfcpci.mst_m = HFCPCI_MASTER;	/* HFC Master Mode */
137
138	Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
139	cs->hw.hfcpci.sctrl = 0x40;	/* set tx_lo mode, error in datasheet ! */
140	Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);
141	cs->hw.hfcpci.sctrl_r = 0;
142	Write_hfc(cs, HFCPCI_SCTRL_R, cs->hw.hfcpci.sctrl_r);
143
144	/* Init GCI/IOM2 in master mode */
145	/* Slots 0 and 1 are set for B-chan 1 and 2 */
146	/* D- and monitor/CI channel are not enabled */
147	/* STIO1 is used as output for data, B1+B2 from ST->IOM+HFC */
148	/* STIO2 is used as data input, B1+B2 from IOM->ST */
149	/* ST B-channel send disabled -> continuous 1s */
150	/* The IOM slots are always enabled */
151	cs->hw.hfcpci.conn = 0x36;	/* set data flow directions */
152	Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
153	Write_hfc(cs, HFCPCI_B1_SSL, 0x80);	/* B1-Slot 0 STIO1 out enabled */
154	Write_hfc(cs, HFCPCI_B2_SSL, 0x81);	/* B2-Slot 1 STIO1 out enabled */
155	Write_hfc(cs, HFCPCI_B1_RSL, 0x80);	/* B1-Slot 0 STIO2 in enabled */
156	Write_hfc(cs, HFCPCI_B2_RSL, 0x81);	/* B2-Slot 1 STIO2 in enabled */
157
158	/* Finally enable IRQ output */
159	cs->hw.hfcpci.int_m2 = HFCPCI_IRQ_ENABLE;
160	Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
161	if (Read_hfc(cs, HFCPCI_INT_S1));
162}
163
164/***************************************************/
165/* Timer function called when kernel timer expires */
166/***************************************************/
167static void
168hfcpci_Timer(struct IsdnCardState *cs)
169{
170	cs->hw.hfcpci.timer.expires = jiffies + 75;
171	/* WD RESET */
172/*      WriteReg(cs, HFCD_DATA, HFCD_CTMT, cs->hw.hfcpci.ctmt | 0x80);
173	add_timer(&cs->hw.hfcpci.timer);
174*/
175}
176
177
178/*********************************/
179/* schedule a new D-channel task */
180/*********************************/
181static void
182sched_event_D_pci(struct IsdnCardState *cs, int event)
183{
184	test_and_set_bit(event, &cs->event);
185	schedule_work(&cs->tqueue);
186}
187
188/*********************************/
189/* schedule a new b_channel task */
190/*********************************/
191static void
192hfcpci_sched_event(struct BCState *bcs, int event)
193{
194	test_and_set_bit(event, &bcs->event);
195	schedule_work(&bcs->tqueue);
196}
197
198/************************************************/
199/* select a b-channel entry matching and active */
200/************************************************/
201static
202struct BCState *
203Sel_BCS(struct IsdnCardState *cs, int channel)
204{
205	if (cs->bcs[0].mode && (cs->bcs[0].channel == channel))
206		return (&cs->bcs[0]);
207	else if (cs->bcs[1].mode && (cs->bcs[1].channel == channel))
208		return (&cs->bcs[1]);
209	else
210		return (NULL);
211}
212
213/***************************************/
214/* clear the desired B-channel rx fifo */
215/***************************************/
216static void hfcpci_clear_fifo_rx(struct IsdnCardState *cs, int fifo)
217{       u_char fifo_state;
218	bzfifo_type *bzr;
219
220	if (fifo) {
221		bzr = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b2;
222		fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B2RX;
223	} else {
224		bzr = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b1;
225		fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B1RX;
226	}
227	if (fifo_state)
228		cs->hw.hfcpci.fifo_en ^= fifo_state;
229	Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
230	cs->hw.hfcpci.last_bfifo_cnt[fifo] = 0;
231	bzr->za[MAX_B_FRAMES].z1 = B_FIFO_SIZE + B_SUB_VAL - 1;
232	bzr->za[MAX_B_FRAMES].z2 = bzr->za[MAX_B_FRAMES].z1;
233	bzr->f1 = MAX_B_FRAMES;
234	bzr->f2 = bzr->f1;	/* init F pointers to remain constant */
235	if (fifo_state)
236		cs->hw.hfcpci.fifo_en |= fifo_state;
237	Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
238}
239
240/***************************************/
241/* clear the desired B-channel tx fifo */
242/***************************************/
243static void hfcpci_clear_fifo_tx(struct IsdnCardState *cs, int fifo)
244{       u_char fifo_state;
245	bzfifo_type *bzt;
246
247	if (fifo) {
248		bzt = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b2;
249		fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B2TX;
250	} else {
251		bzt = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b1;
252		fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B1TX;
253	}
254	if (fifo_state)
255		cs->hw.hfcpci.fifo_en ^= fifo_state;
256	Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
257	bzt->za[MAX_B_FRAMES].z1 = B_FIFO_SIZE + B_SUB_VAL - 1;
258	bzt->za[MAX_B_FRAMES].z2 = bzt->za[MAX_B_FRAMES].z1;
259	bzt->f1 = MAX_B_FRAMES;
260	bzt->f2 = bzt->f1;	/* init F pointers to remain constant */
261	if (fifo_state)
262		cs->hw.hfcpci.fifo_en |= fifo_state;
263	Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
264}
265
266/*********************************************/
267/* read a complete B-frame out of the buffer */
268/*********************************************/
269static struct sk_buff
270*
271hfcpci_empty_fifo(struct BCState *bcs, bzfifo_type *bz, u_char *bdata, int count)
272{
273	u_char *ptr, *ptr1, new_f2;
274	struct sk_buff *skb;
275	struct IsdnCardState *cs = bcs->cs;
276	int total, maxlen, new_z2;
277	z_type *zp;
278
279	if ((cs->debug & L1_DEB_HSCX) && !(cs->debug & L1_DEB_HSCX_FIFO))
280		debugl1(cs, "hfcpci_empty_fifo");
281	zp = &bz->za[bz->f2];	/* point to Z-Regs */
282	new_z2 = zp->z2 + count;	/* new position in fifo */
283	if (new_z2 >= (B_FIFO_SIZE + B_SUB_VAL))
284		new_z2 -= B_FIFO_SIZE;	/* buffer wrap */
285	new_f2 = (bz->f2 + 1) & MAX_B_FRAMES;
286	if ((count > HSCX_BUFMAX + 3) || (count < 4) ||
287	    (*(bdata + (zp->z1 - B_SUB_VAL)))) {
288		if (cs->debug & L1_DEB_WARN)
289			debugl1(cs, "hfcpci_empty_fifo: incoming packet invalid length %d or crc", count);
290#ifdef ERROR_STATISTIC
291		bcs->err_inv++;
292#endif
293		bz->za[new_f2].z2 = new_z2;
294		bz->f2 = new_f2;	/* next buffer */
295		skb = NULL;
296	} else if (!(skb = dev_alloc_skb(count - 3)))
297		printk(KERN_WARNING "HFCPCI: receive out of memory\n");
298	else {
299		total = count;
300		count -= 3;
301		ptr = skb_put(skb, count);
302
303		if (zp->z2 + count <= B_FIFO_SIZE + B_SUB_VAL)
304			maxlen = count;		/* complete transfer */
305		else
306			maxlen = B_FIFO_SIZE + B_SUB_VAL - zp->z2;	/* maximum */
307
308		ptr1 = bdata + (zp->z2 - B_SUB_VAL);	/* start of data */
309		memcpy(ptr, ptr1, maxlen);	/* copy data */
310		count -= maxlen;
311
312		if (count) {	/* rest remaining */
313			ptr += maxlen;
314			ptr1 = bdata;	/* start of buffer */
315			memcpy(ptr, ptr1, count);	/* rest */
316		}
317		bz->za[new_f2].z2 = new_z2;
318		bz->f2 = new_f2;	/* next buffer */
319
320	}
321	return (skb);
322}
323
324/*******************************/
325/* D-channel receive procedure */
326/*******************************/
327static
328int
329receive_dmsg(struct IsdnCardState *cs)
330{
331	struct sk_buff *skb;
332	int maxlen;
333	int rcnt, total;
334	int count = 5;
335	u_char *ptr, *ptr1;
336	dfifo_type *df;
337	z_type *zp;
338
339	df = &((fifo_area *) (cs->hw.hfcpci.fifos))->d_chan.d_rx;
340	if (test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
341		debugl1(cs, "rec_dmsg blocked");
342		return (1);
343	}
344	while (((df->f1 & D_FREG_MASK) != (df->f2 & D_FREG_MASK)) && count--) {
345		zp = &df->za[df->f2 & D_FREG_MASK];
346		rcnt = zp->z1 - zp->z2;
347		if (rcnt < 0)
348			rcnt += D_FIFO_SIZE;
349		rcnt++;
350		if (cs->debug & L1_DEB_ISAC)
351			debugl1(cs, "hfcpci recd f1(%d) f2(%d) z1(%x) z2(%x) cnt(%d)",
352				df->f1, df->f2, zp->z1, zp->z2, rcnt);
353
354		if ((rcnt > MAX_DFRAME_LEN + 3) || (rcnt < 4) ||
355		    (df->data[zp->z1])) {
356			if (cs->debug & L1_DEB_WARN)
357				debugl1(cs, "empty_fifo hfcpci paket inv. len %d or crc %d", rcnt, df->data[zp->z1]);
358#ifdef ERROR_STATISTIC
359			cs->err_rx++;
360#endif
361			df->f2 = ((df->f2 + 1) & MAX_D_FRAMES) | (MAX_D_FRAMES + 1);	/* next buffer */
362			df->za[df->f2 & D_FREG_MASK].z2 = (zp->z2 + rcnt) & (D_FIFO_SIZE - 1);
363		} else if ((skb = dev_alloc_skb(rcnt - 3))) {
364			total = rcnt;
365			rcnt -= 3;
366			ptr = skb_put(skb, rcnt);
367
368			if (zp->z2 + rcnt <= D_FIFO_SIZE)
369				maxlen = rcnt;	/* complete transfer */
370			else
371				maxlen = D_FIFO_SIZE - zp->z2;	/* maximum */
372
373			ptr1 = df->data + zp->z2;	/* start of data */
374			memcpy(ptr, ptr1, maxlen);	/* copy data */
375			rcnt -= maxlen;
376
377			if (rcnt) {	/* rest remaining */
378				ptr += maxlen;
379				ptr1 = df->data;	/* start of buffer */
380				memcpy(ptr, ptr1, rcnt);	/* rest */
381			}
382			df->f2 = ((df->f2 + 1) & MAX_D_FRAMES) | (MAX_D_FRAMES + 1);	/* next buffer */
383			df->za[df->f2 & D_FREG_MASK].z2 = (zp->z2 + total) & (D_FIFO_SIZE - 1);
384
385			skb_queue_tail(&cs->rq, skb);
386			sched_event_D_pci(cs, D_RCVBUFREADY);
387		} else
388			printk(KERN_WARNING "HFC-PCI: D receive out of memory\n");
389	}
390	test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
391	return (1);
392}
393
394/*******************************************************************************/
395/* check for transparent receive data and read max one threshold size if avail */
396/*******************************************************************************/
397static int
398hfcpci_empty_fifo_trans(struct BCState *bcs, bzfifo_type *bz, u_char *bdata)
399{
400	unsigned short *z1r, *z2r;
401	int new_z2, fcnt, maxlen;
402	struct sk_buff *skb;
403	u_char *ptr, *ptr1;
404
405	z1r = &bz->za[MAX_B_FRAMES].z1;		/* pointer to z reg */
406	z2r = z1r + 1;
407
408	if (!(fcnt = *z1r - *z2r))
409		return (0);	/* no data avail */
410
411	if (fcnt <= 0)
412		fcnt += B_FIFO_SIZE;	/* bytes actually buffered */
413	if (fcnt > HFCPCI_BTRANS_THRESHOLD)
414		fcnt = HFCPCI_BTRANS_THRESHOLD;		/* limit size */
415
416	new_z2 = *z2r + fcnt;	/* new position in fifo */
417	if (new_z2 >= (B_FIFO_SIZE + B_SUB_VAL))
418		new_z2 -= B_FIFO_SIZE;	/* buffer wrap */
419
420	if (!(skb = dev_alloc_skb(fcnt)))
421		printk(KERN_WARNING "HFCPCI: receive out of memory\n");
422	else {
423		ptr = skb_put(skb, fcnt);
424		if (*z2r + fcnt <= B_FIFO_SIZE + B_SUB_VAL)
425			maxlen = fcnt;	/* complete transfer */
426		else
427			maxlen = B_FIFO_SIZE + B_SUB_VAL - *z2r;	/* maximum */
428
429		ptr1 = bdata + (*z2r - B_SUB_VAL);	/* start of data */
430		memcpy(ptr, ptr1, maxlen);	/* copy data */
431		fcnt -= maxlen;
432
433		if (fcnt) {	/* rest remaining */
434			ptr += maxlen;
435			ptr1 = bdata;	/* start of buffer */
436			memcpy(ptr, ptr1, fcnt);	/* rest */
437		}
438		skb_queue_tail(&bcs->rqueue, skb);
439		hfcpci_sched_event(bcs, B_RCVBUFREADY);
440	}
441
442	*z2r = new_z2;		/* new position */
443	return (1);
444}				/* hfcpci_empty_fifo_trans */
445
446/**********************************/
447/* B-channel main receive routine */
448/**********************************/
449static void
450main_rec_hfcpci(struct BCState *bcs)
451{
452	struct IsdnCardState *cs = bcs->cs;
453	int rcnt, real_fifo;
454	int receive, count = 5;
455	struct sk_buff *skb;
456	bzfifo_type *bz;
457	u_char *bdata;
458	z_type *zp;
459
460
461	if ((bcs->channel) && (!cs->hw.hfcpci.bswapped)) {
462		bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b2;
463		bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxdat_b2;
464		real_fifo = 1;
465	} else {
466		bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b1;
467		bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxdat_b1;
468		real_fifo = 0;
469	}
470Begin:
471	count--;
472	if (test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
473		debugl1(cs, "rec_data %d blocked", bcs->channel);
474		return;
475	}
476	if (bz->f1 != bz->f2) {
477		if (cs->debug & L1_DEB_HSCX)
478			debugl1(cs, "hfcpci rec %d f1(%d) f2(%d)",
479				bcs->channel, bz->f1, bz->f2);
480		zp = &bz->za[bz->f2];
481
482		rcnt = zp->z1 - zp->z2;
483		if (rcnt < 0)
484			rcnt += B_FIFO_SIZE;
485		rcnt++;
486		if (cs->debug & L1_DEB_HSCX)
487			debugl1(cs, "hfcpci rec %d z1(%x) z2(%x) cnt(%d)",
488				bcs->channel, zp->z1, zp->z2, rcnt);
489		if ((skb = hfcpci_empty_fifo(bcs, bz, bdata, rcnt))) {
490			skb_queue_tail(&bcs->rqueue, skb);
491			hfcpci_sched_event(bcs, B_RCVBUFREADY);
492		}
493		rcnt = bz->f1 - bz->f2;
494		if (rcnt < 0)
495			rcnt += MAX_B_FRAMES + 1;
496		if (cs->hw.hfcpci.last_bfifo_cnt[real_fifo] > rcnt + 1) {
497			rcnt = 0;
498			hfcpci_clear_fifo_rx(cs, real_fifo);
499		}
500		cs->hw.hfcpci.last_bfifo_cnt[real_fifo] = rcnt;
501		if (rcnt > 1)
502			receive = 1;
503		else
504			receive = 0;
505	} else if (bcs->mode == L1_MODE_TRANS)
506		receive = hfcpci_empty_fifo_trans(bcs, bz, bdata);
507	else
508		receive = 0;
509	test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
510	if (count && receive)
511		goto Begin;
512}
513
514/**************************/
515/* D-channel send routine */
516/**************************/
517static void
518hfcpci_fill_dfifo(struct IsdnCardState *cs)
519{
520	int fcnt;
521	int count, new_z1, maxlen;
522	dfifo_type *df;
523	u_char *src, *dst, new_f1;
524
525	if (!cs->tx_skb)
526		return;
527	if (cs->tx_skb->len <= 0)
528		return;
529
530	df = &((fifo_area *) (cs->hw.hfcpci.fifos))->d_chan.d_tx;
531
532	if (cs->debug & L1_DEB_ISAC)
533		debugl1(cs, "hfcpci_fill_Dfifo f1(%d) f2(%d) z1(f1)(%x)",
534			df->f1, df->f2,
535			df->za[df->f1 & D_FREG_MASK].z1);
536	fcnt = df->f1 - df->f2;	/* frame count actually buffered */
537	if (fcnt < 0)
538		fcnt += (MAX_D_FRAMES + 1);	/* if wrap around */
539	if (fcnt > (MAX_D_FRAMES - 1)) {
540		if (cs->debug & L1_DEB_ISAC)
541			debugl1(cs, "hfcpci_fill_Dfifo more as 14 frames");
542#ifdef ERROR_STATISTIC
543		cs->err_tx++;
544#endif
545		return;
546	}
547	/* now determine free bytes in FIFO buffer */
548	count = df->za[df->f2 & D_FREG_MASK].z2 - df->za[df->f1 & D_FREG_MASK].z1 - 1;
549	if (count <= 0)
550		count += D_FIFO_SIZE;	/* count now contains available bytes */
551
552	if (cs->debug & L1_DEB_ISAC)
553		debugl1(cs, "hfcpci_fill_Dfifo count(%u/%d)",
554			cs->tx_skb->len, count);
555	if (count < cs->tx_skb->len) {
556		if (cs->debug & L1_DEB_ISAC)
557			debugl1(cs, "hfcpci_fill_Dfifo no fifo mem");
558		return;
559	}
560	count = cs->tx_skb->len;	/* get frame len */
561	new_z1 = (df->za[df->f1 & D_FREG_MASK].z1 + count) & (D_FIFO_SIZE - 1);
562	new_f1 = ((df->f1 + 1) & D_FREG_MASK) | (D_FREG_MASK + 1);
563	src = cs->tx_skb->data;	/* source pointer */
564	dst = df->data + df->za[df->f1 & D_FREG_MASK].z1;
565	maxlen = D_FIFO_SIZE - df->za[df->f1 & D_FREG_MASK].z1;		/* end fifo */
566	if (maxlen > count)
567		maxlen = count;	/* limit size */
568	memcpy(dst, src, maxlen);	/* first copy */
569
570	count -= maxlen;	/* remaining bytes */
571	if (count) {
572		dst = df->data;	/* start of buffer */
573		src += maxlen;	/* new position */
574		memcpy(dst, src, count);
575	}
576	df->za[new_f1 & D_FREG_MASK].z1 = new_z1;	/* for next buffer */
577	df->za[df->f1 & D_FREG_MASK].z1 = new_z1;	/* new pos actual buffer */
578	df->f1 = new_f1;	/* next frame */
579
580	dev_kfree_skb_any(cs->tx_skb);
581	cs->tx_skb = NULL;
582}
583
584/**************************/
585/* B-channel send routine */
586/**************************/
587static void
588hfcpci_fill_fifo(struct BCState *bcs)
589{
590	struct IsdnCardState *cs = bcs->cs;
591	int maxlen, fcnt;
592	int count, new_z1;
593	bzfifo_type *bz;
594	u_char *bdata;
595	u_char new_f1, *src, *dst;
596	unsigned short *z1t, *z2t;
597
598	if (!bcs->tx_skb)
599		return;
600	if (bcs->tx_skb->len <= 0)
601		return;
602
603	if ((bcs->channel) && (!cs->hw.hfcpci.bswapped)) {
604		bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b2;
605		bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txdat_b2;
606	} else {
607		bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b1;
608		bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txdat_b1;
609	}
610
611	if (bcs->mode == L1_MODE_TRANS) {
612		z1t = &bz->za[MAX_B_FRAMES].z1;
613		z2t = z1t + 1;
614		if (cs->debug & L1_DEB_HSCX)
615			debugl1(cs, "hfcpci_fill_fifo_trans %d z1(%x) z2(%x)",
616				bcs->channel, *z1t, *z2t);
617		fcnt = *z2t - *z1t;
618		if (fcnt <= 0)
619			fcnt += B_FIFO_SIZE;	/* fcnt contains available bytes in fifo */
620		fcnt = B_FIFO_SIZE - fcnt;	/* remaining bytes to send */
621
622		while ((fcnt < 2 * HFCPCI_BTRANS_THRESHOLD) && (bcs->tx_skb)) {
623			if (bcs->tx_skb->len < B_FIFO_SIZE - fcnt) {
624				/* data is suitable for fifo */
625				count = bcs->tx_skb->len;
626
627				new_z1 = *z1t + count;	/* new buffer Position */
628				if (new_z1 >= (B_FIFO_SIZE + B_SUB_VAL))
629					new_z1 -= B_FIFO_SIZE;	/* buffer wrap */
630				src = bcs->tx_skb->data;	/* source pointer */
631				dst = bdata + (*z1t - B_SUB_VAL);
632				maxlen = (B_FIFO_SIZE + B_SUB_VAL) - *z1t;	/* end of fifo */
633				if (maxlen > count)
634					maxlen = count;		/* limit size */
635				memcpy(dst, src, maxlen);	/* first copy */
636
637				count -= maxlen;	/* remaining bytes */
638				if (count) {
639					dst = bdata;	/* start of buffer */
640					src += maxlen;	/* new position */
641					memcpy(dst, src, count);
642				}
643				bcs->tx_cnt -= bcs->tx_skb->len;
644				fcnt += bcs->tx_skb->len;
645				*z1t = new_z1;	/* now send data */
646			} else if (cs->debug & L1_DEB_HSCX)
647				debugl1(cs, "hfcpci_fill_fifo_trans %d frame length %d discarded",
648					bcs->channel, bcs->tx_skb->len);
649
650			if (test_bit(FLG_LLI_L1WAKEUP, &bcs->st->lli.flag) &&
651			    (PACKET_NOACK != bcs->tx_skb->pkt_type)) {
652				u_long	flags;
653				spin_lock_irqsave(&bcs->aclock, flags);
654				bcs->ackcnt += bcs->tx_skb->len;
655				spin_unlock_irqrestore(&bcs->aclock, flags);
656				schedule_event(bcs, B_ACKPENDING);
657			}
658
659			dev_kfree_skb_any(bcs->tx_skb);
660			bcs->tx_skb = skb_dequeue(&bcs->squeue);	/* fetch next data */
661		}
662		test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
663		return;
664	}
665	if (cs->debug & L1_DEB_HSCX)
666		debugl1(cs, "hfcpci_fill_fifo_hdlc %d f1(%d) f2(%d) z1(f1)(%x)",
667			bcs->channel, bz->f1, bz->f2,
668			bz->za[bz->f1].z1);
669
670	fcnt = bz->f1 - bz->f2;	/* frame count actually buffered */
671	if (fcnt < 0)
672		fcnt += (MAX_B_FRAMES + 1);	/* if wrap around */
673	if (fcnt > (MAX_B_FRAMES - 1)) {
674		if (cs->debug & L1_DEB_HSCX)
675			debugl1(cs, "hfcpci_fill_Bfifo more as 14 frames");
676		return;
677	}
678	/* now determine free bytes in FIFO buffer */
679	count = bz->za[bz->f2].z2 - bz->za[bz->f1].z1 - 1;
680	if (count <= 0)
681		count += B_FIFO_SIZE;	/* count now contains available bytes */
682
683	if (cs->debug & L1_DEB_HSCX)
684		debugl1(cs, "hfcpci_fill_fifo %d count(%u/%d),%lx",
685			bcs->channel, bcs->tx_skb->len,
686			count, current->state);
687
688	if (count < bcs->tx_skb->len) {
689		if (cs->debug & L1_DEB_HSCX)
690			debugl1(cs, "hfcpci_fill_fifo no fifo mem");
691		return;
692	}
693	count = bcs->tx_skb->len;	/* get frame len */
694	new_z1 = bz->za[bz->f1].z1 + count;	/* new buffer Position */
695	if (new_z1 >= (B_FIFO_SIZE + B_SUB_VAL))
696		new_z1 -= B_FIFO_SIZE;	/* buffer wrap */
697
698	new_f1 = ((bz->f1 + 1) & MAX_B_FRAMES);
699	src = bcs->tx_skb->data;	/* source pointer */
700	dst = bdata + (bz->za[bz->f1].z1 - B_SUB_VAL);
701	maxlen = (B_FIFO_SIZE + B_SUB_VAL) - bz->za[bz->f1].z1;		/* end fifo */
702	if (maxlen > count)
703		maxlen = count;	/* limit size */
704	memcpy(dst, src, maxlen);	/* first copy */
705
706	count -= maxlen;	/* remaining bytes */
707	if (count) {
708		dst = bdata;	/* start of buffer */
709		src += maxlen;	/* new position */
710		memcpy(dst, src, count);
711	}
712	bcs->tx_cnt -= bcs->tx_skb->len;
713	if (test_bit(FLG_LLI_L1WAKEUP, &bcs->st->lli.flag) &&
714	    (PACKET_NOACK != bcs->tx_skb->pkt_type)) {
715		u_long	flags;
716		spin_lock_irqsave(&bcs->aclock, flags);
717		bcs->ackcnt += bcs->tx_skb->len;
718		spin_unlock_irqrestore(&bcs->aclock, flags);
719		schedule_event(bcs, B_ACKPENDING);
720	}
721
722	bz->za[new_f1].z1 = new_z1;	/* for next buffer */
723	bz->f1 = new_f1;	/* next frame */
724
725	dev_kfree_skb_any(bcs->tx_skb);
726	bcs->tx_skb = NULL;
727	test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
728}
729
730/**********************************************/
731/* D-channel l1 state call for leased NT-mode */
732/**********************************************/
733static void
734dch_nt_l2l1(struct PStack *st, int pr, void *arg)
735{
736	struct IsdnCardState *cs = (struct IsdnCardState *) st->l1.hardware;
737
738	switch (pr) {
739	case (PH_DATA | REQUEST):
740	case (PH_PULL | REQUEST):
741	case (PH_PULL | INDICATION):
742		st->l1.l1hw(st, pr, arg);
743		break;
744	case (PH_ACTIVATE | REQUEST):
745		st->l1.l1l2(st, PH_ACTIVATE | CONFIRM, NULL);
746		break;
747	case (PH_TESTLOOP | REQUEST):
748		if (1 & (long) arg)
749			debugl1(cs, "PH_TEST_LOOP B1");
750		if (2 & (long) arg)
751			debugl1(cs, "PH_TEST_LOOP B2");
752		if (!(3 & (long) arg))
753			debugl1(cs, "PH_TEST_LOOP DISABLED");
754		st->l1.l1hw(st, HW_TESTLOOP | REQUEST, arg);
755		break;
756	default:
757		if (cs->debug)
758			debugl1(cs, "dch_nt_l2l1 msg %04X unhandled", pr);
759		break;
760	}
761}
762
763
764
765/***********************/
766/* set/reset echo mode */
767/***********************/
768static int
769hfcpci_auxcmd(struct IsdnCardState *cs, isdn_ctrl *ic)
770{
771	u_long	flags;
772	int	i = *(unsigned int *) ic->parm.num;
773
774	if ((ic->arg == 98) &&
775	    (!(cs->hw.hfcpci.int_m1 & (HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC + HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC)))) {
776		spin_lock_irqsave(&cs->lock, flags);
777		Write_hfc(cs, HFCPCI_CLKDEL, CLKDEL_NT); /* ST-Bit delay for NT-Mode */
778		Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 0);	/* HFC ST G0 */
779		udelay(10);
780		cs->hw.hfcpci.sctrl |= SCTRL_MODE_NT;
781		Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);	/* set NT-mode */
782		udelay(10);
783		Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 1);	/* HFC ST G1 */
784		udelay(10);
785		Write_hfc(cs, HFCPCI_STATES, 1 | HFCPCI_ACTIVATE | HFCPCI_DO_ACTION);
786		cs->dc.hfcpci.ph_state = 1;
787		cs->hw.hfcpci.nt_mode = 1;
788		cs->hw.hfcpci.nt_timer = 0;
789		cs->stlist->l2.l2l1 = dch_nt_l2l1;
790		spin_unlock_irqrestore(&cs->lock, flags);
791		debugl1(cs, "NT mode activated");
792		return (0);
793	}
794	if ((cs->chanlimit > 1) || (cs->hw.hfcpci.bswapped) ||
795	    (cs->hw.hfcpci.nt_mode) || (ic->arg != 12))
796		return (-EINVAL);
797
798	spin_lock_irqsave(&cs->lock, flags);
799	if (i) {
800		cs->logecho = 1;
801		cs->hw.hfcpci.trm |= 0x20;	/* enable echo chan */
802		cs->hw.hfcpci.int_m1 |= HFCPCI_INTS_B2REC;
803		cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B2RX;
804	} else {
805		cs->logecho = 0;
806		cs->hw.hfcpci.trm &= ~0x20;	/* disable echo chan */
807		cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_B2REC;
808		cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B2RX;
809	}
810	cs->hw.hfcpci.sctrl_r &= ~SCTRL_B2_ENA;
811	cs->hw.hfcpci.sctrl &= ~SCTRL_B2_ENA;
812	cs->hw.hfcpci.conn |= 0x10;	/* B2-IOM -> B2-ST */
813	cs->hw.hfcpci.ctmt &= ~2;
814	Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt);
815	Write_hfc(cs, HFCPCI_SCTRL_R, cs->hw.hfcpci.sctrl_r);
816	Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);
817	Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
818	Write_hfc(cs, HFCPCI_TRM, cs->hw.hfcpci.trm);
819	Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
820	Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
821	spin_unlock_irqrestore(&cs->lock, flags);
822	return (0);
823}				/* hfcpci_auxcmd */
824
825/*****************************/
826/* E-channel receive routine */
827/*****************************/
828static void
829receive_emsg(struct IsdnCardState *cs)
830{
831	int rcnt;
832	int receive, count = 5;
833	bzfifo_type *bz;
834	u_char *bdata;
835	z_type *zp;
836	u_char *ptr, *ptr1, new_f2;
837	int total, maxlen, new_z2;
838	u_char e_buffer[256];
839
840	bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b2;
841	bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxdat_b2;
842Begin:
843	count--;
844	if (test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
845		debugl1(cs, "echo_rec_data blocked");
846		return;
847	}
848	if (bz->f1 != bz->f2) {
849		if (cs->debug & L1_DEB_ISAC)
850			debugl1(cs, "hfcpci e_rec f1(%d) f2(%d)",
851				bz->f1, bz->f2);
852		zp = &bz->za[bz->f2];
853
854		rcnt = zp->z1 - zp->z2;
855		if (rcnt < 0)
856			rcnt += B_FIFO_SIZE;
857		rcnt++;
858		if (cs->debug & L1_DEB_ISAC)
859			debugl1(cs, "hfcpci e_rec z1(%x) z2(%x) cnt(%d)",
860				zp->z1, zp->z2, rcnt);
861		new_z2 = zp->z2 + rcnt;		/* new position in fifo */
862		if (new_z2 >= (B_FIFO_SIZE + B_SUB_VAL))
863			new_z2 -= B_FIFO_SIZE;	/* buffer wrap */
864		new_f2 = (bz->f2 + 1) & MAX_B_FRAMES;
865		if ((rcnt > 256 + 3) || (count < 4) ||
866		    (*(bdata + (zp->z1 - B_SUB_VAL)))) {
867			if (cs->debug & L1_DEB_WARN)
868				debugl1(cs, "hfcpci_empty_echan: incoming packet invalid length %d or crc", rcnt);
869			bz->za[new_f2].z2 = new_z2;
870			bz->f2 = new_f2;	/* next buffer */
871		} else {
872			total = rcnt;
873			rcnt -= 3;
874			ptr = e_buffer;
875
876			if (zp->z2 <= B_FIFO_SIZE + B_SUB_VAL)
877				maxlen = rcnt;	/* complete transfer */
878			else
879				maxlen = B_FIFO_SIZE + B_SUB_VAL - zp->z2;	/* maximum */
880
881			ptr1 = bdata + (zp->z2 - B_SUB_VAL);	/* start of data */
882			memcpy(ptr, ptr1, maxlen);	/* copy data */
883			rcnt -= maxlen;
884
885			if (rcnt) {	/* rest remaining */
886				ptr += maxlen;
887				ptr1 = bdata;	/* start of buffer */
888				memcpy(ptr, ptr1, rcnt);	/* rest */
889			}
890			bz->za[new_f2].z2 = new_z2;
891			bz->f2 = new_f2;	/* next buffer */
892			if (cs->debug & DEB_DLOG_HEX) {
893				ptr = cs->dlog;
894				if ((total - 3) < MAX_DLOG_SPACE / 3 - 10) {
895					*ptr++ = 'E';
896					*ptr++ = 'C';
897					*ptr++ = 'H';
898					*ptr++ = 'O';
899					*ptr++ = ':';
900					ptr += QuickHex(ptr, e_buffer, total - 3);
901					ptr--;
902					*ptr++ = '\n';
903					*ptr = 0;
904					HiSax_putstatus(cs, NULL, cs->dlog);
905				} else
906					HiSax_putstatus(cs, "LogEcho: ", "warning Frame too big (%d)", total - 3);
907			}
908		}
909
910		rcnt = bz->f1 - bz->f2;
911		if (rcnt < 0)
912			rcnt += MAX_B_FRAMES + 1;
913		if (rcnt > 1)
914			receive = 1;
915		else
916			receive = 0;
917	} else
918		receive = 0;
919	test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
920	if (count && receive)
921		goto Begin;
922}				/* receive_emsg */
923
924/*********************/
925/* Interrupt handler */
926/*********************/
927static irqreturn_t
928hfcpci_interrupt(int intno, void *dev_id)
929{
930	u_long flags;
931	struct IsdnCardState *cs = dev_id;
932	u_char exval;
933	struct BCState *bcs;
934	int count = 15;
935	u_char val, stat;
936
937	if (!(cs->hw.hfcpci.int_m2 & 0x08)) {
938		debugl1(cs, "HFC-PCI: int_m2 %x not initialised", cs->hw.hfcpci.int_m2);
939		return IRQ_NONE;	/* not initialised */
940	}
941	spin_lock_irqsave(&cs->lock, flags);
942	if (HFCPCI_ANYINT & (stat = Read_hfc(cs, HFCPCI_STATUS))) {
943		val = Read_hfc(cs, HFCPCI_INT_S1);
944		if (cs->debug & L1_DEB_ISAC)
945			debugl1(cs, "HFC-PCI: stat(%02x) s1(%02x)", stat, val);
946	} else {
947		spin_unlock_irqrestore(&cs->lock, flags);
948		return IRQ_NONE;
949	}
950	if (cs->debug & L1_DEB_ISAC)
951		debugl1(cs, "HFC-PCI irq %x %s", val,
952			test_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags) ?
953			"locked" : "unlocked");
954	val &= cs->hw.hfcpci.int_m1;
955	if (val & 0x40) {	/* state machine irq */
956		exval = Read_hfc(cs, HFCPCI_STATES) & 0xf;
957		if (cs->debug & L1_DEB_ISAC)
958			debugl1(cs, "ph_state chg %d->%d", cs->dc.hfcpci.ph_state,
959				exval);
960		cs->dc.hfcpci.ph_state = exval;
961		sched_event_D_pci(cs, D_L1STATECHANGE);
962		val &= ~0x40;
963	}
964	if (val & 0x80) {	/* timer irq */
965		if (cs->hw.hfcpci.nt_mode) {
966			if ((--cs->hw.hfcpci.nt_timer) < 0)
967				sched_event_D_pci(cs, D_L1STATECHANGE);
968		}
969		val &= ~0x80;
970		Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt | HFCPCI_CLTIMER);
971	}
972	while (val) {
973		if (test_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
974			cs->hw.hfcpci.int_s1 |= val;
975			spin_unlock_irqrestore(&cs->lock, flags);
976			return IRQ_HANDLED;
977		}
978		if (cs->hw.hfcpci.int_s1 & 0x18) {
979			exval = val;
980			val = cs->hw.hfcpci.int_s1;
981			cs->hw.hfcpci.int_s1 = exval;
982		}
983		if (val & 0x08) {
984			if (!(bcs = Sel_BCS(cs, cs->hw.hfcpci.bswapped ? 1 : 0))) {
985				if (cs->debug)
986					debugl1(cs, "hfcpci spurious 0x08 IRQ");
987			} else
988				main_rec_hfcpci(bcs);
989		}
990		if (val & 0x10) {
991			if (cs->logecho)
992				receive_emsg(cs);
993			else if (!(bcs = Sel_BCS(cs, 1))) {
994				if (cs->debug)
995					debugl1(cs, "hfcpci spurious 0x10 IRQ");
996			} else
997				main_rec_hfcpci(bcs);
998		}
999		if (val & 0x01) {
1000			if (!(bcs = Sel_BCS(cs, cs->hw.hfcpci.bswapped ? 1 : 0))) {
1001				if (cs->debug)
1002					debugl1(cs, "hfcpci spurious 0x01 IRQ");
1003			} else {
1004				if (bcs->tx_skb) {
1005					if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1006						hfcpci_fill_fifo(bcs);
1007						test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1008					} else
1009						debugl1(cs, "fill_data %d blocked", bcs->channel);
1010				} else {
1011					if ((bcs->tx_skb = skb_dequeue(&bcs->squeue))) {
1012						if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1013							hfcpci_fill_fifo(bcs);
1014							test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1015						} else
1016							debugl1(cs, "fill_data %d blocked", bcs->channel);
1017					} else {
1018						hfcpci_sched_event(bcs, B_XMTBUFREADY);
1019					}
1020				}
1021			}
1022		}
1023		if (val & 0x02) {
1024			if (!(bcs = Sel_BCS(cs, 1))) {
1025				if (cs->debug)
1026					debugl1(cs, "hfcpci spurious 0x02 IRQ");
1027			} else {
1028				if (bcs->tx_skb) {
1029					if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1030						hfcpci_fill_fifo(bcs);
1031						test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1032					} else
1033						debugl1(cs, "fill_data %d blocked", bcs->channel);
1034				} else {
1035					if ((bcs->tx_skb = skb_dequeue(&bcs->squeue))) {
1036						if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1037							hfcpci_fill_fifo(bcs);
1038							test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1039						} else
1040							debugl1(cs, "fill_data %d blocked", bcs->channel);
1041					} else {
1042						hfcpci_sched_event(bcs, B_XMTBUFREADY);
1043					}
1044				}
1045			}
1046		}
1047		if (val & 0x20) {	/* receive dframe */
1048			receive_dmsg(cs);
1049		}
1050		if (val & 0x04) {	/* dframe transmitted */
1051			if (test_and_clear_bit(FLG_DBUSY_TIMER, &cs->HW_Flags))
1052				del_timer(&cs->dbusytimer);
1053			if (test_and_clear_bit(FLG_L1_DBUSY, &cs->HW_Flags))
1054				sched_event_D_pci(cs, D_CLEARBUSY);
1055			if (cs->tx_skb) {
1056				if (cs->tx_skb->len) {
1057					if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1058						hfcpci_fill_dfifo(cs);
1059						test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1060					} else {
1061						debugl1(cs, "hfcpci_fill_dfifo irq blocked");
1062					}
1063					goto afterXPR;
1064				} else {
1065					dev_kfree_skb_irq(cs->tx_skb);
1066					cs->tx_cnt = 0;
1067					cs->tx_skb = NULL;
1068				}
1069			}
1070			if ((cs->tx_skb = skb_dequeue(&cs->sq))) {
1071				cs->tx_cnt = 0;
1072				if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1073					hfcpci_fill_dfifo(cs);
1074					test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1075				} else {
1076					debugl1(cs, "hfcpci_fill_dfifo irq blocked");
1077				}
1078			} else
1079				sched_event_D_pci(cs, D_XMTBUFREADY);
1080		}
1081	afterXPR:
1082		if (cs->hw.hfcpci.int_s1 && count--) {
1083			val = cs->hw.hfcpci.int_s1;
1084			cs->hw.hfcpci.int_s1 = 0;
1085			if (cs->debug & L1_DEB_ISAC)
1086				debugl1(cs, "HFC-PCI irq %x loop %d", val, 15 - count);
1087		} else
1088			val = 0;
1089	}
1090	spin_unlock_irqrestore(&cs->lock, flags);
1091	return IRQ_HANDLED;
1092}
1093
1094/********************************************************************/
1095/* timer callback for D-chan busy resolution. Currently no function */
1096/********************************************************************/
1097static void
1098hfcpci_dbusy_timer(struct IsdnCardState *cs)
1099{
1100}
1101
1102/*************************************/
1103/* Layer 1 D-channel hardware access */
1104/*************************************/
1105static void
1106HFCPCI_l1hw(struct PStack *st, int pr, void *arg)
1107{
1108	u_long flags;
1109	struct IsdnCardState *cs = (struct IsdnCardState *) st->l1.hardware;
1110	struct sk_buff *skb = arg;
1111
1112	switch (pr) {
1113	case (PH_DATA | REQUEST):
1114		if (cs->debug & DEB_DLOG_HEX)
1115			LogFrame(cs, skb->data, skb->len);
1116		if (cs->debug & DEB_DLOG_VERBOSE)
1117			dlogframe(cs, skb, 0);
1118		spin_lock_irqsave(&cs->lock, flags);
1119		if (cs->tx_skb) {
1120			skb_queue_tail(&cs->sq, skb);
1121#ifdef L2FRAME_DEBUG		/* psa */
1122			if (cs->debug & L1_DEB_LAPD)
1123				Logl2Frame(cs, skb, "PH_DATA Queued", 0);
1124#endif
1125		} else {
1126			cs->tx_skb = skb;
1127			cs->tx_cnt = 0;
1128#ifdef L2FRAME_DEBUG		/* psa */
1129			if (cs->debug & L1_DEB_LAPD)
1130				Logl2Frame(cs, skb, "PH_DATA", 0);
1131#endif
1132			if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1133				hfcpci_fill_dfifo(cs);
1134				test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1135			} else
1136				debugl1(cs, "hfcpci_fill_dfifo blocked");
1137
1138		}
1139		spin_unlock_irqrestore(&cs->lock, flags);
1140		break;
1141	case (PH_PULL | INDICATION):
1142		spin_lock_irqsave(&cs->lock, flags);
1143		if (cs->tx_skb) {
1144			if (cs->debug & L1_DEB_WARN)
1145				debugl1(cs, " l2l1 tx_skb exist this shouldn't happen");
1146			skb_queue_tail(&cs->sq, skb);
1147			spin_unlock_irqrestore(&cs->lock, flags);
1148			break;
1149		}
1150		if (cs->debug & DEB_DLOG_HEX)
1151			LogFrame(cs, skb->data, skb->len);
1152		if (cs->debug & DEB_DLOG_VERBOSE)
1153			dlogframe(cs, skb, 0);
1154		cs->tx_skb = skb;
1155		cs->tx_cnt = 0;
1156#ifdef L2FRAME_DEBUG		/* psa */
1157		if (cs->debug & L1_DEB_LAPD)
1158			Logl2Frame(cs, skb, "PH_DATA_PULLED", 0);
1159#endif
1160		if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1161			hfcpci_fill_dfifo(cs);
1162			test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1163		} else
1164			debugl1(cs, "hfcpci_fill_dfifo blocked");
1165		spin_unlock_irqrestore(&cs->lock, flags);
1166		break;
1167	case (PH_PULL | REQUEST):
1168#ifdef L2FRAME_DEBUG		/* psa */
1169		if (cs->debug & L1_DEB_LAPD)
1170			debugl1(cs, "-> PH_REQUEST_PULL");
1171#endif
1172		if (!cs->tx_skb) {
1173			test_and_clear_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1174			st->l1.l1l2(st, PH_PULL | CONFIRM, NULL);
1175		} else
1176			test_and_set_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1177		break;
1178	case (HW_RESET | REQUEST):
1179		spin_lock_irqsave(&cs->lock, flags);
1180		Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 3);	/* HFC ST 3 */
1181		udelay(6);
1182		Write_hfc(cs, HFCPCI_STATES, 3);	/* HFC ST 2 */
1183		cs->hw.hfcpci.mst_m |= HFCPCI_MASTER;
1184		Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1185		Write_hfc(cs, HFCPCI_STATES, HFCPCI_ACTIVATE | HFCPCI_DO_ACTION);
1186		spin_unlock_irqrestore(&cs->lock, flags);
1187		l1_msg(cs, HW_POWERUP | CONFIRM, NULL);
1188		break;
1189	case (HW_ENABLE | REQUEST):
1190		spin_lock_irqsave(&cs->lock, flags);
1191		Write_hfc(cs, HFCPCI_STATES, HFCPCI_DO_ACTION);
1192		spin_unlock_irqrestore(&cs->lock, flags);
1193		break;
1194	case (HW_DEACTIVATE | REQUEST):
1195		spin_lock_irqsave(&cs->lock, flags);
1196		cs->hw.hfcpci.mst_m &= ~HFCPCI_MASTER;
1197		Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1198		spin_unlock_irqrestore(&cs->lock, flags);
1199		break;
1200	case (HW_INFO3 | REQUEST):
1201		spin_lock_irqsave(&cs->lock, flags);
1202		cs->hw.hfcpci.mst_m |= HFCPCI_MASTER;
1203		Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1204		spin_unlock_irqrestore(&cs->lock, flags);
1205		break;
1206	case (HW_TESTLOOP | REQUEST):
1207		spin_lock_irqsave(&cs->lock, flags);
1208		switch ((long) arg) {
1209		case (1):
1210			Write_hfc(cs, HFCPCI_B1_SSL, 0x80);	/* tx slot */
1211			Write_hfc(cs, HFCPCI_B1_RSL, 0x80);	/* rx slot */
1212			cs->hw.hfcpci.conn = (cs->hw.hfcpci.conn & ~7) | 1;
1213			Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
1214			break;
1215
1216		case (2):
1217			Write_hfc(cs, HFCPCI_B2_SSL, 0x81);	/* tx slot */
1218			Write_hfc(cs, HFCPCI_B2_RSL, 0x81);	/* rx slot */
1219			cs->hw.hfcpci.conn = (cs->hw.hfcpci.conn & ~0x38) | 0x08;
1220			Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
1221			break;
1222
1223		default:
1224			spin_unlock_irqrestore(&cs->lock, flags);
1225			if (cs->debug & L1_DEB_WARN)
1226				debugl1(cs, "hfcpci_l1hw loop invalid %4lx", (long) arg);
1227			return;
1228		}
1229		cs->hw.hfcpci.trm |= 0x80;	/* enable IOM-loop */
1230		Write_hfc(cs, HFCPCI_TRM, cs->hw.hfcpci.trm);
1231		spin_unlock_irqrestore(&cs->lock, flags);
1232		break;
1233	default:
1234		if (cs->debug & L1_DEB_WARN)
1235			debugl1(cs, "hfcpci_l1hw unknown pr %4x", pr);
1236		break;
1237	}
1238}
1239
1240/***********************************************/
1241/* called during init setting l1 stack pointer */
1242/***********************************************/
1243static void
1244setstack_hfcpci(struct PStack *st, struct IsdnCardState *cs)
1245{
1246	st->l1.l1hw = HFCPCI_l1hw;
1247}
1248
1249/**************************************/
1250/* send B-channel data if not blocked */
1251/**************************************/
1252static void
1253hfcpci_send_data(struct BCState *bcs)
1254{
1255	struct IsdnCardState *cs = bcs->cs;
1256
1257	if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1258		hfcpci_fill_fifo(bcs);
1259		test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1260	} else
1261		debugl1(cs, "send_data %d blocked", bcs->channel);
1262}
1263
1264/***************************************************************/
1265/* activate/deactivate hardware for selected channels and mode */
1266/***************************************************************/
1267static void
1268mode_hfcpci(struct BCState *bcs, int mode, int bc)
1269{
1270	struct IsdnCardState *cs = bcs->cs;
1271	int fifo2;
1272
1273	if (cs->debug & L1_DEB_HSCX)
1274		debugl1(cs, "HFCPCI bchannel mode %d bchan %d/%d",
1275			mode, bc, bcs->channel);
1276	bcs->mode = mode;
1277	bcs->channel = bc;
1278	fifo2 = bc;
1279	if (cs->chanlimit > 1) {
1280		cs->hw.hfcpci.bswapped = 0;	/* B1 and B2 normal mode */
1281		cs->hw.hfcpci.sctrl_e &= ~0x80;
1282	} else {
1283		if (bc) {
1284			if (mode != L1_MODE_NULL) {
1285				cs->hw.hfcpci.bswapped = 1;	/* B1 and B2 exchanged */
1286				cs->hw.hfcpci.sctrl_e |= 0x80;
1287			} else {
1288				cs->hw.hfcpci.bswapped = 0;	/* B1 and B2 normal mode */
1289				cs->hw.hfcpci.sctrl_e &= ~0x80;
1290			}
1291			fifo2 = 0;
1292		} else {
1293			cs->hw.hfcpci.bswapped = 0;	/* B1 and B2 normal mode */
1294			cs->hw.hfcpci.sctrl_e &= ~0x80;
1295		}
1296	}
1297	switch (mode) {
1298	case (L1_MODE_NULL):
1299		if (bc) {
1300			cs->hw.hfcpci.sctrl &= ~SCTRL_B2_ENA;
1301			cs->hw.hfcpci.sctrl_r &= ~SCTRL_B2_ENA;
1302		} else {
1303			cs->hw.hfcpci.sctrl &= ~SCTRL_B1_ENA;
1304			cs->hw.hfcpci.sctrl_r &= ~SCTRL_B1_ENA;
1305		}
1306		if (fifo2) {
1307			cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B2;
1308			cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1309		} else {
1310			cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B1;
1311			cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1312		}
1313		break;
1314	case (L1_MODE_TRANS):
1315		hfcpci_clear_fifo_rx(cs, fifo2);
1316		hfcpci_clear_fifo_tx(cs, fifo2);
1317		if (bc) {
1318			cs->hw.hfcpci.sctrl |= SCTRL_B2_ENA;
1319			cs->hw.hfcpci.sctrl_r |= SCTRL_B2_ENA;
1320		} else {
1321			cs->hw.hfcpci.sctrl |= SCTRL_B1_ENA;
1322			cs->hw.hfcpci.sctrl_r |= SCTRL_B1_ENA;
1323		}
1324		if (fifo2) {
1325			cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B2;
1326			cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1327			cs->hw.hfcpci.ctmt |= 2;
1328			cs->hw.hfcpci.conn &= ~0x18;
1329		} else {
1330			cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B1;
1331			cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1332			cs->hw.hfcpci.ctmt |= 1;
1333			cs->hw.hfcpci.conn &= ~0x03;
1334		}
1335		break;
1336	case (L1_MODE_HDLC):
1337		hfcpci_clear_fifo_rx(cs, fifo2);
1338		hfcpci_clear_fifo_tx(cs, fifo2);
1339		if (bc) {
1340			cs->hw.hfcpci.sctrl |= SCTRL_B2_ENA;
1341			cs->hw.hfcpci.sctrl_r |= SCTRL_B2_ENA;
1342		} else {
1343			cs->hw.hfcpci.sctrl |= SCTRL_B1_ENA;
1344			cs->hw.hfcpci.sctrl_r |= SCTRL_B1_ENA;
1345		}
1346		if (fifo2) {
1347			cs->hw.hfcpci.last_bfifo_cnt[1] = 0;
1348			cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B2;
1349			cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1350			cs->hw.hfcpci.ctmt &= ~2;
1351			cs->hw.hfcpci.conn &= ~0x18;
1352		} else {
1353			cs->hw.hfcpci.last_bfifo_cnt[0] = 0;
1354			cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B1;
1355			cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1356			cs->hw.hfcpci.ctmt &= ~1;
1357			cs->hw.hfcpci.conn &= ~0x03;
1358		}
1359		break;
1360	case (L1_MODE_EXTRN):
1361		if (bc) {
1362			cs->hw.hfcpci.conn |= 0x10;
1363			cs->hw.hfcpci.sctrl |= SCTRL_B2_ENA;
1364			cs->hw.hfcpci.sctrl_r |= SCTRL_B2_ENA;
1365			cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B2;
1366			cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1367		} else {
1368			cs->hw.hfcpci.conn |= 0x02;
1369			cs->hw.hfcpci.sctrl |= SCTRL_B1_ENA;
1370			cs->hw.hfcpci.sctrl_r |= SCTRL_B1_ENA;
1371			cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B1;
1372			cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1373		}
1374		break;
1375	}
1376	Write_hfc(cs, HFCPCI_SCTRL_E, cs->hw.hfcpci.sctrl_e);
1377	Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1378	Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
1379	Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);
1380	Write_hfc(cs, HFCPCI_SCTRL_R, cs->hw.hfcpci.sctrl_r);
1381	Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt);
1382	Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
1383}
1384
1385/******************************/
1386/* Layer2 -> Layer 1 Transfer */
1387/******************************/
1388static void
1389hfcpci_l2l1(struct PStack *st, int pr, void *arg)
1390{
1391	struct BCState	*bcs = st->l1.bcs;
1392	u_long		flags;
1393	struct sk_buff	*skb = arg;
1394
1395	switch (pr) {
1396	case (PH_DATA | REQUEST):
1397		spin_lock_irqsave(&bcs->cs->lock, flags);
1398		if (bcs->tx_skb) {
1399			skb_queue_tail(&bcs->squeue, skb);
1400		} else {
1401			bcs->tx_skb = skb;
1402//				test_and_set_bit(BC_FLG_BUSY, &bcs->Flag);
1403			bcs->cs->BC_Send_Data(bcs);
1404		}
1405		spin_unlock_irqrestore(&bcs->cs->lock, flags);
1406		break;
1407	case (PH_PULL | INDICATION):
1408		spin_lock_irqsave(&bcs->cs->lock, flags);
1409		if (bcs->tx_skb) {
1410			spin_unlock_irqrestore(&bcs->cs->lock, flags);
1411			printk(KERN_WARNING "hfc_l2l1: this shouldn't happen\n");
1412			break;
1413		}
1414//			test_and_set_bit(BC_FLG_BUSY, &bcs->Flag);
1415		bcs->tx_skb = skb;
1416		bcs->cs->BC_Send_Data(bcs);
1417		spin_unlock_irqrestore(&bcs->cs->lock, flags);
1418		break;
1419	case (PH_PULL | REQUEST):
1420		if (!bcs->tx_skb) {
1421			test_and_clear_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1422			st->l1.l1l2(st, PH_PULL | CONFIRM, NULL);
1423		} else
1424			test_and_set_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1425		break;
1426	case (PH_ACTIVATE | REQUEST):
1427		spin_lock_irqsave(&bcs->cs->lock, flags);
1428		test_and_set_bit(BC_FLG_ACTIV, &bcs->Flag);
1429		mode_hfcpci(bcs, st->l1.mode, st->l1.bc);
1430		spin_unlock_irqrestore(&bcs->cs->lock, flags);
1431		l1_msg_b(st, pr, arg);
1432		break;
1433	case (PH_DEACTIVATE | REQUEST):
1434		l1_msg_b(st, pr, arg);
1435		break;
1436	case (PH_DEACTIVATE | CONFIRM):
1437		spin_lock_irqsave(&bcs->cs->lock, flags);
1438		test_and_clear_bit(BC_FLG_ACTIV, &bcs->Flag);
1439		test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
1440		mode_hfcpci(bcs, 0, st->l1.bc);
1441		spin_unlock_irqrestore(&bcs->cs->lock, flags);
1442		st->l1.l1l2(st, PH_DEACTIVATE | CONFIRM, NULL);
1443		break;
1444	}
1445}
1446
1447/******************************************/
1448/* deactivate B-channel access and queues */
1449/******************************************/
1450static void
1451close_hfcpci(struct BCState *bcs)
1452{
1453	mode_hfcpci(bcs, 0, bcs->channel);
1454	if (test_and_clear_bit(BC_FLG_INIT, &bcs->Flag)) {
1455		skb_queue_purge(&bcs->rqueue);
1456		skb_queue_purge(&bcs->squeue);
1457		if (bcs->tx_skb) {
1458			dev_kfree_skb_any(bcs->tx_skb);
1459			bcs->tx_skb = NULL;
1460			test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
1461		}
1462	}
1463}
1464
1465/*************************************/
1466/* init B-channel queues and control */
1467/*************************************/
1468static int
1469open_hfcpcistate(struct IsdnCardState *cs, struct BCState *bcs)
1470{
1471	if (!test_and_set_bit(BC_FLG_INIT, &bcs->Flag)) {
1472		skb_queue_head_init(&bcs->rqueue);
1473		skb_queue_head_init(&bcs->squeue);
1474	}
1475	bcs->tx_skb = NULL;
1476	test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
1477	bcs->event = 0;
1478	bcs->tx_cnt = 0;
1479	return (0);
1480}
1481
1482/*********************************/
1483/* inits the stack for B-channel */
1484/*********************************/
1485static int
1486setstack_2b(struct PStack *st, struct BCState *bcs)
1487{
1488	bcs->channel = st->l1.bc;
1489	if (open_hfcpcistate(st->l1.hardware, bcs))
1490		return (-1);
1491	st->l1.bcs = bcs;
1492	st->l2.l2l1 = hfcpci_l2l1;
1493	setstack_manager(st);
1494	bcs->st = st;
1495	setstack_l1_B(st);
1496	return (0);
1497}
1498
1499/***************************/
1500/* handle L1 state changes */
1501/***************************/
1502static void
1503hfcpci_bh(struct work_struct *work)
1504{
1505	struct IsdnCardState *cs =
1506		container_of(work, struct IsdnCardState, tqueue);
1507	u_long	flags;
1508//      struct PStack *stptr;
1509
1510	if (test_and_clear_bit(D_L1STATECHANGE, &cs->event)) {
1511		if (!cs->hw.hfcpci.nt_mode)
1512			switch (cs->dc.hfcpci.ph_state) {
1513			case (0):
1514				l1_msg(cs, HW_RESET | INDICATION, NULL);
1515				break;
1516			case (3):
1517				l1_msg(cs, HW_DEACTIVATE | INDICATION, NULL);
1518				break;
1519			case (8):
1520				l1_msg(cs, HW_RSYNC | INDICATION, NULL);
1521				break;
1522			case (6):
1523				l1_msg(cs, HW_INFO2 | INDICATION, NULL);
1524				break;
1525			case (7):
1526				l1_msg(cs, HW_INFO4_P8 | INDICATION, NULL);
1527				break;
1528			default:
1529				break;
1530			} else {
1531			spin_lock_irqsave(&cs->lock, flags);
1532			switch (cs->dc.hfcpci.ph_state) {
1533			case (2):
1534				if (cs->hw.hfcpci.nt_timer < 0) {
1535					cs->hw.hfcpci.nt_timer = 0;
1536					cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_TIMER;
1537					Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1538					/* Clear already pending ints */
1539					if (Read_hfc(cs, HFCPCI_INT_S1));
1540					Write_hfc(cs, HFCPCI_STATES, 4 | HFCPCI_LOAD_STATE);
1541					udelay(10);
1542					Write_hfc(cs, HFCPCI_STATES, 4);
1543					cs->dc.hfcpci.ph_state = 4;
1544				} else {
1545					cs->hw.hfcpci.int_m1 |= HFCPCI_INTS_TIMER;
1546					Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1547					cs->hw.hfcpci.ctmt &= ~HFCPCI_AUTO_TIMER;
1548					cs->hw.hfcpci.ctmt |= HFCPCI_TIM3_125;
1549					Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt | HFCPCI_CLTIMER);
1550					Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt | HFCPCI_CLTIMER);
1551					cs->hw.hfcpci.nt_timer = NT_T1_COUNT;
1552					Write_hfc(cs, HFCPCI_STATES, 2 | HFCPCI_NT_G2_G3);	/* allow G2 -> G3 transition */
1553				}
1554				break;
1555			case (1):
1556			case (3):
1557			case (4):
1558				cs->hw.hfcpci.nt_timer = 0;
1559				cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_TIMER;
1560				Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1561				break;
1562			default:
1563				break;
1564			}
1565			spin_unlock_irqrestore(&cs->lock, flags);
1566		}
1567	}
1568	if (test_and_clear_bit(D_RCVBUFREADY, &cs->event))
1569		DChannel_proc_rcv(cs);
1570	if (test_and_clear_bit(D_XMTBUFREADY, &cs->event))
1571		DChannel_proc_xmt(cs);
1572}
1573
1574
1575/********************************/
1576/* called for card init message */
1577/********************************/
1578static void
1579inithfcpci(struct IsdnCardState *cs)
1580{
1581	cs->bcs[0].BC_SetStack = setstack_2b;
1582	cs->bcs[1].BC_SetStack = setstack_2b;
1583	cs->bcs[0].BC_Close = close_hfcpci;
1584	cs->bcs[1].BC_Close = close_hfcpci;
1585	cs->dbusytimer.function = (void *) hfcpci_dbusy_timer;
1586	cs->dbusytimer.data = (long) cs;
1587	init_timer(&cs->dbusytimer);
1588	mode_hfcpci(cs->bcs, 0, 0);
1589	mode_hfcpci(cs->bcs + 1, 0, 1);
1590}
1591
1592
1593
1594/*******************************************/
1595/* handle card messages from control layer */
1596/*******************************************/
1597static int
1598hfcpci_card_msg(struct IsdnCardState *cs, int mt, void *arg)
1599{
1600	u_long flags;
1601
1602	if (cs->debug & L1_DEB_ISAC)
1603		debugl1(cs, "HFCPCI: card_msg %x", mt);
1604	switch (mt) {
1605	case CARD_RESET:
1606		spin_lock_irqsave(&cs->lock, flags);
1607		reset_hfcpci(cs);
1608		spin_unlock_irqrestore(&cs->lock, flags);
1609		return (0);
1610	case CARD_RELEASE:
1611		release_io_hfcpci(cs);
1612		return (0);
1613	case CARD_INIT:
1614		spin_lock_irqsave(&cs->lock, flags);
1615		inithfcpci(cs);
1616		reset_hfcpci(cs);
1617		spin_unlock_irqrestore(&cs->lock, flags);
1618		msleep(80);				/* Timeout 80ms */
1619		/* now switch timer interrupt off */
1620		spin_lock_irqsave(&cs->lock, flags);
1621		cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_TIMER;
1622		Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1623		/* reinit mode reg */
1624		Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1625		spin_unlock_irqrestore(&cs->lock, flags);
1626		return (0);
1627	case CARD_TEST:
1628		return (0);
1629	}
1630	return (0);
1631}
1632
1633
1634/* this variable is used as card index when more than one cards are present */
1635static struct pci_dev *dev_hfcpci __devinitdata = NULL;
1636
1637int __devinit
1638setup_hfcpci(struct IsdnCard *card)
1639{
1640	u_long flags;
1641	struct IsdnCardState *cs = card->cs;
1642	char tmp[64];
1643	int i;
1644	struct pci_dev *tmp_hfcpci = NULL;
1645
1646#ifdef __BIG_ENDIAN
1647#error "not running on big endian machines now"
1648#endif
1649
1650	strcpy(tmp, hfcpci_revision);
1651	printk(KERN_INFO "HiSax: HFC-PCI driver Rev. %s\n", HiSax_getrev(tmp));
1652
1653	cs->hw.hfcpci.int_s1 = 0;
1654	cs->dc.hfcpci.ph_state = 0;
1655	cs->hw.hfcpci.fifo = 255;
1656	if (cs->typ != ISDN_CTYPE_HFC_PCI)
1657		return (0);
1658
1659	i = 0;
1660	while (id_list[i].vendor_id) {
1661		tmp_hfcpci = hisax_find_pci_device(id_list[i].vendor_id,
1662						   id_list[i].device_id,
1663						   dev_hfcpci);
1664		i++;
1665		if (tmp_hfcpci) {
1666			dma_addr_t	dma_mask = DMA_BIT_MASK(32) & ~0x7fffUL;
1667			if (pci_enable_device(tmp_hfcpci))
1668				continue;
1669			if (pci_set_dma_mask(tmp_hfcpci, dma_mask)) {
1670				printk(KERN_WARNING
1671				       "HiSax hfc_pci: No suitable DMA available.\n");
1672				continue;
1673			}
1674			if (pci_set_consistent_dma_mask(tmp_hfcpci, dma_mask)) {
1675				printk(KERN_WARNING
1676				       "HiSax hfc_pci: No suitable consistent DMA available.\n");
1677				continue;
1678			}
1679			pci_set_master(tmp_hfcpci);
1680			if ((card->para[0]) && (card->para[0] != (tmp_hfcpci->resource[0].start & PCI_BASE_ADDRESS_IO_MASK)))
1681				continue;
1682			else
1683				break;
1684		}
1685	}
1686
1687	if (!tmp_hfcpci) {
1688		printk(KERN_WARNING "HFC-PCI: No PCI card found\n");
1689		return (0);
1690	}
1691
1692	i--;
1693	dev_hfcpci = tmp_hfcpci;	/* old device */
1694	cs->hw.hfcpci.dev = dev_hfcpci;
1695	cs->irq = dev_hfcpci->irq;
1696	if (!cs->irq) {
1697		printk(KERN_WARNING "HFC-PCI: No IRQ for PCI card found\n");
1698		return (0);
1699	}
1700	cs->hw.hfcpci.pci_io = (char *)(unsigned long)dev_hfcpci->resource[1].start;
1701	printk(KERN_INFO "HiSax: HFC-PCI card manufacturer: %s card name: %s\n", id_list[i].vendor_name, id_list[i].card_name);
1702
1703	if (!cs->hw.hfcpci.pci_io) {
1704		printk(KERN_WARNING "HFC-PCI: No IO-Mem for PCI card found\n");
1705		return (0);
1706	}
1707
1708	/* Allocate memory for FIFOS */
1709	cs->hw.hfcpci.fifos = pci_alloc_consistent(cs->hw.hfcpci.dev,
1710						   0x8000, &cs->hw.hfcpci.dma);
1711	if (!cs->hw.hfcpci.fifos) {
1712		printk(KERN_WARNING "HFC-PCI: Error allocating FIFO memory!\n");
1713		return 0;
1714	}
1715	if (cs->hw.hfcpci.dma & 0x7fff) {
1716		printk(KERN_WARNING
1717		       "HFC-PCI: Error DMA memory not on 32K boundary (%lx)\n",
1718		       (u_long)cs->hw.hfcpci.dma);
1719		pci_free_consistent(cs->hw.hfcpci.dev, 0x8000,
1720				    cs->hw.hfcpci.fifos, cs->hw.hfcpci.dma);
1721		return 0;
1722	}
1723	pci_write_config_dword(cs->hw.hfcpci.dev, 0x80, (u32)cs->hw.hfcpci.dma);
1724	cs->hw.hfcpci.pci_io = ioremap((ulong) cs->hw.hfcpci.pci_io, 256);
1725	printk(KERN_INFO
1726	       "HFC-PCI: defined at mem %p fifo %p(%lx) IRQ %d HZ %d\n",
1727	       cs->hw.hfcpci.pci_io,
1728	       cs->hw.hfcpci.fifos,
1729	       (u_long)cs->hw.hfcpci.dma,
1730	       cs->irq, HZ);
1731
1732	spin_lock_irqsave(&cs->lock, flags);
1733
1734	pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, PCI_ENA_MEMIO);	/* enable memory mapped ports, disable busmaster */
1735	cs->hw.hfcpci.int_m2 = 0;	/* disable alle interrupts */
1736	cs->hw.hfcpci.int_m1 = 0;
1737	Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1738	Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
1739	/* At this point the needed PCI config is done */
1740	/* fifos are still not enabled */
1741
1742	INIT_WORK(&cs->tqueue,  hfcpci_bh);
1743	cs->setstack_d = setstack_hfcpci;
1744	cs->BC_Send_Data = &hfcpci_send_data;
1745	cs->readisac = NULL;
1746	cs->writeisac = NULL;
1747	cs->readisacfifo = NULL;
1748	cs->writeisacfifo = NULL;
1749	cs->BC_Read_Reg = NULL;
1750	cs->BC_Write_Reg = NULL;
1751	cs->irq_func = &hfcpci_interrupt;
1752	cs->irq_flags |= IRQF_SHARED;
1753	cs->hw.hfcpci.timer.function = (void *) hfcpci_Timer;
1754	cs->hw.hfcpci.timer.data = (long) cs;
1755	init_timer(&cs->hw.hfcpci.timer);
1756	cs->cardmsg = &hfcpci_card_msg;
1757	cs->auxcmd = &hfcpci_auxcmd;
1758
1759	spin_unlock_irqrestore(&cs->lock, flags);
1760
1761	return (1);
1762}
1763