[SCSI] ips: sg chaining support to the path to non I/O commands
[linux-2.6] / drivers / atm / iphase.c
1 /******************************************************************************
2          iphase.c: Device driver for Interphase ATM PCI adapter cards 
3                     Author: Peter Wang  <pwang@iphase.com>            
4                    Some fixes: Arnaldo Carvalho de Melo <acme@conectiva.com.br>
5                    Interphase Corporation  <www.iphase.com>           
6                                Version: 1.0                           
7 *******************************************************************************
8       
9       This software may be used and distributed according to the terms
10       of the GNU General Public License (GPL), incorporated herein by reference.
11       Drivers based on this skeleton fall under the GPL and must retain
12       the authorship (implicit copyright) notice.
13
14       This program is distributed in the hope that it will be useful, but
15       WITHOUT ANY WARRANTY; without even the implied warranty of
16       MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17       General Public License for more details.
18       
19       Modified from an incomplete driver for Interphase 5575 1KVC 1M card which 
20       was originally written by Monalisa Agrawal at UNH. Now this driver 
21       supports a variety of varients of Interphase ATM PCI (i)Chip adapter 
22       card family (See www.iphase.com/products/ClassSheet.cfm?ClassID=ATM) 
23       in terms of PHY type, the size of control memory and the size of 
24       packet memory. The followings are the change log and history:
25      
26           Bugfix the Mona's UBR driver.
27           Modify the basic memory allocation and dma logic.
28           Port the driver to the latest kernel from 2.0.46.
29           Complete the ABR logic of the driver, and added the ABR work-
30               around for the hardware anormalies.
31           Add the CBR support.
32           Add the flow control logic to the driver to allow rate-limit VC.
33           Add 4K VC support to the board with 512K control memory.
34           Add the support of all the variants of the Interphase ATM PCI 
35           (i)Chip adapter cards including x575 (155M OC3 and UTP155), x525
36           (25M UTP25) and x531 (DS3 and E3).
37           Add SMP support.
38
39       Support and updates available at: ftp://ftp.iphase.com/pub/atm
40
41 *******************************************************************************/
42
43 #include <linux/module.h>  
44 #include <linux/kernel.h>  
45 #include <linux/mm.h>  
46 #include <linux/pci.h>  
47 #include <linux/errno.h>  
48 #include <linux/atm.h>  
49 #include <linux/atmdev.h>  
50 #include <linux/sonet.h>  
51 #include <linux/skbuff.h>  
52 #include <linux/time.h>  
53 #include <linux/delay.h>  
54 #include <linux/uio.h>  
55 #include <linux/init.h>  
56 #include <linux/wait.h>
57 #include <asm/system.h>  
58 #include <asm/io.h>  
59 #include <asm/atomic.h>  
60 #include <asm/uaccess.h>  
61 #include <asm/string.h>  
62 #include <asm/byteorder.h>  
63 #include <linux/vmalloc.h>  
64 #include "iphase.h"               
65 #include "suni.h"                 
66 #define swap(x) (((x & 0xff) << 8) | ((x & 0xff00) >> 8))  
67 struct suni_priv {
68         struct k_sonet_stats sonet_stats; /* link diagnostics */
69         unsigned char loop_mode;        /* loopback mode */
70         struct atm_dev *dev;            /* device back-pointer */
71         struct suni_priv *next;         /* next SUNI */
72 }; 
73 #define PRIV(dev) ((struct suni_priv *) dev->phy_data)
74
75 static unsigned char ia_phy_get(struct atm_dev *dev, unsigned long addr);
76 static void desc_dbg(IADEV *iadev);
77
78 static IADEV *ia_dev[8];
79 static struct atm_dev *_ia_dev[8];
80 static int iadev_count;
81 static void ia_led_timer(unsigned long arg);
82 static DEFINE_TIMER(ia_timer, ia_led_timer, 0, 0);
83 static int IA_TX_BUF = DFL_TX_BUFFERS, IA_TX_BUF_SZ = DFL_TX_BUF_SZ;
84 static int IA_RX_BUF = DFL_RX_BUFFERS, IA_RX_BUF_SZ = DFL_RX_BUF_SZ;
85 static uint IADebugFlag = /* IF_IADBG_ERR | IF_IADBG_CBR| IF_IADBG_INIT_ADAPTER
86             |IF_IADBG_ABR | IF_IADBG_EVENT*/ 0; 
87
88 module_param(IA_TX_BUF, int, 0);
89 module_param(IA_TX_BUF_SZ, int, 0);
90 module_param(IA_RX_BUF, int, 0);
91 module_param(IA_RX_BUF_SZ, int, 0);
92 module_param(IADebugFlag, uint, 0644);
93
94 MODULE_LICENSE("GPL");
95
96 #if BITS_PER_LONG != 32
97 #  error FIXME: this driver only works on 32-bit platforms
98 #endif
99
100 /**************************** IA_LIB **********************************/
101
102 static void ia_init_rtn_q (IARTN_Q *que) 
103
104    que->next = NULL; 
105    que->tail = NULL; 
106 }
107
108 static void ia_enque_head_rtn_q (IARTN_Q *que, IARTN_Q * data) 
109 {
110    data->next = NULL;
111    if (que->next == NULL) 
112       que->next = que->tail = data;
113    else {
114       data->next = que->next;
115       que->next = data;
116    } 
117    return;
118 }
119
120 static int ia_enque_rtn_q (IARTN_Q *que, struct desc_tbl_t data) {
121    IARTN_Q *entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
122    if (!entry) return -1;
123    entry->data = data;
124    entry->next = NULL;
125    if (que->next == NULL) 
126       que->next = que->tail = entry;
127    else {
128       que->tail->next = entry;
129       que->tail = que->tail->next;
130    }      
131    return 1;
132 }
133
134 static IARTN_Q * ia_deque_rtn_q (IARTN_Q *que) {
135    IARTN_Q *tmpdata;
136    if (que->next == NULL)
137       return NULL;
138    tmpdata = que->next;
139    if ( que->next == que->tail)  
140       que->next = que->tail = NULL;
141    else 
142       que->next = que->next->next;
143    return tmpdata;
144 }
145
146 static void ia_hack_tcq(IADEV *dev) {
147
148   u_short               desc1;
149   u_short               tcq_wr;
150   struct ia_vcc         *iavcc_r = NULL; 
151
152   tcq_wr = readl(dev->seg_reg+TCQ_WR_PTR) & 0xffff;
153   while (dev->host_tcq_wr != tcq_wr) {
154      desc1 = *(u_short *)(dev->seg_ram + dev->host_tcq_wr);
155      if (!desc1) ;
156      else if (!dev->desc_tbl[desc1 -1].timestamp) {
157         IF_ABR(printk(" Desc %d is reset at %ld\n", desc1 -1, jiffies);)
158         *(u_short *) (dev->seg_ram + dev->host_tcq_wr) = 0;
159      }                                 
160      else if (dev->desc_tbl[desc1 -1].timestamp) {
161         if (!(iavcc_r = dev->desc_tbl[desc1 -1].iavcc)) { 
162            printk("IA: Fatal err in get_desc\n");
163            continue;
164         }
165         iavcc_r->vc_desc_cnt--;
166         dev->desc_tbl[desc1 -1].timestamp = 0;
167         IF_EVENT(printk("ia_hack: return_q skb = 0x%x desc = %d\n", 
168                                    (u32)dev->desc_tbl[desc1 -1].txskb, desc1);)
169         if (iavcc_r->pcr < dev->rate_limit) {
170            IA_SKB_STATE (dev->desc_tbl[desc1-1].txskb) |= IA_TX_DONE;
171            if (ia_enque_rtn_q(&dev->tx_return_q, dev->desc_tbl[desc1 -1]) < 0)
172               printk("ia_hack_tcq: No memory available\n");
173         } 
174         dev->desc_tbl[desc1 -1].iavcc = NULL;
175         dev->desc_tbl[desc1 -1].txskb = NULL;
176      }
177      dev->host_tcq_wr += 2;
178      if (dev->host_tcq_wr > dev->ffL.tcq_ed) 
179         dev->host_tcq_wr = dev->ffL.tcq_st;
180   }
181 } /* ia_hack_tcq */
182
183 static u16 get_desc (IADEV *dev, struct ia_vcc *iavcc) {
184   u_short               desc_num, i;
185   struct sk_buff        *skb;
186   struct ia_vcc         *iavcc_r = NULL; 
187   unsigned long delta;
188   static unsigned long timer = 0;
189   int ltimeout;
190
191   ia_hack_tcq (dev);
192   if(((jiffies - timer)>50)||((dev->ffL.tcq_rd==dev->host_tcq_wr))){      
193      timer = jiffies; 
194      i=0;
195      while (i < dev->num_tx_desc) {
196         if (!dev->desc_tbl[i].timestamp) {
197            i++;
198            continue;
199         }
200         ltimeout = dev->desc_tbl[i].iavcc->ltimeout; 
201         delta = jiffies - dev->desc_tbl[i].timestamp;
202         if (delta >= ltimeout) {
203            IF_ABR(printk("RECOVER run!! desc_tbl %d = %d  delta = %ld, time = %ld\n", i,dev->desc_tbl[i].timestamp, delta, jiffies);)
204            if (dev->ffL.tcq_rd == dev->ffL.tcq_st) 
205               dev->ffL.tcq_rd =  dev->ffL.tcq_ed;
206            else 
207               dev->ffL.tcq_rd -= 2;
208            *(u_short *)(dev->seg_ram + dev->ffL.tcq_rd) = i+1;
209            if (!(skb = dev->desc_tbl[i].txskb) || 
210                           !(iavcc_r = dev->desc_tbl[i].iavcc))
211               printk("Fatal err, desc table vcc or skb is NULL\n");
212            else 
213               iavcc_r->vc_desc_cnt--;
214            dev->desc_tbl[i].timestamp = 0;
215            dev->desc_tbl[i].iavcc = NULL;
216            dev->desc_tbl[i].txskb = NULL;
217         }
218         i++;
219      } /* while */
220   }
221   if (dev->ffL.tcq_rd == dev->host_tcq_wr) 
222      return 0xFFFF;
223     
224   /* Get the next available descriptor number from TCQ */
225   desc_num = *(u_short *)(dev->seg_ram + dev->ffL.tcq_rd);
226
227   while (!desc_num || (dev->desc_tbl[desc_num -1]).timestamp) {
228      dev->ffL.tcq_rd += 2;
229      if (dev->ffL.tcq_rd > dev->ffL.tcq_ed) 
230      dev->ffL.tcq_rd = dev->ffL.tcq_st;
231      if (dev->ffL.tcq_rd == dev->host_tcq_wr) 
232         return 0xFFFF; 
233      desc_num = *(u_short *)(dev->seg_ram + dev->ffL.tcq_rd);
234   }
235
236   /* get system time */
237   dev->desc_tbl[desc_num -1].timestamp = jiffies;
238   return desc_num;
239 }
240
241 static void clear_lockup (struct atm_vcc *vcc, IADEV *dev) {
242   u_char                foundLockUp;
243   vcstatus_t            *vcstatus;
244   u_short               *shd_tbl;
245   u_short               tempCellSlot, tempFract;
246   struct main_vc *abr_vc = (struct main_vc *)dev->MAIN_VC_TABLE_ADDR;
247   struct ext_vc *eabr_vc = (struct ext_vc *)dev->EXT_VC_TABLE_ADDR;
248   u_int  i;
249
250   if (vcc->qos.txtp.traffic_class == ATM_ABR) {
251      vcstatus = (vcstatus_t *) &(dev->testTable[vcc->vci]->vc_status);
252      vcstatus->cnt++;
253      foundLockUp = 0;
254      if( vcstatus->cnt == 0x05 ) {
255         abr_vc += vcc->vci;
256         eabr_vc += vcc->vci;
257         if( eabr_vc->last_desc ) {
258            if( (abr_vc->status & 0x07) == ABR_STATE /* 0x2 */ ) {
259               /* Wait for 10 Micro sec */
260               udelay(10);
261               if ((eabr_vc->last_desc)&&((abr_vc->status & 0x07)==ABR_STATE))
262                  foundLockUp = 1;
263            }
264            else {
265               tempCellSlot = abr_vc->last_cell_slot;
266               tempFract    = abr_vc->fraction;
267               if((tempCellSlot == dev->testTable[vcc->vci]->lastTime)
268                          && (tempFract == dev->testTable[vcc->vci]->fract))
269                  foundLockUp = 1;                   
270               dev->testTable[vcc->vci]->lastTime = tempCellSlot;   
271               dev->testTable[vcc->vci]->fract = tempFract; 
272            }        
273         } /* last descriptor */            
274         vcstatus->cnt = 0;      
275      } /* vcstatus->cnt */
276         
277      if (foundLockUp) {
278         IF_ABR(printk("LOCK UP found\n");) 
279         writew(0xFFFD, dev->seg_reg+MODE_REG_0);
280         /* Wait for 10 Micro sec */
281         udelay(10); 
282         abr_vc->status &= 0xFFF8;
283         abr_vc->status |= 0x0001;  /* state is idle */
284         shd_tbl = (u_short *)dev->ABR_SCHED_TABLE_ADDR;                
285         for( i = 0; ((i < dev->num_vc) && (shd_tbl[i])); i++ );
286         if (i < dev->num_vc)
287            shd_tbl[i] = vcc->vci;
288         else
289            IF_ERR(printk("ABR Seg. may not continue on VC %x\n",vcc->vci);)
290         writew(T_ONLINE, dev->seg_reg+MODE_REG_0);
291         writew(~(TRANSMIT_DONE|TCQ_NOT_EMPTY), dev->seg_reg+SEG_MASK_REG);
292         writew(TRANSMIT_DONE, dev->seg_reg+SEG_INTR_STATUS_REG);       
293         vcstatus->cnt = 0;
294      } /* foundLockUp */
295
296   } /* if an ABR VC */
297
298
299 }
300  
301 /*
302 ** Conversion of 24-bit cellrate (cells/sec) to 16-bit floating point format.
303 **
304 **  +----+----+------------------+-------------------------------+
305 **  |  R | NZ |  5-bit exponent  |        9-bit mantissa         |
306 **  +----+----+------------------+-------------------------------+
307 ** 
308 **    R = reserved (written as 0)
309 **    NZ = 0 if 0 cells/sec; 1 otherwise
310 **
311 **    if NZ = 1, rate = 1.mmmmmmmmm x 2^(eeeee) cells/sec
312 */
313 static u16
314 cellrate_to_float(u32 cr)
315 {
316
317 #define NZ              0x4000
318 #define M_BITS          9               /* Number of bits in mantissa */
319 #define E_BITS          5               /* Number of bits in exponent */
320 #define M_MASK          0x1ff           
321 #define E_MASK          0x1f
322   u16   flot;
323   u32   tmp = cr & 0x00ffffff;
324   int   i   = 0;
325   if (cr == 0)
326      return 0;
327   while (tmp != 1) {
328      tmp >>= 1;
329      i++;
330   }
331   if (i == M_BITS)
332      flot = NZ | (i << M_BITS) | (cr & M_MASK);
333   else if (i < M_BITS)
334      flot = NZ | (i << M_BITS) | ((cr << (M_BITS - i)) & M_MASK);
335   else
336      flot = NZ | (i << M_BITS) | ((cr >> (i - M_BITS)) & M_MASK);
337   return flot;
338 }
339
340 #if 0
341 /*
342 ** Conversion of 16-bit floating point format to 24-bit cellrate (cells/sec).
343 */
344 static u32
345 float_to_cellrate(u16 rate)
346 {
347   u32   exp, mantissa, cps;
348   if ((rate & NZ) == 0)
349      return 0;
350   exp = (rate >> M_BITS) & E_MASK;
351   mantissa = rate & M_MASK;
352   if (exp == 0)
353      return 1;
354   cps = (1 << M_BITS) | mantissa;
355   if (exp == M_BITS)
356      cps = cps;
357   else if (exp > M_BITS)
358      cps <<= (exp - M_BITS);
359   else
360      cps >>= (M_BITS - exp);
361   return cps;
362 }
363 #endif 
364
365 static void init_abr_vc (IADEV *dev, srv_cls_param_t *srv_p) {
366   srv_p->class_type = ATM_ABR;
367   srv_p->pcr        = dev->LineRate;
368   srv_p->mcr        = 0;
369   srv_p->icr        = 0x055cb7;
370   srv_p->tbe        = 0xffffff;
371   srv_p->frtt       = 0x3a;
372   srv_p->rif        = 0xf;
373   srv_p->rdf        = 0xb;
374   srv_p->nrm        = 0x4;
375   srv_p->trm        = 0x7;
376   srv_p->cdf        = 0x3;
377   srv_p->adtf       = 50;
378 }
379
380 static int
381 ia_open_abr_vc(IADEV *dev, srv_cls_param_t *srv_p, 
382                                                 struct atm_vcc *vcc, u8 flag)
383 {
384   f_vc_abr_entry  *f_abr_vc;
385   r_vc_abr_entry  *r_abr_vc;
386   u32           icr;
387   u8            trm, nrm, crm;
388   u16           adtf, air, *ptr16;      
389   f_abr_vc =(f_vc_abr_entry *)dev->MAIN_VC_TABLE_ADDR;
390   f_abr_vc += vcc->vci;       
391   switch (flag) {
392      case 1: /* FFRED initialization */
393 #if 0  /* sanity check */
394        if (srv_p->pcr == 0)
395           return INVALID_PCR;
396        if (srv_p->pcr > dev->LineRate)
397           srv_p->pcr = dev->LineRate;
398        if ((srv_p->mcr + dev->sum_mcr) > dev->LineRate)
399           return MCR_UNAVAILABLE;
400        if (srv_p->mcr > srv_p->pcr)
401           return INVALID_MCR;
402        if (!(srv_p->icr))
403           srv_p->icr = srv_p->pcr;
404        if ((srv_p->icr < srv_p->mcr) || (srv_p->icr > srv_p->pcr))
405           return INVALID_ICR;
406        if ((srv_p->tbe < MIN_TBE) || (srv_p->tbe > MAX_TBE))
407           return INVALID_TBE;
408        if ((srv_p->frtt < MIN_FRTT) || (srv_p->frtt > MAX_FRTT))
409           return INVALID_FRTT;
410        if (srv_p->nrm > MAX_NRM)
411           return INVALID_NRM;
412        if (srv_p->trm > MAX_TRM)
413           return INVALID_TRM;
414        if (srv_p->adtf > MAX_ADTF)
415           return INVALID_ADTF;
416        else if (srv_p->adtf == 0)
417           srv_p->adtf = 1;
418        if (srv_p->cdf > MAX_CDF)
419           return INVALID_CDF;
420        if (srv_p->rif > MAX_RIF)
421           return INVALID_RIF;
422        if (srv_p->rdf > MAX_RDF)
423           return INVALID_RDF;
424 #endif
425        memset ((caddr_t)f_abr_vc, 0, sizeof(*f_abr_vc));
426        f_abr_vc->f_vc_type = ABR;
427        nrm = 2 << srv_p->nrm;     /* (2 ** (srv_p->nrm +1)) */
428                                   /* i.e 2**n = 2 << (n-1) */
429        f_abr_vc->f_nrm = nrm << 8 | nrm;
430        trm = 100000/(2 << (16 - srv_p->trm));
431        if ( trm == 0) trm = 1;
432        f_abr_vc->f_nrmexp =(((srv_p->nrm +1) & 0x0f) << 12)|(MRM << 8) | trm;
433        crm = srv_p->tbe / nrm;
434        if (crm == 0) crm = 1;
435        f_abr_vc->f_crm = crm & 0xff;
436        f_abr_vc->f_pcr = cellrate_to_float(srv_p->pcr);
437        icr = min( srv_p->icr, (srv_p->tbe > srv_p->frtt) ?
438                                 ((srv_p->tbe/srv_p->frtt)*1000000) :
439                                 (1000000/(srv_p->frtt/srv_p->tbe)));
440        f_abr_vc->f_icr = cellrate_to_float(icr);
441        adtf = (10000 * srv_p->adtf)/8192;
442        if (adtf == 0) adtf = 1; 
443        f_abr_vc->f_cdf = ((7 - srv_p->cdf) << 12 | adtf) & 0xfff;
444        f_abr_vc->f_mcr = cellrate_to_float(srv_p->mcr);
445        f_abr_vc->f_acr = f_abr_vc->f_icr;
446        f_abr_vc->f_status = 0x0042;
447        break;
448     case 0: /* RFRED initialization */  
449        ptr16 = (u_short *)(dev->reass_ram + REASS_TABLE*dev->memSize); 
450        *(ptr16 + vcc->vci) = NO_AAL5_PKT | REASS_ABR;
451        r_abr_vc = (r_vc_abr_entry*)(dev->reass_ram+ABR_VC_TABLE*dev->memSize);
452        r_abr_vc += vcc->vci;
453        r_abr_vc->r_status_rdf = (15 - srv_p->rdf) & 0x000f;
454        air = srv_p->pcr << (15 - srv_p->rif);
455        if (air == 0) air = 1;
456        r_abr_vc->r_air = cellrate_to_float(air);
457        dev->testTable[vcc->vci]->vc_status = VC_ACTIVE | VC_ABR;
458        dev->sum_mcr        += srv_p->mcr;
459        dev->n_abr++;
460        break;
461     default:
462        break;
463   }
464   return        0;
465 }
466 static int ia_cbr_setup (IADEV *dev, struct atm_vcc *vcc) {
467    u32 rateLow=0, rateHigh, rate;
468    int entries;
469    struct ia_vcc *ia_vcc;
470
471    int   idealSlot =0, testSlot, toBeAssigned, inc;
472    u32   spacing;
473    u16  *SchedTbl, *TstSchedTbl;
474    u16  cbrVC, vcIndex;
475    u32   fracSlot    = 0;
476    u32   sp_mod      = 0;
477    u32   sp_mod2     = 0;
478
479    /* IpAdjustTrafficParams */
480    if (vcc->qos.txtp.max_pcr <= 0) {
481       IF_ERR(printk("PCR for CBR not defined\n");)
482       return -1;
483    }
484    rate = vcc->qos.txtp.max_pcr;
485    entries = rate / dev->Granularity;
486    IF_CBR(printk("CBR: CBR entries=0x%x for rate=0x%x & Gran=0x%x\n",
487                                 entries, rate, dev->Granularity);)
488    if (entries < 1)
489       IF_CBR(printk("CBR: Bandwidth smaller than granularity of CBR table\n");) 
490    rateLow  =  entries * dev->Granularity;
491    rateHigh = (entries + 1) * dev->Granularity;
492    if (3*(rate - rateLow) > (rateHigh - rate))
493       entries++;
494    if (entries > dev->CbrRemEntries) {
495       IF_CBR(printk("CBR: Not enough bandwidth to support this PCR.\n");)
496       IF_CBR(printk("Entries = 0x%x, CbrRemEntries = 0x%x.\n",
497                                        entries, dev->CbrRemEntries);)
498       return -EBUSY;
499    }   
500
501    ia_vcc = INPH_IA_VCC(vcc);
502    ia_vcc->NumCbrEntry = entries; 
503    dev->sum_mcr += entries * dev->Granularity; 
504    /* IaFFrednInsertCbrSched */
505    // Starting at an arbitrary location, place the entries into the table
506    // as smoothly as possible
507    cbrVC   = 0;
508    spacing = dev->CbrTotEntries / entries;
509    sp_mod  = dev->CbrTotEntries % entries; // get modulo
510    toBeAssigned = entries;
511    fracSlot = 0;
512    vcIndex  = vcc->vci;
513    IF_CBR(printk("Vci=0x%x,Spacing=0x%x,Sp_mod=0x%x\n",vcIndex,spacing,sp_mod);)
514    while (toBeAssigned)
515    {
516       // If this is the first time, start the table loading for this connection
517       // as close to entryPoint as possible.
518       if (toBeAssigned == entries)
519       {
520          idealSlot = dev->CbrEntryPt;
521          dev->CbrEntryPt += 2;    // Adding 2 helps to prevent clumping
522          if (dev->CbrEntryPt >= dev->CbrTotEntries) 
523             dev->CbrEntryPt -= dev->CbrTotEntries;// Wrap if necessary
524       } else {
525          idealSlot += (u32)(spacing + fracSlot); // Point to the next location
526          // in the table that would be  smoothest
527          fracSlot = ((sp_mod + sp_mod2) / entries);  // get new integer part
528          sp_mod2  = ((sp_mod + sp_mod2) % entries);  // calc new fractional part
529       }
530       if (idealSlot >= (int)dev->CbrTotEntries) 
531          idealSlot -= dev->CbrTotEntries;  
532       // Continuously check around this ideal value until a null
533       // location is encountered.
534       SchedTbl = (u16*)(dev->seg_ram+CBR_SCHED_TABLE*dev->memSize); 
535       inc = 0;
536       testSlot = idealSlot;
537       TstSchedTbl = (u16*)(SchedTbl+testSlot);  //set index and read in value
538       IF_CBR(printk("CBR Testslot 0x%x AT Location 0x%x, NumToAssign=%d\n",
539                                 testSlot, (u32)TstSchedTbl,toBeAssigned);) 
540       memcpy((caddr_t)&cbrVC,(caddr_t)TstSchedTbl,sizeof(cbrVC));
541       while (cbrVC)  // If another VC at this location, we have to keep looking
542       {
543           inc++;
544           testSlot = idealSlot - inc;
545           if (testSlot < 0) { // Wrap if necessary
546              testSlot += dev->CbrTotEntries;
547              IF_CBR(printk("Testslot Wrap. STable Start=0x%x,Testslot=%d\n",
548                                                        (u32)SchedTbl,testSlot);)
549           }
550           TstSchedTbl = (u16 *)(SchedTbl + testSlot);  // set table index
551           memcpy((caddr_t)&cbrVC,(caddr_t)TstSchedTbl,sizeof(cbrVC)); 
552           if (!cbrVC)
553              break;
554           testSlot = idealSlot + inc;
555           if (testSlot >= (int)dev->CbrTotEntries) { // Wrap if necessary
556              testSlot -= dev->CbrTotEntries;
557              IF_CBR(printk("TotCbrEntries=%d",dev->CbrTotEntries);)
558              IF_CBR(printk(" Testslot=0x%x ToBeAssgned=%d\n", 
559                                             testSlot, toBeAssigned);)
560           } 
561           // set table index and read in value
562           TstSchedTbl = (u16*)(SchedTbl + testSlot);
563           IF_CBR(printk("Reading CBR Tbl from 0x%x, CbrVal=0x%x Iteration %d\n",
564                           (u32)TstSchedTbl,cbrVC,inc);) 
565           memcpy((caddr_t)&cbrVC,(caddr_t)TstSchedTbl,sizeof(cbrVC));
566        } /* while */
567        // Move this VCI number into this location of the CBR Sched table.
568        memcpy((caddr_t)TstSchedTbl, (caddr_t)&vcIndex,sizeof(TstSchedTbl));
569        dev->CbrRemEntries--;
570        toBeAssigned--;
571    } /* while */ 
572
573    /* IaFFrednCbrEnable */
574    dev->NumEnabledCBR++;
575    if (dev->NumEnabledCBR == 1) {
576        writew((CBR_EN | UBR_EN | ABR_EN | (0x23 << 2)), dev->seg_reg+STPARMS);
577        IF_CBR(printk("CBR is enabled\n");)
578    }
579    return 0;
580 }
581 static void ia_cbrVc_close (struct atm_vcc *vcc) {
582    IADEV *iadev;
583    u16 *SchedTbl, NullVci = 0;
584    u32 i, NumFound;
585
586    iadev = INPH_IA_DEV(vcc->dev);
587    iadev->NumEnabledCBR--;
588    SchedTbl = (u16*)(iadev->seg_ram+CBR_SCHED_TABLE*iadev->memSize);
589    if (iadev->NumEnabledCBR == 0) {
590       writew((UBR_EN | ABR_EN | (0x23 << 2)), iadev->seg_reg+STPARMS);
591       IF_CBR (printk("CBR support disabled\n");)
592    }
593    NumFound = 0;
594    for (i=0; i < iadev->CbrTotEntries; i++)
595    {
596       if (*SchedTbl == vcc->vci) {
597          iadev->CbrRemEntries++;
598          *SchedTbl = NullVci;
599          IF_CBR(NumFound++;)
600       }
601       SchedTbl++;   
602    } 
603    IF_CBR(printk("Exit ia_cbrVc_close, NumRemoved=%d\n",NumFound);)
604 }
605
606 static int ia_avail_descs(IADEV *iadev) {
607    int tmp = 0;
608    ia_hack_tcq(iadev);
609    if (iadev->host_tcq_wr >= iadev->ffL.tcq_rd)
610       tmp = (iadev->host_tcq_wr - iadev->ffL.tcq_rd) / 2;
611    else
612       tmp = (iadev->ffL.tcq_ed - iadev->ffL.tcq_rd + 2 + iadev->host_tcq_wr -
613                    iadev->ffL.tcq_st) / 2;
614    return tmp;
615 }    
616
617 static int ia_pkt_tx (struct atm_vcc *vcc, struct sk_buff *skb);
618
619 static int ia_que_tx (IADEV *iadev) { 
620    struct sk_buff *skb;
621    int num_desc;
622    struct atm_vcc *vcc;
623    struct ia_vcc *iavcc;
624    num_desc = ia_avail_descs(iadev);
625
626    while (num_desc && (skb = skb_dequeue(&iadev->tx_backlog))) {
627       if (!(vcc = ATM_SKB(skb)->vcc)) {
628          dev_kfree_skb_any(skb);
629          printk("ia_que_tx: Null vcc\n");
630          break;
631       }
632       if (!test_bit(ATM_VF_READY,&vcc->flags)) {
633          dev_kfree_skb_any(skb);
634          printk("Free the SKB on closed vci %d \n", vcc->vci);
635          break;
636       }
637       iavcc = INPH_IA_VCC(vcc);
638       if (ia_pkt_tx (vcc, skb)) {
639          skb_queue_head(&iadev->tx_backlog, skb);
640       }
641       num_desc--;
642    }
643    return 0;
644 }
645
646 static void ia_tx_poll (IADEV *iadev) {
647    struct atm_vcc *vcc = NULL;
648    struct sk_buff *skb = NULL, *skb1 = NULL;
649    struct ia_vcc *iavcc;
650    IARTN_Q *  rtne;
651
652    ia_hack_tcq(iadev);
653    while ( (rtne = ia_deque_rtn_q(&iadev->tx_return_q))) {
654        skb = rtne->data.txskb;
655        if (!skb) {
656            printk("ia_tx_poll: skb is null\n");
657            goto out;
658        }
659        vcc = ATM_SKB(skb)->vcc;
660        if (!vcc) {
661            printk("ia_tx_poll: vcc is null\n");
662            dev_kfree_skb_any(skb);
663            goto out;
664        }
665
666        iavcc = INPH_IA_VCC(vcc);
667        if (!iavcc) {
668            printk("ia_tx_poll: iavcc is null\n");
669            dev_kfree_skb_any(skb);
670            goto out;
671        }
672
673        skb1 = skb_dequeue(&iavcc->txing_skb);
674        while (skb1 && (skb1 != skb)) {
675           if (!(IA_SKB_STATE(skb1) & IA_TX_DONE)) {
676              printk("IA_tx_intr: Vci %d lost pkt!!!\n", vcc->vci);
677           }
678           IF_ERR(printk("Release the SKB not match\n");)
679           if ((vcc->pop) && (skb1->len != 0))
680           {
681              vcc->pop(vcc, skb1);
682              IF_EVENT(printk("Tansmit Done - skb 0x%lx return\n",
683                                                           (long)skb1);)
684           }
685           else 
686              dev_kfree_skb_any(skb1);
687           skb1 = skb_dequeue(&iavcc->txing_skb);
688        }                                                        
689        if (!skb1) {
690           IF_EVENT(printk("IA: Vci %d - skb not found requed\n",vcc->vci);)
691           ia_enque_head_rtn_q (&iadev->tx_return_q, rtne);
692           break;
693        }
694        if ((vcc->pop) && (skb->len != 0))
695        {
696           vcc->pop(vcc, skb);
697           IF_EVENT(printk("Tx Done - skb 0x%lx return\n",(long)skb);)
698        }
699        else 
700           dev_kfree_skb_any(skb);
701        kfree(rtne);
702     }
703     ia_que_tx(iadev);
704 out:
705     return;
706 }
707 #if 0
708 static void ia_eeprom_put (IADEV *iadev, u32 addr, u_short val)
709 {
710         u32     t;
711         int     i;
712         /*
713          * Issue a command to enable writes to the NOVRAM
714          */
715         NVRAM_CMD (EXTEND + EWEN);
716         NVRAM_CLR_CE;
717         /*
718          * issue the write command
719          */
720         NVRAM_CMD(IAWRITE + addr);
721         /* 
722          * Send the data, starting with D15, then D14, and so on for 16 bits
723          */
724         for (i=15; i>=0; i--) {
725                 NVRAM_CLKOUT (val & 0x8000);
726                 val <<= 1;
727         }
728         NVRAM_CLR_CE;
729         CFG_OR(NVCE);
730         t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS); 
731         while (!(t & NVDO))
732                 t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS); 
733
734         NVRAM_CLR_CE;
735         /*
736          * disable writes again
737          */
738         NVRAM_CMD(EXTEND + EWDS)
739         NVRAM_CLR_CE;
740         CFG_AND(~NVDI);
741 }
742 #endif
743
744 static u16 ia_eeprom_get (IADEV *iadev, u32 addr)
745 {
746         u_short val;
747         u32     t;
748         int     i;
749         /*
750          * Read the first bit that was clocked with the falling edge of the
751          * the last command data clock
752          */
753         NVRAM_CMD(IAREAD + addr);
754         /*
755          * Now read the rest of the bits, the next bit read is D14, then D13,
756          * and so on.
757          */
758         val = 0;
759         for (i=15; i>=0; i--) {
760                 NVRAM_CLKIN(t);
761                 val |= (t << i);
762         }
763         NVRAM_CLR_CE;
764         CFG_AND(~NVDI);
765         return val;
766 }
767
768 static void ia_hw_type(IADEV *iadev) {
769    u_short memType = ia_eeprom_get(iadev, 25);   
770    iadev->memType = memType;
771    if ((memType & MEM_SIZE_MASK) == MEM_SIZE_1M) {
772       iadev->num_tx_desc = IA_TX_BUF;
773       iadev->tx_buf_sz = IA_TX_BUF_SZ;
774       iadev->num_rx_desc = IA_RX_BUF;
775       iadev->rx_buf_sz = IA_RX_BUF_SZ; 
776    } else if ((memType & MEM_SIZE_MASK) == MEM_SIZE_512K) {
777       if (IA_TX_BUF == DFL_TX_BUFFERS)
778         iadev->num_tx_desc = IA_TX_BUF / 2;
779       else 
780         iadev->num_tx_desc = IA_TX_BUF;
781       iadev->tx_buf_sz = IA_TX_BUF_SZ;
782       if (IA_RX_BUF == DFL_RX_BUFFERS)
783         iadev->num_rx_desc = IA_RX_BUF / 2;
784       else
785         iadev->num_rx_desc = IA_RX_BUF;
786       iadev->rx_buf_sz = IA_RX_BUF_SZ;
787    }
788    else {
789       if (IA_TX_BUF == DFL_TX_BUFFERS) 
790         iadev->num_tx_desc = IA_TX_BUF / 8;
791       else
792         iadev->num_tx_desc = IA_TX_BUF;
793       iadev->tx_buf_sz = IA_TX_BUF_SZ;
794       if (IA_RX_BUF == DFL_RX_BUFFERS)
795         iadev->num_rx_desc = IA_RX_BUF / 8;
796       else
797         iadev->num_rx_desc = IA_RX_BUF;
798       iadev->rx_buf_sz = IA_RX_BUF_SZ; 
799    } 
800    iadev->rx_pkt_ram = TX_PACKET_RAM + (iadev->num_tx_desc * iadev->tx_buf_sz); 
801    IF_INIT(printk("BUF: tx=%d,sz=%d rx=%d sz= %d rx_pkt_ram=%d\n",
802          iadev->num_tx_desc, iadev->tx_buf_sz, iadev->num_rx_desc,
803          iadev->rx_buf_sz, iadev->rx_pkt_ram);)
804
805 #if 0
806    if ((memType & FE_MASK) == FE_SINGLE_MODE) {
807       iadev->phy_type = PHY_OC3C_S;
808    else if ((memType & FE_MASK) == FE_UTP_OPTION)
809       iadev->phy_type = PHY_UTP155;
810    else
811      iadev->phy_type = PHY_OC3C_M;
812 #endif
813    
814    iadev->phy_type = memType & FE_MASK;
815    IF_INIT(printk("memType = 0x%x iadev->phy_type = 0x%x\n", 
816                                          memType,iadev->phy_type);)
817    if (iadev->phy_type == FE_25MBIT_PHY) 
818       iadev->LineRate = (u32)(((25600000/8)*26)/(27*53));
819    else if (iadev->phy_type == FE_DS3_PHY)
820       iadev->LineRate = (u32)(((44736000/8)*26)/(27*53));
821    else if (iadev->phy_type == FE_E3_PHY) 
822       iadev->LineRate = (u32)(((34368000/8)*26)/(27*53));
823    else
824        iadev->LineRate = (u32)(ATM_OC3_PCR);
825    IF_INIT(printk("iadev->LineRate = %d \n", iadev->LineRate);)
826
827 }
828
829 static void IaFrontEndIntr(IADEV *iadev) {
830   volatile IA_SUNI *suni;
831   volatile ia_mb25_t *mb25;
832   volatile suni_pm7345_t *suni_pm7345;
833   u32 intr_status;
834   u_int frmr_intr;
835
836   if(iadev->phy_type & FE_25MBIT_PHY) {
837      mb25 = (ia_mb25_t*)iadev->phy;
838      iadev->carrier_detect =  Boolean(mb25->mb25_intr_status & MB25_IS_GSB);
839   } else if (iadev->phy_type & FE_DS3_PHY) {
840      suni_pm7345 = (suni_pm7345_t *)iadev->phy;
841      /* clear FRMR interrupts */
842      frmr_intr   = suni_pm7345->suni_ds3_frm_intr_stat; 
843      iadev->carrier_detect =  
844            Boolean(!(suni_pm7345->suni_ds3_frm_stat & SUNI_DS3_LOSV));
845   } else if (iadev->phy_type & FE_E3_PHY ) {
846      suni_pm7345 = (suni_pm7345_t *)iadev->phy;
847      frmr_intr   = suni_pm7345->suni_e3_frm_maint_intr_ind;
848      iadev->carrier_detect =
849            Boolean(!(suni_pm7345->suni_e3_frm_fram_intr_ind_stat&SUNI_E3_LOS));
850   }
851   else { 
852      suni = (IA_SUNI *)iadev->phy;
853      intr_status = suni->suni_rsop_status & 0xff;
854      iadev->carrier_detect = Boolean(!(suni->suni_rsop_status & SUNI_LOSV));
855   }
856   if (iadev->carrier_detect)
857     printk("IA: SUNI carrier detected\n");
858   else
859     printk("IA: SUNI carrier lost signal\n"); 
860   return;
861 }
862
863 static void ia_mb25_init (IADEV *iadev)
864 {
865    volatile ia_mb25_t  *mb25 = (ia_mb25_t*)iadev->phy;
866 #if 0
867    mb25->mb25_master_ctrl = MB25_MC_DRIC | MB25_MC_DREC | MB25_MC_ENABLED;
868 #endif
869    mb25->mb25_master_ctrl = MB25_MC_DRIC | MB25_MC_DREC;
870    mb25->mb25_diag_control = 0;
871    /*
872     * Initialize carrier detect state
873     */
874    iadev->carrier_detect =  Boolean(mb25->mb25_intr_status & MB25_IS_GSB);
875    return;
876 }                   
877
878 static void ia_suni_pm7345_init (IADEV *iadev)
879 {
880    volatile suni_pm7345_t *suni_pm7345 = (suni_pm7345_t *)iadev->phy;
881    if (iadev->phy_type & FE_DS3_PHY)
882    {
883       iadev->carrier_detect = 
884           Boolean(!(suni_pm7345->suni_ds3_frm_stat & SUNI_DS3_LOSV)); 
885       suni_pm7345->suni_ds3_frm_intr_enbl = 0x17;
886       suni_pm7345->suni_ds3_frm_cfg = 1;
887       suni_pm7345->suni_ds3_tran_cfg = 1;
888       suni_pm7345->suni_config = 0;
889       suni_pm7345->suni_splr_cfg = 0;
890       suni_pm7345->suni_splt_cfg = 0;
891    }
892    else 
893    {
894       iadev->carrier_detect = 
895           Boolean(!(suni_pm7345->suni_e3_frm_fram_intr_ind_stat & SUNI_E3_LOS));
896       suni_pm7345->suni_e3_frm_fram_options = 0x4;
897       suni_pm7345->suni_e3_frm_maint_options = 0x20;
898       suni_pm7345->suni_e3_frm_fram_intr_enbl = 0x1d;
899       suni_pm7345->suni_e3_frm_maint_intr_enbl = 0x30;
900       suni_pm7345->suni_e3_tran_stat_diag_options = 0x0;
901       suni_pm7345->suni_e3_tran_fram_options = 0x1;
902       suni_pm7345->suni_config = SUNI_PM7345_E3ENBL;
903       suni_pm7345->suni_splr_cfg = 0x41;
904       suni_pm7345->suni_splt_cfg = 0x41;
905    } 
906    /*
907     * Enable RSOP loss of signal interrupt.
908     */
909    suni_pm7345->suni_intr_enbl = 0x28;
910  
911    /*
912     * Clear error counters
913     */
914    suni_pm7345->suni_id_reset = 0;
915
916    /*
917     * Clear "PMCTST" in master test register.
918     */
919    suni_pm7345->suni_master_test = 0;
920
921    suni_pm7345->suni_rxcp_ctrl = 0x2c;
922    suni_pm7345->suni_rxcp_fctrl = 0x81;
923  
924    suni_pm7345->suni_rxcp_idle_pat_h1 =
925         suni_pm7345->suni_rxcp_idle_pat_h2 =
926         suni_pm7345->suni_rxcp_idle_pat_h3 = 0;
927    suni_pm7345->suni_rxcp_idle_pat_h4 = 1;
928  
929    suni_pm7345->suni_rxcp_idle_mask_h1 = 0xff;
930    suni_pm7345->suni_rxcp_idle_mask_h2 = 0xff;
931    suni_pm7345->suni_rxcp_idle_mask_h3 = 0xff;
932    suni_pm7345->suni_rxcp_idle_mask_h4 = 0xfe;
933  
934    suni_pm7345->suni_rxcp_cell_pat_h1 =
935         suni_pm7345->suni_rxcp_cell_pat_h2 =
936         suni_pm7345->suni_rxcp_cell_pat_h3 = 0;
937    suni_pm7345->suni_rxcp_cell_pat_h4 = 1;
938  
939    suni_pm7345->suni_rxcp_cell_mask_h1 =
940         suni_pm7345->suni_rxcp_cell_mask_h2 =
941         suni_pm7345->suni_rxcp_cell_mask_h3 =
942         suni_pm7345->suni_rxcp_cell_mask_h4 = 0xff;
943  
944    suni_pm7345->suni_txcp_ctrl = 0xa4;
945    suni_pm7345->suni_txcp_intr_en_sts = 0x10;
946    suni_pm7345->suni_txcp_idle_pat_h5 = 0x55;
947  
948    suni_pm7345->suni_config &= ~(SUNI_PM7345_LLB |
949                                  SUNI_PM7345_CLB |
950                                  SUNI_PM7345_DLB |
951                                   SUNI_PM7345_PLB);
952 #ifdef __SNMP__
953    suni_pm7345->suni_rxcp_intr_en_sts |= SUNI_OOCDE;
954 #endif /* __SNMP__ */
955    return;
956 }
957
958
959 /***************************** IA_LIB END *****************************/
960     
961 #ifdef CONFIG_ATM_IA_DEBUG
962 static int tcnter = 0;
963 static void xdump( u_char*  cp, int  length, char*  prefix )
964 {
965     int col, count;
966     u_char prntBuf[120];
967     u_char*  pBuf = prntBuf;
968     count = 0;
969     while(count < length){
970         pBuf += sprintf( pBuf, "%s", prefix );
971         for(col = 0;count + col < length && col < 16; col++){
972             if (col != 0 && (col % 4) == 0)
973                 pBuf += sprintf( pBuf, " " );
974             pBuf += sprintf( pBuf, "%02X ", cp[count + col] );
975         }
976         while(col++ < 16){      /* pad end of buffer with blanks */
977             if ((col % 4) == 0)
978                 sprintf( pBuf, " " );
979             pBuf += sprintf( pBuf, "   " );
980         }
981         pBuf += sprintf( pBuf, "  " );
982         for(col = 0;count + col < length && col < 16; col++){
983             if (isprint((int)cp[count + col]))
984                 pBuf += sprintf( pBuf, "%c", cp[count + col] );
985             else
986                 pBuf += sprintf( pBuf, "." );
987                 }
988         sprintf( pBuf, "\n" );
989         // SPrint(prntBuf);
990         printk(prntBuf);
991         count += col;
992         pBuf = prntBuf;
993     }
994
995 }  /* close xdump(... */
996 #endif /* CONFIG_ATM_IA_DEBUG */
997
998   
999 static struct atm_dev *ia_boards = NULL;  
1000   
1001 #define ACTUAL_RAM_BASE \
1002         RAM_BASE*((iadev->mem)/(128 * 1024))  
1003 #define ACTUAL_SEG_RAM_BASE \
1004         IPHASE5575_FRAG_CONTROL_RAM_BASE*((iadev->mem)/(128 * 1024))  
1005 #define ACTUAL_REASS_RAM_BASE \
1006         IPHASE5575_REASS_CONTROL_RAM_BASE*((iadev->mem)/(128 * 1024))  
1007   
1008   
1009 /*-- some utilities and memory allocation stuff will come here -------------*/  
1010   
1011 static void desc_dbg(IADEV *iadev) {
1012
1013   u_short tcq_wr_ptr, tcq_st_ptr, tcq_ed_ptr;
1014   u32 i;
1015   void __iomem *tmp;
1016   // regval = readl((u32)ia_cmds->maddr);
1017   tcq_wr_ptr =  readw(iadev->seg_reg+TCQ_WR_PTR);
1018   printk("B_tcq_wr = 0x%x desc = %d last desc = %d\n",
1019                      tcq_wr_ptr, readw(iadev->seg_ram+tcq_wr_ptr),
1020                      readw(iadev->seg_ram+tcq_wr_ptr-2));
1021   printk(" host_tcq_wr = 0x%x  host_tcq_rd = 0x%x \n",  iadev->host_tcq_wr, 
1022                    iadev->ffL.tcq_rd);
1023   tcq_st_ptr =  readw(iadev->seg_reg+TCQ_ST_ADR);
1024   tcq_ed_ptr =  readw(iadev->seg_reg+TCQ_ED_ADR);
1025   printk("tcq_st_ptr = 0x%x    tcq_ed_ptr = 0x%x \n", tcq_st_ptr, tcq_ed_ptr);
1026   i = 0;
1027   while (tcq_st_ptr != tcq_ed_ptr) {
1028       tmp = iadev->seg_ram+tcq_st_ptr;
1029       printk("TCQ slot %d desc = %d  Addr = %p\n", i++, readw(tmp), tmp);
1030       tcq_st_ptr += 2;
1031   }
1032   for(i=0; i <iadev->num_tx_desc; i++)
1033       printk("Desc_tbl[%d] = %d \n", i, iadev->desc_tbl[i].timestamp);
1034
1035   
1036   
1037 /*----------------------------- Recieving side stuff --------------------------*/  
1038  
1039 static void rx_excp_rcvd(struct atm_dev *dev)  
1040 {  
1041 #if 0 /* closing the receiving size will cause too many excp int */  
1042   IADEV *iadev;  
1043   u_short state;  
1044   u_short excpq_rd_ptr;  
1045   //u_short *ptr;  
1046   int vci, error = 1;  
1047   iadev = INPH_IA_DEV(dev);  
1048   state = readl(iadev->reass_reg + STATE_REG) & 0xffff;  
1049   while((state & EXCPQ_EMPTY) != EXCPQ_EMPTY)  
1050   { printk("state = %x \n", state); 
1051         excpq_rd_ptr = readw(iadev->reass_reg + EXCP_Q_RD_PTR) & 0xffff;  
1052  printk("state = %x excpq_rd_ptr = %x \n", state, excpq_rd_ptr); 
1053         if (excpq_rd_ptr == *(u16*)(iadev->reass_reg + EXCP_Q_WR_PTR))
1054             IF_ERR(printk("excpq_rd_ptr is wrong!!!\n");)
1055         // TODO: update exception stat
1056         vci = readw(iadev->reass_ram+excpq_rd_ptr);  
1057         error = readw(iadev->reass_ram+excpq_rd_ptr+2) & 0x0007;  
1058         // pwang_test
1059         excpq_rd_ptr += 4;  
1060         if (excpq_rd_ptr > (readw(iadev->reass_reg + EXCP_Q_ED_ADR)& 0xffff))  
1061             excpq_rd_ptr = readw(iadev->reass_reg + EXCP_Q_ST_ADR)& 0xffff;
1062         writew( excpq_rd_ptr, iadev->reass_reg + EXCP_Q_RD_PTR);  
1063         state = readl(iadev->reass_reg + STATE_REG) & 0xffff;  
1064   }  
1065 #endif
1066 }  
1067   
1068 static void free_desc(struct atm_dev *dev, int desc)  
1069 {  
1070         IADEV *iadev;  
1071         iadev = INPH_IA_DEV(dev);  
1072         writew(desc, iadev->reass_ram+iadev->rfL.fdq_wr); 
1073         iadev->rfL.fdq_wr +=2;
1074         if (iadev->rfL.fdq_wr > iadev->rfL.fdq_ed)
1075                 iadev->rfL.fdq_wr =  iadev->rfL.fdq_st;  
1076         writew(iadev->rfL.fdq_wr, iadev->reass_reg+FREEQ_WR_PTR);  
1077 }  
1078   
1079   
1080 static int rx_pkt(struct atm_dev *dev)  
1081 {  
1082         IADEV *iadev;  
1083         struct atm_vcc *vcc;  
1084         unsigned short status;  
1085         struct rx_buf_desc __iomem *buf_desc_ptr;  
1086         int desc;   
1087         struct dle* wr_ptr;  
1088         int len;  
1089         struct sk_buff *skb;  
1090         u_int buf_addr, dma_addr;  
1091
1092         iadev = INPH_IA_DEV(dev);  
1093         if (iadev->rfL.pcq_rd == (readw(iadev->reass_reg+PCQ_WR_PTR)&0xffff)) 
1094         {  
1095             printk(KERN_ERR DEV_LABEL "(itf %d) Receive queue empty\n", dev->number);  
1096             return -EINVAL;  
1097         }  
1098         /* mask 1st 3 bits to get the actual descno. */  
1099         desc = readw(iadev->reass_ram+iadev->rfL.pcq_rd) & 0x1fff;  
1100         IF_RX(printk("reass_ram = %p iadev->rfL.pcq_rd = 0x%x desc = %d\n", 
1101                                     iadev->reass_ram, iadev->rfL.pcq_rd, desc);
1102               printk(" pcq_wr_ptr = 0x%x\n",
1103                                readw(iadev->reass_reg+PCQ_WR_PTR)&0xffff);)
1104         /* update the read pointer  - maybe we shud do this in the end*/  
1105         if ( iadev->rfL.pcq_rd== iadev->rfL.pcq_ed) 
1106                 iadev->rfL.pcq_rd = iadev->rfL.pcq_st;  
1107         else  
1108                 iadev->rfL.pcq_rd += 2;
1109         writew(iadev->rfL.pcq_rd, iadev->reass_reg+PCQ_RD_PTR);  
1110   
1111         /* get the buffer desc entry.  
1112                 update stuff. - doesn't seem to be any update necessary  
1113         */  
1114         buf_desc_ptr = iadev->RX_DESC_BASE_ADDR;
1115         /* make the ptr point to the corresponding buffer desc entry */  
1116         buf_desc_ptr += desc;     
1117         if (!desc || (desc > iadev->num_rx_desc) || 
1118                       ((buf_desc_ptr->vc_index & 0xffff) > iadev->num_vc)) { 
1119             free_desc(dev, desc);
1120             IF_ERR(printk("IA: bad descriptor desc = %d \n", desc);)
1121             return -1;
1122         }
1123         vcc = iadev->rx_open[buf_desc_ptr->vc_index & 0xffff];  
1124         if (!vcc)  
1125         {      
1126                 free_desc(dev, desc); 
1127                 printk("IA: null vcc, drop PDU\n");  
1128                 return -1;  
1129         }  
1130           
1131   
1132         /* might want to check the status bits for errors */  
1133         status = (u_short) (buf_desc_ptr->desc_mode);  
1134         if (status & (RX_CER | RX_PTE | RX_OFL))  
1135         {  
1136                 atomic_inc(&vcc->stats->rx_err);
1137                 IF_ERR(printk("IA: bad packet, dropping it");)  
1138                 if (status & RX_CER) { 
1139                     IF_ERR(printk(" cause: packet CRC error\n");)
1140                 }
1141                 else if (status & RX_PTE) {
1142                     IF_ERR(printk(" cause: packet time out\n");)
1143                 }
1144                 else {
1145                     IF_ERR(printk(" cause: buffer over flow\n");)
1146                 }
1147                 goto out_free_desc;
1148         }  
1149   
1150         /*  
1151                 build DLE.        
1152         */  
1153   
1154         buf_addr = (buf_desc_ptr->buf_start_hi << 16) | buf_desc_ptr->buf_start_lo;  
1155         dma_addr = (buf_desc_ptr->dma_start_hi << 16) | buf_desc_ptr->dma_start_lo;  
1156         len = dma_addr - buf_addr;  
1157         if (len > iadev->rx_buf_sz) {
1158            printk("Over %d bytes sdu received, dropped!!!\n", iadev->rx_buf_sz);
1159            atomic_inc(&vcc->stats->rx_err);
1160            goto out_free_desc;
1161         }
1162                   
1163         if (!(skb = atm_alloc_charge(vcc, len, GFP_ATOMIC))) {
1164            if (vcc->vci < 32)
1165               printk("Drop control packets\n");
1166               goto out_free_desc;
1167         }
1168         skb_put(skb,len);  
1169         // pwang_test
1170         ATM_SKB(skb)->vcc = vcc;
1171         ATM_DESC(skb) = desc;        
1172         skb_queue_tail(&iadev->rx_dma_q, skb);  
1173
1174         /* Build the DLE structure */  
1175         wr_ptr = iadev->rx_dle_q.write;  
1176         wr_ptr->sys_pkt_addr = pci_map_single(iadev->pci, skb->data,
1177                 len, PCI_DMA_FROMDEVICE);
1178         wr_ptr->local_pkt_addr = buf_addr;  
1179         wr_ptr->bytes = len;    /* We don't know this do we ?? */  
1180         wr_ptr->mode = DMA_INT_ENABLE;  
1181   
1182         /* shud take care of wrap around here too. */  
1183         if(++wr_ptr == iadev->rx_dle_q.end)
1184              wr_ptr = iadev->rx_dle_q.start;
1185         iadev->rx_dle_q.write = wr_ptr;  
1186         udelay(1);  
1187         /* Increment transaction counter */  
1188         writel(1, iadev->dma+IPHASE5575_RX_COUNTER);   
1189 out:    return 0;  
1190 out_free_desc:
1191         free_desc(dev, desc);
1192         goto out;
1193 }  
1194   
1195 static void rx_intr(struct atm_dev *dev)  
1196 {  
1197   IADEV *iadev;  
1198   u_short status;  
1199   u_short state, i;  
1200   
1201   iadev = INPH_IA_DEV(dev);  
1202   status = readl(iadev->reass_reg+REASS_INTR_STATUS_REG) & 0xffff;  
1203   IF_EVENT(printk("rx_intr: status = 0x%x\n", status);)
1204   if (status & RX_PKT_RCVD)  
1205   {  
1206         /* do something */  
1207         /* Basically recvd an interrupt for receving a packet.  
1208         A descriptor would have been written to the packet complete   
1209         queue. Get all the descriptors and set up dma to move the   
1210         packets till the packet complete queue is empty..  
1211         */  
1212         state = readl(iadev->reass_reg + STATE_REG) & 0xffff;  
1213         IF_EVENT(printk("Rx intr status: RX_PKT_RCVD %08x\n", status);) 
1214         while(!(state & PCQ_EMPTY))  
1215         {  
1216              rx_pkt(dev);  
1217              state = readl(iadev->reass_reg + STATE_REG) & 0xffff;  
1218         }  
1219         iadev->rxing = 1;
1220   }  
1221   if (status & RX_FREEQ_EMPT)  
1222   {   
1223      if (iadev->rxing) {
1224         iadev->rx_tmp_cnt = iadev->rx_pkt_cnt;
1225         iadev->rx_tmp_jif = jiffies; 
1226         iadev->rxing = 0;
1227      } 
1228      else if (((jiffies - iadev->rx_tmp_jif) > 50) && 
1229                ((iadev->rx_pkt_cnt - iadev->rx_tmp_cnt) == 0)) {
1230         for (i = 1; i <= iadev->num_rx_desc; i++)
1231                free_desc(dev, i);
1232 printk("Test logic RUN!!!!\n");
1233         writew( ~(RX_FREEQ_EMPT|RX_EXCP_RCVD),iadev->reass_reg+REASS_MASK_REG);
1234         iadev->rxing = 1;
1235      }
1236      IF_EVENT(printk("Rx intr status: RX_FREEQ_EMPT %08x\n", status);)  
1237   }  
1238
1239   if (status & RX_EXCP_RCVD)  
1240   {  
1241         /* probably need to handle the exception queue also. */  
1242         IF_EVENT(printk("Rx intr status: RX_EXCP_RCVD %08x\n", status);)  
1243         rx_excp_rcvd(dev);  
1244   }  
1245
1246
1247   if (status & RX_RAW_RCVD)  
1248   {  
1249         /* need to handle the raw incoming cells. This deepnds on   
1250         whether we have programmed to receive the raw cells or not.  
1251         Else ignore. */  
1252         IF_EVENT(printk("Rx intr status:  RX_RAW_RCVD %08x\n", status);)  
1253   }  
1254 }  
1255   
1256   
1257 static void rx_dle_intr(struct atm_dev *dev)  
1258 {  
1259   IADEV *iadev;  
1260   struct atm_vcc *vcc;   
1261   struct sk_buff *skb;  
1262   int desc;  
1263   u_short state;   
1264   struct dle *dle, *cur_dle;  
1265   u_int dle_lp;  
1266   int len;
1267   iadev = INPH_IA_DEV(dev);  
1268  
1269   /* free all the dles done, that is just update our own dle read pointer   
1270         - do we really need to do this. Think not. */  
1271   /* DMA is done, just get all the recevie buffers from the rx dma queue  
1272         and push them up to the higher layer protocol. Also free the desc  
1273         associated with the buffer. */  
1274   dle = iadev->rx_dle_q.read;  
1275   dle_lp = readl(iadev->dma+IPHASE5575_RX_LIST_ADDR) & (sizeof(struct dle)*DLE_ENTRIES - 1);  
1276   cur_dle = (struct dle*)(iadev->rx_dle_q.start + (dle_lp >> 4));  
1277   while(dle != cur_dle)  
1278   {  
1279       /* free the DMAed skb */  
1280       skb = skb_dequeue(&iadev->rx_dma_q);  
1281       if (!skb)  
1282          goto INCR_DLE;
1283       desc = ATM_DESC(skb);
1284       free_desc(dev, desc);  
1285                
1286       if (!(len = skb->len))
1287       {  
1288           printk("rx_dle_intr: skb len 0\n");  
1289           dev_kfree_skb_any(skb);  
1290       }  
1291       else  
1292       {  
1293           struct cpcs_trailer *trailer;
1294           u_short length;
1295           struct ia_vcc *ia_vcc;
1296
1297           pci_unmap_single(iadev->pci, iadev->rx_dle_q.write->sys_pkt_addr,
1298                 len, PCI_DMA_FROMDEVICE);
1299           /* no VCC related housekeeping done as yet. lets see */  
1300           vcc = ATM_SKB(skb)->vcc;
1301           if (!vcc) {
1302               printk("IA: null vcc\n");  
1303               dev_kfree_skb_any(skb);
1304               goto INCR_DLE;
1305           }
1306           ia_vcc = INPH_IA_VCC(vcc);
1307           if (ia_vcc == NULL)
1308           {
1309              atomic_inc(&vcc->stats->rx_err);
1310              dev_kfree_skb_any(skb);
1311              atm_return(vcc, atm_guess_pdu2truesize(len));
1312              goto INCR_DLE;
1313            }
1314           // get real pkt length  pwang_test
1315           trailer = (struct cpcs_trailer*)((u_char *)skb->data +
1316                                  skb->len - sizeof(*trailer));
1317           length =  swap(trailer->length);
1318           if ((length > iadev->rx_buf_sz) || (length > 
1319                               (skb->len - sizeof(struct cpcs_trailer))))
1320           {
1321              atomic_inc(&vcc->stats->rx_err);
1322              IF_ERR(printk("rx_dle_intr: Bad  AAL5 trailer %d (skb len %d)", 
1323                                                             length, skb->len);)
1324              dev_kfree_skb_any(skb);
1325              atm_return(vcc, atm_guess_pdu2truesize(len));
1326              goto INCR_DLE;
1327           }
1328           skb_trim(skb, length);
1329           
1330           /* Display the packet */  
1331           IF_RXPKT(printk("\nDmad Recvd data: len = %d \n", skb->len);  
1332           xdump(skb->data, skb->len, "RX: ");
1333           printk("\n");)
1334
1335           IF_RX(printk("rx_dle_intr: skb push");)  
1336           vcc->push(vcc,skb);  
1337           atomic_inc(&vcc->stats->rx);
1338           iadev->rx_pkt_cnt++;
1339       }  
1340 INCR_DLE:
1341       if (++dle == iadev->rx_dle_q.end)  
1342           dle = iadev->rx_dle_q.start;  
1343   }  
1344   iadev->rx_dle_q.read = dle;  
1345   
1346   /* if the interrupts are masked because there were no free desc available,  
1347                 unmask them now. */ 
1348   if (!iadev->rxing) {
1349      state = readl(iadev->reass_reg + STATE_REG) & 0xffff;
1350      if (!(state & FREEQ_EMPTY)) {
1351         state = readl(iadev->reass_reg + REASS_MASK_REG) & 0xffff;
1352         writel(state & ~(RX_FREEQ_EMPT |/* RX_EXCP_RCVD |*/ RX_PKT_RCVD),
1353                                       iadev->reass_reg+REASS_MASK_REG);
1354         iadev->rxing++; 
1355      }
1356   }
1357 }  
1358   
1359   
1360 static int open_rx(struct atm_vcc *vcc)  
1361 {  
1362         IADEV *iadev;  
1363         u_short __iomem *vc_table;  
1364         u_short __iomem *reass_ptr;  
1365         IF_EVENT(printk("iadev: open_rx %d.%d\n", vcc->vpi, vcc->vci);)
1366
1367         if (vcc->qos.rxtp.traffic_class == ATM_NONE) return 0;    
1368         iadev = INPH_IA_DEV(vcc->dev);  
1369         if (vcc->qos.rxtp.traffic_class == ATM_ABR) {  
1370            if (iadev->phy_type & FE_25MBIT_PHY) {
1371                printk("IA:  ABR not support\n");
1372                return -EINVAL; 
1373            }
1374         }
1375         /* Make only this VCI in the vc table valid and let all   
1376                 others be invalid entries */  
1377         vc_table = iadev->reass_ram+RX_VC_TABLE*iadev->memSize;
1378         vc_table += vcc->vci;
1379         /* mask the last 6 bits and OR it with 3 for 1K VCs */  
1380
1381         *vc_table = vcc->vci << 6;
1382         /* Also keep a list of open rx vcs so that we can attach them with  
1383                 incoming PDUs later. */  
1384         if ((vcc->qos.rxtp.traffic_class == ATM_ABR) || 
1385                                 (vcc->qos.txtp.traffic_class == ATM_ABR))  
1386         {  
1387                 srv_cls_param_t srv_p;
1388                 init_abr_vc(iadev, &srv_p);
1389                 ia_open_abr_vc(iadev, &srv_p, vcc, 0);
1390         } 
1391         else {  /* for UBR  later may need to add CBR logic */
1392                 reass_ptr = iadev->reass_ram+REASS_TABLE*iadev->memSize;
1393                 reass_ptr += vcc->vci;
1394                 *reass_ptr = NO_AAL5_PKT;
1395         }
1396         
1397         if (iadev->rx_open[vcc->vci])  
1398                 printk(KERN_CRIT DEV_LABEL "(itf %d): VCI %d already open\n",  
1399                         vcc->dev->number, vcc->vci);  
1400         iadev->rx_open[vcc->vci] = vcc;  
1401         return 0;  
1402 }  
1403   
1404 static int rx_init(struct atm_dev *dev)  
1405 {  
1406         IADEV *iadev;  
1407         struct rx_buf_desc __iomem *buf_desc_ptr;  
1408         unsigned long rx_pkt_start = 0;  
1409         void *dle_addr;  
1410         struct abr_vc_table  *abr_vc_table; 
1411         u16 *vc_table;  
1412         u16 *reass_table;  
1413         u16 *ptr16;
1414         int i,j, vcsize_sel;  
1415         u_short freeq_st_adr;  
1416         u_short *freeq_start;  
1417   
1418         iadev = INPH_IA_DEV(dev);  
1419   //    spin_lock_init(&iadev->rx_lock); 
1420   
1421         /* Allocate 4k bytes - more aligned than needed (4k boundary) */
1422         dle_addr = pci_alloc_consistent(iadev->pci, DLE_TOTAL_SIZE,
1423                                         &iadev->rx_dle_dma);  
1424         if (!dle_addr)  {  
1425                 printk(KERN_ERR DEV_LABEL "can't allocate DLEs\n");
1426                 goto err_out;
1427         }
1428         iadev->rx_dle_q.start = (struct dle*)dle_addr;  
1429         iadev->rx_dle_q.read = iadev->rx_dle_q.start;  
1430         iadev->rx_dle_q.write = iadev->rx_dle_q.start;  
1431         iadev->rx_dle_q.end = (struct dle*)((u32)dle_addr+sizeof(struct dle)*DLE_ENTRIES);  
1432         /* the end of the dle q points to the entry after the last  
1433         DLE that can be used. */  
1434   
1435         /* write the upper 20 bits of the start address to rx list address register */  
1436         writel(iadev->rx_dle_dma & 0xfffff000,
1437                iadev->dma + IPHASE5575_RX_LIST_ADDR);  
1438         IF_INIT(printk("Tx Dle list addr: 0x%08x value: 0x%0x\n", 
1439                       (u32)(iadev->dma+IPHASE5575_TX_LIST_ADDR), 
1440                       *(u32*)(iadev->dma+IPHASE5575_TX_LIST_ADDR));  
1441         printk("Rx Dle list addr: 0x%08x value: 0x%0x\n", 
1442                       (u32)(iadev->dma+IPHASE5575_RX_LIST_ADDR), 
1443                       *(u32*)(iadev->dma+IPHASE5575_RX_LIST_ADDR));)  
1444   
1445         writew(0xffff, iadev->reass_reg+REASS_MASK_REG);  
1446         writew(0, iadev->reass_reg+MODE_REG);  
1447         writew(RESET_REASS, iadev->reass_reg+REASS_COMMAND_REG);  
1448   
1449         /* Receive side control memory map  
1450            -------------------------------  
1451   
1452                 Buffer descr    0x0000 (736 - 23K)  
1453                 VP Table        0x5c00 (256 - 512)  
1454                 Except q        0x5e00 (128 - 512)  
1455                 Free buffer q   0x6000 (1K - 2K)  
1456                 Packet comp q   0x6800 (1K - 2K)  
1457                 Reass Table     0x7000 (1K - 2K)  
1458                 VC Table        0x7800 (1K - 2K)  
1459                 ABR VC Table    0x8000 (1K - 32K)  
1460         */  
1461           
1462         /* Base address for Buffer Descriptor Table */  
1463         writew(RX_DESC_BASE >> 16, iadev->reass_reg+REASS_DESC_BASE);  
1464         /* Set the buffer size register */  
1465         writew(iadev->rx_buf_sz, iadev->reass_reg+BUF_SIZE);  
1466   
1467         /* Initialize each entry in the Buffer Descriptor Table */  
1468         iadev->RX_DESC_BASE_ADDR = iadev->reass_ram+RX_DESC_BASE*iadev->memSize;
1469         buf_desc_ptr = iadev->RX_DESC_BASE_ADDR;
1470         memset_io(buf_desc_ptr, 0, sizeof(*buf_desc_ptr));
1471         buf_desc_ptr++;  
1472         rx_pkt_start = iadev->rx_pkt_ram;  
1473         for(i=1; i<=iadev->num_rx_desc; i++)  
1474         {  
1475                 memset_io(buf_desc_ptr, 0, sizeof(*buf_desc_ptr));  
1476                 buf_desc_ptr->buf_start_hi = rx_pkt_start >> 16;  
1477                 buf_desc_ptr->buf_start_lo = rx_pkt_start & 0x0000ffff;  
1478                 buf_desc_ptr++;           
1479                 rx_pkt_start += iadev->rx_buf_sz;  
1480         }  
1481         IF_INIT(printk("Rx Buffer desc ptr: 0x%0x\n", (u32)(buf_desc_ptr));)  
1482         i = FREE_BUF_DESC_Q*iadev->memSize; 
1483         writew(i >> 16,  iadev->reass_reg+REASS_QUEUE_BASE); 
1484         writew(i, iadev->reass_reg+FREEQ_ST_ADR);
1485         writew(i+iadev->num_rx_desc*sizeof(u_short), 
1486                                          iadev->reass_reg+FREEQ_ED_ADR);
1487         writew(i, iadev->reass_reg+FREEQ_RD_PTR);
1488         writew(i+iadev->num_rx_desc*sizeof(u_short), 
1489                                         iadev->reass_reg+FREEQ_WR_PTR);    
1490         /* Fill the FREEQ with all the free descriptors. */  
1491         freeq_st_adr = readw(iadev->reass_reg+FREEQ_ST_ADR);  
1492         freeq_start = (u_short *)(iadev->reass_ram+freeq_st_adr);  
1493         for(i=1; i<=iadev->num_rx_desc; i++)  
1494         {  
1495                 *freeq_start = (u_short)i;  
1496                 freeq_start++;  
1497         }  
1498         IF_INIT(printk("freeq_start: 0x%0x\n", (u32)freeq_start);)  
1499         /* Packet Complete Queue */
1500         i = (PKT_COMP_Q * iadev->memSize) & 0xffff;
1501         writew(i, iadev->reass_reg+PCQ_ST_ADR);
1502         writew(i+iadev->num_vc*sizeof(u_short), iadev->reass_reg+PCQ_ED_ADR);
1503         writew(i, iadev->reass_reg+PCQ_RD_PTR);
1504         writew(i, iadev->reass_reg+PCQ_WR_PTR);
1505
1506         /* Exception Queue */
1507         i = (EXCEPTION_Q * iadev->memSize) & 0xffff;
1508         writew(i, iadev->reass_reg+EXCP_Q_ST_ADR);
1509         writew(i + NUM_RX_EXCP * sizeof(RX_ERROR_Q), 
1510                                              iadev->reass_reg+EXCP_Q_ED_ADR);
1511         writew(i, iadev->reass_reg+EXCP_Q_RD_PTR);
1512         writew(i, iadev->reass_reg+EXCP_Q_WR_PTR); 
1513  
1514         /* Load local copy of FREEQ and PCQ ptrs */
1515         iadev->rfL.fdq_st = readw(iadev->reass_reg+FREEQ_ST_ADR) & 0xffff;
1516         iadev->rfL.fdq_ed = readw(iadev->reass_reg+FREEQ_ED_ADR) & 0xffff ;
1517         iadev->rfL.fdq_rd = readw(iadev->reass_reg+FREEQ_RD_PTR) & 0xffff;
1518         iadev->rfL.fdq_wr = readw(iadev->reass_reg+FREEQ_WR_PTR) & 0xffff;
1519         iadev->rfL.pcq_st = readw(iadev->reass_reg+PCQ_ST_ADR) & 0xffff;
1520         iadev->rfL.pcq_ed = readw(iadev->reass_reg+PCQ_ED_ADR) & 0xffff;
1521         iadev->rfL.pcq_rd = readw(iadev->reass_reg+PCQ_RD_PTR) & 0xffff;
1522         iadev->rfL.pcq_wr = readw(iadev->reass_reg+PCQ_WR_PTR) & 0xffff;
1523         
1524         IF_INIT(printk("INIT:pcq_st:0x%x pcq_ed:0x%x pcq_rd:0x%x pcq_wr:0x%x", 
1525               iadev->rfL.pcq_st, iadev->rfL.pcq_ed, iadev->rfL.pcq_rd, 
1526               iadev->rfL.pcq_wr);)                
1527         /* just for check - no VP TBL */  
1528         /* VP Table */  
1529         /* writew(0x0b80, iadev->reass_reg+VP_LKUP_BASE); */  
1530         /* initialize VP Table for invalid VPIs  
1531                 - I guess we can write all 1s or 0x000f in the entire memory  
1532                   space or something similar.  
1533         */  
1534   
1535         /* This seems to work and looks right to me too !!! */  
1536         i =  REASS_TABLE * iadev->memSize;
1537         writew((i >> 3), iadev->reass_reg+REASS_TABLE_BASE);   
1538         /* initialize Reassembly table to I don't know what ???? */  
1539         reass_table = (u16 *)(iadev->reass_ram+i);  
1540         j = REASS_TABLE_SZ * iadev->memSize;
1541         for(i=0; i < j; i++)  
1542                 *reass_table++ = NO_AAL5_PKT;  
1543        i = 8*1024;
1544        vcsize_sel =  0;
1545        while (i != iadev->num_vc) {
1546           i /= 2;
1547           vcsize_sel++;
1548        }
1549        i = RX_VC_TABLE * iadev->memSize;
1550        writew(((i>>3) & 0xfff8) | vcsize_sel, iadev->reass_reg+VC_LKUP_BASE);
1551        vc_table = (u16 *)(iadev->reass_ram+RX_VC_TABLE*iadev->memSize);  
1552         j = RX_VC_TABLE_SZ * iadev->memSize;
1553         for(i = 0; i < j; i++)  
1554         {  
1555                 /* shift the reassembly pointer by 3 + lower 3 bits of   
1556                 vc_lkup_base register (=3 for 1K VCs) and the last byte   
1557                 is those low 3 bits.   
1558                 Shall program this later.  
1559                 */  
1560                 *vc_table = (i << 6) | 15;      /* for invalid VCI */  
1561                 vc_table++;  
1562         }  
1563         /* ABR VC table */
1564         i =  ABR_VC_TABLE * iadev->memSize;
1565         writew(i >> 3, iadev->reass_reg+ABR_LKUP_BASE);
1566                    
1567         i = ABR_VC_TABLE * iadev->memSize;
1568         abr_vc_table = (struct abr_vc_table *)(iadev->reass_ram+i);  
1569         j = REASS_TABLE_SZ * iadev->memSize;
1570         memset ((char*)abr_vc_table, 0, j * sizeof(*abr_vc_table));
1571         for(i = 0; i < j; i++) {                
1572                 abr_vc_table->rdf = 0x0003;
1573                 abr_vc_table->air = 0x5eb1;
1574                 abr_vc_table++;         
1575         }  
1576
1577         /* Initialize other registers */  
1578   
1579         /* VP Filter Register set for VC Reassembly only */  
1580         writew(0xff00, iadev->reass_reg+VP_FILTER);  
1581         writew(0, iadev->reass_reg+XTRA_RM_OFFSET);
1582         writew(0x1,  iadev->reass_reg+PROTOCOL_ID);
1583
1584         /* Packet Timeout Count  related Registers : 
1585            Set packet timeout to occur in about 3 seconds
1586            Set Packet Aging Interval count register to overflow in about 4 us
1587         */  
1588         writew(0xF6F8, iadev->reass_reg+PKT_TM_CNT );
1589         ptr16 = (u16*)j;
1590         i = ((u32)ptr16 >> 6) & 0xff;
1591         ptr16  += j - 1;
1592         i |=(((u32)ptr16 << 2) & 0xff00);
1593         writew(i, iadev->reass_reg+TMOUT_RANGE);
1594         /* initiate the desc_tble */
1595         for(i=0; i<iadev->num_tx_desc;i++)
1596             iadev->desc_tbl[i].timestamp = 0;
1597
1598         /* to clear the interrupt status register - read it */  
1599         readw(iadev->reass_reg+REASS_INTR_STATUS_REG);   
1600   
1601         /* Mask Register - clear it */  
1602         writew(~(RX_FREEQ_EMPT|RX_PKT_RCVD), iadev->reass_reg+REASS_MASK_REG);  
1603   
1604         skb_queue_head_init(&iadev->rx_dma_q);  
1605         iadev->rx_free_desc_qhead = NULL;   
1606
1607         iadev->rx_open = kzalloc(4 * iadev->num_vc, GFP_KERNEL);
1608         if (!iadev->rx_open) {
1609                 printk(KERN_ERR DEV_LABEL "itf %d couldn't get free page\n",
1610                 dev->number);  
1611                 goto err_free_dle;
1612         }  
1613
1614         iadev->rxing = 1;
1615         iadev->rx_pkt_cnt = 0;
1616         /* Mode Register */  
1617         writew(R_ONLINE, iadev->reass_reg+MODE_REG);  
1618         return 0;  
1619
1620 err_free_dle:
1621         pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->rx_dle_q.start,
1622                             iadev->rx_dle_dma);  
1623 err_out:
1624         return -ENOMEM;
1625 }  
1626   
1627
1628 /*  
1629         The memory map suggested in appendix A and the coding for it.   
1630         Keeping it around just in case we change our mind later.  
1631   
1632                 Buffer descr    0x0000 (128 - 4K)  
1633                 UBR sched       0x1000 (1K - 4K)  
1634                 UBR Wait q      0x2000 (1K - 4K)  
1635                 Commn queues    0x3000 Packet Ready, Trasmit comp(0x3100)  
1636                                         (128 - 256) each  
1637                 extended VC     0x4000 (1K - 8K)  
1638                 ABR sched       0x6000  and ABR wait queue (1K - 2K) each  
1639                 CBR sched       0x7000 (as needed)  
1640                 VC table        0x8000 (1K - 32K)  
1641 */  
1642   
1643 static void tx_intr(struct atm_dev *dev)  
1644 {  
1645         IADEV *iadev;  
1646         unsigned short status;  
1647         unsigned long flags;
1648
1649         iadev = INPH_IA_DEV(dev);  
1650   
1651         status = readl(iadev->seg_reg+SEG_INTR_STATUS_REG);  
1652         if (status & TRANSMIT_DONE){
1653
1654            IF_EVENT(printk("Tansmit Done Intr logic run\n");)
1655            spin_lock_irqsave(&iadev->tx_lock, flags);
1656            ia_tx_poll(iadev);
1657            spin_unlock_irqrestore(&iadev->tx_lock, flags);
1658            writew(TRANSMIT_DONE, iadev->seg_reg+SEG_INTR_STATUS_REG);
1659            if (iadev->close_pending)  
1660                wake_up(&iadev->close_wait);
1661         }         
1662         if (status & TCQ_NOT_EMPTY)  
1663         {  
1664             IF_EVENT(printk("TCQ_NOT_EMPTY int received\n");)  
1665         }  
1666 }  
1667   
1668 static void tx_dle_intr(struct atm_dev *dev)
1669 {
1670         IADEV *iadev;
1671         struct dle *dle, *cur_dle; 
1672         struct sk_buff *skb;
1673         struct atm_vcc *vcc;
1674         struct ia_vcc  *iavcc;
1675         u_int dle_lp;
1676         unsigned long flags;
1677
1678         iadev = INPH_IA_DEV(dev);
1679         spin_lock_irqsave(&iadev->tx_lock, flags);   
1680         dle = iadev->tx_dle_q.read;
1681         dle_lp = readl(iadev->dma+IPHASE5575_TX_LIST_ADDR) & 
1682                                         (sizeof(struct dle)*DLE_ENTRIES - 1);
1683         cur_dle = (struct dle*)(iadev->tx_dle_q.start + (dle_lp >> 4));
1684         while (dle != cur_dle)
1685         {
1686             /* free the DMAed skb */ 
1687             skb = skb_dequeue(&iadev->tx_dma_q); 
1688             if (!skb) break;
1689
1690             /* Revenge of the 2 dle (skb + trailer) used in ia_pkt_tx() */
1691             if (!((dle - iadev->tx_dle_q.start)%(2*sizeof(struct dle)))) {
1692                 pci_unmap_single(iadev->pci, dle->sys_pkt_addr, skb->len,
1693                                  PCI_DMA_TODEVICE);
1694             }
1695             vcc = ATM_SKB(skb)->vcc;
1696             if (!vcc) {
1697                   printk("tx_dle_intr: vcc is null\n");
1698                   spin_unlock_irqrestore(&iadev->tx_lock, flags);
1699                   dev_kfree_skb_any(skb);
1700
1701                   return;
1702             }
1703             iavcc = INPH_IA_VCC(vcc);
1704             if (!iavcc) {
1705                   printk("tx_dle_intr: iavcc is null\n");
1706                   spin_unlock_irqrestore(&iadev->tx_lock, flags);
1707                   dev_kfree_skb_any(skb);
1708                   return;
1709             }
1710             if (vcc->qos.txtp.pcr >= iadev->rate_limit) {
1711                if ((vcc->pop) && (skb->len != 0))
1712                {     
1713                  vcc->pop(vcc, skb);
1714                } 
1715                else {
1716                  dev_kfree_skb_any(skb);
1717                }
1718             }
1719             else { /* Hold the rate-limited skb for flow control */
1720                IA_SKB_STATE(skb) |= IA_DLED;
1721                skb_queue_tail(&iavcc->txing_skb, skb);
1722             }
1723             IF_EVENT(printk("tx_dle_intr: enque skb = 0x%x \n", (u32)skb);)
1724             if (++dle == iadev->tx_dle_q.end)
1725                  dle = iadev->tx_dle_q.start;
1726         }
1727         iadev->tx_dle_q.read = dle;
1728         spin_unlock_irqrestore(&iadev->tx_lock, flags);
1729 }
1730   
1731 static int open_tx(struct atm_vcc *vcc)  
1732 {  
1733         struct ia_vcc *ia_vcc;  
1734         IADEV *iadev;  
1735         struct main_vc *vc;  
1736         struct ext_vc *evc;  
1737         int ret;
1738         IF_EVENT(printk("iadev: open_tx entered vcc->vci = %d\n", vcc->vci);)  
1739         if (vcc->qos.txtp.traffic_class == ATM_NONE) return 0;  
1740         iadev = INPH_IA_DEV(vcc->dev);  
1741         
1742         if (iadev->phy_type & FE_25MBIT_PHY) {
1743            if (vcc->qos.txtp.traffic_class == ATM_ABR) {
1744                printk("IA:  ABR not support\n");
1745                return -EINVAL; 
1746            }
1747           if (vcc->qos.txtp.traffic_class == ATM_CBR) {
1748                printk("IA:  CBR not support\n");
1749                return -EINVAL; 
1750           }
1751         }
1752         ia_vcc =  INPH_IA_VCC(vcc);
1753         memset((caddr_t)ia_vcc, 0, sizeof(*ia_vcc));
1754         if (vcc->qos.txtp.max_sdu > 
1755                          (iadev->tx_buf_sz - sizeof(struct cpcs_trailer))){
1756            printk("IA:  SDU size over (%d) the configured SDU size %d\n",
1757                   vcc->qos.txtp.max_sdu,iadev->tx_buf_sz);
1758            vcc->dev_data = NULL;
1759            kfree(ia_vcc);
1760            return -EINVAL; 
1761         }
1762         ia_vcc->vc_desc_cnt = 0;
1763         ia_vcc->txing = 1;
1764
1765         /* find pcr */
1766         if (vcc->qos.txtp.max_pcr == ATM_MAX_PCR) 
1767            vcc->qos.txtp.pcr = iadev->LineRate;
1768         else if ((vcc->qos.txtp.max_pcr == 0)&&( vcc->qos.txtp.pcr <= 0))
1769            vcc->qos.txtp.pcr = iadev->LineRate;
1770         else if ((vcc->qos.txtp.max_pcr > vcc->qos.txtp.pcr) && (vcc->qos.txtp.max_pcr> 0)) 
1771            vcc->qos.txtp.pcr = vcc->qos.txtp.max_pcr;
1772         if (vcc->qos.txtp.pcr > iadev->LineRate)
1773              vcc->qos.txtp.pcr = iadev->LineRate;
1774         ia_vcc->pcr = vcc->qos.txtp.pcr;
1775
1776         if (ia_vcc->pcr > (iadev->LineRate / 6) ) ia_vcc->ltimeout = HZ / 10;
1777         else if (ia_vcc->pcr > (iadev->LineRate / 130)) ia_vcc->ltimeout = HZ;
1778         else if (ia_vcc->pcr <= 170) ia_vcc->ltimeout = 16 * HZ;
1779         else ia_vcc->ltimeout = 2700 * HZ  / ia_vcc->pcr;
1780         if (ia_vcc->pcr < iadev->rate_limit)
1781            skb_queue_head_init (&ia_vcc->txing_skb);
1782         if (ia_vcc->pcr < iadev->rate_limit) {
1783            struct sock *sk = sk_atm(vcc);
1784
1785            if (vcc->qos.txtp.max_sdu != 0) {
1786                if (ia_vcc->pcr > 60000)
1787                   sk->sk_sndbuf = vcc->qos.txtp.max_sdu * 5;
1788                else if (ia_vcc->pcr > 2000)
1789                   sk->sk_sndbuf = vcc->qos.txtp.max_sdu * 4;
1790                else
1791                  sk->sk_sndbuf = vcc->qos.txtp.max_sdu * 3;
1792            }
1793            else
1794              sk->sk_sndbuf = 24576;
1795         }
1796            
1797         vc = (struct main_vc *)iadev->MAIN_VC_TABLE_ADDR;  
1798         evc = (struct ext_vc *)iadev->EXT_VC_TABLE_ADDR;  
1799         vc += vcc->vci;  
1800         evc += vcc->vci;  
1801         memset((caddr_t)vc, 0, sizeof(*vc));  
1802         memset((caddr_t)evc, 0, sizeof(*evc));  
1803           
1804         /* store the most significant 4 bits of vci as the last 4 bits   
1805                 of first part of atm header.  
1806            store the last 12 bits of vci as first 12 bits of the second  
1807                 part of the atm header.  
1808         */  
1809         evc->atm_hdr1 = (vcc->vci >> 12) & 0x000f;  
1810         evc->atm_hdr2 = (vcc->vci & 0x0fff) << 4;  
1811  
1812         /* check the following for different traffic classes */  
1813         if (vcc->qos.txtp.traffic_class == ATM_UBR)  
1814         {  
1815                 vc->type = UBR;  
1816                 vc->status = CRC_APPEND;
1817                 vc->acr = cellrate_to_float(iadev->LineRate);  
1818                 if (vcc->qos.txtp.pcr > 0) 
1819                    vc->acr = cellrate_to_float(vcc->qos.txtp.pcr);  
1820                 IF_UBR(printk("UBR: txtp.pcr = 0x%x f_rate = 0x%x\n", 
1821                                              vcc->qos.txtp.max_pcr,vc->acr);)
1822         }  
1823         else if (vcc->qos.txtp.traffic_class == ATM_ABR)  
1824         {       srv_cls_param_t srv_p;
1825                 IF_ABR(printk("Tx ABR VCC\n");)  
1826                 init_abr_vc(iadev, &srv_p);
1827                 if (vcc->qos.txtp.pcr > 0) 
1828                    srv_p.pcr = vcc->qos.txtp.pcr;
1829                 if (vcc->qos.txtp.min_pcr > 0) {
1830                    int tmpsum = iadev->sum_mcr+iadev->sum_cbr+vcc->qos.txtp.min_pcr;
1831                    if (tmpsum > iadev->LineRate)
1832                        return -EBUSY;
1833                    srv_p.mcr = vcc->qos.txtp.min_pcr;
1834                    iadev->sum_mcr += vcc->qos.txtp.min_pcr;
1835                 } 
1836                 else srv_p.mcr = 0;
1837                 if (vcc->qos.txtp.icr)
1838                    srv_p.icr = vcc->qos.txtp.icr;
1839                 if (vcc->qos.txtp.tbe)
1840                    srv_p.tbe = vcc->qos.txtp.tbe;
1841                 if (vcc->qos.txtp.frtt)
1842                    srv_p.frtt = vcc->qos.txtp.frtt;
1843                 if (vcc->qos.txtp.rif)
1844                    srv_p.rif = vcc->qos.txtp.rif;
1845                 if (vcc->qos.txtp.rdf)
1846                    srv_p.rdf = vcc->qos.txtp.rdf;
1847                 if (vcc->qos.txtp.nrm_pres)
1848                    srv_p.nrm = vcc->qos.txtp.nrm;
1849                 if (vcc->qos.txtp.trm_pres)
1850                    srv_p.trm = vcc->qos.txtp.trm;
1851                 if (vcc->qos.txtp.adtf_pres)
1852                    srv_p.adtf = vcc->qos.txtp.adtf;
1853                 if (vcc->qos.txtp.cdf_pres)
1854                    srv_p.cdf = vcc->qos.txtp.cdf;    
1855                 if (srv_p.icr > srv_p.pcr)
1856                    srv_p.icr = srv_p.pcr;    
1857                 IF_ABR(printk("ABR:vcc->qos.txtp.max_pcr = %d  mcr = %d\n", 
1858                                                       srv_p.pcr, srv_p.mcr);)
1859                 ia_open_abr_vc(iadev, &srv_p, vcc, 1);
1860         } else if (vcc->qos.txtp.traffic_class == ATM_CBR) {
1861                 if (iadev->phy_type & FE_25MBIT_PHY) {
1862                     printk("IA:  CBR not support\n");
1863                     return -EINVAL; 
1864                 }
1865                 if (vcc->qos.txtp.max_pcr > iadev->LineRate) {
1866                    IF_CBR(printk("PCR is not availble\n");)
1867                    return -1;
1868                 }
1869                 vc->type = CBR;
1870                 vc->status = CRC_APPEND;
1871                 if ((ret = ia_cbr_setup (iadev, vcc)) < 0) {     
1872                     return ret;
1873                 }
1874        } 
1875         else  
1876            printk("iadev:  Non UBR, ABR and CBR traffic not supportedn"); 
1877         
1878         iadev->testTable[vcc->vci]->vc_status |= VC_ACTIVE;
1879         IF_EVENT(printk("ia open_tx returning \n");)  
1880         return 0;  
1881 }  
1882   
1883   
1884 static int tx_init(struct atm_dev *dev)  
1885 {  
1886         IADEV *iadev;  
1887         struct tx_buf_desc *buf_desc_ptr;
1888         unsigned int tx_pkt_start;  
1889         void *dle_addr;  
1890         int i;  
1891         u_short tcq_st_adr;  
1892         u_short *tcq_start;  
1893         u_short prq_st_adr;  
1894         u_short *prq_start;  
1895         struct main_vc *vc;  
1896         struct ext_vc *evc;   
1897         u_short tmp16;
1898         u32 vcsize_sel;
1899  
1900         iadev = INPH_IA_DEV(dev);  
1901         spin_lock_init(&iadev->tx_lock);
1902  
1903         IF_INIT(printk("Tx MASK REG: 0x%0x\n", 
1904                                 readw(iadev->seg_reg+SEG_MASK_REG));)  
1905
1906         /* Allocate 4k (boundary aligned) bytes */
1907         dle_addr = pci_alloc_consistent(iadev->pci, DLE_TOTAL_SIZE,
1908                                         &iadev->tx_dle_dma);  
1909         if (!dle_addr)  {
1910                 printk(KERN_ERR DEV_LABEL "can't allocate DLEs\n");
1911                 goto err_out;
1912         }
1913         iadev->tx_dle_q.start = (struct dle*)dle_addr;  
1914         iadev->tx_dle_q.read = iadev->tx_dle_q.start;  
1915         iadev->tx_dle_q.write = iadev->tx_dle_q.start;  
1916         iadev->tx_dle_q.end = (struct dle*)((u32)dle_addr+sizeof(struct dle)*DLE_ENTRIES);  
1917
1918         /* write the upper 20 bits of the start address to tx list address register */  
1919         writel(iadev->tx_dle_dma & 0xfffff000,
1920                iadev->dma + IPHASE5575_TX_LIST_ADDR);  
1921         writew(0xffff, iadev->seg_reg+SEG_MASK_REG);  
1922         writew(0, iadev->seg_reg+MODE_REG_0);  
1923         writew(RESET_SEG, iadev->seg_reg+SEG_COMMAND_REG);  
1924         iadev->MAIN_VC_TABLE_ADDR = iadev->seg_ram+MAIN_VC_TABLE*iadev->memSize;
1925         iadev->EXT_VC_TABLE_ADDR = iadev->seg_ram+EXT_VC_TABLE*iadev->memSize;
1926         iadev->ABR_SCHED_TABLE_ADDR=iadev->seg_ram+ABR_SCHED_TABLE*iadev->memSize;
1927   
1928         /*  
1929            Transmit side control memory map  
1930            --------------------------------    
1931          Buffer descr   0x0000 (128 - 4K)  
1932          Commn queues   0x1000  Transmit comp, Packet ready(0x1400)   
1933                                         (512 - 1K) each  
1934                                         TCQ - 4K, PRQ - 5K  
1935          CBR Table      0x1800 (as needed) - 6K  
1936          UBR Table      0x3000 (1K - 4K) - 12K  
1937          UBR Wait queue 0x4000 (1K - 4K) - 16K  
1938          ABR sched      0x5000  and ABR wait queue (1K - 2K) each  
1939                                 ABR Tbl - 20K, ABR Wq - 22K   
1940          extended VC    0x6000 (1K - 8K) - 24K  
1941          VC Table       0x8000 (1K - 32K) - 32K  
1942           
1943         Between 0x2000 (8K) and 0x3000 (12K) there is 4K space left for VBR Tbl  
1944         and Wait q, which can be allotted later.  
1945         */  
1946      
1947         /* Buffer Descriptor Table Base address */  
1948         writew(TX_DESC_BASE, iadev->seg_reg+SEG_DESC_BASE);  
1949   
1950         /* initialize each entry in the buffer descriptor table */  
1951         buf_desc_ptr =(struct tx_buf_desc *)(iadev->seg_ram+TX_DESC_BASE);  
1952         memset((caddr_t)buf_desc_ptr, 0, sizeof(*buf_desc_ptr));  
1953         buf_desc_ptr++;  
1954         tx_pkt_start = TX_PACKET_RAM;  
1955         for(i=1; i<=iadev->num_tx_desc; i++)  
1956         {  
1957                 memset((caddr_t)buf_desc_ptr, 0, sizeof(*buf_desc_ptr));  
1958                 buf_desc_ptr->desc_mode = AAL5;  
1959                 buf_desc_ptr->buf_start_hi = tx_pkt_start >> 16;  
1960                 buf_desc_ptr->buf_start_lo = tx_pkt_start & 0x0000ffff;  
1961                 buf_desc_ptr++;           
1962                 tx_pkt_start += iadev->tx_buf_sz;  
1963         }  
1964         iadev->tx_buf = kmalloc(iadev->num_tx_desc*sizeof(struct cpcs_trailer_desc), GFP_KERNEL);
1965         if (!iadev->tx_buf) {
1966             printk(KERN_ERR DEV_LABEL " couldn't get mem\n");
1967             goto err_free_dle;
1968         }
1969         for (i= 0; i< iadev->num_tx_desc; i++)
1970         {
1971             struct cpcs_trailer *cpcs;
1972  
1973             cpcs = kmalloc(sizeof(*cpcs), GFP_KERNEL|GFP_DMA);
1974             if(!cpcs) {                
1975                 printk(KERN_ERR DEV_LABEL " couldn't get freepage\n"); 
1976                 goto err_free_tx_bufs;
1977             }
1978             iadev->tx_buf[i].cpcs = cpcs;
1979             iadev->tx_buf[i].dma_addr = pci_map_single(iadev->pci,
1980                 cpcs, sizeof(*cpcs), PCI_DMA_TODEVICE);
1981         }
1982         iadev->desc_tbl = kmalloc(iadev->num_tx_desc *
1983                                    sizeof(struct desc_tbl_t), GFP_KERNEL);
1984         if (!iadev->desc_tbl) {
1985                 printk(KERN_ERR DEV_LABEL " couldn't get mem\n");
1986                 goto err_free_all_tx_bufs;
1987         }
1988   
1989         /* Communication Queues base address */  
1990         i = TX_COMP_Q * iadev->memSize;
1991         writew(i >> 16, iadev->seg_reg+SEG_QUEUE_BASE);  
1992   
1993         /* Transmit Complete Queue */  
1994         writew(i, iadev->seg_reg+TCQ_ST_ADR);  
1995         writew(i, iadev->seg_reg+TCQ_RD_PTR);  
1996         writew(i+iadev->num_tx_desc*sizeof(u_short),iadev->seg_reg+TCQ_WR_PTR); 
1997         iadev->host_tcq_wr = i + iadev->num_tx_desc*sizeof(u_short);
1998         writew(i+2 * iadev->num_tx_desc * sizeof(u_short), 
1999                                               iadev->seg_reg+TCQ_ED_ADR); 
2000         /* Fill the TCQ with all the free descriptors. */  
2001         tcq_st_adr = readw(iadev->seg_reg+TCQ_ST_ADR);  
2002         tcq_start = (u_short *)(iadev->seg_ram+tcq_st_adr);  
2003         for(i=1; i<=iadev->num_tx_desc; i++)  
2004         {  
2005                 *tcq_start = (u_short)i;  
2006                 tcq_start++;  
2007         }  
2008   
2009         /* Packet Ready Queue */  
2010         i = PKT_RDY_Q * iadev->memSize; 
2011         writew(i, iadev->seg_reg+PRQ_ST_ADR);  
2012         writew(i+2 * iadev->num_tx_desc * sizeof(u_short), 
2013                                               iadev->seg_reg+PRQ_ED_ADR);
2014         writew(i, iadev->seg_reg+PRQ_RD_PTR);  
2015         writew(i, iadev->seg_reg+PRQ_WR_PTR);  
2016          
2017         /* Load local copy of PRQ and TCQ ptrs */
2018         iadev->ffL.prq_st = readw(iadev->seg_reg+PRQ_ST_ADR) & 0xffff;
2019         iadev->ffL.prq_ed = readw(iadev->seg_reg+PRQ_ED_ADR) & 0xffff;
2020         iadev->ffL.prq_wr = readw(iadev->seg_reg+PRQ_WR_PTR) & 0xffff;
2021
2022         iadev->ffL.tcq_st = readw(iadev->seg_reg+TCQ_ST_ADR) & 0xffff;
2023         iadev->ffL.tcq_ed = readw(iadev->seg_reg+TCQ_ED_ADR) & 0xffff;
2024         iadev->ffL.tcq_rd = readw(iadev->seg_reg+TCQ_RD_PTR) & 0xffff;
2025
2026         /* Just for safety initializing the queue to have desc 1 always */  
2027         /* Fill the PRQ with all the free descriptors. */  
2028         prq_st_adr = readw(iadev->seg_reg+PRQ_ST_ADR);  
2029         prq_start = (u_short *)(iadev->seg_ram+prq_st_adr);  
2030         for(i=1; i<=iadev->num_tx_desc; i++)  
2031         {  
2032                 *prq_start = (u_short)0;        /* desc 1 in all entries */  
2033                 prq_start++;  
2034         }  
2035         /* CBR Table */  
2036         IF_INIT(printk("Start CBR Init\n");)
2037 #if 1  /* for 1K VC board, CBR_PTR_BASE is 0 */
2038         writew(0,iadev->seg_reg+CBR_PTR_BASE);
2039 #else /* Charlie's logic is wrong ? */
2040         tmp16 = (iadev->seg_ram+CBR_SCHED_TABLE*iadev->memSize)>>17;
2041         IF_INIT(printk("cbr_ptr_base = 0x%x ", tmp16);)
2042         writew(tmp16,iadev->seg_reg+CBR_PTR_BASE);
2043 #endif
2044
2045         IF_INIT(printk("value in register = 0x%x\n",
2046                                    readw(iadev->seg_reg+CBR_PTR_BASE));)
2047         tmp16 = (CBR_SCHED_TABLE*iadev->memSize) >> 1;
2048         writew(tmp16, iadev->seg_reg+CBR_TAB_BEG);
2049         IF_INIT(printk("cbr_tab_beg = 0x%x in reg = 0x%x \n", tmp16,
2050                                         readw(iadev->seg_reg+CBR_TAB_BEG));)
2051         writew(tmp16, iadev->seg_reg+CBR_TAB_END+1); // CBR_PTR;
2052         tmp16 = (CBR_SCHED_TABLE*iadev->memSize + iadev->num_vc*6 - 2) >> 1;
2053         writew(tmp16, iadev->seg_reg+CBR_TAB_END);
2054         IF_INIT(printk("iadev->seg_reg = 0x%x CBR_PTR_BASE = 0x%x\n",
2055                (u32)iadev->seg_reg, readw(iadev->seg_reg+CBR_PTR_BASE));)
2056         IF_INIT(printk("CBR_TAB_BEG = 0x%x, CBR_TAB_END = 0x%x, CBR_PTR = 0x%x\n",
2057           readw(iadev->seg_reg+CBR_TAB_BEG), readw(iadev->seg_reg+CBR_TAB_END),
2058           readw(iadev->seg_reg+CBR_TAB_END+1));)
2059
2060         /* Initialize the CBR Schedualing Table */
2061         memset_io(iadev->seg_ram+CBR_SCHED_TABLE*iadev->memSize, 
2062                                                           0, iadev->num_vc*6); 
2063         iadev->CbrRemEntries = iadev->CbrTotEntries = iadev->num_vc*3;
2064         iadev->CbrEntryPt = 0;
2065         iadev->Granularity = MAX_ATM_155 / iadev->CbrTotEntries;
2066         iadev->NumEnabledCBR = 0;
2067
2068         /* UBR scheduling Table and wait queue */  
2069         /* initialize all bytes of UBR scheduler table and wait queue to 0   
2070                 - SCHEDSZ is 1K (# of entries).  
2071                 - UBR Table size is 4K  
2072                 - UBR wait queue is 4K  
2073            since the table and wait queues are contiguous, all the bytes   
2074            can be initialized by one memeset.  
2075         */  
2076         
2077         vcsize_sel = 0;
2078         i = 8*1024;
2079         while (i != iadev->num_vc) {
2080           i /= 2;
2081           vcsize_sel++;
2082         }
2083  
2084         i = MAIN_VC_TABLE * iadev->memSize;
2085         writew(vcsize_sel | ((i >> 8) & 0xfff8),iadev->seg_reg+VCT_BASE);
2086         i =  EXT_VC_TABLE * iadev->memSize;
2087         writew((i >> 8) & 0xfffe, iadev->seg_reg+VCTE_BASE);
2088         i = UBR_SCHED_TABLE * iadev->memSize;
2089         writew((i & 0xffff) >> 11,  iadev->seg_reg+UBR_SBPTR_BASE);
2090         i = UBR_WAIT_Q * iadev->memSize; 
2091         writew((i >> 7) & 0xffff,  iadev->seg_reg+UBRWQ_BASE);
2092         memset((caddr_t)(iadev->seg_ram+UBR_SCHED_TABLE*iadev->memSize),
2093                                                        0, iadev->num_vc*8);
2094         /* ABR scheduling Table(0x5000-0x57ff) and wait queue(0x5800-0x5fff)*/  
2095         /* initialize all bytes of ABR scheduler table and wait queue to 0   
2096                 - SCHEDSZ is 1K (# of entries).  
2097                 - ABR Table size is 2K  
2098                 - ABR wait queue is 2K  
2099            since the table and wait queues are contiguous, all the bytes   
2100            can be intialized by one memeset.  
2101         */  
2102         i = ABR_SCHED_TABLE * iadev->memSize;
2103         writew((i >> 11) & 0xffff, iadev->seg_reg+ABR_SBPTR_BASE);
2104         i = ABR_WAIT_Q * iadev->memSize;
2105         writew((i >> 7) & 0xffff, iadev->seg_reg+ABRWQ_BASE);
2106  
2107         i = ABR_SCHED_TABLE*iadev->memSize;
2108         memset((caddr_t)(iadev->seg_ram+i),  0, iadev->num_vc*4);
2109         vc = (struct main_vc *)iadev->MAIN_VC_TABLE_ADDR;  
2110         evc = (struct ext_vc *)iadev->EXT_VC_TABLE_ADDR;  
2111         iadev->testTable = kmalloc(sizeof(long)*iadev->num_vc, GFP_KERNEL); 
2112         if (!iadev->testTable) {
2113            printk("Get freepage  failed\n");
2114            goto err_free_desc_tbl;
2115         }
2116         for(i=0; i<iadev->num_vc; i++)  
2117         {  
2118                 memset((caddr_t)vc, 0, sizeof(*vc));  
2119                 memset((caddr_t)evc, 0, sizeof(*evc));  
2120                 iadev->testTable[i] = kmalloc(sizeof(struct testTable_t),
2121                                                 GFP_KERNEL);
2122                 if (!iadev->testTable[i])
2123                         goto err_free_test_tables;
2124                 iadev->testTable[i]->lastTime = 0;
2125                 iadev->testTable[i]->fract = 0;
2126                 iadev->testTable[i]->vc_status = VC_UBR;
2127                 vc++;  
2128                 evc++;  
2129         }  
2130   
2131         /* Other Initialization */  
2132           
2133         /* Max Rate Register */  
2134         if (iadev->phy_type & FE_25MBIT_PHY) {
2135            writew(RATE25, iadev->seg_reg+MAXRATE);  
2136            writew((UBR_EN | (0x23 << 2)), iadev->seg_reg+STPARMS);  
2137         }
2138         else {
2139            writew(cellrate_to_float(iadev->LineRate),iadev->seg_reg+MAXRATE);
2140            writew((UBR_EN | ABR_EN | (0x23 << 2)), iadev->seg_reg+STPARMS);  
2141         }
2142         /* Set Idle Header Reigisters to be sure */  
2143         writew(0, iadev->seg_reg+IDLEHEADHI);  
2144         writew(0, iadev->seg_reg+IDLEHEADLO);  
2145   
2146         /* Program ABR UBR Priority Register  as  PRI_ABR_UBR_EQUAL */
2147         writew(0xaa00, iadev->seg_reg+ABRUBR_ARB); 
2148
2149         iadev->close_pending = 0;
2150         init_waitqueue_head(&iadev->close_wait);
2151         init_waitqueue_head(&iadev->timeout_wait);
2152         skb_queue_head_init(&iadev->tx_dma_q);  
2153         ia_init_rtn_q(&iadev->tx_return_q);  
2154
2155         /* RM Cell Protocol ID and Message Type */  
2156         writew(RM_TYPE_4_0, iadev->seg_reg+RM_TYPE);  
2157         skb_queue_head_init (&iadev->tx_backlog);
2158   
2159         /* Mode Register 1 */  
2160         writew(MODE_REG_1_VAL, iadev->seg_reg+MODE_REG_1);  
2161   
2162         /* Mode Register 0 */  
2163         writew(T_ONLINE, iadev->seg_reg+MODE_REG_0);  
2164   
2165         /* Interrupt Status Register - read to clear */  
2166         readw(iadev->seg_reg+SEG_INTR_STATUS_REG);  
2167   
2168         /* Interrupt Mask Reg- don't mask TCQ_NOT_EMPTY interrupt generation */  
2169         writew(~(TRANSMIT_DONE | TCQ_NOT_EMPTY), iadev->seg_reg+SEG_MASK_REG);
2170         writew(TRANSMIT_DONE, iadev->seg_reg+SEG_INTR_STATUS_REG);  
2171         iadev->tx_pkt_cnt = 0;
2172         iadev->rate_limit = iadev->LineRate / 3;
2173   
2174         return 0;
2175
2176 err_free_test_tables:
2177         while (--i >= 0)
2178                 kfree(iadev->testTable[i]);
2179         kfree(iadev->testTable);
2180 err_free_desc_tbl:
2181         kfree(iadev->desc_tbl);
2182 err_free_all_tx_bufs:
2183         i = iadev->num_tx_desc;
2184 err_free_tx_bufs:
2185         while (--i >= 0) {
2186                 struct cpcs_trailer_desc *desc = iadev->tx_buf + i;
2187
2188                 pci_unmap_single(iadev->pci, desc->dma_addr,
2189                         sizeof(*desc->cpcs), PCI_DMA_TODEVICE);
2190                 kfree(desc->cpcs);
2191         }
2192         kfree(iadev->tx_buf);
2193 err_free_dle:
2194         pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->tx_dle_q.start,
2195                             iadev->tx_dle_dma);  
2196 err_out:
2197         return -ENOMEM;
2198 }   
2199    
2200 static irqreturn_t ia_int(int irq, void *dev_id)  
2201 {  
2202    struct atm_dev *dev;  
2203    IADEV *iadev;  
2204    unsigned int status;  
2205    int handled = 0;
2206
2207    dev = dev_id;  
2208    iadev = INPH_IA_DEV(dev);  
2209    while( (status = readl(iadev->reg+IPHASE5575_BUS_STATUS_REG) & 0x7f))  
2210    { 
2211         handled = 1;
2212         IF_EVENT(printk("ia_int: status = 0x%x\n", status);) 
2213         if (status & STAT_REASSINT)  
2214         {  
2215            /* do something */  
2216            IF_EVENT(printk("REASSINT Bus status reg: %08x\n", status);) 
2217            rx_intr(dev);  
2218         }  
2219         if (status & STAT_DLERINT)  
2220         {  
2221            /* Clear this bit by writing a 1 to it. */  
2222            *(u_int *)(iadev->reg+IPHASE5575_BUS_STATUS_REG) = STAT_DLERINT;
2223            rx_dle_intr(dev);  
2224         }  
2225         if (status & STAT_SEGINT)  
2226         {  
2227            /* do something */ 
2228            IF_EVENT(printk("IA: tx_intr \n");) 
2229            tx_intr(dev);  
2230         }  
2231         if (status & STAT_DLETINT)  
2232         {  
2233            *(u_int *)(iadev->reg+IPHASE5575_BUS_STATUS_REG) = STAT_DLETINT;  
2234            tx_dle_intr(dev);  
2235         }  
2236         if (status & (STAT_FEINT | STAT_ERRINT | STAT_MARKINT))  
2237         {  
2238            if (status & STAT_FEINT) 
2239                IaFrontEndIntr(iadev);
2240         }  
2241    }
2242    return IRQ_RETVAL(handled);
2243 }  
2244           
2245           
2246           
2247 /*----------------------------- entries --------------------------------*/  
2248 static int get_esi(struct atm_dev *dev)  
2249 {  
2250         IADEV *iadev;  
2251         int i;  
2252         u32 mac1;  
2253         u16 mac2;  
2254           
2255         iadev = INPH_IA_DEV(dev);  
2256         mac1 = cpu_to_be32(le32_to_cpu(readl(  
2257                                 iadev->reg+IPHASE5575_MAC1)));  
2258         mac2 = cpu_to_be16(le16_to_cpu(readl(iadev->reg+IPHASE5575_MAC2)));  
2259         IF_INIT(printk("ESI: 0x%08x%04x\n", mac1, mac2);)  
2260         for (i=0; i<MAC1_LEN; i++)  
2261                 dev->esi[i] = mac1 >>(8*(MAC1_LEN-1-i));  
2262           
2263         for (i=0; i<MAC2_LEN; i++)  
2264                 dev->esi[i+MAC1_LEN] = mac2 >>(8*(MAC2_LEN - 1 -i));  
2265         return 0;  
2266 }  
2267           
2268 static int reset_sar(struct atm_dev *dev)  
2269 {  
2270         IADEV *iadev;  
2271         int i, error = 1;  
2272         unsigned int pci[64];  
2273           
2274         iadev = INPH_IA_DEV(dev);  
2275         for(i=0; i<64; i++)  
2276           if ((error = pci_read_config_dword(iadev->pci,  
2277                                 i*4, &pci[i])) != PCIBIOS_SUCCESSFUL)  
2278               return error;  
2279         writel(0, iadev->reg+IPHASE5575_EXT_RESET);  
2280         for(i=0; i<64; i++)  
2281           if ((error = pci_write_config_dword(iadev->pci,  
2282                                         i*4, pci[i])) != PCIBIOS_SUCCESSFUL)  
2283             return error;  
2284         udelay(5);  
2285         return 0;  
2286 }  
2287           
2288           
2289 static int __devinit ia_init(struct atm_dev *dev)
2290 {  
2291         IADEV *iadev;  
2292         unsigned long real_base;
2293         void __iomem *base;
2294         unsigned short command;  
2295         int error, i; 
2296           
2297         /* The device has been identified and registered. Now we read   
2298            necessary configuration info like memory base address,   
2299            interrupt number etc */  
2300           
2301         IF_INIT(printk(">ia_init\n");)  
2302         dev->ci_range.vpi_bits = 0;  
2303         dev->ci_range.vci_bits = NR_VCI_LD;  
2304
2305         iadev = INPH_IA_DEV(dev);  
2306         real_base = pci_resource_start (iadev->pci, 0);
2307         iadev->irq = iadev->pci->irq;
2308                   
2309         error = pci_read_config_word(iadev->pci, PCI_COMMAND, &command);
2310         if (error) {
2311                 printk(KERN_ERR DEV_LABEL "(itf %d): init error 0x%x\n",  
2312                                 dev->number,error);  
2313                 return -EINVAL;  
2314         }  
2315         IF_INIT(printk(DEV_LABEL "(itf %d): rev.%d,realbase=0x%lx,irq=%d\n",  
2316                         dev->number, iadev->pci->revision, real_base, iadev->irq);)
2317           
2318         /* find mapping size of board */  
2319           
2320         iadev->pci_map_size = pci_resource_len(iadev->pci, 0);
2321
2322         if (iadev->pci_map_size == 0x100000){
2323           iadev->num_vc = 4096;
2324           dev->ci_range.vci_bits = NR_VCI_4K_LD;  
2325           iadev->memSize = 4;
2326         }
2327         else if (iadev->pci_map_size == 0x40000) {
2328           iadev->num_vc = 1024;
2329           iadev->memSize = 1;
2330         }
2331         else {
2332            printk("Unknown pci_map_size = 0x%x\n", iadev->pci_map_size);
2333            return -EINVAL;
2334         }
2335         IF_INIT(printk (DEV_LABEL "map size: %i\n", iadev->pci_map_size);)  
2336           
2337         /* enable bus mastering */
2338         pci_set_master(iadev->pci);
2339
2340         /*  
2341          * Delay at least 1us before doing any mem accesses (how 'bout 10?)  
2342          */  
2343         udelay(10);  
2344           
2345         /* mapping the physical address to a virtual address in address space */  
2346         base = ioremap(real_base,iadev->pci_map_size);  /* ioremap is not resolved ??? */  
2347           
2348         if (!base)  
2349         {  
2350                 printk(DEV_LABEL " (itf %d): can't set up page mapping\n",  
2351                             dev->number);  
2352                 return error;  
2353         }  
2354         IF_INIT(printk(DEV_LABEL " (itf %d): rev.%d,base=%p,irq=%d\n",  
2355                         dev->number, iadev->pci->revision, base, iadev->irq);)
2356           
2357         /* filling the iphase dev structure */  
2358         iadev->mem = iadev->pci_map_size /2;  
2359         iadev->real_base = real_base;  
2360         iadev->base = base;  
2361                   
2362         /* Bus Interface Control Registers */  
2363         iadev->reg = base + REG_BASE;
2364         /* Segmentation Control Registers */  
2365         iadev->seg_reg = base + SEG_BASE;
2366         /* Reassembly Control Registers */  
2367         iadev->reass_reg = base + REASS_BASE;  
2368         /* Front end/ DMA control registers */  
2369         iadev->phy = base + PHY_BASE;  
2370         iadev->dma = base + PHY_BASE;  
2371         /* RAM - Segmentation RAm and Reassembly RAM */  
2372         iadev->ram = base + ACTUAL_RAM_BASE;  
2373         iadev->seg_ram = base + ACTUAL_SEG_RAM_BASE;  
2374         iadev->reass_ram = base + ACTUAL_REASS_RAM_BASE;  
2375   
2376         /* lets print out the above */  
2377         IF_INIT(printk("Base addrs: %p %p %p \n %p %p %p %p\n", 
2378           iadev->reg,iadev->seg_reg,iadev->reass_reg, 
2379           iadev->phy, iadev->ram, iadev->seg_ram, 
2380           iadev->reass_ram);) 
2381           
2382         /* lets try reading the MAC address */  
2383         error = get_esi(dev);  
2384         if (error) {
2385           iounmap(iadev->base);
2386           return error;  
2387         }
2388         printk("IA: ");
2389         for (i=0; i < ESI_LEN; i++)  
2390                 printk("%s%02X",i ? "-" : "",dev->esi[i]);  
2391         printk("\n");  
2392   
2393         /* reset SAR */  
2394         if (reset_sar(dev)) {
2395            iounmap(iadev->base);
2396            printk("IA: reset SAR fail, please try again\n");
2397            return 1;
2398         }
2399         return 0;  
2400 }  
2401
2402 static void ia_update_stats(IADEV *iadev) {
2403     if (!iadev->carrier_detect)
2404         return;
2405     iadev->rx_cell_cnt += readw(iadev->reass_reg+CELL_CTR0)&0xffff;
2406     iadev->rx_cell_cnt += (readw(iadev->reass_reg+CELL_CTR1) & 0xffff) << 16;
2407     iadev->drop_rxpkt +=  readw(iadev->reass_reg + DRP_PKT_CNTR ) & 0xffff;
2408     iadev->drop_rxcell += readw(iadev->reass_reg + ERR_CNTR) & 0xffff;
2409     iadev->tx_cell_cnt += readw(iadev->seg_reg + CELL_CTR_LO_AUTO)&0xffff;
2410     iadev->tx_cell_cnt += (readw(iadev->seg_reg+CELL_CTR_HIGH_AUTO)&0xffff)<<16;
2411     return;
2412 }
2413   
2414 static void ia_led_timer(unsigned long arg) {
2415         unsigned long flags;
2416         static u_char blinking[8] = {0, 0, 0, 0, 0, 0, 0, 0};
2417         u_char i;
2418         static u32 ctrl_reg; 
2419         for (i = 0; i < iadev_count; i++) {
2420            if (ia_dev[i]) {
2421               ctrl_reg = readl(ia_dev[i]->reg+IPHASE5575_BUS_CONTROL_REG);
2422               if (blinking[i] == 0) {
2423                  blinking[i]++;
2424                  ctrl_reg &= (~CTRL_LED);
2425                  writel(ctrl_reg, ia_dev[i]->reg+IPHASE5575_BUS_CONTROL_REG);
2426                  ia_update_stats(ia_dev[i]);
2427               }
2428               else {
2429                  blinking[i] = 0;
2430                  ctrl_reg |= CTRL_LED;
2431                  writel(ctrl_reg, ia_dev[i]->reg+IPHASE5575_BUS_CONTROL_REG);
2432                  spin_lock_irqsave(&ia_dev[i]->tx_lock, flags);
2433                  if (ia_dev[i]->close_pending)  
2434                     wake_up(&ia_dev[i]->close_wait);
2435                  ia_tx_poll(ia_dev[i]);
2436                  spin_unlock_irqrestore(&ia_dev[i]->tx_lock, flags);
2437               }
2438            }
2439         }
2440         mod_timer(&ia_timer, jiffies + HZ / 4);
2441         return;
2442 }
2443
2444 static void ia_phy_put(struct atm_dev *dev, unsigned char value,   
2445         unsigned long addr)  
2446 {  
2447         writel(value, INPH_IA_DEV(dev)->phy+addr);  
2448 }  
2449   
2450 static unsigned char ia_phy_get(struct atm_dev *dev, unsigned long addr)  
2451 {  
2452         return readl(INPH_IA_DEV(dev)->phy+addr);  
2453 }  
2454
2455 static void ia_free_tx(IADEV *iadev)
2456 {
2457         int i;
2458
2459         kfree(iadev->desc_tbl);
2460         for (i = 0; i < iadev->num_vc; i++)
2461                 kfree(iadev->testTable[i]);
2462         kfree(iadev->testTable);
2463         for (i = 0; i < iadev->num_tx_desc; i++) {
2464                 struct cpcs_trailer_desc *desc = iadev->tx_buf + i;
2465
2466                 pci_unmap_single(iadev->pci, desc->dma_addr,
2467                         sizeof(*desc->cpcs), PCI_DMA_TODEVICE);
2468                 kfree(desc->cpcs);
2469         }
2470         kfree(iadev->tx_buf);
2471         pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->tx_dle_q.start,
2472                             iadev->tx_dle_dma);  
2473 }
2474
2475 static void ia_free_rx(IADEV *iadev)
2476 {
2477         kfree(iadev->rx_open);
2478         pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->rx_dle_q.start,
2479                           iadev->rx_dle_dma);  
2480 }
2481
2482 static int __devinit ia_start(struct atm_dev *dev)
2483 {  
2484         IADEV *iadev;  
2485         int error;  
2486         unsigned char phy;  
2487         u32 ctrl_reg;  
2488         IF_EVENT(printk(">ia_start\n");)  
2489         iadev = INPH_IA_DEV(dev);  
2490         if (request_irq(iadev->irq, &ia_int, IRQF_SHARED, DEV_LABEL, dev)) {
2491                 printk(KERN_ERR DEV_LABEL "(itf %d): IRQ%d is already in use\n",  
2492                     dev->number, iadev->irq);  
2493                 error = -EAGAIN;
2494                 goto err_out;
2495         }  
2496         /* @@@ should release IRQ on error */  
2497         /* enabling memory + master */  
2498         if ((error = pci_write_config_word(iadev->pci,   
2499                                 PCI_COMMAND,   
2500                                 PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER )))   
2501         {  
2502                 printk(KERN_ERR DEV_LABEL "(itf %d): can't enable memory+"  
2503                     "master (0x%x)\n",dev->number, error);  
2504                 error = -EIO;  
2505                 goto err_free_irq;
2506         }  
2507         udelay(10);  
2508   
2509         /* Maybe we should reset the front end, initialize Bus Interface Control   
2510                 Registers and see. */  
2511   
2512         IF_INIT(printk("Bus ctrl reg: %08x\n", 
2513                             readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG));)  
2514         ctrl_reg = readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG);  
2515         ctrl_reg = (ctrl_reg & (CTRL_LED | CTRL_FE_RST))  
2516                         | CTRL_B8  
2517                         | CTRL_B16  
2518                         | CTRL_B32  
2519                         | CTRL_B48  
2520                         | CTRL_B64  
2521                         | CTRL_B128  
2522                         | CTRL_ERRMASK  
2523                         | CTRL_DLETMASK         /* shud be removed l8r */  
2524                         | CTRL_DLERMASK  
2525                         | CTRL_SEGMASK  
2526                         | CTRL_REASSMASK          
2527                         | CTRL_FEMASK  
2528                         | CTRL_CSPREEMPT;  
2529   
2530        writel(ctrl_reg, iadev->reg+IPHASE5575_BUS_CONTROL_REG);   
2531   
2532         IF_INIT(printk("Bus ctrl reg after initializing: %08x\n", 
2533                            readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG));  
2534            printk("Bus status reg after init: %08x\n", 
2535                             readl(iadev->reg+IPHASE5575_BUS_STATUS_REG));)  
2536     
2537         ia_hw_type(iadev); 
2538         error = tx_init(dev);  
2539         if (error)
2540                 goto err_free_irq;
2541         error = rx_init(dev);  
2542         if (error)
2543                 goto err_free_tx;
2544   
2545         ctrl_reg = readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG);  
2546         writel(ctrl_reg | CTRL_FE_RST, iadev->reg+IPHASE5575_BUS_CONTROL_REG);   
2547         IF_INIT(printk("Bus ctrl reg after initializing: %08x\n", 
2548                                readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG));)  
2549         phy = 0; /* resolve compiler complaint */
2550         IF_INIT ( 
2551         if ((phy=ia_phy_get(dev,0)) == 0x30)  
2552                 printk("IA: pm5346,rev.%d\n",phy&0x0f);  
2553         else  
2554                 printk("IA: utopia,rev.%0x\n",phy);) 
2555
2556         if (iadev->phy_type &  FE_25MBIT_PHY)
2557            ia_mb25_init(iadev);
2558         else if (iadev->phy_type & (FE_DS3_PHY | FE_E3_PHY))
2559            ia_suni_pm7345_init(iadev);
2560         else {
2561                 error = suni_init(dev);
2562                 if (error)
2563                         goto err_free_rx;
2564                 /* 
2565                  * Enable interrupt on loss of signal
2566                  * SUNI_RSOP_CIE - 0x10
2567                  * SUNI_RSOP_CIE_LOSE - 0x04
2568                  */
2569                 ia_phy_put(dev, ia_phy_get(dev, 0x10) | 0x04, 0x10);
2570 #ifndef MODULE
2571                 error = dev->phy->start(dev);
2572                 if (error)
2573                         goto err_free_rx;
2574 #endif
2575                 /* Get iadev->carrier_detect status */
2576                 IaFrontEndIntr(iadev);
2577         }
2578         return 0;
2579
2580 err_free_rx:
2581         ia_free_rx(iadev);
2582 err_free_tx:
2583         ia_free_tx(iadev);
2584 err_free_irq:
2585         free_irq(iadev->irq, dev);  
2586 err_out:
2587         return error;
2588 }  
2589   
2590 static void ia_close(struct atm_vcc *vcc)  
2591 {
2592         DEFINE_WAIT(wait);
2593         u16 *vc_table;
2594         IADEV *iadev;
2595         struct ia_vcc *ia_vcc;
2596         struct sk_buff *skb = NULL;
2597         struct sk_buff_head tmp_tx_backlog, tmp_vcc_backlog;
2598         unsigned long closetime, flags;
2599
2600         iadev = INPH_IA_DEV(vcc->dev);
2601         ia_vcc = INPH_IA_VCC(vcc);
2602         if (!ia_vcc) return;  
2603
2604         IF_EVENT(printk("ia_close: ia_vcc->vc_desc_cnt = %d  vci = %d\n", 
2605                                               ia_vcc->vc_desc_cnt,vcc->vci);)
2606         clear_bit(ATM_VF_READY,&vcc->flags);
2607         skb_queue_head_init (&tmp_tx_backlog);
2608         skb_queue_head_init (&tmp_vcc_backlog); 
2609         if (vcc->qos.txtp.traffic_class != ATM_NONE) {
2610            iadev->close_pending++;
2611            prepare_to_wait(&iadev->timeout_wait, &wait, TASK_UNINTERRUPTIBLE);
2612            schedule_timeout(50);
2613            finish_wait(&iadev->timeout_wait, &wait);
2614            spin_lock_irqsave(&iadev->tx_lock, flags); 
2615            while((skb = skb_dequeue(&iadev->tx_backlog))) {
2616               if (ATM_SKB(skb)->vcc == vcc){ 
2617                  if (vcc->pop) vcc->pop(vcc, skb);
2618                  else dev_kfree_skb_any(skb);
2619               }
2620               else 
2621                  skb_queue_tail(&tmp_tx_backlog, skb);
2622            } 
2623            while((skb = skb_dequeue(&tmp_tx_backlog))) 
2624              skb_queue_tail(&iadev->tx_backlog, skb);
2625            IF_EVENT(printk("IA TX Done decs_cnt = %d\n", ia_vcc->vc_desc_cnt);) 
2626            closetime = 300000 / ia_vcc->pcr;
2627            if (closetime == 0)
2628               closetime = 1;
2629            spin_unlock_irqrestore(&iadev->tx_lock, flags);
2630            wait_event_timeout(iadev->close_wait, (ia_vcc->vc_desc_cnt <= 0), closetime);
2631            spin_lock_irqsave(&iadev->tx_lock, flags);
2632            iadev->close_pending--;
2633            iadev->testTable[vcc->vci]->lastTime = 0;
2634            iadev->testTable[vcc->vci]->fract = 0; 
2635            iadev->testTable[vcc->vci]->vc_status = VC_UBR; 
2636            if (vcc->qos.txtp.traffic_class == ATM_ABR) {
2637               if (vcc->qos.txtp.min_pcr > 0)
2638                  iadev->sum_mcr -= vcc->qos.txtp.min_pcr;
2639            }
2640            if (vcc->qos.txtp.traffic_class == ATM_CBR) {
2641               ia_vcc = INPH_IA_VCC(vcc); 
2642               iadev->sum_mcr -= ia_vcc->NumCbrEntry*iadev->Granularity;
2643               ia_cbrVc_close (vcc);
2644            }
2645            spin_unlock_irqrestore(&iadev->tx_lock, flags);
2646         }
2647         
2648         if (vcc->qos.rxtp.traffic_class != ATM_NONE) {   
2649            // reset reass table
2650            vc_table = (u16 *)(iadev->reass_ram+REASS_TABLE*iadev->memSize);
2651            vc_table += vcc->vci; 
2652            *vc_table = NO_AAL5_PKT;
2653            // reset vc table
2654            vc_table = (u16 *)(iadev->reass_ram+RX_VC_TABLE*iadev->memSize);
2655            vc_table += vcc->vci;
2656            *vc_table = (vcc->vci << 6) | 15;
2657            if (vcc->qos.rxtp.traffic_class == ATM_ABR) {
2658               struct abr_vc_table __iomem *abr_vc_table = 
2659                                 (iadev->reass_ram+ABR_VC_TABLE*iadev->memSize);
2660               abr_vc_table +=  vcc->vci;
2661               abr_vc_table->rdf = 0x0003;
2662               abr_vc_table->air = 0x5eb1;
2663            }                                 
2664            // Drain the packets
2665            rx_dle_intr(vcc->dev); 
2666            iadev->rx_open[vcc->vci] = NULL;
2667         }
2668         kfree(INPH_IA_VCC(vcc));  
2669         ia_vcc = NULL;
2670         vcc->dev_data = NULL;
2671         clear_bit(ATM_VF_ADDR,&vcc->flags);
2672         return;        
2673 }  
2674   
2675 static int ia_open(struct atm_vcc *vcc)
2676 {  
2677         IADEV *iadev;  
2678         struct ia_vcc *ia_vcc;  
2679         int error;  
2680         if (!test_bit(ATM_VF_PARTIAL,&vcc->flags))  
2681         {  
2682                 IF_EVENT(printk("ia: not partially allocated resources\n");)  
2683                 vcc->dev_data = NULL;
2684         }  
2685         iadev = INPH_IA_DEV(vcc->dev);  
2686         if (vcc->vci != ATM_VPI_UNSPEC && vcc->vpi != ATM_VCI_UNSPEC)  
2687         {  
2688                 IF_EVENT(printk("iphase open: unspec part\n");)  
2689                 set_bit(ATM_VF_ADDR,&vcc->flags);
2690         }  
2691         if (vcc->qos.aal != ATM_AAL5)  
2692                 return -EINVAL;  
2693         IF_EVENT(printk(DEV_LABEL "(itf %d): open %d.%d\n", 
2694                                  vcc->dev->number, vcc->vpi, vcc->vci);)  
2695   
2696         /* Device dependent initialization */  
2697         ia_vcc = kmalloc(sizeof(*ia_vcc), GFP_KERNEL);  
2698         if (!ia_vcc) return -ENOMEM;  
2699         vcc->dev_data = ia_vcc;
2700   
2701         if ((error = open_rx(vcc)))  
2702         {  
2703                 IF_EVENT(printk("iadev: error in open_rx, closing\n");)  
2704                 ia_close(vcc);  
2705                 return error;  
2706         }  
2707   
2708         if ((error = open_tx(vcc)))  
2709         {  
2710                 IF_EVENT(printk("iadev: error in open_tx, closing\n");)  
2711                 ia_close(vcc);  
2712                 return error;  
2713         }  
2714   
2715         set_bit(ATM_VF_READY,&vcc->flags);
2716
2717 #if 0
2718         {
2719            static u8 first = 1; 
2720            if (first) {
2721               ia_timer.expires = jiffies + 3*HZ;
2722               add_timer(&ia_timer);
2723               first = 0;
2724            }           
2725         }
2726 #endif
2727         IF_EVENT(printk("ia open returning\n");)  
2728         return 0;  
2729 }  
2730   
2731 static int ia_change_qos(struct atm_vcc *vcc, struct atm_qos *qos, int flags)  
2732 {  
2733         IF_EVENT(printk(">ia_change_qos\n");)  
2734         return 0;  
2735 }  
2736   
2737 static int ia_ioctl(struct atm_dev *dev, unsigned int cmd, void __user *arg)  
2738 {  
2739    IA_CMDBUF ia_cmds;
2740    IADEV *iadev;
2741    int i, board;
2742    u16 __user *tmps;
2743    IF_EVENT(printk(">ia_ioctl\n");)  
2744    if (cmd != IA_CMD) {
2745       if (!dev->phy->ioctl) return -EINVAL;
2746       return dev->phy->ioctl(dev,cmd,arg);
2747    }
2748    if (copy_from_user(&ia_cmds, arg, sizeof ia_cmds)) return -EFAULT; 
2749    board = ia_cmds.status;
2750    if ((board < 0) || (board > iadev_count))
2751          board = 0;    
2752    iadev = ia_dev[board];
2753    switch (ia_cmds.cmd) {
2754    case MEMDUMP:
2755    {
2756         switch (ia_cmds.sub_cmd) {
2757           case MEMDUMP_DEV:     
2758              if (!capable(CAP_NET_ADMIN)) return -EPERM;
2759              if (copy_to_user(ia_cmds.buf, iadev, sizeof(IADEV)))
2760                 return -EFAULT;
2761              ia_cmds.status = 0;
2762              break;
2763           case MEMDUMP_SEGREG:
2764              if (!capable(CAP_NET_ADMIN)) return -EPERM;
2765              tmps = (u16 __user *)ia_cmds.buf;
2766              for(i=0; i<0x80; i+=2, tmps++)
2767                 if(put_user((u16)(readl(iadev->seg_reg+i) & 0xffff), tmps)) return -EFAULT;
2768              ia_cmds.status = 0;
2769              ia_cmds.len = 0x80;
2770              break;
2771           case MEMDUMP_REASSREG:
2772              if (!capable(CAP_NET_ADMIN)) return -EPERM;
2773              tmps = (u16 __user *)ia_cmds.buf;
2774              for(i=0; i<0x80; i+=2, tmps++)
2775                 if(put_user((u16)(readl(iadev->reass_reg+i) & 0xffff), tmps)) return -EFAULT;
2776              ia_cmds.status = 0;
2777              ia_cmds.len = 0x80;
2778              break;
2779           case MEMDUMP_FFL:
2780           {  
2781              ia_regs_t       *regs_local;
2782              ffredn_t        *ffL;
2783              rfredn_t        *rfL;
2784                      
2785              if (!capable(CAP_NET_ADMIN)) return -EPERM;
2786              regs_local = kmalloc(sizeof(*regs_local), GFP_KERNEL);
2787              if (!regs_local) return -ENOMEM;
2788              ffL = &regs_local->ffredn;
2789              rfL = &regs_local->rfredn;
2790              /* Copy real rfred registers into the local copy */
2791              for (i=0; i<(sizeof (rfredn_t))/4; i++)
2792                 ((u_int *)rfL)[i] = readl(iadev->reass_reg + i) & 0xffff;
2793                 /* Copy real ffred registers into the local copy */
2794              for (i=0; i<(sizeof (ffredn_t))/4; i++)
2795                 ((u_int *)ffL)[i] = readl(iadev->seg_reg + i) & 0xffff;
2796
2797              if (copy_to_user(ia_cmds.buf, regs_local,sizeof(ia_regs_t))) {
2798                 kfree(regs_local);
2799                 return -EFAULT;
2800              }
2801              kfree(regs_local);
2802              printk("Board %d registers dumped\n", board);
2803              ia_cmds.status = 0;                  
2804          }      
2805              break;        
2806          case READ_REG:
2807          {  
2808              if (!capable(CAP_NET_ADMIN)) return -EPERM;
2809              desc_dbg(iadev); 
2810              ia_cmds.status = 0; 
2811          }
2812              break;
2813          case 0x6:
2814          {  
2815              ia_cmds.status = 0; 
2816              printk("skb = 0x%lx\n", (long)skb_peek(&iadev->tx_backlog));
2817              printk("rtn_q: 0x%lx\n",(long)ia_deque_rtn_q(&iadev->tx_return_q));
2818          }
2819              break;
2820          case 0x8:
2821          {
2822              struct k_sonet_stats *stats;
2823              stats = &PRIV(_ia_dev[board])->sonet_stats;
2824              printk("section_bip: %d\n", atomic_read(&stats->section_bip));
2825              printk("line_bip   : %d\n", atomic_read(&stats->line_bip));
2826              printk("path_bip   : %d\n", atomic_read(&stats->path_bip));
2827              printk("line_febe  : %d\n", atomic_read(&stats->line_febe));
2828              printk("path_febe  : %d\n", atomic_read(&stats->path_febe));
2829              printk("corr_hcs   : %d\n", atomic_read(&stats->corr_hcs));
2830              printk("uncorr_hcs : %d\n", atomic_read(&stats->uncorr_hcs));
2831              printk("tx_cells   : %d\n", atomic_read(&stats->tx_cells));
2832              printk("rx_cells   : %d\n", atomic_read(&stats->rx_cells));
2833          }
2834             ia_cmds.status = 0;
2835             break;
2836          case 0x9:
2837             if (!capable(CAP_NET_ADMIN)) return -EPERM;
2838             for (i = 1; i <= iadev->num_rx_desc; i++)
2839                free_desc(_ia_dev[board], i);
2840             writew( ~(RX_FREEQ_EMPT | RX_EXCP_RCVD), 
2841                                             iadev->reass_reg+REASS_MASK_REG);
2842             iadev->rxing = 1;
2843             
2844             ia_cmds.status = 0;
2845             break;
2846
2847          case 0xb:
2848             if (!capable(CAP_NET_ADMIN)) return -EPERM;
2849             IaFrontEndIntr(iadev);
2850             break;
2851          case 0xa:
2852             if (!capable(CAP_NET_ADMIN)) return -EPERM;
2853          {  
2854              ia_cmds.status = 0; 
2855              IADebugFlag = ia_cmds.maddr;
2856              printk("New debug option loaded\n");
2857          }
2858              break;
2859          default:
2860              ia_cmds.status = 0;
2861              break;
2862       } 
2863    }
2864       break;
2865    default:
2866       break;
2867
2868    }    
2869    return 0;  
2870 }  
2871   
2872 static int ia_getsockopt(struct atm_vcc *vcc, int level, int optname,   
2873         void __user *optval, int optlen)  
2874 {  
2875         IF_EVENT(printk(">ia_getsockopt\n");)  
2876         return -EINVAL;  
2877 }  
2878   
2879 static int ia_setsockopt(struct atm_vcc *vcc, int level, int optname,   
2880         void __user *optval, int optlen)  
2881 {  
2882         IF_EVENT(printk(">ia_setsockopt\n");)  
2883         return -EINVAL;  
2884 }  
2885   
2886 static int ia_pkt_tx (struct atm_vcc *vcc, struct sk_buff *skb) {
2887         IADEV *iadev;
2888         struct dle *wr_ptr;
2889         struct tx_buf_desc __iomem *buf_desc_ptr;
2890         int desc;
2891         int comp_code;
2892         int total_len;
2893         struct cpcs_trailer *trailer;
2894         struct ia_vcc *iavcc;
2895
2896         iadev = INPH_IA_DEV(vcc->dev);  
2897         iavcc = INPH_IA_VCC(vcc);
2898         if (!iavcc->txing) {
2899            printk("discard packet on closed VC\n");
2900            if (vcc->pop)
2901                 vcc->pop(vcc, skb);
2902            else
2903                 dev_kfree_skb_any(skb);
2904            return 0;
2905         }
2906
2907         if (skb->len > iadev->tx_buf_sz - 8) {
2908            printk("Transmit size over tx buffer size\n");
2909            if (vcc->pop)
2910                  vcc->pop(vcc, skb);
2911            else
2912                  dev_kfree_skb_any(skb);
2913           return 0;
2914         }
2915         if ((u32)skb->data & 3) {
2916            printk("Misaligned SKB\n");
2917            if (vcc->pop)
2918                  vcc->pop(vcc, skb);
2919            else
2920                  dev_kfree_skb_any(skb);
2921            return 0;
2922         }       
2923         /* Get a descriptor number from our free descriptor queue  
2924            We get the descr number from the TCQ now, since I am using  
2925            the TCQ as a free buffer queue. Initially TCQ will be   
2926            initialized with all the descriptors and is hence, full.  
2927         */
2928         desc = get_desc (iadev, iavcc);
2929         if (desc == 0xffff) 
2930             return 1;
2931         comp_code = desc >> 13;  
2932         desc &= 0x1fff;  
2933   
2934         if ((desc == 0) || (desc > iadev->num_tx_desc))  
2935         {  
2936                 IF_ERR(printk(DEV_LABEL "invalid desc for send: %d\n", desc);) 
2937                 atomic_inc(&vcc->stats->tx);
2938                 if (vcc->pop)   
2939                     vcc->pop(vcc, skb);   
2940                 else  
2941                     dev_kfree_skb_any(skb);
2942                 return 0;   /* return SUCCESS */
2943         }  
2944   
2945         if (comp_code)  
2946         {  
2947             IF_ERR(printk(DEV_LABEL "send desc:%d completion code %d error\n", 
2948                                                             desc, comp_code);)  
2949         }  
2950        
2951         /* remember the desc and vcc mapping */
2952         iavcc->vc_desc_cnt++;
2953         iadev->desc_tbl[desc-1].iavcc = iavcc;
2954         iadev->desc_tbl[desc-1].txskb = skb;
2955         IA_SKB_STATE(skb) = 0;
2956
2957         iadev->ffL.tcq_rd += 2;
2958         if (iadev->ffL.tcq_rd > iadev->ffL.tcq_ed)
2959                 iadev->ffL.tcq_rd  = iadev->ffL.tcq_st;
2960         writew(iadev->ffL.tcq_rd, iadev->seg_reg+TCQ_RD_PTR);
2961   
2962         /* Put the descriptor number in the packet ready queue  
2963                 and put the updated write pointer in the DLE field   
2964         */   
2965         *(u16*)(iadev->seg_ram+iadev->ffL.prq_wr) = desc; 
2966
2967         iadev->ffL.prq_wr += 2;
2968         if (iadev->ffL.prq_wr > iadev->ffL.prq_ed)
2969                 iadev->ffL.prq_wr = iadev->ffL.prq_st;
2970           
2971         /* Figure out the exact length of the packet and padding required to 
2972            make it  aligned on a 48 byte boundary.  */
2973         total_len = skb->len + sizeof(struct cpcs_trailer);  
2974         total_len = ((total_len + 47) / 48) * 48;
2975         IF_TX(printk("ia packet len:%d padding:%d\n", total_len, total_len - skb->len);)  
2976  
2977         /* Put the packet in a tx buffer */   
2978         trailer = iadev->tx_buf[desc-1].cpcs;
2979         IF_TX(printk("Sent: skb = 0x%x skb->data: 0x%x len: %d, desc: %d\n",
2980                   (u32)skb, (u32)skb->data, skb->len, desc);)
2981         trailer->control = 0; 
2982         /*big endian*/ 
2983         trailer->length = ((skb->len & 0xff) << 8) | ((skb->len & 0xff00) >> 8);
2984         trailer->crc32 = 0;     /* not needed - dummy bytes */  
2985
2986         /* Display the packet */  
2987         IF_TXPKT(printk("Sent data: len = %d MsgNum = %d\n", 
2988                                                         skb->len, tcnter++);  
2989         xdump(skb->data, skb->len, "TX: ");
2990         printk("\n");)
2991
2992         /* Build the buffer descriptor */  
2993         buf_desc_ptr = iadev->seg_ram+TX_DESC_BASE;
2994         buf_desc_ptr += desc;   /* points to the corresponding entry */  
2995         buf_desc_ptr->desc_mode = AAL5 | EOM_EN | APP_CRC32 | CMPL_INT;   
2996         /* Huh ? p.115 of users guide describes this as a read-only register */
2997         writew(TRANSMIT_DONE, iadev->seg_reg+SEG_INTR_STATUS_REG);
2998         buf_desc_ptr->vc_index = vcc->vci;
2999         buf_desc_ptr->bytes = total_len;  
3000
3001         if (vcc->qos.txtp.traffic_class == ATM_ABR)  
3002            clear_lockup (vcc, iadev);
3003
3004         /* Build the DLE structure */  
3005         wr_ptr = iadev->tx_dle_q.write;  
3006         memset((caddr_t)wr_ptr, 0, sizeof(*wr_ptr));  
3007         wr_ptr->sys_pkt_addr = pci_map_single(iadev->pci, skb->data,
3008                 skb->len, PCI_DMA_TODEVICE);
3009         wr_ptr->local_pkt_addr = (buf_desc_ptr->buf_start_hi << 16) | 
3010                                                   buf_desc_ptr->buf_start_lo;  
3011         /* wr_ptr->bytes = swap(total_len);     didn't seem to affect ?? */  
3012         wr_ptr->bytes = skb->len;  
3013
3014         /* hw bug - DLEs of 0x2d, 0x2e, 0x2f cause DMA lockup */
3015         if ((wr_ptr->bytes >> 2) == 0xb)
3016            wr_ptr->bytes = 0x30;
3017
3018         wr_ptr->mode = TX_DLE_PSI; 
3019         wr_ptr->prq_wr_ptr_data = 0;
3020   
3021         /* end is not to be used for the DLE q */  
3022         if (++wr_ptr == iadev->tx_dle_q.end)  
3023                 wr_ptr = iadev->tx_dle_q.start;  
3024         
3025         /* Build trailer dle */
3026         wr_ptr->sys_pkt_addr = iadev->tx_buf[desc-1].dma_addr;
3027         wr_ptr->local_pkt_addr = ((buf_desc_ptr->buf_start_hi << 16) | 
3028           buf_desc_ptr->buf_start_lo) + total_len - sizeof(struct cpcs_trailer);
3029
3030         wr_ptr->bytes = sizeof(struct cpcs_trailer);
3031         wr_ptr->mode = DMA_INT_ENABLE; 
3032         wr_ptr->prq_wr_ptr_data = iadev->ffL.prq_wr;
3033         
3034         /* end is not to be used for the DLE q */
3035         if (++wr_ptr == iadev->tx_dle_q.end)  
3036                 wr_ptr = iadev->tx_dle_q.start;
3037
3038         iadev->tx_dle_q.write = wr_ptr;  
3039         ATM_DESC(skb) = vcc->vci;
3040         skb_queue_tail(&iadev->tx_dma_q, skb);
3041
3042         atomic_inc(&vcc->stats->tx);
3043         iadev->tx_pkt_cnt++;
3044         /* Increment transaction counter */  
3045         writel(2, iadev->dma+IPHASE5575_TX_COUNTER);  
3046         
3047 #if 0        
3048         /* add flow control logic */ 
3049         if (atomic_read(&vcc->stats->tx) % 20 == 0) {
3050           if (iavcc->vc_desc_cnt > 10) {
3051              vcc->tx_quota =  vcc->tx_quota * 3 / 4;
3052             printk("Tx1:  vcc->tx_quota = %d \n", (u32)vcc->tx_quota );
3053               iavcc->flow_inc = -1;
3054               iavcc->saved_tx_quota = vcc->tx_quota;
3055            } else if ((iavcc->flow_inc < 0) && (iavcc->vc_desc_cnt < 3)) {
3056              // vcc->tx_quota = 3 * iavcc->saved_tx_quota / 4;
3057              printk("Tx2:  vcc->tx_quota = %d \n", (u32)vcc->tx_quota ); 
3058               iavcc->flow_inc = 0;
3059            }
3060         }
3061 #endif
3062         IF_TX(printk("ia send done\n");)  
3063         return 0;  
3064 }  
3065
3066 static int ia_send(struct atm_vcc *vcc, struct sk_buff *skb)
3067 {
3068         IADEV *iadev; 
3069         struct ia_vcc *iavcc;
3070         unsigned long flags;
3071
3072         iadev = INPH_IA_DEV(vcc->dev);
3073         iavcc = INPH_IA_VCC(vcc); 
3074         if ((!skb)||(skb->len>(iadev->tx_buf_sz-sizeof(struct cpcs_trailer))))
3075         {
3076             if (!skb)
3077                 printk(KERN_CRIT "null skb in ia_send\n");
3078             else dev_kfree_skb_any(skb);
3079             return -EINVAL;
3080         }                         
3081         spin_lock_irqsave(&iadev->tx_lock, flags); 
3082         if (!test_bit(ATM_VF_READY,&vcc->flags)){ 
3083             dev_kfree_skb_any(skb);
3084             spin_unlock_irqrestore(&iadev->tx_lock, flags);
3085             return -EINVAL; 
3086         }
3087         ATM_SKB(skb)->vcc = vcc;
3088  
3089         if (skb_peek(&iadev->tx_backlog)) {
3090            skb_queue_tail(&iadev->tx_backlog, skb);
3091         }
3092         else {
3093            if (ia_pkt_tx (vcc, skb)) {
3094               skb_queue_tail(&iadev->tx_backlog, skb);
3095            }
3096         }
3097         spin_unlock_irqrestore(&iadev->tx_lock, flags);
3098         return 0;
3099
3100 }
3101
3102 static int ia_proc_read(struct atm_dev *dev,loff_t *pos,char *page)
3103
3104   int   left = *pos, n;   
3105   char  *tmpPtr;
3106   IADEV *iadev = INPH_IA_DEV(dev);
3107   if(!left--) {
3108      if (iadev->phy_type == FE_25MBIT_PHY) {
3109        n = sprintf(page, "  Board Type         :  Iphase5525-1KVC-128K\n");
3110        return n;
3111      }
3112      if (iadev->phy_type == FE_DS3_PHY)
3113         n = sprintf(page, "  Board Type         :  Iphase-ATM-DS3");
3114      else if (iadev->phy_type == FE_E3_PHY)
3115         n = sprintf(page, "  Board Type         :  Iphase-ATM-E3");
3116      else if (iadev->phy_type == FE_UTP_OPTION)
3117          n = sprintf(page, "  Board Type         :  Iphase-ATM-UTP155"); 
3118      else
3119         n = sprintf(page, "  Board Type         :  Iphase-ATM-OC3");
3120      tmpPtr = page + n;
3121      if (iadev->pci_map_size == 0x40000)
3122         n += sprintf(tmpPtr, "-1KVC-");
3123      else
3124         n += sprintf(tmpPtr, "-4KVC-");  
3125      tmpPtr = page + n; 
3126      if ((iadev->memType & MEM_SIZE_MASK) == MEM_SIZE_1M)
3127         n += sprintf(tmpPtr, "1M  \n");
3128      else if ((iadev->memType & MEM_SIZE_MASK) == MEM_SIZE_512K)
3129         n += sprintf(tmpPtr, "512K\n");
3130      else
3131        n += sprintf(tmpPtr, "128K\n");
3132      return n;
3133   }
3134   if (!left) {
3135      return  sprintf(page, "  Number of Tx Buffer:  %u\n"
3136                            "  Size of Tx Buffer  :  %u\n"
3137                            "  Number of Rx Buffer:  %u\n"
3138                            "  Size of Rx Buffer  :  %u\n"
3139                            "  Packets Receiverd  :  %u\n"
3140                            "  Packets Transmitted:  %u\n"
3141                            "  Cells Received     :  %u\n"
3142                            "  Cells Transmitted  :  %u\n"
3143                            "  Board Dropped Cells:  %u\n"
3144                            "  Board Dropped Pkts :  %u\n",
3145                            iadev->num_tx_desc,  iadev->tx_buf_sz,
3146                            iadev->num_rx_desc,  iadev->rx_buf_sz,
3147                            iadev->rx_pkt_cnt,   iadev->tx_pkt_cnt,
3148                            iadev->rx_cell_cnt, iadev->tx_cell_cnt,
3149                            iadev->drop_rxcell, iadev->drop_rxpkt);                        
3150   }
3151   return 0;
3152 }
3153   
3154 static const struct atmdev_ops ops = {  
3155         .open           = ia_open,  
3156         .close          = ia_close,  
3157         .ioctl          = ia_ioctl,  
3158         .getsockopt     = ia_getsockopt,  
3159         .setsockopt     = ia_setsockopt,  
3160         .send           = ia_send,  
3161         .phy_put        = ia_phy_put,  
3162         .phy_get        = ia_phy_get,  
3163         .change_qos     = ia_change_qos,  
3164         .proc_read      = ia_proc_read,
3165         .owner          = THIS_MODULE,
3166 };  
3167           
3168 static int __devinit ia_init_one(struct pci_dev *pdev,
3169                                  const struct pci_device_id *ent)
3170 {  
3171         struct atm_dev *dev;  
3172         IADEV *iadev;  
3173         unsigned long flags;
3174         int ret;
3175
3176         iadev = kzalloc(sizeof(*iadev), GFP_KERNEL);
3177         if (!iadev) {
3178                 ret = -ENOMEM;
3179                 goto err_out;
3180         }
3181
3182         iadev->pci = pdev;
3183
3184         IF_INIT(printk("ia detected at bus:%d dev: %d function:%d\n",
3185                 pdev->bus->number, PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn));)
3186         if (pci_enable_device(pdev)) {
3187                 ret = -ENODEV;
3188                 goto err_out_free_iadev;
3189         }
3190         dev = atm_dev_register(DEV_LABEL, &ops, -1, NULL);
3191         if (!dev) {
3192                 ret = -ENOMEM;
3193                 goto err_out_disable_dev;
3194         }
3195         dev->dev_data = iadev;
3196         IF_INIT(printk(DEV_LABEL "registered at (itf :%d)\n", dev->number);)
3197         IF_INIT(printk("dev_id = 0x%x iadev->LineRate = %d \n", (u32)dev,
3198                 iadev->LineRate);)
3199
3200         ia_dev[iadev_count] = iadev;
3201         _ia_dev[iadev_count] = dev;
3202         iadev_count++;
3203         spin_lock_init(&iadev->misc_lock);
3204         /* First fixes first. I don't want to think about this now. */
3205         spin_lock_irqsave(&iadev->misc_lock, flags); 
3206         if (ia_init(dev) || ia_start(dev)) {  
3207                 IF_INIT(printk("IA register failed!\n");)
3208                 iadev_count--;
3209                 ia_dev[iadev_count] = NULL;
3210                 _ia_dev[iadev_count] = NULL;
3211                 spin_unlock_irqrestore(&iadev->misc_lock, flags); 
3212                 ret = -EINVAL;
3213                 goto err_out_deregister_dev;
3214         }
3215         spin_unlock_irqrestore(&iadev->misc_lock, flags); 
3216         IF_EVENT(printk("iadev_count = %d\n", iadev_count);)
3217
3218         iadev->next_board = ia_boards;  
3219         ia_boards = dev;  
3220
3221         pci_set_drvdata(pdev, dev);
3222
3223         return 0;
3224
3225 err_out_deregister_dev:
3226         atm_dev_deregister(dev);  
3227 err_out_disable_dev:
3228         pci_disable_device(pdev);
3229 err_out_free_iadev:
3230         kfree(iadev);
3231 err_out:
3232         return ret;
3233 }
3234
3235 static void __devexit ia_remove_one(struct pci_dev *pdev)
3236 {
3237         struct atm_dev *dev = pci_get_drvdata(pdev);
3238         IADEV *iadev = INPH_IA_DEV(dev);
3239
3240         ia_phy_put(dev, ia_phy_get(dev,0x10) & ~(0x4), 0x10); 
3241         udelay(1);
3242
3243         /* De-register device */  
3244         free_irq(iadev->irq, dev);
3245         iadev_count--;
3246         ia_dev[iadev_count] = NULL;
3247         _ia_dev[iadev_count] = NULL;
3248         IF_EVENT(printk("deregistering iav at (itf:%d)\n", dev->number);)
3249         atm_dev_deregister(dev);
3250
3251         iounmap(iadev->base);  
3252         pci_disable_device(pdev);
3253
3254         ia_free_rx(iadev);
3255         ia_free_tx(iadev);
3256
3257         kfree(iadev);
3258 }
3259
3260 static struct pci_device_id ia_pci_tbl[] = {
3261         { PCI_VENDOR_ID_IPHASE, 0x0008, PCI_ANY_ID, PCI_ANY_ID, },
3262         { PCI_VENDOR_ID_IPHASE, 0x0009, PCI_ANY_ID, PCI_ANY_ID, },
3263         { 0,}
3264 };
3265 MODULE_DEVICE_TABLE(pci, ia_pci_tbl);
3266
3267 static struct pci_driver ia_driver = {
3268         .name =         DEV_LABEL,
3269         .id_table =     ia_pci_tbl,
3270         .probe =        ia_init_one,
3271         .remove =       __devexit_p(ia_remove_one),
3272 };
3273
3274 static int __init ia_module_init(void)
3275 {
3276         int ret;
3277
3278         ret = pci_register_driver(&ia_driver);
3279         if (ret >= 0) {
3280                 ia_timer.expires = jiffies + 3*HZ;
3281                 add_timer(&ia_timer); 
3282         } else
3283                 printk(KERN_ERR DEV_LABEL ": no adapter found\n");  
3284         return ret;
3285 }
3286
3287 static void __exit ia_module_exit(void)
3288 {
3289         pci_unregister_driver(&ia_driver);
3290
3291         del_timer(&ia_timer);
3292 }
3293
3294 module_init(ia_module_init);
3295 module_exit(ia_module_exit);