/* * TE410P Quad-T1/E1 PCI Driver version 0.1, 12/16/02 * * Written by Mark Spencer * Based on previous works, designs, and archetectures conceived and * written by Jim Dixon . * * Copyright (C) 2001 Jim Dixon / Zapata Telephony. * Copyright (C) 2001-2009, Digium, Inc. * * All rights reserved. * */ /* * See http://www.asterisk.org for more information about * the Asterisk project. Please do not directly contact * any of the maintainers of this project for assistance; * the project provides a web site, mailing lists and IRC * channels for your use. * * This program is free software, distributed under the terms of * the GNU General Public License Version 2 as published by the * Free Software Foundation. See the LICENSE file included with * this program for more details. */ #include #include #include #include #include #include #include #include #include #include #include #include "zaptel.h" #ifdef LINUX26 #include #endif #include "wct4xxp.h" #include "vpm450m.h" /* Work queues are a way to better distribute load on SMP systems */ #if defined(LINUX26) && (LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)) /* * Work queues can significantly improve performance and scalability * on multi-processor machines, but requires bypassing some kernel * API's, so it's not guaranteed to be compatible with all kernels. */ /* #define ENABLE_WORKQUEUES */ #endif /* Enable prefetching may help performance */ #define ENABLE_PREFETCH /* Support first generation cards? */ #define SUPPORT_GEN1 /* Define to get more attention-grabbing but slightly more I/O using alarm status */ #define FANCY_ALARM /* Define to support Digium Voice Processing Module expansion card */ #define VPM_SUPPORT #define DEBUG_MAIN (1 << 0) #define DEBUG_DTMF (1 << 1) #define DEBUG_REGS (1 << 2) #define DEBUG_TSI (1 << 3) #define DEBUG_ECHOCAN (1 << 4) #define DEBUG_RBS (1 << 5) #define DEBUG_FRAMER (1 << 6) #ifdef ENABLE_WORKQUEUES #include /* XXX UGLY!!!! XXX We have to access the direct structures of the workqueue which are only defined within workqueue.c because they don't give us a routine to allow us to nail a work to a particular thread of the CPU. Nailing to threads gives us substantially higher scalability in multi-CPU environments though! */ /* * The per-CPU workqueue (if single thread, we always use cpu 0's). * * The sequence counters are for flush_scheduled_work(). It wants to wait * until until all currently-scheduled works are completed, but it doesn't * want to be livelocked by new, incoming ones. So it waits until * remove_sequence is >= the insert_sequence which pertained when * flush_scheduled_work() was called. */ struct cpu_workqueue_struct { spinlock_t lock; long remove_sequence; /* Least-recently added (next to run) */ long insert_sequence; /* Next to add */ struct list_head worklist; wait_queue_head_t more_work; wait_queue_head_t work_done; struct workqueue_struct *wq; task_t *thread; int run_depth; /* Detect run_workqueue() recursion depth */ } ____cacheline_aligned; /* * The externally visible workqueue abstraction is an array of * per-CPU workqueues: */ struct workqueue_struct { /* TODO: Find out exactly where the API changed */ #if LINUX_VERSION_CODE > KERNEL_VERSION(2,6,15) struct cpu_workqueue_struct *cpu_wq; #else struct cpu_workqueue_struct cpu_wq[NR_CPUS]; #endif const char *name; struct list_head list; /* Empty if single thread */ }; /* Preempt must be disabled. */ static void __t4_queue_work(struct cpu_workqueue_struct *cwq, struct work_struct *work) { unsigned long flags; spin_lock_irqsave(&cwq->lock, flags); work->wq_data = cwq; list_add_tail(&work->entry, &cwq->worklist); cwq->insert_sequence++; wake_up(&cwq->more_work); spin_unlock_irqrestore(&cwq->lock, flags); } /* * Queue work on a workqueue. Return non-zero if it was successfully * added. * * We queue the work to the CPU it was submitted, but there is no * guarantee that it will be processed by that CPU. */ static inline int t4_queue_work(struct workqueue_struct *wq, struct work_struct *work, int cpu) { int ret = 0; get_cpu(); if (!test_and_set_bit(0, &work->pending)) { BUG_ON(!list_empty(&work->entry)); __t4_queue_work(wq->cpu_wq + cpu, work); ret = 1; } put_cpu(); return ret; } #endif static int pedanticpci = 1; static int debug=0; static int timingcable = 0; static int highestorder; static int t1e1override = -1; //0xFF; // -1 = jumper; 0xFF = E1 static int j1mode = 0; static int sigmode = FRMR_MODE_NO_ADDR_CMP; static int loopback = 0; static int alarmdebounce = 0; #ifdef VPM_SUPPORT static int vpmsupport = 1; /* If set to auto, vpmdtmfsupport is enabled for VPM400M and disabled for VPM450M */ static int vpmdtmfsupport = -1; /* -1=auto, 0=disabled, 1=enabled*/ static int vpmspans = 4; #define VPM_DEFAULT_DTMFTHRESHOLD 1000 static int dtmfthreshold = VPM_DEFAULT_DTMFTHRESHOLD; static int lastdtmfthreshold = VPM_DEFAULT_DTMFTHRESHOLD; #endif /* Enabling bursting can more efficiently utilize PCI bus bandwidth, but can also cause PCI bus starvation, especially in combination with other aggressive cards. Please note that burst mode has no effect on CPU utilization / max number of calls / etc. */ static int noburst = 1; /* For 56kbps links, set this module parameter to 0x7f */ static int hardhdlcmode = 0xff; #ifdef FANCY_ALARM static int altab[] = { 0, 0, 0, 1, 2, 3, 4, 6, 8, 9, 11, 13, 16, 18, 20, 22, 24, 25, 27, 28, 29, 30, 31, 31, 32, 31, 31, 30, 29, 28, 27, 25, 23, 22, 20, 18, 16, 13, 11, 9, 8, 6, 4, 3, 2, 1, 0, 0, }; #endif #define MAX_SPANS 16 #define FLAG_STARTED (1 << 0) #define FLAG_NMF (1 << 1) #define FLAG_SENDINGYELLOW (1 << 2) #define TYPE_T1 1 /* is a T1 card */ #define TYPE_E1 2 /* is an E1 card */ #define TYPE_J1 3 /* is a running J1 */ #define FLAG_2NDGEN (1 << 3) #define FLAG_2PORT (1 << 4) #define FLAG_VPM2GEN (1 << 5) #define FLAG_OCTOPT (1 << 6) #define FLAG_3RDGEN (1 << 7) #define FLAG_BURST (1 << 8) #define FLAG_EXPRESS (1 << 9) #define CANARY 0xc0de struct devtype { char *desc; unsigned int flags; }; static struct devtype wct4xxp = { "Wildcard TE410P/TE405P (1st Gen)", 0 }; static struct devtype wct420p4 = { "Wildcard TE420 (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_EXPRESS }; static struct devtype wct410p4 = { "Wildcard TE410P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN }; static struct devtype wct410p3 = { "Wildcard TE410P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN }; static struct devtype wct405p4 = { "Wildcard TE405P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN }; static struct devtype wct405p3 = { "Wildcard TE405P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN }; static struct devtype wct410p2 = { "Wildcard TE410P (2nd Gen)", FLAG_2NDGEN }; static struct devtype wct405p2 = { "Wildcard TE405P (2nd Gen)", FLAG_2NDGEN }; static struct devtype wct220p4 = { "Wildcard TE220 (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT | FLAG_EXPRESS }; static struct devtype wct205p4 = { "Wildcard TE205P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT }; static struct devtype wct205p3 = { "Wildcard TE205P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT }; static struct devtype wct210p4 = { "Wildcard TE210P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT }; static struct devtype wct210p3 = { "Wildcard TE210P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT }; static struct devtype wct205 = { "Wildcard TE205P ", FLAG_2NDGEN | FLAG_2PORT }; static struct devtype wct210 = { "Wildcard TE210P ", FLAG_2NDGEN | FLAG_2PORT }; struct t4; struct t4_span { struct t4 *owner; unsigned int *writechunk; /* Double-word aligned write memory */ unsigned int *readchunk; /* Double-word aligned read memory */ int spantype; /* card type, T1 or E1 or J1 */ int sync; int psync; int alarmtimer; int redalarms; int notclear; int alarmcount; int spanflags; int syncpos; #ifdef SUPPORT_GEN1 int e1check; /* E1 check */ #endif struct zt_span span; unsigned char txsigs[16]; /* Transmit sigs */ int loopupcnt; int loopdowncnt; #ifdef SUPPORT_GEN1 unsigned char ec_chunk1[31][ZT_CHUNKSIZE]; /* first EC chunk buffer */ unsigned char ec_chunk2[31][ZT_CHUNKSIZE]; /* second EC chunk buffer */ #endif int irqmisses; /* HDLC controller fields */ struct zt_chan *sigchan; unsigned char sigmode; int sigactive; int frames_out; int frames_in; #ifdef VPM_SUPPORT unsigned long dtmfactive; unsigned long dtmfmask; unsigned long dtmfmutemask; short dtmfenergy[31]; short dtmfdigit[31]; #endif #ifdef ENABLE_WORKQUEUES struct work_struct swork; #endif struct zt_chan chans[0]; /* Individual channels */ }; struct t4 { /* This structure exists one per card */ struct pci_dev *dev; /* Pointer to PCI device */ unsigned int intcount; int num; /* Which card we are */ int t1e1; /* T1/E1 select pins */ int globalconfig; /* Whether global setup has been done */ int syncsrc; /* active sync source */ struct t4_span *tspans[4]; /* Individual spans */ int numspans; /* Number of spans on the card */ int blinktimer; #ifdef FANCY_ALARM int alarmpos; #endif int irq; /* IRQ used by device */ int order; /* Order */ int flags; /* Device flags */ int master; /* Are we master */ int ledreg; /* LED Register */ unsigned int gpio; unsigned int gpioctl; int e1recover; /* E1 recovery timer */ spinlock_t reglock; /* lock register access */ int spansstarted; /* number of spans started */ volatile unsigned int *writechunk; /* Double-word aligned write memory */ volatile unsigned int *readchunk; /* Double-word aligned read memory */ unsigned short canary; #ifdef ENABLE_WORKQUEUES atomic_t worklist; struct workqueue_struct *workq; #endif unsigned int passno; /* number of interrupt passes */ char *variety; int last0; /* for detecting double-missed IRQ */ /* DMA related fields */ unsigned int dmactrl; dma_addr_t readdma; dma_addr_t writedma; unsigned long memaddr; /* Base address of card */ unsigned long memlen; volatile unsigned int *membase; /* Base address of card */ /* Add this for our softlockup protector */ unsigned int oct_rw_count; /* Flags for our bottom half */ unsigned long checkflag; struct tasklet_struct t4_tlet; unsigned int vpm400checkstatus; #ifdef VPM_SUPPORT struct vpm450m *vpm450m; int vpm; #endif }; #define T4_VPM_PRESENT (1 << 28) #ifdef VPM_SUPPORT static void t4_vpm400_init(struct t4 *wc); static void t4_vpm450_init(struct t4 *wc); static void t4_vpm_set_dtmf_threshold(struct t4 *wc, unsigned int threshold); #endif static void __set_clear(struct t4 *wc, int span); static int t4_startup(struct zt_span *span); static int t4_shutdown(struct zt_span *span); static int t4_rbsbits(struct zt_chan *chan, int bits); static int t4_maint(struct zt_span *span, int cmd); #ifdef SUPPORT_GEN1 static int t4_reset_dma(struct t4 *wc); #endif static void t4_hdlc_hard_xmit(struct zt_chan *chan); static int t4_ioctl(struct zt_chan *chan, unsigned int cmd, unsigned long data); static void t4_tsi_assign(struct t4 *wc, int fromspan, int fromchan, int tospan, int tochan); static void t4_tsi_unassign(struct t4 *wc, int tospan, int tochan); static void __t4_set_timing_source(struct t4 *wc, int unit, int master, int slave); static void t4_check_alarms(struct t4 *wc, int span); static void t4_check_sigbits(struct t4 *wc, int span); #define WC_RDADDR 0 #define WC_WRADDR 1 #define WC_COUNT 2 #define WC_DMACTRL 3 #define WC_INTR 4 /* #define WC_GPIO 5 */ #define WC_VERSION 6 #define WC_LEDS 7 #define WC_GPIOCTL 8 #define WC_GPIO 9 #define WC_LADDR 10 #define WC_LDATA 11 #define WC_LCS (1 << 11) #define WC_LCS2 (1 << 12) #define WC_LALE (1 << 13) #define WC_LFRMR_CS (1 << 10) /* Framer's ChipSelect signal */ #define WC_ACTIVATE (1 << 12) #define WC_LREAD (1 << 15) #define WC_LWRITE (1 << 16) #define WC_OFF (0) #define WC_RED (1) #define WC_GREEN (2) #define WC_YELLOW (3) #define MAX_T4_CARDS 64 static void t4_isr_bh(unsigned long data); static struct t4 *cards[MAX_T4_CARDS]; #define MAX_TDM_CHAN 32 #define MAX_DTMF_DET 16 #define HDLC_IMR0_MASK (FRMR_IMR0_RME | FRMR_IMR0_RPF) #if 0 #define HDLC_IMR1_MASK (FRMR_IMR1_ALLS | FRMR_IMR1_XDU | FRMR_IMR1_XPR) #else #define HDLC_IMR1_MASK (FRMR_IMR1_XDU | FRMR_IMR1_XPR) #endif static inline unsigned int __t4_pci_in(struct t4 *wc, const unsigned int addr) { unsigned int res = readl(&wc->membase[addr]); return res; } static inline void __t4_pci_out(struct t4 *wc, const unsigned int addr, const unsigned int value) { unsigned int tmp; writel(value, &wc->membase[addr]); if (pedanticpci) { /* Even though we do not support fast back-to-back * transactions, some host bridges appear to generate them. * This delay prevents this. */ tmp = __t4_pci_in(wc, WC_VERSION); if ((tmp & 0xffff0000) != 0xc01a0000) printk("TE4XXP: Version Synchronization Error!\n"); } #if 0 tmp = __t4_pci_in(wc, addr); if ((value != tmp) && (addr != WC_LEDS) && (addr != WC_LDATA) && (addr != WC_GPIO) && (addr != WC_INTR)) printk("Tried to load %08x into %08x, but got %08x instead\n", value, addr, tmp); #endif } static inline void __t4_gpio_set(struct t4 *wc, unsigned bits, unsigned int val) { unsigned int newgpio; newgpio = wc->gpio & (~bits); newgpio |= val; if (newgpio != wc->gpio) { wc->gpio = newgpio; __t4_pci_out(wc, WC_GPIO, wc->gpio); } } static inline void __t4_gpio_setdir(struct t4 *wc, unsigned int bits, unsigned int val) { unsigned int newgpioctl; newgpioctl = wc->gpioctl & (~bits); newgpioctl |= val; if (newgpioctl != wc->gpioctl) { wc->gpioctl = newgpioctl; __t4_pci_out(wc, WC_GPIOCTL, wc->gpioctl); } } static inline void t4_gpio_setdir(struct t4 *wc, unsigned int bits, unsigned int val) { unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); __t4_gpio_setdir(wc, bits, val); spin_unlock_irqrestore(&wc->reglock, flags); } static inline void t4_gpio_set(struct t4 *wc, unsigned int bits, unsigned int val) { unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); __t4_gpio_set(wc, bits, val); spin_unlock_irqrestore(&wc->reglock, flags); } static inline void t4_pci_out(struct t4 *wc, const unsigned int addr, const unsigned int value) { unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); __t4_pci_out(wc, addr, value); spin_unlock_irqrestore(&wc->reglock, flags); } static inline void __t4_set_led(struct t4 *wc, int span, int color) { int oldreg = wc->ledreg; wc->ledreg &= ~(0x3 << (span << 1)); wc->ledreg |= (color << (span << 1)); if (oldreg != wc->ledreg) __t4_pci_out(wc, WC_LEDS, wc->ledreg); } static inline void t4_activate(struct t4 *wc) { wc->ledreg |= WC_ACTIVATE; t4_pci_out(wc, WC_LEDS, wc->ledreg); } static inline unsigned int t4_pci_in(struct t4 *wc, const unsigned int addr) { unsigned int ret; unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); ret = __t4_pci_in(wc, addr); spin_unlock_irqrestore(&wc->reglock, flags); return ret; } static inline unsigned int __t4_framer_in(struct t4 *wc, int unit, const unsigned int addr) { unsigned int ret; unit &= 0x3; __t4_pci_out(wc, WC_LADDR, (unit << 8) | (addr & 0xff)); __t4_pci_out(wc, WC_LADDR, (unit << 8) | (addr & 0xff) | WC_LFRMR_CS | WC_LREAD); if (pedanticpci) { __t4_pci_out(wc, WC_VERSION, 0); } ret = __t4_pci_in(wc, WC_LDATA); __t4_pci_out(wc, WC_LADDR, (unit << 8) | (addr & 0xff)); return ret & 0xff; } static inline unsigned int t4_framer_in(struct t4 *wc, int unit, const unsigned int addr) { unsigned long flags; unsigned int ret; spin_lock_irqsave(&wc->reglock, flags); ret = __t4_framer_in(wc, unit, addr); spin_unlock_irqrestore(&wc->reglock, flags); return ret; } static inline void __t4_framer_out(struct t4 *wc, int unit, const unsigned int addr, const unsigned int value) { unit &= 0x3; if (unlikely(debug & DEBUG_REGS)) printk("Writing %02x to address %02x of unit %d\n", value, addr, unit); __t4_pci_out(wc, WC_LADDR, (unit << 8) | (addr & 0xff)); __t4_pci_out(wc, WC_LDATA, value); __t4_pci_out(wc, WC_LADDR, (unit << 8) | (addr & 0xff) | WC_LFRMR_CS | WC_LWRITE); __t4_pci_out(wc, WC_LADDR, (unit << 8) | (addr & 0xff)); if (unlikely(debug & DEBUG_REGS)) printk("Write complete\n"); #if 0 if ((addr != FRMR_TXFIFO) && (addr != FRMR_CMDR) && (addr != 0xbc)) { unsigned int tmp; tmp = __t4_framer_in(wc, unit, addr); if (tmp != value) { printk("Expected %d from unit %d register %d but got %d instead\n", value, unit, addr, tmp); } } #endif } static inline void t4_framer_out(struct t4 *wc, int unit, const unsigned int addr, const unsigned int value) { unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); __t4_framer_out(wc, unit, addr, value); spin_unlock_irqrestore(&wc->reglock, flags); } #ifdef VPM_SUPPORT static inline void wait_a_little(void) { unsigned long newjiffies=jiffies+2; while(jiffies < newjiffies); } static inline unsigned int __t4_vpm_in(struct t4 *wc, int unit, const unsigned int addr) { unsigned int ret; unit &= 0x7; __t4_pci_out(wc, WC_LADDR, (addr & 0x1ff) | ( unit << 12)); __t4_pci_out(wc, WC_LADDR, (addr & 0x1ff) | ( unit << 12) | (1 << 11) | WC_LREAD); ret = __t4_pci_in(wc, WC_LDATA); __t4_pci_out(wc, WC_LADDR, 0); return ret & 0xff; } static inline void __t4_raw_oct_out(struct t4 *wc, const unsigned int addr, const unsigned int value) { int octopt = wc->tspans[0]->spanflags & FLAG_OCTOPT; if (!octopt) __t4_gpio_set(wc, 0xff, (addr >> 8)); __t4_pci_out(wc, WC_LDATA, 0x10000 | (addr & 0xffff)); if (!octopt) __t4_pci_out(wc, WC_LADDR, (WC_LWRITE)); __t4_pci_out(wc, WC_LADDR, (WC_LWRITE | WC_LALE)); if (!octopt) __t4_gpio_set(wc, 0xff, (value >> 8)); __t4_pci_out(wc, WC_LDATA, (value & 0xffff)); __t4_pci_out(wc, WC_LADDR, (WC_LWRITE | WC_LALE | WC_LCS)); __t4_pci_out(wc, WC_LADDR, (0)); } static inline unsigned int __t4_raw_oct_in(struct t4 *wc, const unsigned int addr) { unsigned int ret; int octopt = wc->tspans[0]->spanflags & FLAG_OCTOPT; if (!octopt) __t4_gpio_set(wc, 0xff, (addr >> 8)); __t4_pci_out(wc, WC_LDATA, 0x10000 | (addr & 0xffff)); if (!octopt) __t4_pci_out(wc, WC_LADDR, (WC_LWRITE)); __t4_pci_out(wc, WC_LADDR, (WC_LWRITE | WC_LALE)); #ifdef PEDANTIC_OCTASIC_CHECKING __t4_pci_out(wc, WC_LADDR, (WC_LALE)); #endif if (!octopt) { __t4_gpio_setdir(wc, 0xff, 0x00); __t4_gpio_set(wc, 0xff, 0x00); } __t4_pci_out(wc, WC_LADDR, (WC_LREAD | WC_LALE | WC_LCS)); if (octopt) { ret = __t4_pci_in(wc, WC_LDATA) & 0xffff; } else { ret = __t4_pci_in(wc, WC_LDATA) & 0xff; ret |= (__t4_pci_in(wc, WC_GPIO) & 0xff) << 8; } __t4_pci_out(wc, WC_LADDR, (0)); if (!octopt) __t4_gpio_setdir(wc, 0xff, 0xff); return ret & 0xffff; } static inline unsigned int __t4_oct_in(struct t4 *wc, unsigned int addr) { #ifdef PEDANTIC_OCTASIC_CHECKING int count = 1000; #endif __t4_raw_oct_out(wc, 0x0008, (addr >> 20)); __t4_raw_oct_out(wc, 0x000a, (addr >> 4) & ((1 << 16) - 1)); __t4_raw_oct_out(wc, 0x0000, (((addr >> 1) & 0x7) << 9) | (1 << 8) | (1)); #ifdef PEDANTIC_OCTASIC_CHECKING while((__t4_raw_oct_in(wc, 0x0000) & (1 << 8)) && --count); if (count != 1000) printk("Yah, read can be slow...\n"); if (!count) printk("Read timed out!\n"); #endif return __t4_raw_oct_in(wc, 0x0004); } static inline unsigned int t4_oct_in(struct t4 *wc, const unsigned int addr) { unsigned long flags; unsigned int ret; spin_lock_irqsave(&wc->reglock, flags); ret = __t4_oct_in(wc, addr); spin_unlock_irqrestore(&wc->reglock, flags); return ret; } static inline unsigned int t4_vpm_in(struct t4 *wc, int unit, const unsigned int addr) { unsigned long flags; unsigned int ret; spin_lock_irqsave(&wc->reglock, flags); ret = __t4_vpm_in(wc, unit, addr); spin_unlock_irqrestore(&wc->reglock, flags); return ret; } static inline void __t4_vpm_out(struct t4 *wc, int unit, const unsigned int addr, const unsigned int value) { unit &= 0x7; if (debug & DEBUG_REGS) printk("Writing %02x to address %02x of ec unit %d\n", value, addr, unit); __t4_pci_out(wc, WC_LADDR, (addr & 0xff)); __t4_pci_out(wc, WC_LDATA, value); __t4_pci_out(wc, WC_LADDR, (unit << 12) | (addr & 0x1ff) | (1 << 11)); __t4_pci_out(wc, WC_LADDR, (unit << 12) | (addr & 0x1ff) | (1 << 11) | WC_LWRITE); __t4_pci_out(wc, WC_LADDR, (unit << 12) | (addr & 0x1ff) | (1 << 11)); __t4_pci_out(wc, WC_LADDR, (unit << 12) | (addr & 0x1ff)); __t4_pci_out(wc, WC_LADDR, 0); if (debug & DEBUG_REGS) printk("Write complete\n"); #if 0 { unsigned int tmp; tmp = t4_vpm_in(wc, unit, addr); if (tmp != value) { printk("Expected %d from unit %d echo register %d but got %d instead\n", value, unit, addr, tmp); } } #endif } static inline void __t4_oct_out(struct t4 *wc, unsigned int addr, unsigned int value) { #ifdef PEDANTIC_OCTASIC_CHECKING int count = 1000; #endif __t4_raw_oct_out(wc, 0x0008, (addr >> 20)); __t4_raw_oct_out(wc, 0x000a, (addr >> 4) & ((1 << 16) - 1)); __t4_raw_oct_out(wc, 0x0004, value); __t4_raw_oct_out(wc, 0x0000, (((addr >> 1) & 0x7) << 9) | (1 << 8) | (3 << 12) | 1); #ifdef PEDANTIC_OCTASIC_CHECKING while((__t4_raw_oct_in(wc, 0x0000) & (1 << 8)) && --count); if (count != 1000) printk("Yah, write can be slow\n"); if (!count) printk("Write timed out!\n"); #endif } static inline void t4_oct_out(struct t4 *wc, const unsigned int addr, const unsigned int value) { unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); __t4_oct_out(wc, addr, value); spin_unlock_irqrestore(&wc->reglock, flags); } static inline void t4_vpm_out(struct t4 *wc, int unit, const unsigned int addr, const unsigned int value) { unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); __t4_vpm_out(wc, unit, addr, value); spin_unlock_irqrestore(&wc->reglock, flags); } static const char vpm_digits[] = {'0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'A', 'B', 'C', 'D', '*', '#'}; static void t4_check_vpm450(struct t4 *wc) { int channel, tone, start, span; if (vpm450m_checkirq(wc->vpm450m)) { while(vpm450m_getdtmf(wc->vpm450m, &channel, &tone, &start)) { span = channel & 0x3; channel >>= 2; if (!wc->t1e1) channel -= 5; else channel -= 1; if (unlikely(debug)) printk("Got tone %s of '%c' on channel %d of span %d\n", (start ? "START" : "STOP"), tone, channel, span + 1); if (test_bit(channel, &wc->tspans[span]->dtmfmask) && (tone != 'u')) { if (start) { /* The octasic is supposed to mute us, but... Yah, you guessed it. */ if (test_bit(channel, &wc->tspans[span]->dtmfmutemask)) { unsigned long flags; struct zt_chan *chan = &wc->tspans[span]->span.chans[channel]; int y; spin_lock_irqsave(&chan->lock, flags); for (y=0;ynumbufs;y++) { if ((chan->inreadbuf > -1) && (chan->readidx[y])) memset(chan->readbuf[chan->inreadbuf], ZT_XLAW(0, chan), chan->readidx[y]); } spin_unlock_irqrestore(&chan->lock, flags); } set_bit(channel, &wc->tspans[span]->dtmfactive); zt_qevent_lock(&wc->tspans[span]->span.chans[channel], (ZT_EVENT_DTMFDOWN | tone)); } else { clear_bit(channel, &wc->tspans[span]->dtmfactive); zt_qevent_lock(&wc->tspans[span]->span.chans[channel], (ZT_EVENT_DTMFUP | tone)); } } } } } static void t4_check_vpm400(struct t4 *wc, unsigned int newio) { unsigned int digit, regval = 0; unsigned int regbyte; int x, i; short energy=0; static unsigned int lastio = 0; struct t4_span *ts; if (debug && (newio != lastio)) printk("Last was %08x, new is %08x\n", lastio, newio); lastio = newio; for(x = 0; x < 8; x++) { if (newio & (1 << (7 - x))) continue; ts = wc->tspans[x%4]; /* Start of DTMF detection process */ regbyte = t4_vpm_in(wc, x, 0xb8); t4_vpm_out(wc, x, 0xb8, regbyte); /* Write 1 to clear */ regval = regbyte << 8; regbyte = t4_vpm_in(wc, x, 0xb9); t4_vpm_out(wc, x, 0xb9, regbyte); regval |= regbyte; for(i = 0; (i < MAX_DTMF_DET) && regval; i++) { if(regval & 0x0001) { int channel = (i << 1) + (x >> 2); int base = channel - 1; if (!wc->t1e1) base -= 4; regbyte = t4_vpm_in(wc, x, 0xa8 + i); digit = vpm_digits[regbyte]; if (!(wc->tspans[0]->spanflags & FLAG_VPM2GEN)) { energy = t4_vpm_in(wc, x, 0x58 + channel); energy = ZT_XLAW(energy, ts->chans); ts->dtmfenergy[base] = energy; } set_bit(base, &ts->dtmfactive); if (ts->dtmfdigit[base]) { if (ts->dtmfmask & (1 << base)) zt_qevent_lock(&ts->span.chans[base], (ZT_EVENT_DTMFUP | ts->dtmfdigit[base])); } ts->dtmfdigit[base] = digit; if (test_bit(base, &ts->dtmfmask)) zt_qevent_lock(&ts->span.chans[base], (ZT_EVENT_DTMFDOWN | digit)); if (test_bit(base, &ts->dtmfmutemask)) { /* Mute active receive buffer*/ unsigned long flags; struct zt_chan *chan = &ts->span.chans[base]; int y; spin_lock_irqsave(&chan->lock, flags); for (y=0;ynumbufs;y++) { if ((chan->inreadbuf > -1) && (chan->readidx[y])) memset(chan->readbuf[chan->inreadbuf], ZT_XLAW(0, chan), chan->readidx[y]); } spin_unlock_irqrestore(&chan->lock, flags); } if (debug) printk("Digit Seen: %d, Span: %d, channel: %d, energy: %02x, 'channel %d' chip %d\n", digit, x % 4, base + 1, energy, channel, x); } regval = regval >> 1; } if (!(wc->tspans[0]->spanflags & FLAG_VPM2GEN)) continue; /* Start of DTMF off detection process */ regbyte = t4_vpm_in(wc, x, 0xbc); t4_vpm_out(wc, x, 0xbc, regbyte); /* Write 1 to clear */ regval = regbyte << 8; regbyte = t4_vpm_in(wc, x, 0xbd); t4_vpm_out(wc, x, 0xbd, regbyte); regval |= regbyte; for(i = 0; (i < MAX_DTMF_DET) && regval; i++) { if(regval & 0x0001) { int channel = (i << 1) + (x >> 2); int base = channel - 1; if (!wc->t1e1) base -= 4; clear_bit(base, &ts->dtmfactive); if (ts->dtmfdigit[base]) { if (test_bit(base, &ts->dtmfmask)) zt_qevent_lock(&ts->span.chans[base], (ZT_EVENT_DTMFUP | ts->dtmfdigit[base])); } digit = ts->dtmfdigit[base]; ts->dtmfdigit[base] = 0; if (debug) printk("Digit Gone: %d, Span: %d, channel: %d, energy: %02x, 'channel %d' chip %d\n", digit, x % 4, base + 1, energy, channel, x); } regval = regval >> 1; } } } #endif static void hdlc_stop(struct t4 *wc, unsigned int span) { struct t4_span *t = wc->tspans[span]; unsigned char imr0, imr1, mode; int i = 0; if (debug & DEBUG_FRAMER) printk("Stopping HDLC controller on span %d\n", span+1); /* Clear receive and transmit timeslots */ for (i = 0; i < 4; i++) { t4_framer_out(wc, span, FRMR_RTR_BASE + i, 0x00); t4_framer_out(wc, span, FRMR_TTR_BASE + i, 0x00); } imr0 = t4_framer_in(wc, span, FRMR_IMR0); imr1 = t4_framer_in(wc, span, FRMR_IMR1); /* Disable HDLC interrupts */ imr0 |= HDLC_IMR0_MASK; t4_framer_out(wc, span, FRMR_IMR0, imr0); imr1 |= HDLC_IMR1_MASK; t4_framer_out(wc, span, FRMR_IMR1, imr1); mode = t4_framer_in(wc, span, FRMR_MODE); mode &= ~FRMR_MODE_HRAC; t4_framer_out(wc, span, FRMR_MODE, mode); t->sigactive = 0; } static inline void __t4_framer_cmd(struct t4 *wc, unsigned int span, int cmd) { __t4_framer_out(wc, span, FRMR_CMDR, cmd); } static inline void t4_framer_cmd_wait(struct t4 *wc, unsigned int span, int cmd) { int sis; int loops = 0; /* XXX could be time consuming XXX */ for (;;) { sis = t4_framer_in(wc, span, FRMR_SIS); if (!(sis & 0x04)) break; if (!loops++ && (debug & DEBUG_FRAMER)) { printk("!!!SIS Waiting before cmd %02x\n", cmd); } } if (loops && (debug & DEBUG_FRAMER)) printk("!!!SIS waited %d loops\n", loops); t4_framer_out(wc, span, FRMR_CMDR, cmd); } static int hdlc_start(struct t4 *wc, unsigned int span, struct zt_chan *chan, unsigned char mode) { struct t4_span *t = wc->tspans[span]; unsigned char imr0, imr1; int offset = chan->chanpos; unsigned long flags; if (debug & DEBUG_FRAMER) printk("Starting HDLC controller for channel %d span %d\n", offset, span+1); if (mode != FRMR_MODE_NO_ADDR_CMP) return -1; mode |= FRMR_MODE_HRAC; /* Make sure we're in the right mode */ t4_framer_out(wc, span, FRMR_MODE, mode); t4_framer_out(wc, span, FRMR_TSEO, 0x00); t4_framer_out(wc, span, FRMR_TSBS1, hardhdlcmode); /* Set the interframe gaps, etc */ t4_framer_out(wc, span, FRMR_CCR1, FRMR_CCR1_ITF|FRMR_CCR1_EITS); t4_framer_out(wc, span, FRMR_CCR2, FRMR_CCR2_RCRC); /* Set up the time slot that we want to tx/rx on */ t4_framer_out(wc, span, FRMR_TTR_BASE + (offset / 8), (0x80 >> (offset % 8))); t4_framer_out(wc, span, FRMR_RTR_BASE + (offset / 8), (0x80 >> (offset % 8))); imr0 = t4_framer_in(wc, span, FRMR_IMR0); imr1 = t4_framer_in(wc, span, FRMR_IMR1); /* Enable our interrupts again */ imr0 &= ~HDLC_IMR0_MASK; t4_framer_out(wc, span, FRMR_IMR0, imr0); imr1 &= ~HDLC_IMR1_MASK; t4_framer_out(wc, span, FRMR_IMR1, imr1); /* Reset the signaling controller */ t4_framer_cmd_wait(wc, span, FRMR_CMDR_SRES); spin_lock_irqsave(&wc->reglock, flags); t->sigchan = chan; spin_unlock_irqrestore(&wc->reglock, flags); t->sigactive = 0; return 0; } static void __set_clear(struct t4 *wc, int span) { int i,j; int oldnotclear; unsigned short val=0; struct t4_span *ts = wc->tspans[span]; oldnotclear = ts->notclear; if ((ts->spantype == TYPE_T1) || (ts->spantype == TYPE_J1)) { for (i=0;i<24;i++) { j = (i/8); if (ts->span.chans[i].flags & ZT_FLAG_CLEAR) { val |= 1 << (7 - (i % 8)); ts->notclear &= ~(1 << i); } else ts->notclear |= (1 << i); if ((i % 8)==7) { if (debug) printk("Putting %d in register %02x on span %d\n", val, 0x2f + j, span + 1); __t4_framer_out(wc, span, 0x2f + j, val); val = 0; } } } else { for (i=0;i<31;i++) { if (ts->span.chans[i].flags & ZT_FLAG_CLEAR) ts->notclear &= ~(1 << i); else ts->notclear |= (1 << i); } } if (ts->notclear != oldnotclear) { unsigned char reg; reg = __t4_framer_in(wc, span, FRMR_IMR0); if (ts->notclear) reg &= ~0x08; else reg |= 0x08; __t4_framer_out(wc, span, FRMR_IMR0, reg); } } #if 0 static void set_clear(struct t4 *wc, int span) { unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); __set_clear(wc, span); spin_unlock_irqrestore(&wc->reglock, flags); } #endif static int t4_dacs(struct zt_chan *dst, struct zt_chan *src) { struct t4 *wc; struct t4_span *ts; wc = dst->pvt; ts = wc->tspans[dst->span->offset]; if (src && (src->pvt != dst->pvt)) { if (ts->spanflags & FLAG_2NDGEN) t4_tsi_unassign(wc, dst->span->offset, dst->chanpos); wc = src->pvt; if (ts->spanflags & FLAG_2NDGEN) t4_tsi_unassign(wc, src->span->offset, src->chanpos); if (debug) printk("Unassigning %d/%d by default and...\n", src->span->offset, src->chanpos); if (debug) printk("Unassigning %d/%d by default\n", dst->span->offset, dst->chanpos); return -1; } if (src) { t4_tsi_assign(wc, src->span->offset, src->chanpos, dst->span->offset, dst->chanpos); if (debug) printk("Assigning channel %d/%d -> %d/%d!\n", src->span->offset, src->chanpos, dst->span->offset, dst->chanpos); } else { t4_tsi_unassign(wc, dst->span->offset, dst->chanpos); if (debug) printk("Unassigning channel %d/%d!\n", dst->span->offset, dst->chanpos); } return 0; } #ifdef VPM_SUPPORT void oct_set_reg(void *data, unsigned int reg, unsigned int val) { struct t4 *wc = data; t4_oct_out(wc, reg, val); } unsigned int oct_get_reg(void *data, unsigned int reg) { struct t4 *wc = data; unsigned int ret; ret = t4_oct_in(wc, reg); return ret; } static int t4_vpm_unit(int span, int channel) { int unit = 0; switch(vpmspans) { case 4: unit = span; unit += (channel & 1) << 2; break; case 2: unit = span; unit += (channel & 0x3) << 1; break; case 1: unit = span; unit += (channel & 0x7); } return unit; } static int t4_echocan(struct zt_chan *chan, int eclen) { struct t4 *wc = chan->pvt; int channel; int unit; if (!wc->vpm) return -ENODEV; if (chan->span->offset >= vpmspans) return -ENODEV; if (wc->t1e1) channel = chan->chanpos; else channel = chan->chanpos + 4; if (wc->vpm450m) { channel = channel << 2; channel |= chan->span->offset; if(debug & DEBUG_ECHOCAN) printk("echocan: Card is %d, Channel is %d, Span is %d, offset is %d length %d\n", wc->num, chan->chanpos, chan->span->offset, channel, eclen); vpm450m_setec(wc->vpm450m, channel, eclen); // Mark msleep(10); // msleep(100); // longer test } else { unit = t4_vpm_unit(chan->span->offset, channel); if(debug & DEBUG_ECHOCAN) printk("echocan: Card is %d, Channel is %d, Span is %d, unit is %d, unit offset is %d length %d\n", wc->num, chan->chanpos, chan->span->offset, unit, channel, eclen); if (eclen) t4_vpm_out(wc,unit,channel,0x3e); else t4_vpm_out(wc,unit,channel,0x01); } return 0; } #endif static int t4_ioctl(struct zt_chan *chan, unsigned int cmd, unsigned long data) { struct t4_regs regs; int x; struct t4 *wc = chan->pvt; #ifdef VPM_SUPPORT int j; int channel; struct t4_span *ts = wc->tspans[chan->span->offset]; #endif #ifdef VPM_SUPPORT if (dtmfthreshold == 0) dtmfthreshold = VPM_DEFAULT_DTMFTHRESHOLD; if (lastdtmfthreshold != dtmfthreshold) { lastdtmfthreshold = dtmfthreshold; t4_vpm_set_dtmf_threshold(wc, dtmfthreshold); } #endif switch(cmd) { case WCT4_GET_REGS: for (x=0;xspan->offset, x); if (copy_to_user((struct t4_regs *)data, ®s, sizeof(regs))) return -EFAULT; break; #ifdef VPM_SUPPORT case ZT_TONEDETECT: if (get_user(j, (int *)data)) return -EFAULT; if (!wc->vpm) return -ENOSYS; if (j && (vpmdtmfsupport == 0)) return -ENOSYS; if (j & ZT_TONEDETECT_ON) set_bit(chan->chanpos - 1, &ts->dtmfmask); else clear_bit(chan->chanpos - 1, &ts->dtmfmask); if (j & ZT_TONEDETECT_MUTE) set_bit(chan->chanpos - 1, &ts->dtmfmutemask); else clear_bit(chan->chanpos - 1, &ts->dtmfmutemask); if (wc->vpm450m) { channel = (chan->chanpos) << 2; if (!wc->t1e1) channel += (4 << 2); channel |= chan->span->offset; vpm450m_setdtmf(wc->vpm450m, channel, j & ZT_TONEDETECT_ON, j & ZT_TONEDETECT_MUTE); } return 0; #endif default: return -ENOTTY; } return 0; } static void inline t4_hdlc_xmit_fifo(struct t4 *wc, unsigned int span, struct t4_span *ts) { int res, i, size = 32; unsigned char buf[32]; res = zt_hdlc_getbuf(ts->sigchan, buf, &size); if (debug & DEBUG_FRAMER) printk("Got buffer sized %d and res %d for %d\n", size, res, span); if (size > 0) { ts->sigactive = 1; if (debug & DEBUG_FRAMER) { printk("TX("); for (i = 0; i < size; i++) printk((i ? " %02x" : "%02x"), buf[i]); printk(")\n"); } for (i = 0; i < size; i++) t4_framer_out(wc, span, FRMR_TXFIFO, buf[i]); if (res) /* End of message */ { if (debug & DEBUG_FRAMER) printk("transmiting XHF|XME\n"); t4_framer_cmd_wait(wc, span, FRMR_CMDR_XHF | FRMR_CMDR_XME); #if 0 ts->sigactive = (__t4_framer_in(wc, span, FRMR_SIS) & FRMR_SIS_XFW) ? 0 : 1; #endif ++ts->frames_out; if ((debug & DEBUG_FRAMER) && !(ts->frames_out & 0x0f)) printk("Transmitted %d frames on span %d\n", ts->frames_out, span); } else { /* Still more to transmit */ if (debug & DEBUG_FRAMER) printk("transmiting XHF\n"); t4_framer_cmd_wait(wc, span, FRMR_CMDR_XHF); } } else if (res < 0) ts->sigactive = 0; } static void t4_hdlc_hard_xmit(struct zt_chan *chan) { struct t4 *wc = chan->pvt; int span = chan->span->offset; struct t4_span *ts = wc->tspans[span]; unsigned long flags; spin_lock_irqsave(&wc->reglock, flags); if (!ts->sigchan) { printk("t4_hdlc_hard_xmit: Invalid (NULL) signalling channel\n"); spin_unlock_irqrestore(&wc->reglock, flags); return; } spin_unlock_irqrestore(&wc->reglock, flags); if (debug & DEBUG_FRAMER) printk("t4_hdlc_hard_xmit on channel %s (sigchan %s), sigactive=%d\n", chan->name, ts->sigchan->name, ts->sigactive); if ((ts->sigchan == chan) && !ts->sigactive) t4_hdlc_xmit_fifo(wc, span, ts); } static int t4_maint(struct zt_span *span, int cmd) { struct t4_span *ts = span->pvt; struct t4 *wc = ts->owner; if (ts->spantype == TYPE_E1) { switch(cmd) { case ZT_MAINT_NONE: printk("XXX Turn off local and remote loops E1 XXX\n"); break; case ZT_MAINT_LOCALLOOP: printk("XXX Turn on local loopback E1 XXX\n"); break; case ZT_MAINT_REMOTELOOP: printk("XXX Turn on remote loopback E1 XXX\n"); break; case ZT_MAINT_LOOPUP: printk("XXX Send loopup code E1 XXX\n"); break; case ZT_MAINT_LOOPDOWN: printk("XXX Send loopdown code E1 XXX\n"); break; case ZT_MAINT_LOOPSTOP: printk("XXX Stop sending loop codes E1 XXX\n"); break; default: printk("TE%dXXP: Unknown E1 maint command: %d\n", wc->numspans, cmd); break; } } else { switch(cmd) { case ZT_MAINT_NONE: printk("XXX Turn off local and remote loops T1 XXX\n"); break; case ZT_MAINT_LOCALLOOP: printk("XXX Turn on local loop and no remote loop XXX\n"); break; case ZT_MAINT_REMOTELOOP: printk("XXX Turn on remote loopup XXX\n"); break; case ZT_MAINT_LOOPUP: t4_framer_out(wc, span->offset, 0x21, 0x50); /* FMR5: Nothing but RBS mode */ break; case ZT_MAINT_LOOPDOWN: t4_framer_out(wc, span->offset, 0x21, 0x60); /* FMR5: Nothing but RBS mode */ break; case ZT_MAINT_LOOPSTOP: t4_framer_out(wc, span->offset, 0x21, 0x40); /* FMR5: Nothing but RBS mode */ break; default: printk("TE%dXXP: Unknown T1 maint command: %d\n", wc->numspans, cmd); break; } } return 0; } static int t4_rbsbits(struct zt_chan *chan, int bits) { u_char m,c; int k,n,b; struct t4 *wc = chan->pvt; struct t4_span *ts = wc->tspans[chan->span->offset]; unsigned long flags; if(debug & DEBUG_RBS) printk("Setting bits to %d on channel %s\n", bits, chan->name); spin_lock_irqsave(&wc->reglock, flags); k = chan->span->offset; if (ts->spantype == TYPE_E1) { /* do it E1 way */ if (chan->chanpos == 16) { spin_unlock_irqrestore(&wc->reglock, flags); return 0; } n = chan->chanpos - 1; if (chan->chanpos > 15) n--; b = (n % 15); c = ts->txsigs[b]; m = (n / 15) << 2; /* nibble selector */ c &= (0xf << m); /* keep the other nibble */ c |= (bits & 0xf) << (4 - m); /* put our new nibble here */ ts->txsigs[b] = c; /* output them to the chip */ __t4_framer_out(wc,k,0x71 + b,c); } else if (ts->span.lineconfig & ZT_CONFIG_D4) { n = chan->chanpos - 1; b = (n/4); c = ts->txsigs[b]; m = ((3 - (n % 4)) << 1); /* nibble selector */ c &= ~(0x3 << m); /* keep the other nibble */ c |= ((bits >> 2) & 0x3) << m; /* put our new nibble here */ ts->txsigs[b] = c; /* output them to the chip */ __t4_framer_out(wc,k,0x70 + b,c); __t4_framer_out(wc,k,0x70 + b + 6,c); } else if (ts->span.lineconfig & ZT_CONFIG_ESF) { n = chan->chanpos - 1; b = (n/2); c = ts->txsigs[b]; m = ((n % 2) << 2); /* nibble selector */ c &= (0xf << m); /* keep the other nibble */ c |= (bits & 0xf) << (4 - m); /* put our new nibble here */ ts->txsigs[b] = c; /* output them to the chip */ __t4_framer_out(wc,k,0x70 + b,c); } spin_unlock_irqrestore(&wc->reglock, flags); if (debug & DEBUG_RBS) printk("Finished setting RBS bits\n"); return 0; } static int t4_shutdown(struct zt_span *span) { int tspan; int wasrunning; unsigned long flags; struct t4_span *ts = span->pvt; struct t4 *wc = ts->owner; tspan = span->offset + 1; if (tspan < 0) { printk("T%dXXP: Span '%d' isn't us?\n", wc->numspans, span->spanno); return -1; } if (debug & DEBUG_MAIN) printk("Shutting down span %d (%s)\n", span->spanno, span->name); /* Stop HDLC controller if runned */ if (ts->sigchan) hdlc_stop(wc, span->offset); spin_lock_irqsave(&wc->reglock, flags); wasrunning = span->flags & ZT_FLAG_RUNNING; span->flags &= ~ZT_FLAG_RUNNING; __t4_set_led(wc, span->offset, WC_OFF); if (((wc->numspans == 4) && (!(wc->tspans[0]->span.flags & ZT_FLAG_RUNNING)) && (!(wc->tspans[1]->span.flags & ZT_FLAG_RUNNING)) && (!(wc->tspans[2]->span.flags & ZT_FLAG_RUNNING)) && (!(wc->tspans[3]->span.flags & ZT_FLAG_RUNNING))) || ((wc->numspans == 2) && (!(wc->tspans[0]->span.flags & ZT_FLAG_RUNNING)) && (!(wc->tspans[1]->span.flags & ZT_FLAG_RUNNING)))) { /* No longer in use, disable interrupts */ printk("TE%dXXP: Disabling interrupts since there are no active spans\n", wc->numspans); set_bit(T4_STOP_DMA, &wc->checkflag); } else set_bit(T4_CHECK_TIMING, &wc->checkflag); spin_unlock_irqrestore(&wc->reglock, flags); /* Wait for interrupt routine to shut itself down */ msleep(10); if (wasrunning) wc->spansstarted--; if (debug & DEBUG_MAIN) printk("Span %d (%s) shutdown\n", span->spanno, span->name); return 0; } static int t4_spanconfig(struct zt_span *span, struct zt_lineconfig *lc) { int i; struct t4_span *ts = span->pvt; struct t4 *wc = ts->owner; printk("About to enter spanconfig!\n"); if (debug & DEBUG_MAIN) printk("TE%dXXP: Configuring span %d\n", wc->numspans, span->spanno); if (lc->sync < 0) lc->sync = 0; if (lc->sync > 4) lc->sync = 0; /* remove this span number from the current sync sources, if there */ for(i = 0; i < wc->numspans; i++) { if (wc->tspans[i]->sync == span->spanno) { wc->tspans[i]->sync = 0; wc->tspans[i]->psync = 0; } } wc->tspans[span->offset]->syncpos = lc->sync; /* if a sync src, put it in proper place */ if (lc->sync) { wc->tspans[lc->sync - 1]->sync = span->spanno; wc->tspans[lc->sync - 1]->psync = span->offset + 1; } set_bit(T4_CHECK_TIMING, &wc->checkflag); /* Make sure this is clear in case of multiple startup and shutdown * iterations */ clear_bit(T4_STOP_DMA, &wc->checkflag); /* If we're already running, then go ahead and apply the changes */ if (span->flags & ZT_FLAG_RUNNING) return t4_startup(span); printk("Done with spanconfig!\n"); return 0; } static int t4_chanconfig(struct zt_chan *chan, int sigtype) { int alreadyrunning; unsigned long flags; struct t4 *wc = chan->pvt; struct t4_span *ts = wc->tspans[chan->span->offset]; alreadyrunning = ts->span.flags & ZT_FLAG_RUNNING; if (debug & DEBUG_MAIN) { if (alreadyrunning) printk("TE%dXXP: Reconfigured channel %d (%s) sigtype %d\n", wc->numspans, chan->channo, chan->name, sigtype); else printk("TE%dXXP: Configured channel %d (%s) sigtype %d\n", wc->numspans, chan->channo, chan->name, sigtype); } spin_lock_irqsave(&wc->reglock, flags); if (alreadyrunning) __set_clear(wc, chan->span->offset); spin_unlock_irqrestore(&wc->reglock, flags); /* (re)configure signalling channel */ if ((sigtype == ZT_SIG_HARDHDLC) || (ts->sigchan == chan)) { if (debug & DEBUG_FRAMER) printk("%sonfiguring hardware HDLC on %s\n", ((sigtype == ZT_SIG_HARDHDLC) ? "C" : "Unc"), chan->name); if (alreadyrunning) { if (ts->sigchan) hdlc_stop(wc, ts->sigchan->span->offset); if (sigtype == ZT_SIG_HARDHDLC) { if (hdlc_start(wc, chan->span->offset, chan, ts->sigmode)) { printk("Error initializing signalling controller\n"); return -1; } } else { spin_lock_irqsave(&wc->reglock, flags); ts->sigchan = NULL; spin_unlock_irqrestore(&wc->reglock, flags); } } else { spin_lock_irqsave(&wc->reglock, flags); ts->sigchan = (sigtype == ZT_SIG_HARDHDLC) ? chan : NULL; spin_unlock_irqrestore(&wc->reglock, flags); ts->sigactive = 0; } } return 0; } static int t4_open(struct zt_chan *chan) { #ifndef LINUX26 MOD_INC_USE_COUNT; #else try_module_get(THIS_MODULE); #endif return 0; } static int t4_close(struct zt_chan *chan) { #ifndef LINUX26 MOD_DEC_USE_COUNT; #else module_put(THIS_MODULE); #endif return 0; } /* The number of cards we have seen with each possible 'order' switch setting. */ static unsigned int order_index[16]; static void init_spans(struct t4 *wc) { int x,y; int gen2; int offset = 1; struct t4_span *ts; gen2 = (wc->tspans[0]->spanflags & FLAG_2NDGEN); if (!wc->t1e1) offset += 4; for (x = 0; x < wc->numspans; x++) { ts = wc->tspans[x]; sprintf(ts->span.name, "TE%d/%d/%d", wc->numspans, wc->num, x + 1); snprintf(ts->span.desc, sizeof(ts->span.desc) - 1, "T%dXXP (PCI) Card %d Span %d", wc->numspans, wc->num, x+1); ts->span.manufacturer = "Digium"; zap_copy_string(ts->span.devicetype, wc->variety, sizeof(ts->span.devicetype)); if (wc->vpm == T4_VPM_PRESENT) { if (!wc->vpm450m) strncat(ts->span.devicetype, " with VPM400M", sizeof(ts->span.devicetype) - 1); else strncat(ts->span.devicetype, (wc->numspans > 2) ? " with VPMOCT128" : " with VPMOCT064", sizeof(ts->span.devicetype) - 1); } if (order_index[wc->order] == 1) snprintf(ts->span.location, sizeof(ts->span.location) - 1, "Board ID Switch %d", wc->order); else snprintf(ts->span.location, sizeof(ts->span.location) - 1, "PCI%s Bus %02d Slot %02d", (ts->spanflags & FLAG_EXPRESS) ? " Express" : " ", wc->dev->bus->number, PCI_SLOT(wc->dev->devfn) + 1); switch (ts->spantype) { case TYPE_T1: ts->span.spantype = "T1"; break; case TYPE_E1: ts->span.spantype = "E1"; break; case TYPE_J1: ts->span.spantype = "J1"; break; } ts->span.spanconfig = t4_spanconfig; ts->span.chanconfig = t4_chanconfig; ts->span.irq = wc->dev->irq; ts->span.startup = t4_startup; ts->span.shutdown = t4_shutdown; ts->span.rbsbits = t4_rbsbits; ts->span.maint = t4_maint; ts->span.open = t4_open; ts->span.close = t4_close; /* HDLC Specific init */ ts->sigchan = NULL; ts->sigmode = sigmode; ts->sigactive = 0; if (ts->spantype == TYPE_T1 || ts->spantype == TYPE_J1) { ts->span.channels = 24; ts->span.deflaw = ZT_LAW_MULAW; ts->span.linecompat = ZT_CONFIG_AMI | ZT_CONFIG_B8ZS | ZT_CONFIG_D4 | ZT_CONFIG_ESF; } else { ts->span.channels = 31; ts->span.deflaw = ZT_LAW_ALAW; ts->span.linecompat = ZT_CONFIG_HDB3 | ZT_CONFIG_CCS | ZT_CONFIG_CRC4; } ts->span.chans = ts->chans; ts->span.flags = ZT_FLAG_RBS; ts->span.ioctl = t4_ioctl; ts->span.hdlc_hard_xmit = t4_hdlc_hard_xmit; if (gen2) { #ifdef VPM_SUPPORT ts->span.echocan = t4_echocan; #endif ts->span.dacs = t4_dacs; } ts->span.pvt = ts; ts->owner = wc; ts->span.offset = x; ts->writechunk = (void *)(wc->writechunk + x * 32 * 2); ts->readchunk = (void *)(wc->readchunk + x * 32 * 2); init_waitqueue_head(&ts->span.maintq); for (y=0;ytspans[x]->span.channels;y++) { struct zt_chan *mychans = ts->chans + y; sprintf(mychans->name, "TE%d/%d/%d/%d", wc->numspans, wc->num, x + 1, y + 1); mychans->sigcap = ZT_SIG_EM | ZT_SIG_CLEAR | ZT_SIG_FXSLS | ZT_SIG_FXSGS | ZT_SIG_FXSKS | ZT_SIG_HARDHDLC | ZT_SIG_MTP2 | ZT_SIG_FXOLS | ZT_SIG_FXOGS | ZT_SIG_FXOKS | ZT_SIG_CAS | ZT_SIG_EM_E1 | ZT_SIG_DACS_RBS; mychans->pvt = wc; mychans->chanpos = y + 1; if (gen2) { mychans->writechunk = (void *)(wc->writechunk + (x * 32 + y + offset) * 2); mychans->readchunk = (void *)(wc->readchunk + (x * 32 + y + offset) * 2); } } } } static void t4_serial_setup(struct t4 *wc, int unit) { if (!wc->globalconfig) { wc->globalconfig = 1; printk("TE%dXXP: Setting up global serial parameters\n", wc->numspans); t4_framer_out(wc, 0, 0x85, 0xe0); /* GPC1: Multiplex mode enabled, FSC is output, active low, RCLK from channel 0 */ t4_framer_out(wc, 0, 0x08, 0x01); /* IPC: Interrupt push/pull active low */ /* Global clocks (8.192 Mhz CLK) */ t4_framer_out(wc, 0, 0x92, 0x00); t4_framer_out(wc, 0, 0x93, 0x18); t4_framer_out(wc, 0, 0x94, 0xfb); t4_framer_out(wc, 0, 0x95, 0x0b); t4_framer_out(wc, 0, 0x96, 0x00); t4_framer_out(wc, 0, 0x97, 0x0b); t4_framer_out(wc, 0, 0x98, 0xdb); t4_framer_out(wc, 0, 0x99, 0xdf); } /* Configure interrupts */ t4_framer_out(wc, unit, FRMR_GCR, 0x00); /* GCR: Interrupt on Activation/Deactivation of each */ /* Configure system interface */ t4_framer_out(wc, unit, FRMR_SIC1, 0xc2); /* SIC1: 8.192 Mhz clock/bus, double buffer receive / transmit, byte interleaved */ t4_framer_out(wc, unit, FRMR_SIC2, 0x20 | (unit << 1)); /* SIC2: No FFS, no center receive eliastic buffer, phase */ t4_framer_out(wc, unit, FRMR_SIC3, 0x04); /* SIC3: Edges for capture */ t4_framer_out(wc, unit, FRMR_CMR2, 0x00); /* CMR2: We provide sync and clock for tx and rx. */ if (!wc->t1e1) { /* T1 mode */ t4_framer_out(wc, unit, FRMR_XC0, 0x03); /* XC0: Normal operation of Sa-bits */ t4_framer_out(wc, unit, FRMR_XC1, 0x84); /* XC1: 0 offset */ if (wc->tspans[unit]->spantype == TYPE_J1) t4_framer_out(wc, unit, FRMR_RC0, 0x83); /* RC0: Just shy of 1023 */ else t4_framer_out(wc, unit, FRMR_RC0, 0x03); /* RC0: Just shy of 1023 */ t4_framer_out(wc, unit, FRMR_RC1, 0x84); /* RC1: The rest of RC0 */ } else { /* E1 mode */ t4_framer_out(wc, unit, FRMR_XC0, 0x00); /* XC0: Normal operation of Sa-bits */ t4_framer_out(wc, unit, FRMR_XC1, 0x04); /* XC1: 0 offset */ t4_framer_out(wc, unit, FRMR_RC0, 0x04); /* RC0: Just shy of 1023 */ t4_framer_out(wc, unit, FRMR_RC1, 0x04); /* RC1: The rest of RC0 */ } /* Configure ports */ t4_framer_out(wc, unit, 0x80, 0x00); /* PC1: SPYR/SPYX input on RPA/XPA */ t4_framer_out(wc, unit, 0x81, 0x22); /* PC2: RMFB/XSIG output/input on RPB/XPB */ t4_framer_out(wc, unit, 0x82, 0x65); /* PC3: Some unused stuff */ t4_framer_out(wc, unit, 0x83, 0x35); /* PC4: Some more unused stuff */ t4_framer_out(wc, unit, 0x84, 0x01); /* PC5: XMFS active low, SCLKR is input, RCLK is output */ if (debug & DEBUG_MAIN) printk("Successfully initialized serial bus for unit %d\n", unit); } static int syncsrc = 0; static int syncnum = 0 /* -1 */; static int syncspan = 0; #ifdef DEFINE_SPINLOCK static DEFINE_SPINLOCK(synclock); #else static spinlock_t synclock = SPIN_LOCK_UNLOCKED; #endif static void __t4_set_timing_source(struct t4 *wc, int unit, int master, int slave) { unsigned int timing; int x; if (unit != wc->syncsrc) { timing = 0x34; /* CMR1: RCLK unit, 8.192 Mhz TCLK, RCLK is 8.192 Mhz */ if ((unit > -1) && (unit < 4)) { timing |= (unit << 6); for (x=0;xnumspans;x++) /* set all 4 receive reference clocks to unit */ __t4_framer_out(wc, x, 0x44, timing); wc->dmactrl |= (1 << 29); } else { for (x=0;xnumspans;x++) /* set each receive reference clock to itself */ __t4_framer_out(wc, x, 0x44, timing | (x << 6)); wc->dmactrl &= ~(1 << 29); } if (slave) wc->dmactrl |= (1 << 25); else wc->dmactrl &= ~(1 << 25); if (master) wc->dmactrl |= (1 << 24); else wc->dmactrl &= ~(1 << 24); __t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); if (!master && !slave) wc->syncsrc = unit; if ((unit < 0) || (unit > 3)) unit = 0; else unit++; if (!master && !slave) { for (x=0;xnumspans;x++) wc->tspans[x]->span.syncsrc = unit; } } else { if (debug & DEBUG_MAIN) printk("TE%dXXP: Timing source already set to %d\n", wc->numspans, unit); } #if 0 printk("wct4xxp: Timing source set to %d\n",unit); #endif } static inline void __t4_update_timing(struct t4 *wc) { int i; /* update sync src info */ if (wc->syncsrc != syncsrc) { printk("Swapping card %d from %d to %d\n", wc->num, wc->syncsrc, syncsrc); wc->syncsrc = syncsrc; /* Update sync sources */ for (i = 0; i < wc->numspans; i++) { wc->tspans[i]->span.syncsrc = wc->syncsrc; } if (syncnum == wc->num) { __t4_set_timing_source(wc, syncspan-1, 1, 0); if (debug) printk("Card %d, using sync span %d, master\n", wc->num, syncspan); } else { __t4_set_timing_source(wc, syncspan-1, 0, 1); if (debug) printk("Card %d, using Timing Bus, NOT master\n", wc->num); } } } static int __t4_findsync(struct t4 *wc) { int i; int x; unsigned long flags; int p; int nonzero; int newsyncsrc = 0; /* Zaptel span number */ int newsyncnum = 0; /* wct4xxp card number */ int newsyncspan = 0; /* span on given wct4xxp card */ spin_lock_irqsave(&synclock, flags); #if 1 if (!wc->num) { /* If we're the first card, go through all the motions, up to 8 levels of sync source */ p = 1; while (p < 8) { nonzero = 0; for (x=0;cards[x];x++) { for (i = 0; i < wc->numspans; i++) { if (cards[x]->tspans[i]->syncpos) { nonzero = 1; if ((cards[x]->tspans[i]->syncpos == p) && !(cards[x]->tspans[i]->span.alarms & (ZT_ALARM_RED | ZT_ALARM_BLUE | ZT_ALARM_LOOPBACK)) && (cards[x]->tspans[i]->span.flags & ZT_FLAG_RUNNING)) { /* This makes a good sync source */ newsyncsrc = cards[x]->tspans[i]->span.spanno; newsyncnum = x; newsyncspan = i + 1; /* Jump out */ goto found; } } } } if (nonzero) p++; else break; } found: if ((syncnum != newsyncnum) || (syncsrc != newsyncsrc) || (newsyncspan != syncspan)) { if (debug) printk("New syncnum: %d (was %d), syncsrc: %d (was %d), syncspan: %d (was %d)\n", newsyncnum, syncnum, newsyncsrc, syncsrc, newsyncspan, syncspan); syncnum = newsyncnum; syncsrc = newsyncsrc; syncspan = newsyncspan; for (x=0;cards[x];x++) { __t4_update_timing(cards[x]); } } } else set_bit(T4_CHECK_TIMING, &cards[0]->checkflag); #endif spin_unlock_irqrestore(&synclock, flags); return 0; } static void __t4_set_timing_source_auto(struct t4 *wc) { int x; printk("timing source auto card %d!\n", wc->num); clear_bit(T4_CHECK_TIMING, &wc->checkflag); if (timingcable) { __t4_findsync(wc); } else { for (x=0;xnumspans;x++) { if (wc->tspans[x]->sync) { if ((wc->tspans[wc->tspans[x]->psync - 1]->span.flags & ZT_FLAG_RUNNING) && !(wc->tspans[wc->tspans[x]->psync - 1]->span.alarms & (ZT_ALARM_RED | ZT_ALARM_BLUE) )) { /* Valid timing source */ __t4_set_timing_source(wc, wc->tspans[x]->psync - 1, 0, 0); return; } } } __t4_set_timing_source(wc, 4, 0, 0); } } static void __t4_configure_t1(struct t4 *wc, int unit, int lineconfig, int txlevel) { unsigned int fmr4, fmr2, fmr1, fmr0, lim2; char *framing, *line; int mytxlevel; if ((txlevel > 7) || (txlevel < 4)) mytxlevel = 0; else mytxlevel = txlevel - 4; fmr1 = 0x9c; /* FMR1: Mode 1, T1 mode, CRC on for ESF, 8.192 Mhz system data rate, no XAIS */ fmr2 = 0x22; /* FMR2: no payload loopback, auto send yellow alarm */ if (loopback) fmr2 |= 0x4; fmr4 = 0x0c; /* FMR4: Lose sync on 2 out of 5 framing bits, auto resync */ lim2 = 0x21; /* LIM2: 50% peak is a "1", Advanced Loss recovery */ lim2 |= (mytxlevel << 6); /* LIM2: Add line buildout */ __t4_framer_out(wc, unit, 0x1d, fmr1); __t4_framer_out(wc, unit, 0x1e, fmr2); /* Configure line interface */ if (lineconfig & ZT_CONFIG_AMI) { line = "AMI"; fmr0 = 0xa0; } else { line = "B8ZS"; fmr0 = 0xf0; } if (lineconfig & ZT_CONFIG_D4) { framing = "D4"; } else { framing = "ESF"; fmr4 |= 0x2; fmr2 |= 0xc0; } __t4_framer_out(wc, unit, 0x1c, fmr0); __t4_framer_out(wc, unit, 0x20, fmr4); __t4_framer_out(wc, unit, 0x21, 0x40); /* FMR5: Enable RBS mode */ __t4_framer_out(wc, unit, 0x37, 0xf0 ); /* LIM1: Clear data in case of LOS, Set receiver threshold (0.5V), No remote loop, no DRS */ __t4_framer_out(wc, unit, 0x36, 0x08); /* LIM0: Enable auto long haul mode, no local loop (must be after LIM1) */ __t4_framer_out(wc, unit, 0x02, 0x50); /* CMDR: Reset the receiver and transmitter line interface */ __t4_framer_out(wc, unit, 0x02, 0x00); /* CMDR: Reset the receiver and transmitter line interface */ __t4_framer_out(wc, unit, 0x3a, lim2); /* LIM2: 50% peak amplitude is a "1" */ __t4_framer_out(wc, unit, 0x38, 0x0a); /* PCD: LOS after 176 consecutive "zeros" */ __t4_framer_out(wc, unit, 0x39, 0x15); /* PCR: 22 "ones" clear LOS */ /* Generate pulse mask for T1 */ switch(mytxlevel) { case 3: __t4_framer_out(wc, unit, 0x26, 0x07); /* XPM0 */ __t4_framer_out(wc, unit, 0x27, 0x01); /* XPM1 */ __t4_framer_out(wc, unit, 0x28, 0x00); /* XPM2 */ break; case 2: __t4_framer_out(wc, unit, 0x26, 0x8c); /* XPM0 */ __t4_framer_out(wc, unit, 0x27, 0x11); /* XPM1 */ __t4_framer_out(wc, unit, 0x28, 0x01); /* XPM2 */ break; case 1: __t4_framer_out(wc, unit, 0x26, 0x8c); /* XPM0 */ __t4_framer_out(wc, unit, 0x27, 0x01); /* XPM1 */ __t4_framer_out(wc, unit, 0x28, 0x00); /* XPM2 */ break; case 0: default: __t4_framer_out(wc, unit, 0x26, 0xd7); /* XPM0 */ __t4_framer_out(wc, unit, 0x27, 0x22); /* XPM1 */ __t4_framer_out(wc, unit, 0x28, 0x01); /* XPM2 */ break; } /* Don't mask framer interrupts if hardware HDLC is in use */ __t4_framer_out(wc, unit, FRMR_IMR0, 0xff & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR0_MASK : 0)); /* IMR0: We care about CAS changes, etc */ __t4_framer_out(wc, unit, FRMR_IMR1, 0xff & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR1_MASK : 0)); /* IMR1: We care about nothing */ __t4_framer_out(wc, unit, 0x16, 0x00); /* IMR2: We care about all the alarm stuff! */ __t4_framer_out(wc, unit, 0x17, 0xf4); /* IMR3: We care about AIS and friends */ __t4_framer_out(wc, unit, 0x18, 0x3f); /* IMR4: We care about slips on transmit */ printk("TE%dXXP: Span %d configured for %s/%s\n", wc->numspans, unit + 1, framing, line); } static void __t4_configure_e1(struct t4 *wc, int unit, int lineconfig) { unsigned int fmr2, fmr1, fmr0; unsigned int cas = 0; unsigned int imr3extra=0; char *crc4 = ""; char *framing, *line; fmr1 = 0x44; /* FMR1: E1 mode, Automatic force resync, PCM30 mode, 8.192 Mhz backplane, no XAIS */ fmr2 = 0x03; /* FMR2: Auto transmit remote alarm, auto loss of multiframe recovery, no payload loopback */ if (loopback) fmr2 |= 0x4; if (lineconfig & ZT_CONFIG_CRC4) { fmr1 |= 0x08; /* CRC4 transmit */ fmr2 |= 0xc0; /* CRC4 receive */ crc4 = "/CRC4"; } __t4_framer_out(wc, unit, 0x1d, fmr1); __t4_framer_out(wc, unit, 0x1e, fmr2); /* Configure line interface */ if (lineconfig & ZT_CONFIG_AMI) { line = "AMI"; fmr0 = 0xa0; } else { line = "HDB3"; fmr0 = 0xf0; } if (lineconfig & ZT_CONFIG_CCS) { framing = "CCS"; imr3extra = 0x28; } else { framing = "CAS"; cas = 0x40; } __t4_framer_out(wc, unit, 0x1c, fmr0); __t4_framer_out(wc, unit, 0x37, 0xf0 /*| 0x6 */ ); /* LIM1: Clear data in case of LOS, Set receiver threshold (0.5V), No remote loop, no DRS */ __t4_framer_out(wc, unit, 0x36, 0x08); /* LIM0: Enable auto long haul mode, no local loop (must be after LIM1) */ __t4_framer_out(wc, unit, 0x02, 0x50); /* CMDR: Reset the receiver and transmitter line interface */ __t4_framer_out(wc, unit, 0x02, 0x00); /* CMDR: Reset the receiver and transmitter line interface */ /* Condition receive line interface for E1 after reset */ __t4_framer_out(wc, unit, 0xbb, 0x17); __t4_framer_out(wc, unit, 0xbc, 0x55); __t4_framer_out(wc, unit, 0xbb, 0x97); __t4_framer_out(wc, unit, 0xbb, 0x11); __t4_framer_out(wc, unit, 0xbc, 0xaa); __t4_framer_out(wc, unit, 0xbb, 0x91); __t4_framer_out(wc, unit, 0xbb, 0x12); __t4_framer_out(wc, unit, 0xbc, 0x55); __t4_framer_out(wc, unit, 0xbb, 0x92); __t4_framer_out(wc, unit, 0xbb, 0x0c); __t4_framer_out(wc, unit, 0xbb, 0x00); __t4_framer_out(wc, unit, 0xbb, 0x8c); __t4_framer_out(wc, unit, 0x3a, 0x20); /* LIM2: 50% peak amplitude is a "1" */ __t4_framer_out(wc, unit, 0x38, 0x0a); /* PCD: LOS after 176 consecutive "zeros" */ __t4_framer_out(wc, unit, 0x39, 0x15); /* PCR: 22 "ones" clear LOS */ __t4_framer_out(wc, unit, 0x20, 0x9f); /* XSW: Spare bits all to 1 */ __t4_framer_out(wc, unit, 0x21, 0x1c|cas); /* XSP: E-bit set when async. AXS auto, XSIF to 1 */ /* Generate pulse mask for E1 */ __t4_framer_out(wc, unit, 0x26, 0x54); /* XPM0 */ __t4_framer_out(wc, unit, 0x27, 0x02); /* XPM1 */ __t4_framer_out(wc, unit, 0x28, 0x00); /* XPM2 */ /* Don't mask framer interrupts if hardware HDLC is in use */ __t4_framer_out(wc, unit, FRMR_IMR0, 0xff & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR0_MASK : 0)); /* IMR0: We care about CRC errors, CAS changes, etc */ __t4_framer_out(wc, unit, FRMR_IMR1, 0x3f & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR1_MASK : 0)); /* IMR1: We care about loopup / loopdown */ __t4_framer_out(wc, unit, 0x16, 0x00); /* IMR2: We care about all the alarm stuff! */ __t4_framer_out(wc, unit, 0x17, 0xc4 | imr3extra); /* IMR3: We care about AIS and friends */ __t4_framer_out(wc, unit, 0x18, 0x3f); /* IMR4: We care about slips on transmit */ printk("TE%dXXP: Span %d configured for %s/%s%s\n", wc->numspans, unit + 1, framing, line, crc4); } static int t4_startup(struct zt_span *span) { #ifdef SUPPORT_GEN1 int i; #endif int tspan; unsigned long flags; int alreadyrunning; struct t4_span *ts = span->pvt; struct t4 *wc = ts->owner; printk("About to enter startup!\n"); tspan = span->offset + 1; if (tspan < 0) { printk("TE%dXXP: Span '%d' isn't us?\n", wc->numspans, span->spanno); return -1; } spin_lock_irqsave(&wc->reglock, flags); alreadyrunning = span->flags & ZT_FLAG_RUNNING; #ifdef SUPPORT_GEN1 /* initialize the start value for the entire chunk of last ec buffer */ for(i = 0; i < span->channels; i++) { memset(ts->ec_chunk1[i], ZT_LIN2X(0,&span->chans[i]),ZT_CHUNKSIZE); memset(ts->ec_chunk2[i], ZT_LIN2X(0,&span->chans[i]),ZT_CHUNKSIZE); } #endif /* Force re-evaluation fo timing source */ if (timingcable) wc->syncsrc = -1; if (ts->spantype == TYPE_E1) { /* if this is an E1 card */ __t4_configure_e1(wc, span->offset, span->lineconfig); } else { /* is a T1 card */ __t4_configure_t1(wc, span->offset, span->lineconfig, span->txlevel); } /* Note clear channel status */ wc->tspans[span->offset]->notclear = 0; __set_clear(wc, span->offset); if (!alreadyrunning) { span->flags |= ZT_FLAG_RUNNING; wc->spansstarted++; /* enable interrupts */ /* Start DMA, enabling DMA interrupts on read only */ wc->dmactrl = 1 << 29; #if 0 /* Enable framer only interrupts */ wc->dmactrl |= 1 << 27; #endif wc->dmactrl |= (ts->spanflags & FLAG_2NDGEN) ? 0xc0000000 : 0xc0000003; #ifdef VPM_SUPPORT wc->dmactrl |= wc->vpm; #endif /* Seed interrupt register */ __t4_pci_out(wc, WC_INTR, 0x0c); if (noburst && !(ts->spanflags & FLAG_BURST)) wc->dmactrl |= (1 << 26); __t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); /* Startup HDLC controller too */ } if (ts->sigchan) { struct zt_chan *sigchan = ts->sigchan; spin_unlock_irqrestore(&wc->reglock, flags); if (hdlc_start(wc, span->offset, sigchan, ts->sigmode)) { printk("Error initializing signalling controller\n"); return -1; } spin_lock_irqsave(&wc->reglock, flags); } spin_unlock_irqrestore(&wc->reglock, flags); t4_check_alarms(wc, span->offset); t4_check_sigbits(wc, span->offset); if (wc->tspans[0]->sync == span->spanno) printk("SPAN %d: Primary Sync Source\n",span->spanno); if (wc->tspans[1]->sync == span->spanno) printk("SPAN %d: Secondary Sync Source\n",span->spanno); if (wc->numspans == 4) { if (wc->tspans[2]->sync == span->spanno) printk("SPAN %d: Tertiary Sync Source\n",span->spanno); if (wc->tspans[3]->sync == span->spanno) printk("SPAN %d: Quaternary Sync Source\n",span->spanno); } #ifdef VPM_SUPPORT if (!alreadyrunning && !wc->vpm) { wait_a_little(); t4_vpm400_init(wc); if (!wc->vpm) t4_vpm450_init(wc); wc->dmactrl |= wc->vpm; t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); } #endif printk("Completed startup!\n"); return 0; } #ifdef SUPPORT_GEN1 static inline void e1_check(struct t4 *wc, int span, int val) { struct t4_span *ts = wc->tspans[span]; if ((ts->span.channels > 24) && (ts->span.flags & ZT_FLAG_RUNNING) && !(ts->span.alarms) && (!wc->e1recover)) { if (val != 0x1b) { ts->e1check++; } else ts->e1check = 0; if (ts->e1check > 100) { /* Wait 1000 ms */ wc->e1recover = 1000 * 8; wc->tspans[0]->e1check = wc->tspans[1]->e1check = 0; if (wc->numspans == 4) wc->tspans[2]->e1check = wc->tspans[3]->e1check = 0; if (debug & DEBUG_MAIN) printk("Detected loss of E1 alignment on span %d!\n", span); t4_reset_dma(wc); } } } static void t4_receiveprep(struct t4 *wc, int irq) { volatile unsigned int *readchunk; int dbl = 0; int x,y,z; unsigned int tmp; int offset=0; if (!wc->t1e1) offset = 4; if (irq & 1) { /* First part */ readchunk = wc->readchunk; if (!wc->last0) dbl = 1; wc->last0 = 0; } else { readchunk = wc->readchunk + ZT_CHUNKSIZE * 32; if (wc->last0) dbl = 1; wc->last0 = 1; } if (dbl) { for (x=0;xnumspans;x++) wc->tspans[x]->irqmisses++; if (debug & DEBUG_MAIN) printk("TE%dXXP: Double/missed interrupt detected\n", wc->numspans); } for (x=0;xnumspans == 4) { wc->tspans[3]->span.chans[z].readchunk[x] = tmp & 0xff; wc->tspans[2]->span.chans[z].readchunk[x] = (tmp & 0xff00) >> 8; } wc->tspans[1]->span.chans[z].readchunk[x] = (tmp & 0xff0000) >> 16; wc->tspans[0]->span.chans[z].readchunk[x] = tmp >> 24; } if (wc->t1e1) { if (wc->e1recover > 0) wc->e1recover--; tmp = readchunk[0]; if (wc->numspans == 4) { e1_check(wc, 3, (tmp & 0x7f)); e1_check(wc, 2, (tmp & 0x7f00) >> 8); } e1_check(wc, 1, (tmp & 0x7f0000) >> 16); e1_check(wc, 0, (tmp & 0x7f000000) >> 24); for (z=24;z<31;z++) { /* Only E1 channels now */ tmp = readchunk[z+1]; if (wc->numspans == 4) { if (wc->tspans[3]->span.channels > 24) wc->tspans[3]->span.chans[z].readchunk[x] = tmp & 0xff; if (wc->tspans[2]->span.channels > 24) wc->tspans[2]->span.chans[z].readchunk[x] = (tmp & 0xff00) >> 8; } if (wc->tspans[1]->span.channels > 24) wc->tspans[1]->span.chans[z].readchunk[x] = (tmp & 0xff0000) >> 16; if (wc->tspans[0]->span.channels > 24) wc->tspans[0]->span.chans[z].readchunk[x] = tmp >> 24; } } /* Advance pointer by 4 TDM frame lengths */ readchunk += 32; } for (x=0;xnumspans;x++) { if (wc->tspans[x]->span.flags & ZT_FLAG_RUNNING) { for (y=0;ytspans[x]->span.channels;y++) { /* Echo cancel double buffered data */ zt_ec_chunk(&wc->tspans[x]->span.chans[y], wc->tspans[x]->span.chans[y].readchunk, wc->tspans[x]->ec_chunk2[y]); memcpy(wc->tspans[x]->ec_chunk2[y],wc->tspans[x]->ec_chunk1[y], ZT_CHUNKSIZE); memcpy(wc->tspans[x]->ec_chunk1[y], wc->tspans[x]->span.chans[y].writechunk, ZT_CHUNKSIZE); } zt_receive(&wc->tspans[x]->span); } } } #endif #if (ZT_CHUNKSIZE != 8) #error Sorry, nextgen does not support chunksize != 8 #endif static inline void __receive_span(struct t4_span *ts) { #ifdef VPM_SUPPORT int y; unsigned long merged; merged = ts->dtmfactive & ts->dtmfmutemask; if (merged) { for (y=0;yspan.channels;y++) { /* Mute any DTMFs which are supposed to be muted */ if (test_bit(y, &merged)) { memset(ts->span.chans[y].readchunk, ZT_XLAW(0, (ts->span.chans + y)), ZT_CHUNKSIZE); } } } #endif #ifdef ENABLE_PREFETCH prefetch((void *)(ts->readchunk)); prefetch((void *)(ts->writechunk)); prefetch((void *)(ts->readchunk + 8)); prefetch((void *)(ts->writechunk + 8)); prefetch((void *)(ts->readchunk + 16)); prefetch((void *)(ts->writechunk + 16)); prefetch((void *)(ts->readchunk + 24)); prefetch((void *)(ts->writechunk + 24)); prefetch((void *)(ts->readchunk + 32)); prefetch((void *)(ts->writechunk + 32)); prefetch((void *)(ts->readchunk + 40)); prefetch((void *)(ts->writechunk + 40)); prefetch((void *)(ts->readchunk + 48)); prefetch((void *)(ts->writechunk + 48)); prefetch((void *)(ts->readchunk + 56)); prefetch((void *)(ts->writechunk + 56)); #endif zt_ec_span(&ts->span); zt_receive(&ts->span); } static inline void __transmit_span(struct t4_span *ts) { zt_transmit(&ts->span); } #ifdef ENABLE_WORKQUEUES static void workq_handlespan(void *data) { struct t4_span *ts = data; struct t4 *wc = ts->owner; __receive_span(ts); __transmit_span(ts); atomic_dec(&wc->worklist); if (!atomic_read(&wc->worklist)) t4_pci_out(wc, WC_INTR, 0); } #else static void t4_prep_gen2(struct t4 *wc) { int x; for (x=0;xnumspans;x++) { if (wc->tspans[x]->span.flags & ZT_FLAG_RUNNING) { __receive_span(wc->tspans[x]); __transmit_span(wc->tspans[x]); } } } #endif #ifdef SUPPORT_GEN1 static void t4_transmitprep(struct t4 *wc, int irq) { volatile unsigned int *writechunk; int x,y,z; unsigned int tmp; int offset=0; if (!wc->t1e1) offset = 4; if (irq & 1) { /* First part */ writechunk = wc->writechunk + 1; } else { writechunk = wc->writechunk + ZT_CHUNKSIZE * 32 + 1; } for (y=0;ynumspans;y++) { if (wc->tspans[y]->span.flags & ZT_FLAG_RUNNING) zt_transmit(&wc->tspans[y]->span); } for (x=0;xtspans[3]->span.chans[z].writechunk[x]) | (wc->tspans[2]->span.chans[z].writechunk[x] << 8) | (wc->tspans[1]->span.chans[z].writechunk[x] << 16) | (wc->tspans[0]->span.chans[z].writechunk[x] << 24); writechunk[z+offset] = tmp; } if (wc->t1e1) { for (z=24;z<31;z++) { /* Only E1 channels now */ tmp = 0; if (wc->numspans == 4) { if (wc->tspans[3]->span.channels > 24) tmp |= wc->tspans[3]->span.chans[z].writechunk[x]; if (wc->tspans[2]->span.channels > 24) tmp |= (wc->tspans[2]->span.chans[z].writechunk[x] << 8); } if (wc->tspans[1]->span.channels > 24) tmp |= (wc->tspans[1]->span.chans[z].writechunk[x] << 16); if (wc->tspans[0]->span.channels > 24) tmp |= (wc->tspans[0]->span.chans[z].writechunk[x] << 24); writechunk[z] = tmp; } } /* Advance pointer by 4 TDM frame lengths */ writechunk += 32; } } #endif static void t4_check_sigbits(struct t4 *wc, int span) { int a,i,rxs; struct t4_span *ts = wc->tspans[span]; if (debug & DEBUG_RBS) printk("Checking sigbits on span %d\n", span + 1); if (!(ts->span.flags & ZT_FLAG_RUNNING)) return; if (ts->spantype == TYPE_E1) { for (i = 0; i < 15; i++) { a = t4_framer_in(wc, span, 0x71 + i); /* Get high channel in low bits */ rxs = (a & 0xf); if (!(ts->span.chans[i+16].sig & ZT_SIG_CLEAR)) { if (ts->span.chans[i+16].rxsig != rxs) zt_rbsbits(&ts->span.chans[i+16], rxs); } rxs = (a >> 4) & 0xf; if (!(ts->span.chans[i].sig & ZT_SIG_CLEAR)) { if (ts->span.chans[i].rxsig != rxs) zt_rbsbits(&ts->span.chans[i], rxs); } } } else if (ts->span.lineconfig & ZT_CONFIG_D4) { for (i = 0; i < 24; i+=4) { a = t4_framer_in(wc, span, 0x70 + (i>>2)); /* Get high channel in low bits */ rxs = (a & 0x3) << 2; if (!(ts->span.chans[i+3].sig & ZT_SIG_CLEAR)) { if (ts->span.chans[i+3].rxsig != rxs) zt_rbsbits(&ts->span.chans[i+3], rxs); } rxs = (a & 0xc); if (!(ts->span.chans[i+2].sig & ZT_SIG_CLEAR)) { if (ts->span.chans[i+2].rxsig != rxs) zt_rbsbits(&ts->span.chans[i+2], rxs); } rxs = (a >> 2) & 0xc; if (!(ts->span.chans[i+1].sig & ZT_SIG_CLEAR)) { if (ts->span.chans[i+1].rxsig != rxs) zt_rbsbits(&ts->span.chans[i+1], rxs); } rxs = (a >> 4) & 0xc; if (!(ts->span.chans[i].sig & ZT_SIG_CLEAR)) { if (ts->span.chans[i].rxsig != rxs) zt_rbsbits(&ts->span.chans[i], rxs); } } } else { for (i = 0; i < 24; i+=2) { a = t4_framer_in(wc, span, 0x70 + (i>>1)); /* Get high channel in low bits */ rxs = (a & 0xf); if (!(ts->span.chans[i+1].sig & ZT_SIG_CLEAR)) { /* XXX Not really reset on every trans! XXX */ if (ts->span.chans[i+1].rxsig != rxs) { zt_rbsbits(&ts->span.chans[i+1], rxs); } } rxs = (a >> 4) & 0xf; if (!(ts->span.chans[i].sig & ZT_SIG_CLEAR)) { /* XXX Not really reset on every trans! XXX */ if (ts->span.chans[i].rxsig != rxs) { zt_rbsbits(&ts->span.chans[i], rxs); } } } } } static void t4_check_alarms(struct t4 *wc, int span) { unsigned char c,d; int alarms; int x,j; struct t4_span *ts = wc->tspans[span]; unsigned long flags; if (!(ts->span.flags & ZT_FLAG_RUNNING)) return; spin_lock_irqsave(&wc->reglock, flags); c = __t4_framer_in(wc, span, 0x4c); d = __t4_framer_in(wc, span, 0x4d); /* Assume no alarms */ alarms = 0; /* And consider only carrier alarms */ ts->span.alarms &= (ZT_ALARM_RED | ZT_ALARM_BLUE | ZT_ALARM_NOTOPEN); if (ts->spantype == TYPE_E1) { if (c & 0x04) { /* No multiframe found, force RAI high after 400ms only if we haven't found a multiframe since last loss of frame */ if (!(ts->spanflags & FLAG_NMF)) { __t4_framer_out(wc, span, 0x20, 0x9f | 0x20); /* LIM0: Force RAI High */ ts->spanflags |= FLAG_NMF; printk("NMF workaround on!\n"); } __t4_framer_out(wc, span, 0x1e, 0xc3); /* Reset to CRC4 mode */ __t4_framer_out(wc, span, 0x1c, 0xf2); /* Force Resync */ __t4_framer_out(wc, span, 0x1c, 0xf0); /* Force Resync */ } else if (!(c & 0x02)) { if ((ts->spanflags & FLAG_NMF)) { __t4_framer_out(wc, span, 0x20, 0x9f); /* LIM0: Clear forced RAI */ ts->spanflags &= ~FLAG_NMF; printk("NMF workaround off!\n"); } } } else { /* Detect loopup code if we're not sending one */ if ((!ts->span.mainttimer) && (d & 0x08)) { /* Loop-up code detected */ if ((ts->loopupcnt++ > 80) && (ts->span.maintstat != ZT_MAINT_REMOTELOOP)) { __t4_framer_out(wc, span, 0x36, 0x08); /* LIM0: Disable any local loop */ __t4_framer_out(wc, span, 0x37, 0xf6 ); /* LIM1: Enable remote loop */ ts->span.maintstat = ZT_MAINT_REMOTELOOP; } } else ts->loopupcnt = 0; /* Same for loopdown code */ if ((!ts->span.mainttimer) && (d & 0x10)) { /* Loop-down code detected */ if ((ts->loopdowncnt++ > 80) && (ts->span.maintstat == ZT_MAINT_REMOTELOOP)) { __t4_framer_out(wc, span, 0x36, 0x08); /* LIM0: Disable any local loop */ __t4_framer_out(wc, span, 0x37, 0xf0 ); /* LIM1: Disable remote loop */ ts->span.maintstat = ZT_MAINT_NONE; } } else ts->loopdowncnt = 0; } if (ts->span.lineconfig & ZT_CONFIG_NOTOPEN) { for (x=0,j=0;x < ts->span.channels;x++) if ((ts->span.chans[x].flags & ZT_FLAG_OPEN) || (ts->span.chans[x].flags & ZT_FLAG_NETDEV)) j++; if (!j) alarms |= ZT_ALARM_NOTOPEN; } if (c & 0xa0) { if (ts->alarmcount >= alarmdebounce) alarms |= ZT_ALARM_RED; else ts->alarmcount++; } else ts->alarmcount = 0; if (c & 0x4) alarms |= ZT_ALARM_BLUE; if (((!ts->span.alarms) && alarms) || (ts->span.alarms && (!alarms))) set_bit(T4_CHECK_TIMING, &wc->checkflag); /* Keep track of recovering */ if ((!alarms) && ts->span.alarms) ts->alarmtimer = ZT_ALARMSETTLE_TIME; if (ts->alarmtimer) alarms |= ZT_ALARM_RECOVER; /* If receiving alarms, go into Yellow alarm state */ if (alarms && !(ts->spanflags & FLAG_SENDINGYELLOW)) { unsigned char fmr4; #if 1 printk("wct%dxxp: Setting yellow alarm on span %d\n", wc->numspans, span + 1); #endif /* We manually do yellow alarm to handle RECOVER and NOTOPEN, otherwise it's auto anyway */ fmr4 = __t4_framer_in(wc, span, 0x20); __t4_framer_out(wc, span, 0x20, fmr4 | 0x20); ts->spanflags |= FLAG_SENDINGYELLOW; } else if ((!alarms) && (ts->spanflags & FLAG_SENDINGYELLOW)) { unsigned char fmr4; #if 1 printk("wct%dxxp: Clearing yellow alarm on span %d\n", wc->numspans, span + 1); #endif /* We manually do yellow alarm to handle RECOVER */ fmr4 = __t4_framer_in(wc, span, 0x20); __t4_framer_out(wc, span, 0x20, fmr4 & ~0x20); ts->spanflags &= ~FLAG_SENDINGYELLOW; } /* Re-check the timing source when we enter/leave alarm, not withstanding yellow alarm */ if (c & 0x10) alarms |= ZT_ALARM_YELLOW; if (ts->span.mainttimer || ts->span.maintstat) alarms |= ZT_ALARM_LOOPBACK; ts->span.alarms = alarms; spin_unlock_irqrestore(&wc->reglock, flags); zt_alarm_notify(&ts->span); } static void t4_do_counters(struct t4 *wc) { int span; for (span=0;spannumspans;span++) { struct t4_span *ts = wc->tspans[span]; int docheck=0; spin_lock(&wc->reglock); if (ts->loopupcnt || ts->loopdowncnt) docheck++; if (ts->alarmtimer) { if (!--ts->alarmtimer) { docheck++; ts->span.alarms &= ~(ZT_ALARM_RECOVER); } } spin_unlock(&wc->reglock); if (docheck) { t4_check_alarms(wc, span); zt_alarm_notify(&ts->span); } } } static inline void __handle_leds(struct t4 *wc) { int x; wc->blinktimer++; for (x=0;xnumspans;x++) { struct t4_span *ts = wc->tspans[x]; if (ts->span.flags & ZT_FLAG_RUNNING) { if (ts->span.alarms & (ZT_ALARM_RED | ZT_ALARM_BLUE)) { #ifdef FANCY_ALARM if (wc->blinktimer == (altab[wc->alarmpos] >> 1)) { __t4_set_led(wc, x, WC_RED); } if (wc->blinktimer == 0xf) { __t4_set_led(wc, x, WC_OFF); } #else if (wc->blinktimer == 160) { __t4_set_led(wc, x, WC_RED); } else if (wc->blinktimer == 480) { __t4_set_led(wc, x, WC_OFF); } #endif } else if (ts->span.alarms & ZT_ALARM_YELLOW) { /* Yellow Alarm */ __t4_set_led(wc, x, WC_YELLOW); } else if (ts->span.mainttimer || ts->span.maintstat) { #ifdef FANCY_ALARM if (wc->blinktimer == (altab[wc->alarmpos] >> 1)) { __t4_set_led(wc, x, WC_GREEN); } if (wc->blinktimer == 0xf) { __t4_set_led(wc, x, WC_OFF); } #else if (wc->blinktimer == 160) { __t4_set_led(wc, x, WC_GREEN); } else if (wc->blinktimer == 480) { __t4_set_led(wc, x, WC_OFF); } #endif } else { /* No Alarm */ __t4_set_led(wc, x, WC_GREEN); } } else __t4_set_led(wc, x, WC_OFF); } #ifdef FANCY_ALARM if (wc->blinktimer == 0xf) { wc->blinktimer = -1; wc->alarmpos++; if (wc->alarmpos >= (sizeof(altab) / sizeof(altab[0]))) wc->alarmpos = 0; } #else if (wc->blinktimer == 480) wc->blinktimer = 0; #endif } static inline void t4_framer_interrupt(struct t4 *wc, int span) { /* Check interrupts for a given span */ unsigned char gis, isr0, isr1, isr2, isr3, isr4; int readsize = -1; struct t4_span *ts = wc->tspans[span]; struct zt_chan *sigchan; unsigned long flags; if (debug & DEBUG_FRAMER) printk("framer interrupt span %d:%d!\n", wc->num, span + 1); /* 1st gen cards isn't used interrupts */ gis = t4_framer_in(wc, span, FRMR_GIS); isr0 = (gis & FRMR_GIS_ISR0) ? t4_framer_in(wc, span, FRMR_ISR0) : 0; isr1 = (gis & FRMR_GIS_ISR1) ? t4_framer_in(wc, span, FRMR_ISR1) : 0; isr2 = (gis & FRMR_GIS_ISR2) ? t4_framer_in(wc, span, FRMR_ISR2) : 0; isr3 = (gis & FRMR_GIS_ISR3) ? t4_framer_in(wc, span, FRMR_ISR3) : 0; isr4 = (gis & FRMR_GIS_ISR4) ? t4_framer_in(wc, span, FRMR_ISR4) : 0; if (debug & DEBUG_FRAMER) printk("gis: %02x, isr0: %02x, isr1: %02x, isr2: %02x, isr3: %02x, isr4: %02x\n", gis, isr0, isr1, isr2, isr3, isr4); if (isr0) t4_check_sigbits(wc, span); if (ts->spantype == TYPE_E1) { /* E1 checks */ if ((isr3 & 0x38) || isr2 || isr1) t4_check_alarms(wc, span); } else { /* T1 checks */ if (isr2 || (isr3 & 0x08)) t4_check_alarms(wc, span); } if (!ts->span.alarms) { if ((isr3 & 0x3) || (isr4 & 0xc0)) ts->span.timingslips++; if (debug & DEBUG_MAIN) { if (isr3 & 0x02) printk("TE%d10P: RECEIVE slip NEGATIVE on span %d\n", wc->numspans, span + 1); if (isr3 & 0x01) printk("TE%d10P: RECEIVE slip POSITIVE on span %d\n", wc->numspans, span + 1); if (isr4 & 0x80) printk("TE%dXXP: TRANSMIT slip POSITIVE on span %d\n", wc->numspans, span + 1); if (isr4 & 0x40) printk("TE%d10P: TRANSMIT slip NEGATIVE on span %d\n", wc->numspans, span + 1); } } else ts->span.timingslips = 0; spin_lock_irqsave(&wc->reglock, flags); /* HDLC controller checks - receive side */ if (!ts->sigchan) { spin_unlock_irqrestore(&wc->reglock, flags); return; } sigchan = ts->sigchan; spin_unlock_irqrestore(&wc->reglock, flags); if (isr0 & FRMR_ISR0_RME) { readsize = (t4_framer_in(wc, span, FRMR_RBCH) << 8) | t4_framer_in(wc, span, FRMR_RBCL); if (debug & DEBUG_FRAMER) printk("Received data length is %d (%d)\n", readsize, readsize & FRMR_RBCL_MAX_SIZE); /* RPF isn't set on last part of frame */ if ((readsize > 0) && ((readsize &= FRMR_RBCL_MAX_SIZE) == 0)) readsize = 32; } else if (isr0 & FRMR_ISR0_RPF) readsize = 32; if (readsize > 0) { int i; unsigned char readbuf[readsize]; if (debug & DEBUG_FRAMER) printk("Framer %d: Got RPF/RME! readsize is %d\n", sigchan->span->offset, readsize); for (i = 0; i < readsize; i++) readbuf[i] = t4_framer_in(wc, span, FRMR_RXFIFO); /* Tell the framer to clear the RFIFO */ t4_framer_cmd_wait(wc, span, FRMR_CMDR_RMC); if (debug & DEBUG_FRAMER) { printk("RX("); for (i = 0; i < readsize; i++) printk((i ? " %02x" : "%02x"), readbuf[i]); printk(")\n"); } if (isr0 & FRMR_ISR0_RME) { /* Do checks for HDLC problems */ unsigned char rsis = readbuf[readsize-1]; #if 0 unsigned int olddebug = debug; #endif unsigned char rsis_reg = t4_framer_in(wc, span, FRMR_RSIS); #if 0 if ((rsis != 0xA2) || (rsis != rsis_reg)) debug |= DEBUG_FRAMER; #endif ++ts->frames_in; if ((debug & DEBUG_FRAMER) && !(ts->frames_in & 0x0f)) printk("Received %d frames on span %d\n", ts->frames_in, span); if (debug & DEBUG_FRAMER) printk("Received HDLC frame %d. RSIS = 0x%x (%x)\n", ts->frames_in, rsis, rsis_reg); if (!(rsis & FRMR_RSIS_CRC16)) { if (debug & DEBUG_FRAMER) printk("CRC check failed %d\n", span); zt_hdlc_abort(sigchan, ZT_EVENT_BADFCS); } else if (rsis & FRMR_RSIS_RAB) { if (debug & DEBUG_FRAMER) printk("ABORT of current frame due to overflow %d\n", span); zt_hdlc_abort(sigchan, ZT_EVENT_ABORT); } else if (rsis & FRMR_RSIS_RDO) { if (debug & DEBUG_FRAMER) printk("HDLC overflow occured %d\n", span); zt_hdlc_abort(sigchan, ZT_EVENT_OVERRUN); } else if (!(rsis & FRMR_RSIS_VFR)) { if (debug & DEBUG_FRAMER) printk("Valid Frame check failed on span %d\n", span); zt_hdlc_abort(sigchan, ZT_EVENT_ABORT); } else { zt_hdlc_putbuf(sigchan, readbuf, readsize - 1); zt_hdlc_finish(sigchan); if (debug & DEBUG_FRAMER) printk("Received valid HDLC frame on span %d\n", span); } #if 0 debug = olddebug; #endif } else if (isr0 & FRMR_ISR0_RPF) zt_hdlc_putbuf(sigchan, readbuf, readsize); } /* Transmit side */ if (isr1 & FRMR_ISR1_XDU) { if (debug & DEBUG_FRAMER) printk("XDU: Resetting signal controler!\n"); t4_framer_cmd_wait(wc, span, FRMR_CMDR_SRES); } else if (isr1 & FRMR_ISR1_XPR) { if (debug & DEBUG_FRAMER) printk("Sigchan %d is %p\n", sigchan->chanpos, sigchan); if (debug & DEBUG_FRAMER) printk("Framer %d: Got XPR!\n", sigchan->span->offset); t4_hdlc_xmit_fifo(wc, span, ts); } if (isr1 & FRMR_ISR1_ALLS) { if (debug & DEBUG_FRAMER) printk("ALLS received\n"); } } #ifdef SUPPORT_GEN1 ZAP_IRQ_HANDLER(t4_interrupt) { struct t4 *wc = dev_id; unsigned long flags; int x; unsigned int status; unsigned int status2; #if 0 if (wc->intcount < 20) printk("Pre-interrupt\n"); #endif /* Make sure it's really for us */ status = __t4_pci_in(wc, WC_INTR); /* Process framer interrupts */ status2 = t4_framer_in(wc, 0, FRMR_CIS); if (status2 & 0x0f) { for (x = 0; x < wc->numspans; ++x) { if (status2 & (1 << x)) t4_framer_interrupt(wc, x); } } /* Ignore if it's not for us */ if (!status) #ifdef LINUX26 return IRQ_NONE; #else return; #endif __t4_pci_out(wc, WC_INTR, 0); if (!wc->spansstarted) { printk("Not prepped yet!\n"); #ifdef LINUX26 return IRQ_NONE; #else return; #endif } wc->intcount++; #if 0 if (wc->intcount < 20) printk("Got interrupt, status = %08x\n", status); #endif if (status & 0x3) { t4_receiveprep(wc, status); t4_transmitprep(wc, status); } #if 0 if ((wc->intcount < 10) || !(wc->intcount % 1000)) { status2 = t4_framer_in(wc, 0, FRMR_CIS); printk("Status2: %04x\n", status2); for (x = 0;x<4;x++) { status2 = t4_framer_in(wc, x, FRMR_FRS0); printk("FRS0/%d: %04x\n", x, status2); } } #endif t4_do_counters(wc); x = wc->intcount & 15 /* 63 */; switch(x) { case 0: case 1: case 2: case 3: t4_check_sigbits(wc, x); break; case 4: case 5: case 6: case 7: t4_check_alarms(wc, x - 4); break; } spin_lock_irqsave(&wc->reglock, flags); __handle_leds(wc); if (test_bit(T4_CHECK_TIMING, &wc->checkflag)) __t4_set_timing_source_auto(wc); spin_unlock_irqrestore(&wc->reglock, flags); #ifdef LINUX26 return IRQ_RETVAL(1); #endif } #endif static void t4_isr_bh(unsigned long data) { struct t4 *wc = (struct t4 *)data; #ifdef VPM_SUPPORT if (wc->vpm) { if (test_and_clear_bit(T4_CHECK_VPM, &wc->checkflag)) { if (wc->vpm450m) { /* How stupid is it that the octasic can't generate an interrupt when there's a tone, in spite of what their documentation says? */ t4_check_vpm450(wc); } else t4_check_vpm400(wc, wc->vpm400checkstatus); } } #endif } ZAP_IRQ_HANDLER(t4_interrupt_gen2) { struct t4 *wc = dev_id; unsigned int status; /* Check this first in case we get a spurious interrupt */ if (unlikely(test_bit(T4_STOP_DMA, &wc->checkflag))) { /* Stop DMA cleanly if requested */ wc->dmactrl = 0x0; t4_pci_out(wc, WC_DMACTRL, 0x00000000); /* Acknowledge any pending interrupts */ t4_pci_out(wc, WC_INTR, 0x00000000); spin_lock(&wc->reglock); __t4_set_timing_source(wc, 4, 0, 0); spin_unlock(&wc->reglock); return IRQ_RETVAL(1); } /* Make sure it's really for us */ status = __t4_pci_in(wc, WC_INTR); /* Ignore if it's not for us */ if (!(status & 0x7)) { #ifdef LINUX26 return IRQ_NONE; #else return; #endif } #ifdef ENABLE_WORKQUEUES __t4_pci_out(wc, WC_INTR, status & 0x00000008); #endif if (unlikely(!wc->spansstarted)) { printk("Not prepped yet!\n"); #ifdef LINUX26 return IRQ_NONE; #else return; #endif } wc->intcount++; if (unlikely((wc->intcount < 20) && debug)) printk("2G: Got interrupt, status = %08x, CIS = %04x\n", status, t4_framer_in(wc, 0, FRMR_CIS)); if (likely(status & 0x2)) { #ifdef ENABLE_WORKQUEUES int cpus = num_online_cpus(); atomic_set(&wc->worklist, wc->numspans); if (wc->tspans[0]->span.flags & ZT_FLAG_RUNNING) t4_queue_work(wc->workq, &wc->tspans[0]->swork, 0); else atomic_dec(&wc->worklist); if (wc->tspans[1]->span.flags & ZT_FLAG_RUNNING) t4_queue_work(wc->workq, &wc->tspans[1]->swork, 1 % cpus); else atomic_dec(&wc->worklist); if (wc->numspans == 4) { if (wc->tspans[2]->span.flags & ZT_FLAG_RUNNING) t4_queue_work(wc->workq, &wc->tspans[2]->swork, 2 % cpus); else atomic_dec(&wc->worklist); if (wc->tspans[3]->span.flags & ZT_FLAG_RUNNING) t4_queue_work(wc->workq, &wc->tspans[3]->swork, 3 % cpus); else atomic_dec(&wc->worklist); } #else t4_prep_gen2(wc); #endif t4_do_counters(wc); spin_lock(&wc->reglock); __handle_leds(wc); spin_unlock(&wc->reglock); } if (unlikely(status & 0x1)) { unsigned char cis; cis = t4_framer_in(wc, 0, FRMR_CIS); if (cis & FRMR_CIS_GIS1) t4_framer_interrupt(wc, 0); if (cis & FRMR_CIS_GIS2) t4_framer_interrupt(wc, 1); if (cis & FRMR_CIS_GIS3) t4_framer_interrupt(wc, 2); if (cis & FRMR_CIS_GIS4) t4_framer_interrupt(wc, 3); } if (wc->vpm && vpmdtmfsupport) { if (wc->vpm450m) { /* How stupid is it that the octasic can't generate an interrupt when there's a tone, in spite of what their documentation says? */ if (!(wc->intcount & 0xf)) { set_bit(T4_CHECK_VPM, &wc->checkflag); } } else if ((status & 0xff00) != 0xff00) { wc->vpm400checkstatus = (status & 0xff00) >> 8; set_bit(T4_CHECK_VPM, &wc->checkflag); } } spin_lock(&wc->reglock); if (unlikely(test_bit(T4_CHECK_TIMING, &wc->checkflag))) { __t4_set_timing_source_auto(wc); } spin_unlock(&wc->reglock); if (unlikely(test_bit(T4_CHECK_VPM, &wc->checkflag))) tasklet_schedule(&wc->t4_tlet); #ifndef ENABLE_WORKQUEUES __t4_pci_out(wc, WC_INTR, 0); #endif #ifdef LINUX26 return IRQ_RETVAL(1); #endif } #ifdef SUPPORT_GEN1 static int t4_reset_dma(struct t4 *wc) { /* Turn off DMA and such */ wc->dmactrl = 0x0; t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); t4_pci_out(wc, WC_COUNT, 0); t4_pci_out(wc, WC_RDADDR, 0); t4_pci_out(wc, WC_WRADDR, 0); t4_pci_out(wc, WC_INTR, 0); /* Turn it all back on */ t4_pci_out(wc, WC_RDADDR, wc->readdma); t4_pci_out(wc, WC_WRADDR, wc->writedma); t4_pci_out(wc, WC_COUNT, ((ZT_MAX_CHUNKSIZE * 2 * 32 - 1) << 18) | ((ZT_MAX_CHUNKSIZE * 2 * 32 - 1) << 2)); t4_pci_out(wc, WC_INTR, 0); #ifdef VPM_SUPPORT wc->dmactrl = 0xc0000000 | (1 << 29) | wc->vpm; #else wc->dmactrl = 0xc0000000 | (1 << 29); #endif if (noburst) wc->dmactrl |= (1 << 26); t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); return 0; } #endif #ifdef VPM_SUPPORT static void t4_vpm_set_dtmf_threshold(struct t4 *wc, unsigned int threshold) { unsigned int x; for (x = 0; x < 8; x++) { t4_vpm_out(wc, x, 0xC4, (threshold >> 8) & 0xFF); t4_vpm_out(wc, x, 0xC5, (threshold & 0xFF)); } printk("VPM: DTMF threshold set to %d\n", threshold); } static unsigned int t4_vpm_mask(int chip) { unsigned int mask=0; switch(vpmspans) { case 4: mask = 0x55555555 << (chip >> 2); break; case 2: mask = 0x11111111 << (chip >> 1); break; case 1: mask = 0x01010101 << chip; break; } return mask; } static int t4_vpm_spanno(int chip) { int spanno = 0; switch(vpmspans) { case 4: spanno = chip & 0x3; break; case 2: spanno = chip & 0x1; break; /* Case 1 is implicit */ } return spanno; } static int t4_vpm_echotail(void) { int echotail = 0x01ff; switch(vpmspans) { case 4: echotail = 0x007f; break; case 2: echotail = 0x00ff; break; /* Case 1 is implicit */ } return echotail; } static void t4_vpm450_init(struct t4 *wc) { unsigned int check1, check2; int laws[4] = { 0, }; int x; unsigned int vpm_capacity; struct firmware embedded_firmware; const struct firmware *firmware = &embedded_firmware; #if !defined(HOTPLUG_FIRMWARE) extern void _binary_zaptel_fw_oct6114_064_bin_size; extern void _binary_zaptel_fw_oct6114_128_bin_size; extern u8 _binary_zaptel_fw_oct6114_064_bin_start[]; extern u8 _binary_zaptel_fw_oct6114_128_bin_start[]; #else static const char oct064_firmware[] = "zaptel-fw-oct6114-064.bin"; static const char oct128_firmware[] = "zaptel-fw-oct6114-128.bin"; #endif if (!vpmsupport) { printk("VPM450: Support Disabled\n"); return; } /* Turn on GPIO/DATA mux if supported */ t4_gpio_setdir(wc, (1 << 24), (1 << 24)); __t4_raw_oct_out(wc, 0x000a, 0x5678); __t4_raw_oct_out(wc, 0x0004, 0x1234); check1 = __t4_raw_oct_in(wc, 0x0004); check2 = __t4_raw_oct_in(wc, 0x000a); if (debug) printk("OCT Result: %04x/%04x\n", __t4_raw_oct_in(wc, 0x0004), __t4_raw_oct_in(wc, 0x000a)); if (__t4_raw_oct_in(wc, 0x0004) != 0x1234) { printk("VPM450: Not Present\n"); return; } /* Setup alaw vs ulaw rules */ for (x = 0;x < wc->numspans; x++) { if (wc->tspans[x]->span.channels > 24) laws[x] = 1; } switch ((vpm_capacity = get_vpm450m_capacity(wc))) { case 64: #if defined(HOTPLUG_FIRMWARE) if ((request_firmware(&firmware, oct064_firmware, &wc->dev->dev) != 0) || !firmware) { printk("VPM450: firmware %s not available from userspace\n", oct064_firmware); return; } #else embedded_firmware.data = _binary_zaptel_fw_oct6114_064_bin_start; /* Yes... this is weird. objcopy gives us a symbol containing the size of the firmware, not a pointer a variable containing the size. The only way we can get the value of the symbol is to take its address, so we define it as a pointer and then cast that value to the proper type. */ embedded_firmware.size = (size_t) &_binary_zaptel_fw_oct6114_064_bin_size; #endif break; case 128: #if defined(HOTPLUG_FIRMWARE) if ((request_firmware(&firmware, oct128_firmware, &wc->dev->dev) != 0) || !firmware) { printk("VPM450: firmware %s not available from userspace\n", oct128_firmware); return; } #else embedded_firmware.data = _binary_zaptel_fw_oct6114_128_bin_start; /* Yes... this is weird. objcopy gives us a symbol containing the size of the firmware, not a pointer a variable containing the size. The only way we can get the value of the symbol is to take its address, so we define it as a pointer and then cast that value to the proper type. */ embedded_firmware.size = (size_t) &_binary_zaptel_fw_oct6114_128_bin_size; #endif break; default: printk("Unsupported channel capacity found on VPM module (%d).\n", vpm_capacity); return; } if (!(wc->vpm450m = init_vpm450m(wc, laws, wc->numspans, firmware))) { printk("VPM450: Failed to initialize\n"); if (firmware != &embedded_firmware) release_firmware(firmware); return; } if (firmware != &embedded_firmware) release_firmware(firmware); if (vpmdtmfsupport == -1) { printk("VPM450: hardware DTMF disabled.\n"); vpmdtmfsupport = 0; } wc->vpm = T4_VPM_PRESENT; printk("VPM450: Present and operational servicing %d span(s)\n", wc->numspans); } static void t4_vpm400_init(struct t4 *wc) { unsigned char reg; unsigned int mask; unsigned int ver; unsigned int i, x, y, gen2vpm=0; if (!vpmsupport) { printk("VPM400: Support Disabled\n"); return; } switch(vpmspans) { case 4: case 2: case 1: break; default: printk("VPM400: %d is not a valid vpmspans value, using 4\n", vpmspans); vpmspans = 4; } for (x=0;x<8;x++) { int spanno = t4_vpm_spanno(x); struct t4_span *ts = wc->tspans[spanno]; int echotail = t4_vpm_echotail(); ver = t4_vpm_in(wc, x, 0x1a0); /* revision */ if ((ver != 0x26) && (ver != 0x33)) { printk("VPM400: %s\n", x ? "Inoperable" : "Not Present"); return; } if (ver == 0x33) { if (x && !gen2vpm) { printk("VPM400: Inconsistent\n"); return; } ts->spanflags |= FLAG_VPM2GEN; gen2vpm++; } else if (gen2vpm) { printk("VPM400: Inconsistent\n"); return; } /* Setup GPIO's */ for (y=0;y<4;y++) { t4_vpm_out(wc, x, 0x1a8 + y, 0x00); /* GPIO out */ t4_vpm_out(wc, x, 0x1ac + y, 0x00); /* GPIO dir */ t4_vpm_out(wc, x, 0x1b0 + y, 0x00); /* GPIO sel */ } /* Setup TDM path - sets fsync and tdm_clk as inputs */ reg = t4_vpm_in(wc, x, 0x1a3); /* misc_con */ t4_vpm_out(wc, x, 0x1a3, reg & ~2); /* Setup timeslots */ t4_vpm_out(wc, x, 0x02f, 0x20 | (spanno << 3)); /* Setup Echo length (128 taps) */ t4_vpm_out(wc, x, 0x022, (echotail >> 8)); t4_vpm_out(wc, x, 0x023, (echotail & 0xff)); /* Setup the tdm channel masks for all chips*/ mask = t4_vpm_mask(x); for (i = 0; i < 4; i++) t4_vpm_out(wc, x, 0x30 + i, (mask >> (i << 3)) & 0xff); /* Setup convergence rate */ reg = t4_vpm_in(wc,x,0x20); reg &= 0xE0; if (ts->spantype == TYPE_E1) { if (x < vpmspans) printk("VPM400: Span %d A-law mode\n", spanno); reg |= 0x01; } else { if (x < vpmspans) printk("VPM400: Span %d U-law mode\n", spanno); reg &= ~0x01; } t4_vpm_out(wc,x,0x20,(reg | 0x20)); /* Initialize echo cans */ for (i = 0 ; i < MAX_TDM_CHAN; i++) { if (mask & (0x00000001 << i)) t4_vpm_out(wc,x,i,0x00); } wait_a_little(); /* Put in bypass mode */ for (i = 0 ; i < MAX_TDM_CHAN ; i++) { if (mask & (0x00000001 << i)) { t4_vpm_out(wc,x,i,0x01); } } /* Enable bypass */ for (i = 0 ; i < MAX_TDM_CHAN ; i++) { if (mask & (0x00000001 << i)) t4_vpm_out(wc,x,0x78 + i,0x01); } /* set DTMF detection threshold */ t4_vpm_set_dtmf_threshold(wc, dtmfthreshold); /* Enable DTMF detectors (always DTMF detect all spans) */ for (i = 0; i < MAX_DTMF_DET; i++) { t4_vpm_out(wc, x, 0x98 + i, 0x40 | (i * 2) | ((x < 4) ? 0 : 1)); } for (i = 0x34; i < 0x38; i++) t4_vpm_out(wc, x, i, 0x00); for (i = 0x3C; i < 0x40; i++) t4_vpm_out(wc, x, i, 0x00); for (i = 0x48; i < 0x4B; i++) t4_vpm_out(wc, x, i, 0x00); for (i = 0x50; i < 0x53; i++) t4_vpm_out(wc, x, i, 0x00); for (i = 0xB8; i < 0xBE; i++) t4_vpm_out(wc, x, i, 0xFF); if (gen2vpm) { for (i = 0xBE; i < 0xC0; i++) t4_vpm_out(wc, x, i, 0xFF); } else { for (i = 0xBE; i < 0xC0; i++) t4_vpm_out(wc, x, i, 0x00); } for (i = 0xC0; i < 0xC4; i++) t4_vpm_out(wc, x, i, (x < 4) ? 0x55 : 0xAA); } if (vpmdtmfsupport == -1) { printk("VPM400: hardware DTMF enabled.\n"); vpmdtmfsupport = 0; } printk("VPM400%s: Present and operational servicing %d span(s)\n", (gen2vpm ? " (2nd Gen)" : ""), wc->numspans); wc->vpm = T4_VPM_PRESENT; } #endif static void t4_tsi_reset(struct t4 *wc) { int x; for (x=0;x<128;x++) { wc->dmactrl &= ~0x00007fff; wc->dmactrl |= (0x00004000 | (x << 7)); t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); } wc->dmactrl &= ~0x00007fff; t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); } /* Note that channels here start from 1 */ static void t4_tsi_assign(struct t4 *wc, int fromspan, int fromchan, int tospan, int tochan) { unsigned long flags; int fromts, tots; fromts = (fromspan << 5) |(fromchan); tots = (tospan << 5) | (tochan); if (!wc->t1e1) { fromts += 4; tots += 4; } spin_lock_irqsave(&wc->reglock, flags); wc->dmactrl &= ~0x00007fff; wc->dmactrl |= (0x00004000 | (tots << 7) | (fromts)); __t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); wc->dmactrl &= ~0x00007fff; __t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); spin_unlock_irqrestore(&wc->reglock, flags); } static void t4_tsi_unassign(struct t4 *wc, int tospan, int tochan) { unsigned long flags; int tots; tots = (tospan << 5) | (tochan); if (!wc->t1e1) tots += 4; spin_lock_irqsave(&wc->reglock, flags); wc->dmactrl &= ~0x00007fff; wc->dmactrl |= (0x00004000 | (tots << 7)); __t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); if (debug & DEBUG_TSI) printk("Sending '%08x\n", wc->dmactrl); wc->dmactrl &= ~0x00007fff; __t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); spin_unlock_irqrestore(&wc->reglock, flags); } static int t4_hardware_init_1(struct t4 *wc, unsigned int cardflags) { unsigned int version; version = t4_pci_in(wc, WC_VERSION); printk("TE%dXXP version %08x, burst %s\n", wc->numspans, version, (!(cardflags & FLAG_BURST) && noburst) ? "OFF" : "ON"); #ifdef ENABLE_WORKQUEUES printk("TE%dXXP running with work queues.\n", wc->numspans); #endif /* Make sure DMA engine is not running and interrupts are acknowledged */ wc->dmactrl = 0x0; t4_pci_out(wc, WC_DMACTRL, wc->dmactrl); /* Reset Framer and friends */ t4_pci_out(wc, WC_LEDS, 0x00000000); /* Set DMA addresses */ t4_pci_out(wc, WC_RDADDR, wc->readdma); t4_pci_out(wc, WC_WRADDR, wc->writedma); /* Setup counters, interrupt flags (ignored in Gen2) */ if (cardflags & FLAG_2NDGEN) { t4_tsi_reset(wc); } else { t4_pci_out(wc, WC_COUNT, ((ZT_MAX_CHUNKSIZE * 2 * 32 - 1) << 18) | ((ZT_MAX_CHUNKSIZE * 2 * 32 - 1) << 2)); } /* Reset pending interrupts */ t4_pci_out(wc, WC_INTR, 0x00000000); /* Read T1/E1 status */ if (t1e1override > -1) wc->t1e1 = t1e1override; else wc->t1e1 = ((t4_pci_in(wc, WC_LEDS)) & 0x0f00) >> 8; wc->order = ((t4_pci_in(wc, WC_LEDS)) & 0xf0000000) >> 28; order_index[wc->order]++; return 0; } static int t4_hardware_init_2(struct t4 *wc) { int x; unsigned int falcver; if (t4_pci_in(wc, WC_VERSION) >= 0xc01a0165) { wc->tspans[0]->spanflags |= FLAG_OCTOPT; printk("Octasic optimized!\n"); } /* Setup LEDS, take out of reset */ t4_pci_out(wc, WC_LEDS, 0x000000ff); t4_activate(wc); t4_framer_out(wc, 0, 0x4a, 0xaa); falcver = t4_framer_in(wc, 0 ,0x4a); printk("FALC version: %08x, Board ID: %02x\n", falcver, wc->order); for (x=0;x< 11;x++) printk("Reg %d: 0x%08x\n", x, t4_pci_in(wc, x)); return 0; } static int __devinit t4_launch(struct t4 *wc) { int x; unsigned long flags; if (wc->tspans[0]->span.flags & ZT_FLAG_REGISTERED) return 0; printk("TE%dXXP: Launching card: %d\n", wc->numspans, wc->order); /* Setup serial parameters and system interface */ for (x=0;x<4;x++) t4_serial_setup(wc, x); if (zt_register(&wc->tspans[0]->span, 0)) { printk(KERN_ERR "Unable to register span %s\n", wc->tspans[0]->span.name); return -1; } if (zt_register(&wc->tspans[1]->span, 0)) { printk(KERN_ERR "Unable to register span %s\n", wc->tspans[1]->span.name); zt_unregister(&wc->tspans[0]->span); return -1; } if (wc->numspans == 4) { if (zt_register(&wc->tspans[2]->span, 0)) { printk(KERN_ERR "Unable to register span %s\n", wc->tspans[2]->span.name); zt_unregister(&wc->tspans[0]->span); zt_unregister(&wc->tspans[1]->span); return -1; } if (zt_register(&wc->tspans[3]->span, 0)) { printk(KERN_ERR "Unable to register span %s\n", wc->tspans[3]->span.name); zt_unregister(&wc->tspans[0]->span); zt_unregister(&wc->tspans[1]->span); zt_unregister(&wc->tspans[2]->span); return -1; } } /* Start the first span on the card with the default configuration so * that it may provide timing to Asterisk before being properly * configured. */ t4_startup(&wc->tspans[0]->span); set_bit(T4_CHECK_TIMING, &wc->checkflag); spin_lock_irqsave(&wc->reglock, flags); __t4_set_timing_source(wc,4, 0, 0); spin_unlock_irqrestore(&wc->reglock, flags); tasklet_init(&wc->t4_tlet, t4_isr_bh, (unsigned long)wc); return 0; } static int __devinit t4_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) { int res; struct t4 *wc; struct devtype *dt; int x,f; int basesize; #if 0 int y; unsigned int *canary; #endif if (pci_enable_device(pdev)) { res = -EIO; } else { wc = kmalloc(sizeof(struct t4), GFP_KERNEL); if (wc) { memset(wc, 0x0, sizeof(struct t4)); spin_lock_init(&wc->reglock); dt = (struct devtype *)(ent->driver_data); if (dt->flags & FLAG_2NDGEN) basesize = ZT_MAX_CHUNKSIZE * 32 * 4; else basesize = ZT_MAX_CHUNKSIZE * 32 * 2 * 4; if (dt->flags & FLAG_2PORT) wc->numspans = 2; else wc->numspans = 4; wc->variety = dt->desc; wc->memaddr = pci_resource_start(pdev, 0); wc->memlen = pci_resource_len(pdev, 0); wc->membase = ioremap(wc->memaddr, wc->memlen); /* This rids of the Double missed interrupt message after loading */ wc->last0 = 1; #if 0 if (!request_mem_region(wc->memaddr, wc->memlen, wc->variety)) printk("wct4: Unable to request memory region :(, using anyway...\n"); #endif if (pci_request_regions(pdev, wc->variety)) printk("wct%dxxp: Unable to request regions\n", wc->numspans); printk("Found TE%dXXP at base address %08lx, remapped to %p\n", wc->numspans, wc->memaddr, wc->membase); wc->dev = pdev; wc->writechunk = /* 32 channels, Double-buffer, Read/Write, 4 spans */ (unsigned int *)pci_alloc_consistent(pdev, basesize * 2, &wc->writedma); if (!wc->writechunk) { printk("wct%dxxp: Unable to allocate DMA-able memory\n", wc->numspans); return -ENOMEM; } /* Read is after the whole write piece (in words) */ wc->readchunk = wc->writechunk + basesize / 4; /* Same thing but in bytes... */ wc->readdma = wc->writedma + basesize; /* Initialize Write/Buffers to all blank data */ memset((void *)wc->writechunk,0x00, basesize); memset((void *)wc->readchunk,0xff, basesize); #if 0 memset((void *)wc->readchunk,0xff,ZT_MAX_CHUNKSIZE * 2 * 32 * 4); /* Initialize canary */ canary = (unsigned int *)(wc->readchunk + ZT_CHUNKSIZE * 64 * 4 - 4); *canary = (CANARY << 16) | (0xffff); #endif /* Enable bus mastering */ pci_set_master(pdev); /* Keep track of which device we are */ pci_set_drvdata(pdev, wc); /* Initialize hardware */ t4_hardware_init_1(wc, dt->flags); for(x = 0; x < MAX_T4_CARDS; x++) { if (!cards[x]) break; } if (x >= MAX_T4_CARDS) { printk("No cards[] slot available!!\n"); return -ENOMEM; } wc->num = x; cards[x] = wc; #ifdef ENABLE_WORKQUEUES if (dt->flags & FLAG_2NDGEN) { char tmp[20]; sprintf(tmp, "te%dxxp[%d]", wc->numspans, wc->num); wc->workq = create_workqueue(tmp); } #endif /* Allocate pieces we need here */ for (x=0;x<4;x++) { if (wc->t1e1 & (1 << x)) { wc->tspans[x] = kmalloc(sizeof(struct t4_span) + sizeof(struct zt_chan) * 31, GFP_KERNEL); if (wc->tspans[x]) { memset(wc->tspans[x], 0, sizeof(struct t4_span) + sizeof(struct zt_chan) * 31); wc->tspans[x]->spantype = TYPE_E1; } } else { wc->tspans[x] = kmalloc(sizeof(struct t4_span) + sizeof(struct zt_chan) * 24, GFP_KERNEL); if (wc->tspans[x]) { memset(wc->tspans[x], 0, sizeof(struct t4_span) + sizeof(struct zt_chan) * 24); if (j1mode) wc->tspans[x]->spantype = TYPE_J1; else wc->tspans[x]->spantype = TYPE_T1; } } if (!wc->tspans[x]) return -ENOMEM; #ifdef ENABLE_WORKQUEUES INIT_WORK(&wc->tspans[x]->swork, workq_handlespan, wc->tspans[x]); #endif wc->tspans[x]->spanflags |= dt->flags; } /* Continue hardware intiialization */ t4_hardware_init_2(wc); #ifdef SUPPORT_GEN1 if (request_irq(pdev->irq, (dt->flags & FLAG_2NDGEN) ? t4_interrupt_gen2 :t4_interrupt, ZAP_IRQ_SHARED_DISABLED, (wc->numspans == 2) ? "wct2xxp" : "wct4xxp", wc)) #else if (!(wc->tspans[0]->spanflags & FLAG_2NDGEN)) { printk("This driver does not support 1st gen modules\n"); kfree(wc); return -ENODEV; } if (request_irq(pdev->irq, t4_interrupt_gen2, ZAP_IRQ_SHARED_DISABLED, "t4xxp", wc)) #endif { printk("t4xxp: Unable to request IRQ %d\n", pdev->irq); kfree(wc); return -EIO; } init_spans(wc); /* Launch cards as appropriate */ for (;;) { /* Find a card to activate */ f = 0; for (x=0;cards[x];x++) { if (cards[x]->order <= highestorder) { t4_launch(cards[x]); if (cards[x]->order == highestorder) f = 1; } } /* If we found at least one, increment the highest order and search again, otherwise stop */ if (f) highestorder++; else break; } printk("Found a Wildcard: %s\n", wc->variety); wc->gpio = 0x00000000; t4_pci_out(wc, WC_GPIO, wc->gpio); t4_gpio_setdir(wc, (1 << 17), (1 << 17)); t4_gpio_setdir(wc, (0xff), (0xff)); #if 0 for (x=0;x<0x10000;x++) { __t4_raw_oct_out(wc, 0x0004, x); __t4_raw_oct_out(wc, 0x000a, x ^ 0xffff); if (__t4_raw_oct_in(wc, 0x0004) != x) printk("Register 4 failed %04x\n", x); if (__t4_raw_oct_in(wc, 0x000a) != (x ^ 0xffff)) printk("Register 10 failed %04x\n", x); } #endif res = 0; } else res = -ENOMEM; } return res; } static int t4_hardware_stop(struct t4 *wc) { /* Turn off DMA, leave interrupts enabled */ set_bit(T4_STOP_DMA, &wc->checkflag); /* Wait for interrupts to stop */ msleep(25); /* Turn off counter, address, etc */ if (wc->tspans[0]->spanflags & FLAG_2NDGEN) { t4_tsi_reset(wc); } else { t4_pci_out(wc, WC_COUNT, 0x000000); } t4_pci_out(wc, WC_RDADDR, 0x0000000); t4_pci_out(wc, WC_WRADDR, 0x0000000); wc->gpio = 0x00000000; t4_pci_out(wc, WC_GPIO, wc->gpio); t4_pci_out(wc, WC_LEDS, 0x00000000); printk("\nStopped TE%dXXP, Turned off DMA\n", wc->numspans); return 0; } static void __devexit t4_remove_one(struct pci_dev *pdev) { struct t4 *wc = pci_get_drvdata(pdev); int x; if (wc) { /* Stop hardware */ t4_hardware_stop(wc); /* Release vpm450m */ if (wc->vpm450m) release_vpm450m(wc->vpm450m); wc->vpm450m = NULL; /* Unregister spans */ if (wc->tspans[0]->span.flags & ZT_FLAG_REGISTERED) zt_unregister(&wc->tspans[0]->span); if (wc->tspans[1]->span.flags & ZT_FLAG_REGISTERED) zt_unregister(&wc->tspans[1]->span); if (wc->numspans == 4) { if (wc->tspans[2]->span.flags & ZT_FLAG_REGISTERED) zt_unregister(&wc->tspans[2]->span); if (wc->tspans[3]->span.flags & ZT_FLAG_REGISTERED) zt_unregister(&wc->tspans[3]->span); } #ifdef ENABLE_WORKQUEUES if (wc->workq) { flush_workqueue(wc->workq); destroy_workqueue(wc->workq); } #endif free_irq(pdev->irq, wc); if (wc->membase) iounmap((void *)wc->membase); pci_release_regions(pdev); /* Immediately free resources */ pci_free_consistent(pdev, ZT_MAX_CHUNKSIZE * 2 * 2 * 32 * 4, (void *)wc->writechunk, wc->writedma); order_index[wc->order]--; cards[wc->num] = NULL; pci_set_drvdata(pdev, NULL); for (x=0;xnumspans;x++) { if (wc->tspans[x]) kfree(wc->tspans[x]); } kfree(wc); } } static struct pci_device_id t4_pci_tbl[] __devinitdata = { { 0x10ee, 0x0314, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct4xxp }, { 0xd161, 0x0420, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct420p4 }, { 0xd161, 0x0410, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct410p4 }, { 0xd161, 0x0405, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct405p4 }, { 0xd161, 0x0410, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct410p3 }, { 0xd161, 0x0405, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct405p3 }, { 0xd161, 0x0410, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct410p2 }, { 0xd161, 0x0405, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct405p2 }, { 0xd161, 0x0220, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct220p4 }, { 0xd161, 0x0205, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct205p4 }, { 0xd161, 0x0210, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct210p4 }, { 0xd161, 0x0205, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct205p3 }, { 0xd161, 0x0210, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct210p3 }, { 0xd161, 0x0205, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct205 }, { 0xd161, 0x0210, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct210 }, { 0, } }; static struct pci_driver t4_driver = { name: "wct4xxp", probe: t4_init_one, #ifdef LINUX26 remove: __devexit_p(t4_remove_one), #else remove: t4_remove_one, #endif suspend: NULL, resume: NULL, id_table: t4_pci_tbl, }; static int __init t4_init(void) { int res; res = zap_pci_module(&t4_driver); if (res) return -ENODEV; return 0; } static void __exit t4_cleanup(void) { pci_unregister_driver(&t4_driver); } MODULE_AUTHOR("Mark Spencer"); MODULE_DESCRIPTION("Unified TE4XXP-TE2XXP PCI Driver"); #if defined(MODULE_ALIAS) MODULE_ALIAS("wct2xxp"); #endif #ifdef MODULE_LICENSE MODULE_LICENSE("GPL"); #endif #ifdef LINUX26 module_param(pedanticpci, int, 0600); module_param(debug, int, 0600); module_param(loopback, int, 0600); module_param(noburst, int, 0600); module_param(timingcable, int, 0600); module_param(t1e1override, int, 0600); module_param(alarmdebounce, int, 0600); module_param(j1mode, int, 0600); module_param(sigmode, int, 0600); #ifdef VPM_SUPPORT module_param(vpmsupport, int, 0600); module_param(vpmdtmfsupport, int, 0600); module_param(vpmspans, int, 0600); module_param(dtmfthreshold, int, 0600); #endif #else MODULE_PARM(pedanticpci, "i"); MODULE_PARM(debug, "i"); MODULE_PARM(loopback, "i"); MODULE_PARM(noburst, "i"); MODULE_PARM(hardhdlcmode, "i"); MODULE_PARM(timingcable, "i"); MODULE_PARM(t1e1override, "i"); MODULE_PARM(alarmdebounce, "i"); MODULE_PARM(j1mode, "i"); MODULE_PARM(sigmode, "i"); #ifdef VPM_SUPPORT MODULE_PARM(vpmsupport, "i"); MODULE_PARM(vpmdtmfsupport, "i"); MODULE_PARM(vpmspans, "i"); MODULE_PARM(dtmfthreshold, "i"); #endif #endif MODULE_DEVICE_TABLE(pci, t4_pci_tbl); module_init(t4_init); module_exit(t4_cleanup);