64a98af676
These flags are direct mappings to the IRQF_[SHARED|DISABLED] flags and no longer need to be kept separate. Signed-off-by: Shaun Ruffell <sruffell@digium.com> Signed-off-by: Russ Meyerriecks <rmeyerriecks@digium.com>
5635 lines
154 KiB
C
5635 lines
154 KiB
C
/*
|
|
* TE410P Quad-T1/E1 PCI Driver version 0.1, 12/16/02
|
|
*
|
|
* Written by Mark Spencer <markster@digium.com>
|
|
* Based on previous works, designs, and archetectures conceived and
|
|
* written by Jim Dixon <jim@lambdatel.com>.
|
|
* Further modified, optimized, and maintained by
|
|
* Matthew Fredrickson <creslin@digium.com> and
|
|
* Russ Meyerriecks <rmeyerriecks@digium.com>
|
|
*
|
|
* Copyright (C) 2001 Jim Dixon / Zapata Telephony.
|
|
* Copyright (C) 2001-2012, Digium, Inc.
|
|
*
|
|
* All rights reserved.
|
|
*
|
|
*/
|
|
|
|
/*
|
|
* See http://www.asterisk.org for more information about
|
|
* the Asterisk project. Please do not directly contact
|
|
* any of the maintainers of this project for assistance;
|
|
* the project provides a web site, mailing lists and IRC
|
|
* channels for your use.
|
|
*
|
|
* This program is free software, distributed under the terms of
|
|
* the GNU General Public License Version 2 as published by the
|
|
* Free Software Foundation. See the LICENSE file included with
|
|
* this program for more details.
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/module.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/init.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/spinlock.h>
|
|
#include <asm/io.h>
|
|
#include <linux/version.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/crc32.h>
|
|
#include <linux/slab.h>
|
|
|
|
#include <stdbool.h>
|
|
#include <dahdi/kernel.h>
|
|
|
|
#include "wct4xxp.h"
|
|
#include "vpm450m.h"
|
|
|
|
/* Work queues are a way to better distribute load on SMP systems */
|
|
#if (LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20))
|
|
/*
|
|
* Work queues can significantly improve performance and scalability
|
|
* on multi-processor machines, but requires bypassing some kernel
|
|
* API's, so it's not guaranteed to be compatible with all kernels.
|
|
*/
|
|
/* #define ENABLE_WORKQUEUES */
|
|
#endif
|
|
|
|
/* Support first generation cards? */
|
|
#define SUPPORT_GEN1
|
|
|
|
/* Define to get more attention-grabbing but slightly more I/O using
|
|
alarm status */
|
|
#undef FANCY_ALARM
|
|
|
|
/* Define to support Digium Voice Processing Module expansion card */
|
|
#define VPM_SUPPORT
|
|
|
|
#define DEBUG_MAIN (1 << 0)
|
|
#define DEBUG_DTMF (1 << 1)
|
|
#define DEBUG_REGS (1 << 2)
|
|
#define DEBUG_TSI (1 << 3)
|
|
#define DEBUG_ECHOCAN (1 << 4)
|
|
#define DEBUG_RBS (1 << 5)
|
|
#define DEBUG_FRAMER (1 << 6)
|
|
|
|
/* Maximum latency to be used with Gen 5 */
|
|
#define GEN5_MAX_LATENCY 127
|
|
|
|
#ifdef ENABLE_WORKQUEUES
|
|
#include <linux/cpu.h>
|
|
|
|
/* XXX UGLY!!!! XXX We have to access the direct structures of the workqueue which
|
|
are only defined within workqueue.c because they don't give us a routine to allow us
|
|
to nail a work to a particular thread of the CPU. Nailing to threads gives us substantially
|
|
higher scalability in multi-CPU environments though! */
|
|
|
|
/*
|
|
* The per-CPU workqueue (if single thread, we always use cpu 0's).
|
|
*
|
|
* The sequence counters are for flush_scheduled_work(). It wants to wait
|
|
* until until all currently-scheduled works are completed, but it doesn't
|
|
* want to be livelocked by new, incoming ones. So it waits until
|
|
* remove_sequence is >= the insert_sequence which pertained when
|
|
* flush_scheduled_work() was called.
|
|
*/
|
|
|
|
struct cpu_workqueue_struct {
|
|
|
|
spinlock_t lock;
|
|
|
|
long remove_sequence; /* Least-recently added (next to run) */
|
|
long insert_sequence; /* Next to add */
|
|
|
|
struct list_head worklist;
|
|
wait_queue_head_t more_work;
|
|
wait_queue_head_t work_done;
|
|
|
|
struct workqueue_struct *wq;
|
|
task_t *thread;
|
|
|
|
int run_depth; /* Detect run_workqueue() recursion depth */
|
|
} ____cacheline_aligned;
|
|
|
|
/*
|
|
* The externally visible workqueue abstraction is an array of
|
|
* per-CPU workqueues:
|
|
*/
|
|
struct workqueue_struct {
|
|
/* TODO: Find out exactly where the API changed */
|
|
struct cpu_workqueue_struct *cpu_wq;
|
|
const char *name;
|
|
struct list_head list; /* Empty if single thread */
|
|
};
|
|
|
|
/* Preempt must be disabled. */
|
|
static void __t4_queue_work(struct cpu_workqueue_struct *cwq,
|
|
struct work_struct *work)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&cwq->lock, flags);
|
|
work->wq_data = cwq;
|
|
list_add_tail(&work->entry, &cwq->worklist);
|
|
cwq->insert_sequence++;
|
|
wake_up(&cwq->more_work);
|
|
spin_unlock_irqrestore(&cwq->lock, flags);
|
|
}
|
|
|
|
/*
|
|
* Queue work on a workqueue. Return non-zero if it was successfully
|
|
* added.
|
|
*
|
|
* We queue the work to the CPU it was submitted, but there is no
|
|
* guarantee that it will be processed by that CPU.
|
|
*/
|
|
static inline int t4_queue_work(struct workqueue_struct *wq, struct work_struct *work, int cpu)
|
|
{
|
|
int ret = 0;
|
|
get_cpu();
|
|
if (!test_and_set_bit(0, &work->pending)) {
|
|
BUG_ON(!list_empty(&work->entry));
|
|
__t4_queue_work(wq->cpu_wq + cpu, work);
|
|
ret = 1;
|
|
}
|
|
put_cpu();
|
|
return ret;
|
|
}
|
|
|
|
#endif
|
|
|
|
/*
|
|
* Define CONFIG_FORCE_EXTENDED_RESET to allow the qfalc framer extra time
|
|
* to reset itself upon hardware initialization. This exits for rare
|
|
* cases for customers who are seeing the qfalc returning unexpected
|
|
* information at initialization
|
|
*/
|
|
/* #define CONFIG_FORCE_EXTENDED_RESET */
|
|
/* #define CONFIG_NOEXTENDED_RESET */
|
|
|
|
/*
|
|
* Uncomment the following definition in order to disable Active-State Power
|
|
* Management on the PCIe bridge for PCIe cards. This has been known to work
|
|
* around issues where the BIOS enables it on the cards even though the
|
|
* platform does not support it.
|
|
*
|
|
*/
|
|
/* #define CONFIG_WCT4XXP_DISABLE_ASPM */
|
|
|
|
#if defined(CONFIG_FORCE_EXTENDED_RESET) && defined(CONFIG_NOEXTENDED_RESET)
|
|
#error "You cannot define both CONFIG_FORCE_EXTENDED_RESET and " \
|
|
"CONFIG_NOEXTENDED_RESET."
|
|
#endif
|
|
|
|
int debug = 0;
|
|
static int timingcable = 0;
|
|
static int t1e1override = -1; /* deprecated */
|
|
static char *default_linemode = "auto";
|
|
static int j1mode = 0;
|
|
static int sigmode = FRMR_MODE_NO_ADDR_CMP;
|
|
static int alarmdebounce = 2500; /* LOF/LFA def to 2.5s AT&T TR54016*/
|
|
static int losalarmdebounce = 2500;/* LOS def to 2.5s AT&T TR54016*/
|
|
static int aisalarmdebounce = 2500;/* AIS(blue) def to 2.5s AT&T TR54016*/
|
|
static int yelalarmdebounce = 500;/* RAI(yellow) def to 0.5s AT&T devguide */
|
|
static int max_latency = GEN5_MAX_LATENCY; /* Used to set a maximum latency (if you don't wish it to hard cap it at a certain value) in milliseconds */
|
|
#ifdef VPM_SUPPORT
|
|
static int vpmsupport = 1;
|
|
/* If set to auto, vpmdtmfsupport is enabled for VPM400M and disabled for VPM450M */
|
|
static int vpmdtmfsupport = -1; /* -1=auto, 0=disabled, 1=enabled*/
|
|
#endif /* VPM_SUPPORT */
|
|
|
|
/* Enabling bursting can more efficiently utilize PCI bus bandwidth, but
|
|
can also cause PCI bus starvation, especially in combination with other
|
|
aggressive cards. Please note that burst mode has no effect on CPU
|
|
utilization / max number of calls / etc. */
|
|
static int noburst;
|
|
/* For 56kbps links, set this module parameter to 0x7f */
|
|
static int hardhdlcmode = 0xff;
|
|
|
|
static int latency = 1;
|
|
|
|
static int ms_per_irq = 1;
|
|
static int ignore_rotary;
|
|
|
|
#ifdef FANCY_ALARM
|
|
static int altab[] = {
|
|
0, 0, 0, 1, 2, 3, 4, 6, 8, 9, 11, 13, 16, 18, 20, 22, 24, 25, 27, 28, 29, 30, 31, 31, 32, 31, 31, 30, 29, 28, 27, 25, 23, 22, 20, 18, 16, 13, 11, 9, 8, 6, 4, 3, 2, 1, 0, 0,
|
|
};
|
|
#endif
|
|
|
|
#define MAX_SPANS 16
|
|
|
|
#define FLAG_STARTED (1 << 0)
|
|
#define FLAG_NMF (1 << 1)
|
|
#define FLAG_SENDINGYELLOW (1 << 2)
|
|
|
|
#define FLAG_2NDGEN (1 << 3)
|
|
#define FLAG_2PORT (1 << 4)
|
|
#define FLAG_VPM2GEN (1 << 5)
|
|
#define FLAG_OCTOPT (1 << 6)
|
|
#define FLAG_3RDGEN (1 << 7)
|
|
#define FLAG_BURST (1 << 8)
|
|
#define FLAG_EXPRESS (1 << 9)
|
|
#define FLAG_5THGEN (1 << 10)
|
|
#define FLAG_8PORT (1 << 11)
|
|
|
|
#define CANARY 0xc0de
|
|
|
|
/* names of available HWEC modules */
|
|
#ifdef VPM_SUPPORT
|
|
#define T4_VPM_PRESENT (1 << 28)
|
|
static const char *vpmoct064_name = "VPMOCT064";
|
|
static const char *vpmoct128_name = "VPMOCT128";
|
|
static const char *vpmoct256_name = "VPMOCT256";
|
|
#endif
|
|
|
|
struct devtype {
|
|
char *desc;
|
|
unsigned int flags;
|
|
};
|
|
|
|
static struct devtype wct820p5 = { "Wildcard TE820 (5th Gen)", FLAG_8PORT | FLAG_5THGEN | FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_EXPRESS };
|
|
static struct devtype wct420p5 = { "Wildcard TE420 (5th Gen)", FLAG_5THGEN | FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_EXPRESS };
|
|
static struct devtype wct410p5 = { "Wildcard TE410P (5th Gen)", FLAG_5THGEN | FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN };
|
|
static struct devtype wct405p5 = { "Wildcard TE405P (5th Gen)", FLAG_5THGEN | FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN };
|
|
static struct devtype wct220p5 = { "Wildcard TE220 (5th Gen)", FLAG_5THGEN | FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT | FLAG_EXPRESS };
|
|
static struct devtype wct210p5 = { "Wildcard TE210P (5th Gen)", FLAG_5THGEN | FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT };
|
|
static struct devtype wct205p5 = { "Wildcard TE205P (5th Gen)", FLAG_5THGEN | FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT };
|
|
|
|
static struct devtype wct4xxp = { "Wildcard TE410P/TE405P (1st Gen)", 0 };
|
|
static struct devtype wct420p4 = { "Wildcard TE420 (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_EXPRESS };
|
|
static struct devtype wct410p4 = { "Wildcard TE410P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN };
|
|
static struct devtype wct410p3 = { "Wildcard TE410P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN };
|
|
static struct devtype wct405p4 = { "Wildcard TE405P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN };
|
|
static struct devtype wct405p3 = { "Wildcard TE405P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN };
|
|
static struct devtype wct410p2 = { "Wildcard TE410P (2nd Gen)", FLAG_2NDGEN };
|
|
static struct devtype wct405p2 = { "Wildcard TE405P (2nd Gen)", FLAG_2NDGEN };
|
|
static struct devtype wct220p4 = { "Wildcard TE220 (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT | FLAG_EXPRESS };
|
|
static struct devtype wct205p4 = { "Wildcard TE205P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT };
|
|
static struct devtype wct205p3 = { "Wildcard TE205P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT };
|
|
static struct devtype wct210p4 = { "Wildcard TE210P (4th Gen)", FLAG_BURST | FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT };
|
|
static struct devtype wct210p3 = { "Wildcard TE210P (3rd Gen)", FLAG_2NDGEN | FLAG_3RDGEN | FLAG_2PORT };
|
|
static struct devtype wct205 = { "Wildcard TE205P ", FLAG_2NDGEN | FLAG_2PORT };
|
|
static struct devtype wct210 = { "Wildcard TE210P ", FLAG_2NDGEN | FLAG_2PORT };
|
|
|
|
|
|
struct t4;
|
|
|
|
enum linemode {T1, E1, J1};
|
|
|
|
struct spi_state {
|
|
int wrreg;
|
|
int rdreg;
|
|
};
|
|
|
|
struct t4_span {
|
|
struct t4 *owner;
|
|
u32 *writechunk; /* Double-word aligned write memory */
|
|
u32 *readchunk; /* Double-word aligned read memory */
|
|
enum linemode linemode;
|
|
int sync;
|
|
int alarmtimer;
|
|
int notclear;
|
|
unsigned long alarm_time;
|
|
unsigned long losalarm_time;
|
|
unsigned long aisalarm_time;
|
|
unsigned long yelalarm_time;
|
|
unsigned long alarmcheck_time;
|
|
int spanflags;
|
|
int syncpos;
|
|
|
|
#ifdef SUPPORT_GEN1
|
|
int e1check; /* E1 check */
|
|
#endif
|
|
struct dahdi_span span;
|
|
unsigned char txsigs[16]; /* Transmit sigs */
|
|
int loopupcnt;
|
|
int loopdowncnt;
|
|
#ifdef SUPPORT_GEN1
|
|
unsigned char ec_chunk1[31][DAHDI_CHUNKSIZE]; /* first EC chunk buffer */
|
|
unsigned char ec_chunk2[31][DAHDI_CHUNKSIZE]; /* second EC chunk buffer */
|
|
#endif
|
|
/* HDLC controller fields */
|
|
struct dahdi_chan *sigchan;
|
|
unsigned char sigmode;
|
|
int sigactive;
|
|
int frames_out;
|
|
int frames_in;
|
|
|
|
#ifdef VPM_SUPPORT
|
|
unsigned long dtmfactive;
|
|
unsigned long dtmfmask;
|
|
unsigned long dtmfmutemask;
|
|
#endif
|
|
#ifdef ENABLE_WORKQUEUES
|
|
struct work_struct swork;
|
|
#endif
|
|
struct dahdi_chan *chans[32]; /* Individual channels */
|
|
struct dahdi_echocan_state *ec[32]; /* Echocan state for each channel */
|
|
};
|
|
|
|
struct t4 {
|
|
/* This structure exists one per card */
|
|
struct pci_dev *dev; /* Pointer to PCI device */
|
|
unsigned int intcount;
|
|
int num; /* Which card we are */
|
|
int syncsrc; /* active sync source */
|
|
struct dahdi_device *ddev;
|
|
struct t4_span *tspans[8]; /* Individual spans */
|
|
int numspans; /* Number of spans on the card */
|
|
int blinktimer;
|
|
#ifdef FANCY_ALARM
|
|
int alarmpos;
|
|
#endif
|
|
int irq; /* IRQ used by device */
|
|
int order; /* Order */
|
|
const struct devtype *devtype;
|
|
unsigned int reset_required:1; /* If reset needed in serial_setup */
|
|
unsigned int falc31:1; /* are we falc v3.1 (atomic not necessary) */
|
|
unsigned int t1e1:8; /* T1 / E1 select pins */
|
|
int ledreg; /* LED Register */
|
|
int ledreg2; /* LED Register2 */
|
|
unsigned int gpio;
|
|
unsigned int gpioctl;
|
|
int e1recover; /* E1 recovery timer */
|
|
spinlock_t reglock; /* lock register access */
|
|
int spansstarted; /* number of spans started */
|
|
u32 *writechunk; /* Double-word aligned write memory */
|
|
u32 *readchunk; /* Double-word aligned read memory */
|
|
#ifdef ENABLE_WORKQUEUES
|
|
atomic_t worklist;
|
|
struct workqueue_struct *workq;
|
|
#endif
|
|
int last0; /* for detecting double-missed IRQ */
|
|
|
|
/* DMA related fields */
|
|
unsigned int dmactrl;
|
|
dma_addr_t readdma;
|
|
dma_addr_t writedma;
|
|
void __iomem *membase; /* Base address of card */
|
|
|
|
#define T4_CHECK_VPM 0
|
|
#define T4_LOADING_FW 1
|
|
#define T4_STOP_DMA 2
|
|
#define T4_CHECK_TIMING 3
|
|
#define T4_CHANGE_LATENCY 4
|
|
#define T4_IGNORE_LATENCY 5
|
|
unsigned long checkflag;
|
|
struct work_struct bh_work;
|
|
/* Latency related additions */
|
|
unsigned char rxident;
|
|
unsigned char lastindex;
|
|
int numbufs;
|
|
int needed_latency;
|
|
|
|
#ifdef VPM_SUPPORT
|
|
struct vpm450m *vpm;
|
|
#endif
|
|
struct spi_state st;
|
|
};
|
|
|
|
static inline bool is_pcie(const struct t4 *wc)
|
|
{
|
|
return (wc->devtype->flags & FLAG_EXPRESS) > 0;
|
|
}
|
|
|
|
static inline bool has_e1_span(const struct t4 *wc)
|
|
{
|
|
return (wc->t1e1 > 0);
|
|
}
|
|
|
|
static inline bool is_octal(const struct t4 *wc)
|
|
{
|
|
return (wc->devtype->flags & FLAG_8PORT) > 0;
|
|
}
|
|
|
|
static inline int T4_BASE_SIZE(struct t4 *wc)
|
|
{
|
|
if (is_octal(wc))
|
|
return DAHDI_MAX_CHUNKSIZE * 32 * 8;
|
|
else
|
|
return DAHDI_MAX_CHUNKSIZE * 32 * 4;
|
|
}
|
|
|
|
/**
|
|
* ports_on_framer - The number of ports on the framers.
|
|
* @wc: Board to check.
|
|
*
|
|
* The framer ports could be different the the number of ports on the card
|
|
* since the dual spans have four ports internally but two ports extenally.
|
|
*
|
|
*/
|
|
static inline unsigned int ports_on_framer(const struct t4 *wc)
|
|
{
|
|
return (is_octal(wc)) ? 8 : 4;
|
|
}
|
|
|
|
#ifdef VPM_SUPPORT
|
|
static void t4_vpm_init(struct t4 *wc);
|
|
|
|
static void echocan_free(struct dahdi_chan *chan, struct dahdi_echocan_state *ec);
|
|
|
|
static const struct dahdi_echocan_features vpm_ec_features = {
|
|
.NLP_automatic = 1,
|
|
.CED_tx_detect = 1,
|
|
.CED_rx_detect = 1,
|
|
};
|
|
|
|
static const struct dahdi_echocan_ops vpm_ec_ops = {
|
|
.echocan_free = echocan_free,
|
|
};
|
|
#endif
|
|
|
|
static void __set_clear(struct t4 *wc, int span);
|
|
static int _t4_startup(struct file *file, struct dahdi_span *span);
|
|
static int t4_startup(struct file *file, struct dahdi_span *span);
|
|
static int t4_shutdown(struct dahdi_span *span);
|
|
static int t4_rbsbits(struct dahdi_chan *chan, int bits);
|
|
static int t4_maint(struct dahdi_span *span, int cmd);
|
|
static int t4_clear_maint(struct dahdi_span *span);
|
|
static int t4_reset_counters(struct dahdi_span *span);
|
|
#ifdef SUPPORT_GEN1
|
|
static int t4_reset_dma(struct t4 *wc);
|
|
#endif
|
|
static void t4_hdlc_hard_xmit(struct dahdi_chan *chan);
|
|
static int t4_ioctl(struct dahdi_chan *chan, unsigned int cmd, unsigned long data);
|
|
static void t4_tsi_assign(struct t4 *wc, int fromspan, int fromchan, int tospan, int tochan);
|
|
static void t4_tsi_unassign(struct t4 *wc, int tospan, int tochan);
|
|
static void __t4_set_rclk_src(struct t4 *wc, int span);
|
|
static void __t4_set_sclk_src(struct t4 *wc, int mode, int master, int slave);
|
|
static void t4_check_alarms(struct t4 *wc, int span);
|
|
static void t4_check_sigbits(struct t4 *wc, int span);
|
|
|
|
#define WC_RDADDR 0
|
|
#define WC_WRADDR 1
|
|
#define WC_COUNT 2
|
|
#define WC_DMACTRL 3
|
|
#define WC_INTR 4
|
|
/* #define WC_GPIO 5 */
|
|
#define WC_VERSION 6
|
|
#define WC_LEDS 7
|
|
#define WC_GPIOCTL 8
|
|
#define WC_GPIO 9
|
|
#define WC_LADDR 10
|
|
#define WC_LDATA 11
|
|
#define WC_LEDS2 12
|
|
|
|
#define WC_SET_AUTH (1 << 20)
|
|
#define WC_GET_AUTH (1 << 12)
|
|
|
|
#define WC_LFRMR_CS (1 << 10) /* Framer's ChipSelect signal */
|
|
#define WC_LCS (1 << 11)
|
|
#define WC_LCS2 (1 << 12)
|
|
#define WC_LALE (1 << 13)
|
|
#define WC_LFRMR_CS2 (1 << 14) /* Framer's ChipSelect signal 2 */
|
|
#define WC_LREAD (1 << 15)
|
|
#define WC_LWRITE (1 << 16)
|
|
|
|
#define WC_ACTIVATE (1 << 12)
|
|
|
|
#define WC_OFF (0)
|
|
#define WC_RED (1)
|
|
#define WC_GREEN (2)
|
|
#define WC_YELLOW (3)
|
|
|
|
#define WC_RECOVER 0
|
|
#define WC_SELF 1
|
|
|
|
#define LIM0_T 0x36 /* Line interface mode 0 register */
|
|
#define LIM0_LL (1 << 1) /* Local Loop */
|
|
#define LIM1_T 0x37 /* Line interface mode 1 register */
|
|
#define LIM1_RL (1 << 1) /* Remote Loop */
|
|
|
|
#define FMR0 0x1C /* Framer Mode Register 0 */
|
|
#define FMR0_SIM (1 << 0) /* Alarm Simulation */
|
|
#define FMR1_T 0x1D /* Framer Mode Register 1 */
|
|
#define FMR1_ECM (1 << 2) /* Error Counter 1sec Interrupt Enable */
|
|
#define FMR5 0x21 /* Framer Mode Register 5 */
|
|
#define FMR5_XLU (1 << 4) /* Transmit loopup code */
|
|
#define FMR5_XLD (1 << 5) /* Transmit loopdown code */
|
|
#define FMR5_EIBR (1 << 6) /* Internal Bit Robbing Access */
|
|
#define DEC_T 0x60 /* Diable Error Counter */
|
|
#define IERR_T 0x1B /* Single Bit Defect Insertion Register */
|
|
#define IBV (1 << 0) /* Bipolar violation */
|
|
#define IPE (1 << 1) /* PRBS defect */
|
|
#define ICASE (1 << 2) /* CAS defect */
|
|
#define ICRCE (1 << 3) /* CRC defect */
|
|
#define IMFE (1 << 4) /* Multiframe defect */
|
|
#define IFASE (1 << 5) /* FAS defect */
|
|
#define ISR3_SEC (1 << 6) /* Internal one-second interrupt bit mask */
|
|
#define ISR3_ES (1 << 7) /* Errored Second interrupt bit mask */
|
|
#define ESM 0x47 /* Errored Second mask register */
|
|
|
|
#define FMR2_T 0x1E /* Framer Mode Register 2 */
|
|
#define FMR2_PLB (1 << 2) /* Framer Mode Register 2 */
|
|
|
|
#define FECL_T 0x50 /* Framing Error Counter Lower Byte */
|
|
#define FECH_T 0x51 /* Framing Error Counter Higher Byte */
|
|
#define CVCL_T 0x52 /* Code Violation Counter Lower Byte */
|
|
#define CVCH_T 0x53 /* Code Violation Counter Higher Byte */
|
|
#define CEC1L_T 0x54 /* CRC Error Counter 1 Lower Byte */
|
|
#define CEC1H_T 0x55 /* CRC Error Counter 1 Higher Byte */
|
|
#define EBCL_T 0x56 /* E-Bit Error Counter Lower Byte */
|
|
#define EBCH_T 0x57 /* E-Bit Error Counter Higher Byte */
|
|
#define BECL_T 0x58 /* Bit Error Counter Lower Byte */
|
|
#define BECH_T 0x59 /* Bit Error Counter Higher Byte */
|
|
#define COEC_T 0x5A /* COFA Event Counter */
|
|
#define PRBSSTA_T 0xDA /* PRBS Status Register */
|
|
|
|
#define LCR1_T 0x3B /* Loop Code Register 1 */
|
|
#define EPRM (1 << 7) /* Enable PRBS rx */
|
|
#define XPRBS (1 << 6) /* Enable PRBS tx */
|
|
#define FLLB (1 << 1) /* Framed line loop/Invert */
|
|
#define LLBP (1 << 0) /* Line Loopback Pattern */
|
|
#define TPC0_T 0xA8 /* Test Pattern Control Register */
|
|
#define FRA (1 << 6) /* Framed/Unframed Selection */
|
|
#define PRBS23 (3 << 4) /* Pattern selection (23 poly) */
|
|
#define PRM (1 << 2) /* Non framed mode */
|
|
#define FRS1_T 0x4D /* Framer Receive Status Reg 1 */
|
|
#define LLBDD (1 << 4)
|
|
#define LLBAD (1 << 3)
|
|
|
|
#define MAX_T4_CARDS 64
|
|
|
|
static struct t4 *cards[MAX_T4_CARDS];
|
|
|
|
struct t8_firm_header {
|
|
u8 header[6];
|
|
__le32 chksum;
|
|
u8 pad[18];
|
|
__le32 version;
|
|
} __packed;
|
|
|
|
#define MAX_TDM_CHAN 32
|
|
#define MAX_DTMF_DET 16
|
|
|
|
#define HDLC_IMR0_MASK (FRMR_IMR0_RME | FRMR_IMR0_RPF)
|
|
#define HDLC_IMR1_MASK (FRMR_IMR1_XDU | FRMR_IMR1_XPR)
|
|
|
|
static inline unsigned int __t4_pci_in(struct t4 *wc, const unsigned int addr)
|
|
{
|
|
unsigned int res = readl(wc->membase + (addr * sizeof(u32)));
|
|
return res;
|
|
}
|
|
|
|
static inline void __t4_pci_out(struct t4 *wc, const unsigned int addr, const unsigned int value)
|
|
{
|
|
#ifdef DEBUG
|
|
unsigned int tmp;
|
|
#endif
|
|
writel(value, wc->membase + (addr * sizeof(u32)));
|
|
#ifdef DEBUG
|
|
tmp = __t4_pci_in(wc, WC_VERSION);
|
|
if ((tmp & 0xffff0000) != 0xc01a0000)
|
|
dev_notice(&wc->dev->dev,
|
|
"Version Synchronization Error!\n");
|
|
#else
|
|
__t4_pci_in(wc, WC_VERSION);
|
|
#endif
|
|
}
|
|
|
|
static inline void __t4_gpio_set(struct t4 *wc, unsigned bits, unsigned int val)
|
|
{
|
|
unsigned int newgpio;
|
|
newgpio = wc->gpio & (~bits);
|
|
newgpio |= val;
|
|
if (newgpio != wc->gpio) {
|
|
wc->gpio = newgpio;
|
|
__t4_pci_out(wc, WC_GPIO, wc->gpio);
|
|
}
|
|
}
|
|
|
|
static inline void __t4_gpio_setdir(struct t4 *wc, unsigned int bits, unsigned int val)
|
|
{
|
|
unsigned int newgpioctl;
|
|
newgpioctl = wc->gpioctl & (~bits);
|
|
newgpioctl |= val;
|
|
if (newgpioctl != wc->gpioctl) {
|
|
wc->gpioctl = newgpioctl;
|
|
__t4_pci_out(wc, WC_GPIOCTL, wc->gpioctl);
|
|
}
|
|
}
|
|
|
|
static inline void t4_gpio_setdir(struct t4 *wc, unsigned int bits, unsigned int val)
|
|
{
|
|
unsigned long flags;
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
__t4_gpio_setdir(wc, bits, val);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
static inline void t4_gpio_set(struct t4 *wc, unsigned int bits, unsigned int val)
|
|
{
|
|
unsigned long flags;
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
__t4_gpio_set(wc, bits, val);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
static inline void t4_pci_out(struct t4 *wc, const unsigned int addr, const unsigned int value)
|
|
{
|
|
unsigned long flags;
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
__t4_pci_out(wc, addr, value);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
static inline void __t4_set_led(struct t4 *wc, int span, int color)
|
|
{
|
|
if (span <= 3) {
|
|
int oldreg = wc->ledreg;
|
|
|
|
wc->ledreg &= ~(0x3 << (span << 1));
|
|
wc->ledreg |= (color << (span << 1));
|
|
if (oldreg != wc->ledreg)
|
|
__t4_pci_out(wc, WC_LEDS, wc->ledreg);
|
|
} else {
|
|
int oldreg = wc->ledreg2;
|
|
|
|
span &= 3;
|
|
wc->ledreg2 &= ~(0x3 << (span << 1));
|
|
wc->ledreg2 |= (color << (span << 1));
|
|
if (oldreg != wc->ledreg2)
|
|
__t4_pci_out(wc, WC_LEDS2, wc->ledreg2);
|
|
}
|
|
}
|
|
|
|
static inline void t4_activate(struct t4 *wc)
|
|
{
|
|
wc->ledreg |= WC_ACTIVATE;
|
|
t4_pci_out(wc, WC_LEDS, wc->ledreg);
|
|
}
|
|
|
|
static inline unsigned int t4_pci_in(struct t4 *wc, const unsigned int addr)
|
|
{
|
|
unsigned int ret;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
ret = __t4_pci_in(wc, addr);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
return ret;
|
|
}
|
|
|
|
static unsigned int __t4_framer_in(const struct t4 *wc, int unit,
|
|
const unsigned int addr)
|
|
{
|
|
unsigned int ret;
|
|
register u32 val;
|
|
void __iomem *const wc_laddr = wc->membase + (WC_LADDR*sizeof(u32));
|
|
void __iomem *const wc_version = wc->membase + (WC_VERSION*sizeof(u32));
|
|
void __iomem *const wc_ldata = wc->membase + (WC_LDATA*sizeof(u32));
|
|
int haddr = (((unit & 4) ? 0 : WC_LFRMR_CS2));
|
|
unit &= 0x3;
|
|
|
|
val = ((unit & 0x3) << 8) | (addr & 0xff) | haddr;
|
|
writel(val, wc_laddr);
|
|
readl(wc_version);
|
|
writel(val | WC_LFRMR_CS | WC_LREAD, wc_laddr);
|
|
readl(wc_version);
|
|
ret = readb(wc_ldata);
|
|
writel(val, wc_laddr);
|
|
readl(wc_version);
|
|
return ret;
|
|
}
|
|
|
|
static unsigned int
|
|
t4_framer_in(struct t4 *wc, int unit, const unsigned int addr)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int ret;
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
ret = __t4_framer_in(wc, unit, addr);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
return ret;
|
|
}
|
|
|
|
static void __t4_framer_out(const struct t4 *wc, int unit, const u8 addr,
|
|
const unsigned int value)
|
|
{
|
|
register u32 val;
|
|
void __iomem *const wc_laddr = wc->membase + (WC_LADDR*sizeof(u32));
|
|
void __iomem *const wc_version = wc->membase + (WC_VERSION*sizeof(u32));
|
|
void __iomem *const wc_ldata = wc->membase + (WC_LDATA*sizeof(u32));
|
|
int haddr = (((unit & 4) ? 0 : WC_LFRMR_CS2));
|
|
|
|
val = ((unit & 0x3) << 8) | (addr & 0xff) | haddr;
|
|
writel(val, wc_laddr);
|
|
readl(wc_version);
|
|
writel(value, wc_ldata);
|
|
readl(wc_version);
|
|
writel(val | WC_LFRMR_CS | WC_LWRITE, wc_laddr);
|
|
readl(wc_version);
|
|
writel(val, wc_laddr);
|
|
readl(wc_version);
|
|
}
|
|
|
|
static void t4_framer_out(struct t4 *wc, int unit,
|
|
const unsigned int addr,
|
|
const unsigned int value)
|
|
{
|
|
unsigned long flags;
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
__t4_framer_out(wc, unit, addr, value);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
#ifdef VPM_SUPPORT
|
|
|
|
static inline void __t4_raw_oct_out(struct t4 *wc, const unsigned int addr, const unsigned int value)
|
|
{
|
|
int octopt = wc->tspans[0]->spanflags & FLAG_OCTOPT;
|
|
if (!octopt)
|
|
__t4_gpio_set(wc, 0xff, (addr >> 8));
|
|
__t4_pci_out(wc, WC_LDATA, 0x10000 | (addr & 0xffff));
|
|
if (!octopt)
|
|
__t4_pci_out(wc, WC_LADDR, (WC_LWRITE));
|
|
__t4_pci_out(wc, WC_LADDR, (WC_LWRITE | WC_LALE));
|
|
if (!octopt)
|
|
__t4_gpio_set(wc, 0xff, (value >> 8));
|
|
__t4_pci_out(wc, WC_LDATA, (value & 0xffff));
|
|
__t4_pci_out(wc, WC_LADDR, (WC_LWRITE | WC_LALE | WC_LCS));
|
|
__t4_pci_out(wc, WC_LADDR, (0));
|
|
}
|
|
|
|
static inline unsigned int __t4_raw_oct_in(struct t4 *wc, const unsigned int addr)
|
|
{
|
|
unsigned int ret;
|
|
int octopt = wc->tspans[0]->spanflags & FLAG_OCTOPT;
|
|
if (!octopt)
|
|
__t4_gpio_set(wc, 0xff, (addr >> 8));
|
|
__t4_pci_out(wc, WC_LDATA, 0x10000 | (addr & 0xffff));
|
|
if (!octopt)
|
|
__t4_pci_out(wc, WC_LADDR, (WC_LWRITE));
|
|
__t4_pci_out(wc, WC_LADDR, (WC_LWRITE | WC_LALE));
|
|
__t4_pci_out(wc, WC_LADDR, (WC_LALE));
|
|
if (!octopt) {
|
|
__t4_gpio_setdir(wc, 0xff, 0x00);
|
|
__t4_gpio_set(wc, 0xff, 0x00);
|
|
}
|
|
__t4_pci_out(wc, WC_LADDR, (WC_LREAD | WC_LALE | WC_LCS));
|
|
if (octopt) {
|
|
ret = __t4_pci_in(wc, WC_LDATA) & 0xffff;
|
|
} else {
|
|
ret = __t4_pci_in(wc, WC_LDATA) & 0xff;
|
|
ret |= (__t4_pci_in(wc, WC_GPIO) & 0xff) << 8;
|
|
}
|
|
__t4_pci_out(wc, WC_LADDR, (0));
|
|
if (!octopt)
|
|
__t4_gpio_setdir(wc, 0xff, 0xff);
|
|
return ret & 0xffff;
|
|
}
|
|
|
|
static inline unsigned int __t4_oct_in(struct t4 *wc, unsigned int addr)
|
|
{
|
|
#ifdef PEDANTIC_OCTASIC_CHECKING
|
|
int count = 1000;
|
|
#endif
|
|
__t4_raw_oct_out(wc, 0x0008, (addr >> 20));
|
|
__t4_raw_oct_out(wc, 0x000a, (addr >> 4) & ((1 << 16) - 1));
|
|
__t4_raw_oct_out(wc, 0x0000, (((addr >> 1) & 0x7) << 9) | (1 << 8) | (1));
|
|
#ifdef PEDANTIC_OCTASIC_CHECKING
|
|
while((__t4_raw_oct_in(wc, 0x0000) & (1 << 8)) && --count);
|
|
if (count != 1000)
|
|
dev_notice(&wc->dev->dev, "Yah, read can be slow...\n");
|
|
if (!count)
|
|
dev_notice(&wc->dev->dev, "Read timed out!\n");
|
|
#endif
|
|
return __t4_raw_oct_in(wc, 0x0004);
|
|
}
|
|
|
|
static inline unsigned int t4_oct_in(struct t4 *wc, const unsigned int addr)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int ret;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
ret = __t4_oct_in(wc, addr);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
return ret;
|
|
}
|
|
|
|
static inline void __t4_oct_out(struct t4 *wc, unsigned int addr, unsigned int value)
|
|
{
|
|
#ifdef PEDANTIC_OCTASIC_CHECKING
|
|
int count = 1000;
|
|
#endif
|
|
__t4_raw_oct_out(wc, 0x0008, (addr >> 20));
|
|
__t4_raw_oct_out(wc, 0x000a, (addr >> 4) & ((1 << 16) - 1));
|
|
__t4_raw_oct_out(wc, 0x0004, value);
|
|
__t4_raw_oct_out(wc, 0x0000, (((addr >> 1) & 0x7) << 9) | (1 << 8) | (3 << 12) | 1);
|
|
#ifdef PEDANTIC_OCTASIC_CHECKING
|
|
while((__t4_raw_oct_in(wc, 0x0000) & (1 << 8)) && --count);
|
|
if (count != 1000)
|
|
dev_notice(&wc->dev->dev, "Yah, write can be slow\n");
|
|
if (!count)
|
|
dev_notice(&wc->dev->dev, "Write timed out!\n");
|
|
#endif
|
|
}
|
|
|
|
static inline void t4_oct_out(struct t4 *wc, const unsigned int addr, const unsigned int value)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
__t4_oct_out(wc, addr, value);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
static void t4_check_vpm(struct t4 *wc)
|
|
{
|
|
int channel, tone, start, span;
|
|
|
|
if (vpm450m_checkirq(wc->vpm)) {
|
|
while(vpm450m_getdtmf(wc->vpm, &channel, &tone, &start)) {
|
|
span = channel & 0x3;
|
|
channel >>= 2;
|
|
if (!has_e1_span(wc))
|
|
channel -= 5;
|
|
else
|
|
channel -= 1;
|
|
if (unlikely(debug))
|
|
dev_info(&wc->dev->dev, "Got tone %s of '%c' "
|
|
"on channel %d of span %d\n",
|
|
(start ? "START" : "STOP"),
|
|
tone, channel, span + 1);
|
|
if (test_bit(channel, &wc->tspans[span]->dtmfmask) && (tone != 'u')) {
|
|
if (start) {
|
|
/* The octasic is supposed to mute us, but... Yah, you
|
|
guessed it. */
|
|
if (test_bit(channel, &wc->tspans[span]->dtmfmutemask)) {
|
|
unsigned long flags;
|
|
struct dahdi_chan *chan = wc->tspans[span]->span.chans[channel];
|
|
int y;
|
|
spin_lock_irqsave(&chan->lock, flags);
|
|
for (y=0;y<chan->numbufs;y++) {
|
|
if ((chan->inreadbuf > -1) && (chan->readidx[y]))
|
|
memset(chan->readbuf[chan->inreadbuf], DAHDI_XLAW(0, chan), chan->readidx[y]);
|
|
}
|
|
spin_unlock_irqrestore(&chan->lock, flags);
|
|
}
|
|
set_bit(channel, &wc->tspans[span]->dtmfactive);
|
|
dahdi_qevent_lock(wc->tspans[span]->span.chans[channel], (DAHDI_EVENT_DTMFDOWN | tone));
|
|
} else {
|
|
clear_bit(channel, &wc->tspans[span]->dtmfactive);
|
|
dahdi_qevent_lock(wc->tspans[span]->span.chans[channel], (DAHDI_EVENT_DTMFUP | tone));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
#endif /* VPM_SUPPORT */
|
|
|
|
static void hdlc_stop(struct t4 *wc, unsigned int span)
|
|
{
|
|
struct t4_span *t = wc->tspans[span];
|
|
unsigned char imr0, imr1, mode;
|
|
unsigned long flags;
|
|
int i = 0;
|
|
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Stopping HDLC controller on span "
|
|
"%d\n", span+1);
|
|
|
|
/* Clear receive and transmit timeslots */
|
|
for (i = 0; i < 4; i++) {
|
|
t4_framer_out(wc, span, FRMR_RTR_BASE + i, 0x00);
|
|
t4_framer_out(wc, span, FRMR_TTR_BASE + i, 0x00);
|
|
}
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
imr0 = __t4_framer_in(wc, span, FRMR_IMR0);
|
|
imr1 = __t4_framer_in(wc, span, FRMR_IMR1);
|
|
|
|
/* Disable HDLC interrupts */
|
|
imr0 |= HDLC_IMR0_MASK;
|
|
__t4_framer_out(wc, span, FRMR_IMR0, imr0);
|
|
|
|
imr1 |= HDLC_IMR1_MASK;
|
|
__t4_framer_out(wc, span, FRMR_IMR1, imr1);
|
|
|
|
mode = __t4_framer_in(wc, span, FRMR_MODE);
|
|
mode &= ~FRMR_MODE_HRAC;
|
|
__t4_framer_out(wc, span, FRMR_MODE, mode);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
t->sigactive = 0;
|
|
}
|
|
|
|
static inline void __t4_framer_cmd(struct t4 *wc, unsigned int span, int cmd)
|
|
{
|
|
__t4_framer_out(wc, span, FRMR_CMDR, cmd);
|
|
}
|
|
|
|
static inline void t4_framer_cmd_wait(struct t4 *wc, unsigned int span, int cmd)
|
|
{
|
|
int sis;
|
|
int loops = 0;
|
|
|
|
/* XXX could be time consuming XXX */
|
|
for (;;) {
|
|
sis = t4_framer_in(wc, span, FRMR_SIS);
|
|
if (!(sis & 0x04))
|
|
break;
|
|
if (!loops++ && (debug & DEBUG_FRAMER)) {
|
|
dev_notice(&wc->dev->dev, "!!!SIS Waiting before cmd "
|
|
"%02x\n", cmd);
|
|
}
|
|
}
|
|
if (loops && (debug & DEBUG_FRAMER))
|
|
dev_notice(&wc->dev->dev, "!!!SIS waited %d loops\n", loops);
|
|
|
|
t4_framer_out(wc, span, FRMR_CMDR, cmd);
|
|
}
|
|
|
|
static int hdlc_start(struct t4 *wc, unsigned int span, struct dahdi_chan *chan, unsigned char mode)
|
|
{
|
|
struct t4_span *t = wc->tspans[span];
|
|
unsigned char imr0, imr1;
|
|
int offset = chan->chanpos;
|
|
unsigned long flags;
|
|
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_info(&wc->dev->dev, "Starting HDLC controller for channel "
|
|
"%d span %d\n", offset, span+1);
|
|
|
|
if (mode != FRMR_MODE_NO_ADDR_CMP)
|
|
return -1;
|
|
|
|
mode |= FRMR_MODE_HRAC;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
|
|
/* Make sure we're in the right mode */
|
|
__t4_framer_out(wc, span, FRMR_MODE, mode);
|
|
__t4_framer_out(wc, span, FRMR_TSEO, 0x00);
|
|
__t4_framer_out(wc, span, FRMR_TSBS1, hardhdlcmode);
|
|
|
|
/* Set the interframe gaps, etc */
|
|
__t4_framer_out(wc, span, FRMR_CCR1, FRMR_CCR1_ITF|FRMR_CCR1_EITS);
|
|
|
|
__t4_framer_out(wc, span, FRMR_CCR2, FRMR_CCR2_RCRC);
|
|
|
|
/* Set up the time slot that we want to tx/rx on */
|
|
__t4_framer_out(wc, span, FRMR_TTR_BASE + (offset / 8), (0x80 >> (offset % 8)));
|
|
__t4_framer_out(wc, span, FRMR_RTR_BASE + (offset / 8), (0x80 >> (offset % 8)));
|
|
|
|
imr0 = __t4_framer_in(wc, span, FRMR_IMR0);
|
|
imr1 = __t4_framer_in(wc, span, FRMR_IMR1);
|
|
|
|
/* Enable our interrupts again */
|
|
imr0 &= ~HDLC_IMR0_MASK;
|
|
__t4_framer_out(wc, span, FRMR_IMR0, imr0);
|
|
|
|
imr1 &= ~HDLC_IMR1_MASK;
|
|
__t4_framer_out(wc, span, FRMR_IMR1, imr1);
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
/* Reset the signaling controller */
|
|
t4_framer_cmd_wait(wc, span, FRMR_CMDR_SRES);
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
t->sigchan = chan;
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
t->sigactive = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void __set_clear(struct t4 *wc, int span)
|
|
{
|
|
int i,j;
|
|
int oldnotclear;
|
|
unsigned short val=0;
|
|
struct t4_span *ts = wc->tspans[span];
|
|
|
|
oldnotclear = ts->notclear;
|
|
if (E1 != ts->linemode) {
|
|
for (i=0;i<24;i++) {
|
|
j = (i/8);
|
|
if (ts->span.chans[i]->flags & DAHDI_FLAG_CLEAR) {
|
|
val |= 1 << (7 - (i % 8));
|
|
ts->notclear &= ~(1 << i);
|
|
} else
|
|
ts->notclear |= (1 << i);
|
|
if ((i % 8)==7) {
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "Putting %d "
|
|
"in register %02x on span %d"
|
|
"\n", val, 0x2f + j, span + 1);
|
|
__t4_framer_out(wc, span, 0x2f + j, val);
|
|
val = 0;
|
|
}
|
|
}
|
|
} else {
|
|
for (i=0;i<31;i++) {
|
|
if (ts->span.chans[i]->flags & DAHDI_FLAG_CLEAR)
|
|
ts->notclear &= ~(1 << i);
|
|
else
|
|
ts->notclear |= (1 << i);
|
|
}
|
|
}
|
|
if (ts->notclear != oldnotclear) {
|
|
unsigned char reg;
|
|
reg = __t4_framer_in(wc, span, FRMR_IMR0);
|
|
if (ts->notclear)
|
|
reg &= ~0x08;
|
|
else
|
|
reg |= 0x08;
|
|
__t4_framer_out(wc, span, FRMR_IMR0, reg);
|
|
}
|
|
}
|
|
|
|
static int t4_dacs(struct dahdi_chan *dst, struct dahdi_chan *src)
|
|
{
|
|
struct t4 *wc;
|
|
struct t4_span *ts;
|
|
wc = dst->pvt;
|
|
ts = wc->tspans[dst->span->offset];
|
|
if (src && (src->pvt != dst->pvt)) {
|
|
if (ts->spanflags & FLAG_2NDGEN)
|
|
t4_tsi_unassign(wc, dst->span->offset, dst->chanpos);
|
|
wc = src->pvt;
|
|
if (ts->spanflags & FLAG_2NDGEN)
|
|
t4_tsi_unassign(wc, src->span->offset, src->chanpos);
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "Unassigning %d/%d by "
|
|
"default and...\n", src->span->offset,
|
|
src->chanpos);
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "Unassigning %d/%d by "
|
|
"default\n", dst->span->offset, dst->chanpos);
|
|
return -1;
|
|
}
|
|
if (src) {
|
|
t4_tsi_assign(wc, src->span->offset, src->chanpos, dst->span->offset, dst->chanpos);
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "Assigning channel %d/%d -> "
|
|
"%d/%d!\n", src->span->offset, src->chanpos,
|
|
dst->span->offset, dst->chanpos);
|
|
} else {
|
|
t4_tsi_unassign(wc, dst->span->offset, dst->chanpos);
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "Unassigning channel %d/%d!"
|
|
"\n", dst->span->offset, dst->chanpos);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#ifdef VPM_SUPPORT
|
|
|
|
void oct_set_reg(void *data, unsigned int reg, unsigned int val)
|
|
{
|
|
struct t4 *wc = data;
|
|
t4_oct_out(wc, reg, val);
|
|
}
|
|
|
|
unsigned int oct_get_reg(void *data, unsigned int reg)
|
|
{
|
|
struct t4 *wc = data;
|
|
unsigned int ret;
|
|
ret = t4_oct_in(wc, reg);
|
|
return ret;
|
|
}
|
|
|
|
static const char *__t4_echocan_name(struct t4 *wc)
|
|
{
|
|
if (wc->vpm) {
|
|
if (wc->numspans == 2)
|
|
return vpmoct064_name;
|
|
else if (wc->numspans == 4)
|
|
return vpmoct128_name;
|
|
else if (wc->numspans == 8)
|
|
return vpmoct256_name;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static const char *t4_echocan_name(const struct dahdi_chan *chan)
|
|
{
|
|
struct t4 *wc = chan->pvt;
|
|
return __t4_echocan_name(wc);
|
|
}
|
|
|
|
static int t4_echocan_create(struct dahdi_chan *chan,
|
|
struct dahdi_echocanparams *ecp,
|
|
struct dahdi_echocanparam *p,
|
|
struct dahdi_echocan_state **ec)
|
|
{
|
|
struct t4 *wc = chan->pvt;
|
|
struct t4_span *tspan = container_of(chan->span, struct t4_span, span);
|
|
int channel;
|
|
const bool alaw = (chan->span->deflaw == 2);
|
|
|
|
if (!vpmsupport || !wc->vpm)
|
|
return -ENODEV;
|
|
|
|
if (ecp->param_count > 0) {
|
|
dev_warn(&wc->dev->dev, "%s echo canceller does not support "
|
|
"parameters; failing request\n",
|
|
chan->ec_factory->get_name(chan));
|
|
return -EINVAL;
|
|
}
|
|
|
|
*ec = tspan->ec[chan->chanpos - 1];
|
|
(*ec)->ops = &vpm_ec_ops;
|
|
(*ec)->features = vpm_ec_features;
|
|
|
|
channel = has_e1_span(wc) ? chan->chanpos : chan->chanpos + 4;
|
|
|
|
if (is_octal(wc))
|
|
channel = channel << 3;
|
|
else
|
|
channel = channel << 2;
|
|
channel |= chan->span->offset;
|
|
if (debug & DEBUG_ECHOCAN) {
|
|
dev_notice(&wc->dev->dev,
|
|
"echocan: Card is %d, Channel is %d, Span is %d, offset is %d length %d\n",
|
|
wc->num, chan->chanpos, chan->span->offset,
|
|
channel, ecp->tap_length);
|
|
}
|
|
vpm450m_set_alaw_companding(wc->vpm, channel, alaw);
|
|
vpm450m_setec(wc->vpm, channel, ecp->tap_length);
|
|
return 0;
|
|
}
|
|
|
|
static void echocan_free(struct dahdi_chan *chan, struct dahdi_echocan_state *ec)
|
|
{
|
|
struct t4 *wc = chan->pvt;
|
|
int channel;
|
|
|
|
if (!wc->vpm)
|
|
return;
|
|
|
|
memset(ec, 0, sizeof(*ec));
|
|
channel = has_e1_span(wc) ? chan->chanpos : chan->chanpos + 4;
|
|
|
|
if (is_octal(wc))
|
|
channel = channel << 3;
|
|
else
|
|
channel = channel << 2;
|
|
channel |= chan->span->offset;
|
|
if (debug & DEBUG_ECHOCAN) {
|
|
dev_notice(&wc->dev->dev,
|
|
"echocan: Card is %d, Channel is %d, Span is %d, offset is %d length 0\n",
|
|
wc->num, chan->chanpos, chan->span->offset, channel);
|
|
}
|
|
vpm450m_setec(wc->vpm, channel, 0);
|
|
}
|
|
#endif
|
|
|
|
static int t4_ioctl(struct dahdi_chan *chan, unsigned int cmd, unsigned long data)
|
|
{
|
|
struct t4_regs regs;
|
|
struct t4_reg reg;
|
|
int x;
|
|
struct t4 *wc = chan->pvt;
|
|
#ifdef VPM_SUPPORT
|
|
int j;
|
|
int channel;
|
|
struct t4_span *ts = wc->tspans[chan->span->offset];
|
|
#endif
|
|
|
|
switch(cmd) {
|
|
case WCT4_SET_REG:
|
|
if (copy_from_user(®, (struct t4_reg __user *)data,
|
|
sizeof(reg)))
|
|
return -EFAULT;
|
|
t4_pci_out(wc, reg.reg, reg.val);
|
|
break;
|
|
case WCT4_GET_REG:
|
|
if (copy_from_user(®, (struct t4_reg __user *)data,
|
|
sizeof(reg)))
|
|
return -EFAULT;
|
|
reg.val = t4_pci_in(wc, reg.reg);
|
|
if (copy_to_user((struct t4_reg __user *)data,
|
|
®, sizeof(reg)))
|
|
return -EFAULT;
|
|
break;
|
|
case WCT4_GET_REGS:
|
|
for (x=0;x<NUM_PCI;x++)
|
|
regs.pci[x] = t4_pci_in(wc, x);
|
|
for (x=0;x<NUM_REGS;x++)
|
|
regs.regs[x] = t4_framer_in(wc, chan->span->offset, x);
|
|
if (copy_to_user((void __user *) data,
|
|
®s, sizeof(regs)))
|
|
return -EFAULT;
|
|
break;
|
|
#ifdef VPM_SUPPORT
|
|
case DAHDI_TONEDETECT:
|
|
if (get_user(j, (__user int *) data))
|
|
return -EFAULT;
|
|
if (!wc->vpm)
|
|
return -ENOSYS;
|
|
if (j && (vpmdtmfsupport == 0))
|
|
return -ENOSYS;
|
|
if (j & DAHDI_TONEDETECT_ON)
|
|
set_bit(chan->chanpos - 1, &ts->dtmfmask);
|
|
else
|
|
clear_bit(chan->chanpos - 1, &ts->dtmfmask);
|
|
if (j & DAHDI_TONEDETECT_MUTE)
|
|
set_bit(chan->chanpos - 1, &ts->dtmfmutemask);
|
|
else
|
|
clear_bit(chan->chanpos - 1, &ts->dtmfmutemask);
|
|
|
|
channel = has_e1_span(wc) ? chan->chanpos : chan->chanpos + 4;
|
|
if (is_octal(wc))
|
|
channel = channel << 3;
|
|
else
|
|
channel = channel << 2;
|
|
channel |= chan->span->offset;
|
|
vpm450m_setdtmf(wc->vpm, channel, j & DAHDI_TONEDETECT_ON,
|
|
j & DAHDI_TONEDETECT_MUTE);
|
|
return 0;
|
|
#endif
|
|
default:
|
|
return -ENOTTY;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void inline t4_hdlc_xmit_fifo(struct t4 *wc, unsigned int span, struct t4_span *ts)
|
|
{
|
|
int res, i;
|
|
unsigned int size = 32;
|
|
unsigned char buf[32];
|
|
|
|
res = dahdi_hdlc_getbuf(ts->sigchan, buf, &size);
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Got buffer sized %d and res %d "
|
|
"for %d\n", size, res, span);
|
|
if (size > 0) {
|
|
ts->sigactive = 1;
|
|
|
|
if (debug & DEBUG_FRAMER) {
|
|
dev_notice(&wc->dev->dev, "TX(");
|
|
for (i = 0; i < size; i++)
|
|
dev_notice(&wc->dev->dev, "%s%02x",
|
|
(i ? " " : ""), buf[i]);
|
|
dev_notice(&wc->dev->dev, ")\n");
|
|
}
|
|
|
|
for (i = 0; i < size; i++)
|
|
t4_framer_out(wc, span, FRMR_TXFIFO, buf[i]);
|
|
|
|
if (res) /* End of message */ {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev,
|
|
"transmiting XHF|XME\n");
|
|
t4_framer_cmd_wait(wc, span, FRMR_CMDR_XHF | FRMR_CMDR_XME);
|
|
++ts->frames_out;
|
|
if ((debug & DEBUG_FRAMER) && !(ts->frames_out & 0x0f))
|
|
dev_notice(&wc->dev->dev, "Transmitted %d "
|
|
"frames on span %d\n", ts->frames_out,
|
|
span);
|
|
} else { /* Still more to transmit */
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "transmiting XHF\n");
|
|
t4_framer_cmd_wait(wc, span, FRMR_CMDR_XHF);
|
|
}
|
|
}
|
|
else if (res < 0)
|
|
ts->sigactive = 0;
|
|
}
|
|
|
|
static void t4_hdlc_hard_xmit(struct dahdi_chan *chan)
|
|
{
|
|
struct t4 *wc = chan->pvt;
|
|
int span = chan->span->offset;
|
|
struct t4_span *ts = wc->tspans[span];
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
if (!ts->sigchan) {
|
|
dev_notice(&wc->dev->dev, "t4_hdlc_hard_xmit: Invalid (NULL) "
|
|
"signalling channel\n");
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
return;
|
|
}
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "t4_hdlc_hard_xmit on channel %s "
|
|
"(sigchan %s), sigactive=%d\n", chan->name,
|
|
ts->sigchan->name, ts->sigactive);
|
|
|
|
if ((ts->sigchan == chan) && !ts->sigactive)
|
|
t4_hdlc_xmit_fifo(wc, span, ts);
|
|
}
|
|
|
|
/**
|
|
* t4_set_framer_bits - Atomically set bits in a framer register.
|
|
*/
|
|
static void t4_set_framer_bits(struct t4 *wc, unsigned int spanno,
|
|
unsigned int const addr, u16 bits)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int reg;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
reg = __t4_framer_in(wc, spanno, addr);
|
|
__t4_framer_out(wc, spanno, addr, (reg | bits));
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
static int t4_maint(struct dahdi_span *span, int cmd)
|
|
{
|
|
struct t4_span *ts = container_of(span, struct t4_span, span);
|
|
struct t4 *wc = ts->owner;
|
|
unsigned int reg;
|
|
unsigned long flags;
|
|
|
|
if (E1 == ts->linemode) {
|
|
switch(cmd) {
|
|
case DAHDI_MAINT_NONE:
|
|
dev_info(&wc->dev->dev, "Clearing all maint modes\n");
|
|
t4_clear_maint(span);
|
|
break;
|
|
case DAHDI_MAINT_LOCALLOOP:
|
|
dev_info(&wc->dev->dev,
|
|
"Turning on local loopback\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, LIM0_T, LIM0_LL);
|
|
break;
|
|
case DAHDI_MAINT_NETWORKLINELOOP:
|
|
dev_info(&wc->dev->dev,
|
|
"Turning on network line loopback\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, LIM1_T, LIM1_RL);
|
|
break;
|
|
case DAHDI_MAINT_NETWORKPAYLOADLOOP:
|
|
dev_info(&wc->dev->dev,
|
|
"Turning on network payload loopback\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, FMR2_T, FMR2_PLB);
|
|
break;
|
|
case DAHDI_MAINT_LOOPUP:
|
|
case DAHDI_MAINT_LOOPDOWN:
|
|
dev_info(&wc->dev->dev,
|
|
"Loopup & loopdown not supported in E1 mode\n");
|
|
return -ENOSYS;
|
|
case DAHDI_MAINT_FAS_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IFASE);
|
|
break;
|
|
case DAHDI_MAINT_MULTI_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IMFE);
|
|
break;
|
|
case DAHDI_MAINT_CRC_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, ICRCE);
|
|
break;
|
|
case DAHDI_MAINT_CAS_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, ICASE);
|
|
break;
|
|
case DAHDI_MAINT_PRBS_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IPE);
|
|
break;
|
|
case DAHDI_MAINT_BIPOLAR_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IBV);
|
|
break;
|
|
case DAHDI_RESET_COUNTERS:
|
|
t4_reset_counters(span);
|
|
break;
|
|
case DAHDI_MAINT_ALARM_SIM:
|
|
dev_info(&wc->dev->dev, "Invoking alarm state");
|
|
t4_set_framer_bits(wc, span->offset, FMR0, FMR0_SIM);
|
|
break;
|
|
default:
|
|
dev_info(&wc->dev->dev,
|
|
"Unknown E1 maint command: %d\n", cmd);
|
|
return -ENOSYS;
|
|
}
|
|
} else {
|
|
switch(cmd) {
|
|
case DAHDI_MAINT_NONE:
|
|
dev_info(&wc->dev->dev, "Clearing all maint modes\n");
|
|
t4_clear_maint(span);
|
|
break;
|
|
case DAHDI_MAINT_LOCALLOOP:
|
|
dev_info(&wc->dev->dev,
|
|
"Turning on local loopback\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, LIM0_T, LIM0_LL);
|
|
break;
|
|
case DAHDI_MAINT_NETWORKLINELOOP:
|
|
dev_info(&wc->dev->dev,
|
|
"Turning on network line loopback\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, LIM1_T, LIM1_RL);
|
|
break;
|
|
case DAHDI_MAINT_NETWORKPAYLOADLOOP:
|
|
dev_info(&wc->dev->dev,
|
|
"Turning on network payload loopback\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, FMR2_T, FMR2_PLB);
|
|
break;
|
|
case DAHDI_MAINT_LOOPUP:
|
|
dev_info(&wc->dev->dev, "Transmitting loopup code\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, FMR5, FMR5_XLU);
|
|
ts->span.maintstat = DAHDI_MAINT_REMOTELOOP;
|
|
break;
|
|
case DAHDI_MAINT_LOOPDOWN:
|
|
dev_info(&wc->dev->dev, "Transmitting loopdown code\n");
|
|
t4_clear_maint(span);
|
|
t4_set_framer_bits(wc, span->offset, FMR5, FMR5_XLD);
|
|
break;
|
|
case DAHDI_MAINT_FAS_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IFASE);
|
|
break;
|
|
case DAHDI_MAINT_MULTI_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IMFE);
|
|
break;
|
|
case DAHDI_MAINT_CRC_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, ICRCE);
|
|
break;
|
|
case DAHDI_MAINT_CAS_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, ICASE);
|
|
break;
|
|
case DAHDI_MAINT_PRBS_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IPE);
|
|
break;
|
|
case DAHDI_MAINT_BIPOLAR_DEFECT:
|
|
t4_framer_out(wc, span->offset, IERR_T, IBV);
|
|
break;
|
|
case DAHDI_MAINT_PRBS:
|
|
dev_info(&wc->dev->dev, "PRBS not supported\n");
|
|
return -ENOSYS;
|
|
case DAHDI_RESET_COUNTERS:
|
|
t4_reset_counters(span);
|
|
break;
|
|
case DAHDI_MAINT_ALARM_SIM:
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
reg = __t4_framer_in(wc, span->offset, FMR0);
|
|
|
|
/*
|
|
* The alarm simulation state machine requires us to
|
|
* bring this bit up and down for at least 1 clock cycle
|
|
*/
|
|
__t4_framer_out(wc, span->offset,
|
|
FMR0, (reg | FMR0_SIM));
|
|
udelay(1);
|
|
__t4_framer_out(wc, span->offset,
|
|
FMR0, (reg & ~FMR0_SIM));
|
|
udelay(1);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
reg = t4_framer_in(wc, span->offset, 0x4e);
|
|
if (debug & DEBUG_MAIN) {
|
|
dev_info(&wc->dev->dev,
|
|
"FRS2(alarm state): %d\n",
|
|
((reg & 0xe0) >> 5));
|
|
}
|
|
break;
|
|
default:
|
|
dev_info(&wc->dev->dev, "Unknown T1 maint command:%d\n",
|
|
cmd);
|
|
break;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int t4_clear_maint(struct dahdi_span *span)
|
|
{
|
|
struct t4_span *ts = container_of(span, struct t4_span, span);
|
|
struct t4 *wc = ts->owner;
|
|
unsigned int reg;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
|
|
/* Clear local loop */
|
|
reg = __t4_framer_in(wc, span->offset, LIM0_T);
|
|
__t4_framer_out(wc, span->offset, LIM0_T, (reg & ~LIM0_LL));
|
|
|
|
/* Clear Remote Loop */
|
|
reg = __t4_framer_in(wc, span->offset, LIM1_T);
|
|
__t4_framer_out(wc, span->offset, LIM1_T, (reg & ~LIM1_RL));
|
|
|
|
/* Clear Remote Payload Loop */
|
|
reg = __t4_framer_in(wc, span->offset, FMR2_T);
|
|
__t4_framer_out(wc, span->offset, FMR2_T, (reg & ~FMR2_PLB));
|
|
|
|
/* Clear PRBS */
|
|
reg = __t4_framer_in(wc, span->offset, LCR1_T);
|
|
__t4_framer_out(wc, span->offset, LCR1_T, (reg & ~(XPRBS | EPRM)));
|
|
|
|
/* Clear loopup/loopdown signals on the line */
|
|
reg = __t4_framer_in(wc, span->offset, FMR5);
|
|
__t4_framer_out(wc, span->offset, FMR5, (reg & ~(FMR5_XLU | FMR5_XLD)));
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
span->mainttimer = 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int t4_reset_counters(struct dahdi_span *span)
|
|
{
|
|
struct t4_span *ts = container_of(span, struct t4_span, span);
|
|
memset(&ts->span.count, 0, sizeof(ts->span.count));
|
|
return 0;
|
|
}
|
|
|
|
static int t4_rbsbits(struct dahdi_chan *chan, int bits)
|
|
{
|
|
u_char m,c;
|
|
int k,n,b;
|
|
struct t4 *wc = chan->pvt;
|
|
struct t4_span *ts = wc->tspans[chan->span->offset];
|
|
unsigned long flags;
|
|
|
|
if (debug & DEBUG_RBS)
|
|
dev_notice(&wc->dev->dev, "Setting bits to %d on channel %s\n",
|
|
bits, chan->name);
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
k = chan->span->offset;
|
|
if (E1 == ts->linemode) {
|
|
if (chan->chanpos == 16) {
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
return 0;
|
|
}
|
|
n = chan->chanpos - 1;
|
|
if (chan->chanpos > 15) n--;
|
|
b = (n % 15);
|
|
c = ts->txsigs[b];
|
|
m = (n / 15) << 2; /* nibble selector */
|
|
c &= (0xf << m); /* keep the other nibble */
|
|
c |= (bits & 0xf) << (4 - m); /* put our new nibble here */
|
|
ts->txsigs[b] = c;
|
|
/* output them to the chip */
|
|
__t4_framer_out(wc,k,0x71 + b,c);
|
|
} else if (ts->span.lineconfig & DAHDI_CONFIG_D4) {
|
|
n = chan->chanpos - 1;
|
|
b = (n/4);
|
|
c = ts->txsigs[b];
|
|
m = ((3 - (n % 4)) << 1); /* nibble selector */
|
|
c &= ~(0x3 << m); /* keep the other nibble */
|
|
c |= ((bits >> 2) & 0x3) << m; /* put our new nibble here */
|
|
ts->txsigs[b] = c;
|
|
/* output them to the chip */
|
|
__t4_framer_out(wc,k,0x70 + b,c);
|
|
__t4_framer_out(wc,k,0x70 + b + 6,c);
|
|
} else if (ts->span.lineconfig & DAHDI_CONFIG_ESF) {
|
|
n = chan->chanpos - 1;
|
|
b = (n/2);
|
|
c = ts->txsigs[b];
|
|
m = ((n % 2) << 2); /* nibble selector */
|
|
c &= (0xf << m); /* keep the other nibble */
|
|
c |= (bits & 0xf) << (4 - m); /* put our new nibble here */
|
|
ts->txsigs[b] = c;
|
|
/* output them to the chip */
|
|
__t4_framer_out(wc,k,0x70 + b,c);
|
|
}
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
if (debug & DEBUG_RBS)
|
|
dev_notice(&wc->dev->dev, "Finished setting RBS bits\n");
|
|
return 0;
|
|
}
|
|
|
|
static int t4_shutdown(struct dahdi_span *span)
|
|
{
|
|
int tspan;
|
|
int wasrunning;
|
|
unsigned long flags;
|
|
struct t4_span *ts = container_of(span, struct t4_span, span);
|
|
struct t4 *wc = ts->owner;
|
|
|
|
tspan = span->offset + 1;
|
|
if (tspan < 0) {
|
|
dev_notice(&wc->dev->dev, "T%dXXP: Span '%d' isn't us?\n",
|
|
wc->numspans, span->spanno);
|
|
return -1;
|
|
}
|
|
|
|
if (debug & DEBUG_MAIN)
|
|
dev_notice(&wc->dev->dev, "Shutting down span %d (%s)\n",
|
|
span->spanno, span->name);
|
|
|
|
/* Stop HDLC controller if runned */
|
|
if (ts->sigchan)
|
|
hdlc_stop(wc, span->offset);
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
wasrunning = span->flags & DAHDI_FLAG_RUNNING;
|
|
|
|
span->flags &= ~DAHDI_FLAG_RUNNING;
|
|
__t4_set_led(wc, span->offset, WC_OFF);
|
|
if (((wc->numspans == 8) &&
|
|
(!(wc->tspans[0]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[1]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[2]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[3]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[4]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[5]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[6]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[7]->span.flags & DAHDI_FLAG_RUNNING)))
|
|
||
|
|
((wc->numspans == 4) &&
|
|
(!(wc->tspans[0]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[1]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[2]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[3]->span.flags & DAHDI_FLAG_RUNNING)))
|
|
||
|
|
((wc->numspans == 2) &&
|
|
(!(wc->tspans[0]->span.flags & DAHDI_FLAG_RUNNING)) &&
|
|
(!(wc->tspans[1]->span.flags & DAHDI_FLAG_RUNNING)))) {
|
|
/* No longer in use, disable interrupts */
|
|
dev_info(&wc->dev->dev, "TE%dXXP: Disabling interrupts since "
|
|
"there are no active spans\n", wc->numspans);
|
|
set_bit(T4_STOP_DMA, &wc->checkflag);
|
|
} else
|
|
set_bit(T4_CHECK_TIMING, &wc->checkflag);
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
/* Wait for interrupt routine to shut itself down */
|
|
msleep(10);
|
|
if (wasrunning)
|
|
wc->spansstarted--;
|
|
|
|
if (debug & DEBUG_MAIN)
|
|
dev_notice(&wc->dev->dev, "Span %d (%s) shutdown\n",
|
|
span->spanno, span->name);
|
|
return 0;
|
|
}
|
|
|
|
static void t4_chan_set_sigcap(struct dahdi_span *span, int x)
|
|
{
|
|
struct t4_span *wc = container_of(span, struct t4_span, span);
|
|
struct dahdi_chan *chan = wc->chans[x];
|
|
chan->sigcap = DAHDI_SIG_CLEAR;
|
|
/* E&M variant supported depends on span type */
|
|
if (E1 == wc->linemode) {
|
|
/* E1 sigcap setup */
|
|
if (span->lineconfig & DAHDI_CONFIG_CCS) {
|
|
/* CCS setup */
|
|
chan->sigcap |= DAHDI_SIG_MTP2 | DAHDI_SIG_SF |
|
|
DAHDI_SIG_HARDHDLC;
|
|
return;
|
|
}
|
|
/* clear out sig and sigcap for channel 16 on E1 CAS
|
|
* lines, otherwise, set it correctly */
|
|
if (x == 15) {
|
|
/* CAS signaling channel setup */
|
|
wc->chans[15]->sigcap = 0;
|
|
wc->chans[15]->sig = 0;
|
|
return;
|
|
}
|
|
/* normal CAS setup */
|
|
chan->sigcap |= DAHDI_SIG_EM_E1 | DAHDI_SIG_FXSLS |
|
|
DAHDI_SIG_FXSGS | DAHDI_SIG_FXSKS | DAHDI_SIG_SF |
|
|
DAHDI_SIG_FXOLS | DAHDI_SIG_FXOGS | DAHDI_SIG_FXOKS |
|
|
DAHDI_SIG_CAS | DAHDI_SIG_DACS_RBS;
|
|
} else {
|
|
/* T1 sigcap setup */
|
|
chan->sigcap |= DAHDI_SIG_EM | DAHDI_SIG_FXSLS |
|
|
DAHDI_SIG_FXSGS | DAHDI_SIG_FXSKS | DAHDI_SIG_MTP2 |
|
|
DAHDI_SIG_SF | DAHDI_SIG_FXOLS | DAHDI_SIG_FXOGS |
|
|
DAHDI_SIG_FXOKS | DAHDI_SIG_CAS | DAHDI_SIG_DACS_RBS |
|
|
DAHDI_SIG_HARDHDLC;
|
|
}
|
|
}
|
|
|
|
static int
|
|
_t4_spanconfig(struct file *file, struct dahdi_span *span,
|
|
struct dahdi_lineconfig *lc)
|
|
{
|
|
int i;
|
|
struct t4_span *ts = container_of(span, struct t4_span, span);
|
|
struct t4 *wc = ts->owner;
|
|
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "About to enter spanconfig!\n");
|
|
if (debug & DEBUG_MAIN)
|
|
dev_notice(&wc->dev->dev, "TE%dXXP: Configuring span %d\n",
|
|
wc->numspans, span->spanno);
|
|
|
|
if (lc->sync < 0)
|
|
lc->sync = 0;
|
|
if (lc->sync > wc->numspans) {
|
|
dev_warn(&wc->dev->dev, "WARNING: Cannot set priority on span %d to %d. Please set to a number between 1 and %d\n",
|
|
span->spanno, lc->sync, wc->numspans);
|
|
lc->sync = 0;
|
|
}
|
|
|
|
/* remove this span number from the current sync sources, if there */
|
|
for(i = 0; i < wc->numspans; i++) {
|
|
if (wc->tspans[i]->sync == span->spanno)
|
|
wc->tspans[i]->sync = 0;
|
|
}
|
|
wc->tspans[span->offset]->syncpos = lc->sync;
|
|
/* if a sync src, put it in proper place */
|
|
if (lc->sync)
|
|
wc->tspans[lc->sync - 1]->sync = span->spanno;
|
|
|
|
set_bit(T4_CHECK_TIMING, &wc->checkflag);
|
|
|
|
/* Make sure this is clear in case of multiple startup and shutdown
|
|
* iterations */
|
|
clear_bit(T4_STOP_DMA, &wc->checkflag);
|
|
|
|
/* make sure that sigcaps gets updated if necessary */
|
|
for (i = 0; i < span->channels; i++)
|
|
t4_chan_set_sigcap(span, i);
|
|
|
|
/* If we're already running, then go ahead and apply the changes */
|
|
if (span->flags & DAHDI_FLAG_RUNNING)
|
|
return _t4_startup(file, span);
|
|
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "Done with spanconfig!\n");
|
|
return 0;
|
|
}
|
|
|
|
static int
|
|
t4_spanconfig(struct file *file, struct dahdi_span *span,
|
|
struct dahdi_lineconfig *lc)
|
|
{
|
|
int ret;
|
|
struct dahdi_device *const ddev = span->parent;
|
|
struct dahdi_span *s;
|
|
|
|
ret = _t4_spanconfig(file, span, lc);
|
|
|
|
/* Make sure all the spans have a basic configuration in case they are
|
|
* not all specified in the configuration files. */
|
|
lc->sync = 0;
|
|
list_for_each_entry(s, &ddev->spans, device_node) {
|
|
WARN_ON(!s->channels);
|
|
if (!s->channels)
|
|
continue;
|
|
if (!s->chans[0]->sigcap)
|
|
_t4_spanconfig(file, s, lc);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
t4_chanconfig(struct file *file, struct dahdi_chan *chan, int sigtype)
|
|
{
|
|
int alreadyrunning;
|
|
unsigned long flags;
|
|
struct t4 *wc = chan->pvt;
|
|
struct t4_span *ts = wc->tspans[chan->span->offset];
|
|
|
|
alreadyrunning = ts->span.flags & DAHDI_FLAG_RUNNING;
|
|
if (debug & DEBUG_MAIN) {
|
|
if (alreadyrunning)
|
|
dev_notice(&wc->dev->dev, "TE%dXXP: Reconfigured "
|
|
"channel %d (%s) sigtype %d\n", wc->numspans,
|
|
chan->channo, chan->name, sigtype);
|
|
else
|
|
dev_notice(&wc->dev->dev, "TE%dXXP: Configured channel"
|
|
" %d (%s) sigtype %d\n", wc->numspans,
|
|
chan->channo, chan->name, sigtype);
|
|
}
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
|
|
if (alreadyrunning)
|
|
__set_clear(wc, chan->span->offset);
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
/* (re)configure signalling channel */
|
|
if ((sigtype == DAHDI_SIG_HARDHDLC) || (ts->sigchan == chan)) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "%sonfiguring hardware HDLC "
|
|
"on %s\n",
|
|
((sigtype == DAHDI_SIG_HARDHDLC) ? "C" : "Unc"),
|
|
chan->name);
|
|
if (alreadyrunning) {
|
|
if (ts->sigchan)
|
|
hdlc_stop(wc, ts->sigchan->span->offset);
|
|
if (sigtype == DAHDI_SIG_HARDHDLC) {
|
|
if (hdlc_start(wc, chan->span->offset, chan, ts->sigmode)) {
|
|
dev_notice(&wc->dev->dev, "Error "
|
|
"initializing signalling "
|
|
"controller\n");
|
|
return -1;
|
|
}
|
|
} else {
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
ts->sigchan = NULL;
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
}
|
|
else {
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
ts->sigchan = (sigtype == DAHDI_SIG_HARDHDLC) ? chan : NULL;
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
ts->sigactive = 0;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int set_span_devicetype(struct t4 *wc)
|
|
{
|
|
#ifdef VPM_SUPPORT
|
|
const char *vpmstring = __t4_echocan_name(wc);
|
|
|
|
if (vpmstring) {
|
|
wc->ddev->devicetype = kasprintf(GFP_KERNEL, "%s (%s)",
|
|
wc->devtype->desc, vpmstring);
|
|
} else {
|
|
wc->ddev->devicetype = kasprintf(GFP_KERNEL, wc->devtype->desc);
|
|
}
|
|
#else
|
|
wc->ddev->devicetype = kasprintf(GFP_KERNEL, wc->devtype->desc);
|
|
#endif
|
|
|
|
if (!wc->ddev->devicetype)
|
|
return -ENOMEM;
|
|
return 0;
|
|
}
|
|
|
|
/* The number of cards we have seen with each
|
|
possible 'order' switch setting.
|
|
*/
|
|
static unsigned int order_index[16];
|
|
|
|
static void setup_chunks(struct t4 *wc, int which)
|
|
{
|
|
struct t4_span *ts;
|
|
int offset = 1;
|
|
int x, y;
|
|
int gen2;
|
|
int basesize = T4_BASE_SIZE(wc) >> 2;
|
|
|
|
if (!has_e1_span(wc))
|
|
offset += 4;
|
|
|
|
gen2 = (wc->tspans[0]->spanflags & FLAG_2NDGEN);
|
|
|
|
for (x = 0; x < wc->numspans; x++) {
|
|
ts = wc->tspans[x];
|
|
ts->writechunk = (void *)(wc->writechunk + (x * 32 * 2) + (which * (basesize)));
|
|
ts->readchunk = (void *)(wc->readchunk + (x * 32 * 2) + (which * (basesize)));
|
|
for (y=0;y<wc->tspans[x]->span.channels;y++) {
|
|
struct dahdi_chan *mychans = ts->chans[y];
|
|
if (gen2) {
|
|
mychans->writechunk = (void *)(wc->writechunk + ((x * 32 + y + offset) * 2) + (which * (basesize)));
|
|
mychans->readchunk = (void *)(wc->readchunk + ((x * 32 + y + offset) * 2) + (which * (basesize)));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
static int __t4_hardware_init_1(struct t4 *wc, unsigned int cardflags,
|
|
bool first_time);
|
|
static int __t4_hardware_init_2(struct t4 *wc, bool first_time);
|
|
|
|
static int t4_hardware_stop(struct t4 *wc);
|
|
|
|
static void t4_framer_reset(struct t4 *wc)
|
|
{
|
|
const bool first_time = false;
|
|
bool have_vpm = wc->vpm != NULL;
|
|
if (have_vpm) {
|
|
release_vpm450m(wc->vpm);
|
|
wc->vpm = NULL;
|
|
}
|
|
t4_hardware_stop(wc);
|
|
__t4_set_sclk_src(wc, WC_SELF, 0, 0);
|
|
__t4_hardware_init_1(wc, wc->devtype->flags, first_time);
|
|
__t4_hardware_init_2(wc, first_time);
|
|
if (have_vpm) {
|
|
t4_vpm_init(wc);
|
|
wc->dmactrl |= (wc->vpm) ? T4_VPM_PRESENT : 0;
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
}
|
|
setup_chunks(wc, 0);
|
|
wc->lastindex = 0;
|
|
}
|
|
|
|
/**
|
|
* t4_serial_setup - Setup serial parameters and system interface.
|
|
* @wc: The card to configure.
|
|
*
|
|
*/
|
|
static void t4_serial_setup(struct t4 *wc)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int unit;
|
|
bool reset_required = false;
|
|
|
|
if (debug) {
|
|
dev_info(&wc->dev->dev,
|
|
"TE%dXXP: Setting up global serial parameters\n",
|
|
wc->numspans);
|
|
}
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
reset_required = wc->reset_required > 0;
|
|
wc->reset_required = 0;
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
if (reset_required)
|
|
t4_framer_reset(wc);
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
/* GPC1: Multiplex mode enabled, FSC is output, active low, RCLK from
|
|
* channel 0 */
|
|
__t4_framer_out(wc, 0, 0x85, 0xe0);
|
|
if (is_octal(wc))
|
|
__t4_framer_out(wc, 0, FRMR_GPC2, 0x00);
|
|
|
|
/* IPC: Interrupt push/pull active low */
|
|
__t4_framer_out(wc, 0, 0x08, 0x01);
|
|
|
|
if (is_octal(wc)) {
|
|
/* Global clocks (16.384 Mhz CLK) */
|
|
__t4_framer_out(wc, 0, 0x92, 0x00); /* GCM1 */
|
|
__t4_framer_out(wc, 0, 0x93, 0x18);
|
|
__t4_framer_out(wc, 0, 0x94, 0xfb);
|
|
__t4_framer_out(wc, 0, 0x95, 0x0b);
|
|
__t4_framer_out(wc, 0, 0x96, 0x01);
|
|
__t4_framer_out(wc, 0, 0x97, 0x0b);
|
|
__t4_framer_out(wc, 0, 0x98, 0xdb);
|
|
__t4_framer_out(wc, 0, 0x99, 0xdf);
|
|
} else {
|
|
/* Global clocks (8.192 Mhz CLK) */
|
|
__t4_framer_out(wc, 0, 0x92, 0x00);
|
|
__t4_framer_out(wc, 0, 0x93, 0x18);
|
|
__t4_framer_out(wc, 0, 0x94, 0xfb);
|
|
__t4_framer_out(wc, 0, 0x95, 0x0b);
|
|
__t4_framer_out(wc, 0, 0x96, 0x00);
|
|
__t4_framer_out(wc, 0, 0x97, 0x0b);
|
|
__t4_framer_out(wc, 0, 0x98, 0xdb);
|
|
__t4_framer_out(wc, 0, 0x99, 0xdf);
|
|
}
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
for (unit = 0; unit < ports_on_framer(wc); ++unit) {
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
|
|
/* Configure interrupts */
|
|
/* GCR: Interrupt on Activation/Deactivation of each */
|
|
__t4_framer_out(wc, unit, FRMR_GCR, 0x00);
|
|
|
|
/* Configure system interface */
|
|
if (is_octal(wc)) {
|
|
/* SIC1: 16.384 Mhz clock/bus, double buffer receive /
|
|
* transmit, byte interleaved */
|
|
__t4_framer_out(wc, unit, FRMR_SIC1, 0xc2 | 0x08);
|
|
} else {
|
|
/* SIC1: 8.192 Mhz clock/bus, double buffer receive /
|
|
* transmit, byte interleaved */
|
|
__t4_framer_out(wc, unit, FRMR_SIC1, 0xc2);
|
|
}
|
|
/* SIC2: No FFS, no center receive eliastic buffer, phase */
|
|
__t4_framer_out(wc, unit, FRMR_SIC2, 0x20 | (unit << 1));
|
|
/* SIC3: Edges for capture */
|
|
if (is_octal(wc)) {
|
|
__t4_framer_out(wc, unit, FRMR_SIC3, 0x04 | (1 << 4));
|
|
} else {
|
|
__t4_framer_out(wc, unit, FRMR_SIC3, 0x04);
|
|
}
|
|
/* CMR2: We provide sync and clock for tx and rx. */
|
|
__t4_framer_out(wc, unit, FRMR_CMR2, 0x00);
|
|
|
|
if (is_octal(wc)) {
|
|
/* Set RCLK to 16 MHz */
|
|
__t4_framer_out(wc, unit, FRMR_CMR4, 0x5);
|
|
|
|
if (!has_e1_span(wc)) { /* T1/J1 mode */
|
|
__t4_framer_out(wc, unit, FRMR_XC0, 0x07);
|
|
__t4_framer_out(wc, unit, FRMR_XC1, 0x04);
|
|
if (wc->tspans[unit]->linemode == J1)
|
|
__t4_framer_out(wc, unit, FRMR_RC0, 0x87);
|
|
else
|
|
__t4_framer_out(wc, unit, FRMR_RC0, 0x07);
|
|
__t4_framer_out(wc, unit, FRMR_RC1, 0x04);
|
|
} else { /* E1 mode */
|
|
__t4_framer_out(wc, unit, FRMR_XC0, 0x00);
|
|
__t4_framer_out(wc, unit, FRMR_XC1, 0x04);
|
|
__t4_framer_out(wc, unit, FRMR_RC0, 0x00);
|
|
__t4_framer_out(wc, unit, FRMR_RC1, 0x04);
|
|
}
|
|
|
|
} else {
|
|
if (!has_e1_span(wc)) { /* T1/J1 mode */
|
|
__t4_framer_out(wc, unit, FRMR_XC0, 0x03);
|
|
__t4_framer_out(wc, unit, FRMR_XC1, 0x84);
|
|
if (J1 == wc->tspans[unit]->linemode)
|
|
__t4_framer_out(wc, unit, FRMR_RC0, 0x83);
|
|
else
|
|
__t4_framer_out(wc, unit, FRMR_RC0, 0x03);
|
|
__t4_framer_out(wc, unit, FRMR_RC1, 0x84);
|
|
} else { /* E1 mode */
|
|
__t4_framer_out(wc, unit, FRMR_XC0, 0x00);
|
|
__t4_framer_out(wc, unit, FRMR_XC1, 0x04);
|
|
__t4_framer_out(wc, unit, FRMR_RC0, 0x04);
|
|
__t4_framer_out(wc, unit, FRMR_RC1, 0x04);
|
|
}
|
|
}
|
|
|
|
/* Configure ports */
|
|
|
|
/* PC1: SPYR/SPYX input on RPA/XPA */
|
|
__t4_framer_out(wc, unit, 0x80, 0x00);
|
|
|
|
/* PC2: RMFB/XSIG output/input on RPB/XPB */
|
|
/* PC3: Some unused stuff */
|
|
/* PC4: Some more unused stuff */
|
|
if (is_octal(wc)) {
|
|
__t4_framer_out(wc, unit, 0x81, 0xBB);
|
|
__t4_framer_out(wc, unit, 0x82, 0xf5);
|
|
__t4_framer_out(wc, unit, 0x83, 0x35);
|
|
} else if (wc->falc31) {
|
|
__t4_framer_out(wc, unit, 0x81, 0xBB);
|
|
__t4_framer_out(wc, unit, 0x82, 0xBB);
|
|
__t4_framer_out(wc, unit, 0x83, 0xBB);
|
|
} else {
|
|
__t4_framer_out(wc, unit, 0x81, 0x22);
|
|
__t4_framer_out(wc, unit, 0x82, 0x65);
|
|
__t4_framer_out(wc, unit, 0x83, 0x35);
|
|
}
|
|
|
|
/* PC5: XMFS active low, SCLKR is input, RCLK is output */
|
|
__t4_framer_out(wc, unit, 0x84, 0x01);
|
|
|
|
if (debug & DEBUG_MAIN) {
|
|
dev_notice(&wc->dev->dev,
|
|
"Successfully initialized serial bus "
|
|
"for unit %d\n", unit);
|
|
}
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* t4_span_assigned - Called when the span is assigned by DAHDI.
|
|
* @span: Span that has been assigned.
|
|
*
|
|
* When this function is called, the span has a valid spanno and all the
|
|
* channels on the span have valid channel numbers assigned.
|
|
*
|
|
* This function is necessary because a device may be registered, and
|
|
* then user space may then later decide to assign span numbers and the
|
|
* channel numbers.
|
|
*
|
|
*/
|
|
static void t4_span_assigned(struct dahdi_span *span)
|
|
{
|
|
struct t4_span *tspan = container_of(span, struct t4_span, span);
|
|
struct t4 *wc = tspan->owner;
|
|
struct dahdi_span *pos;
|
|
unsigned int unassigned_spans = 0;
|
|
unsigned long flags;
|
|
|
|
/* We use this to make sure all the spans are assigned before
|
|
* running the serial setup. */
|
|
list_for_each_entry(pos, &wc->ddev->spans, device_node) {
|
|
if (!test_bit(DAHDI_FLAGBIT_REGISTERED, &pos->flags))
|
|
++unassigned_spans;
|
|
}
|
|
|
|
if (0 == unassigned_spans) {
|
|
t4_serial_setup(wc);
|
|
|
|
set_bit(T4_CHECK_TIMING, &wc->checkflag);
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
__t4_set_sclk_src(wc, WC_SELF, 0, 0);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
}
|
|
|
|
static void free_wc(struct t4 *wc)
|
|
{
|
|
unsigned int x, y;
|
|
|
|
flush_scheduled_work();
|
|
|
|
for (x = 0; x < ARRAY_SIZE(wc->tspans); x++) {
|
|
if (!wc->tspans[x])
|
|
continue;
|
|
for (y = 0; y < ARRAY_SIZE(wc->tspans[x]->chans); y++) {
|
|
kfree(wc->tspans[x]->chans[y]);
|
|
kfree(wc->tspans[x]->ec[y]);
|
|
}
|
|
kfree(wc->tspans[x]);
|
|
}
|
|
|
|
kfree(wc->ddev->devicetype);
|
|
kfree(wc->ddev->location);
|
|
kfree(wc->ddev->hardware_id);
|
|
dahdi_free_device(wc->ddev);
|
|
kfree(wc);
|
|
}
|
|
|
|
/**
|
|
* t4_alloc_channels - Allocate the channels on a span.
|
|
* @wc: The board we're allocating for.
|
|
* @ts: The span we're allocating for.
|
|
* @linemode: Which mode (T1/E1/J1) to use for this span.
|
|
*
|
|
* This function must only be called before the span is assigned it's
|
|
* possible for user processes to have an open reference to the
|
|
* channels.
|
|
*
|
|
*/
|
|
static int t4_alloc_channels(struct t4 *wc, struct t4_span *ts,
|
|
enum linemode linemode)
|
|
{
|
|
int i;
|
|
|
|
if (test_bit(DAHDI_FLAGBIT_REGISTERED, &ts->span.flags)) {
|
|
dev_dbg(&wc->dev->dev,
|
|
"Cannot allocate channels on a span that is already "
|
|
"assigned.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Cleanup any previously allocated channels. */
|
|
for (i = 0; i < ARRAY_SIZE(ts->chans); ++i) {
|
|
kfree(ts->chans[i]);
|
|
kfree(ts->ec[i]);
|
|
ts->chans[i] = NULL;
|
|
ts->ec[i] = NULL;
|
|
}
|
|
|
|
ts->linemode = linemode;
|
|
for (i = 0; i < ((E1 == ts->linemode) ? 31 : 24); i++) {
|
|
struct dahdi_chan *chan;
|
|
struct dahdi_echocan_state *ec;
|
|
|
|
chan = kzalloc(sizeof(*chan), GFP_KERNEL);
|
|
if (!chan) {
|
|
free_wc(wc);
|
|
return -ENOMEM;
|
|
}
|
|
ts->chans[i] = chan;
|
|
|
|
ec = kzalloc(sizeof(*ec), GFP_KERNEL);
|
|
if (!ec) {
|
|
free_wc(wc);
|
|
return -ENOMEM;
|
|
}
|
|
ts->ec[i] = ec;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void t4_init_one_span(struct t4 *wc, struct t4_span *ts)
|
|
{
|
|
unsigned long flags;
|
|
unsigned int reg;
|
|
int i;
|
|
|
|
snprintf(ts->span.name, sizeof(ts->span.name) - 1,
|
|
"TE%d/%d/%d", wc->numspans, wc->num, ts->span.offset + 1);
|
|
snprintf(ts->span.desc, sizeof(ts->span.desc) - 1,
|
|
"T%dXXP (PCI) Card %d Span %d", wc->numspans, wc->num,
|
|
ts->span.offset + 1);
|
|
|
|
switch (ts->linemode) {
|
|
case T1:
|
|
ts->span.spantype = SPANTYPE_DIGITAL_T1;
|
|
break;
|
|
case E1:
|
|
ts->span.spantype = SPANTYPE_DIGITAL_E1;
|
|
break;
|
|
case J1:
|
|
ts->span.spantype = SPANTYPE_DIGITAL_J1;
|
|
break;
|
|
}
|
|
|
|
/* HDLC Specific init */
|
|
ts->sigchan = NULL;
|
|
ts->sigmode = sigmode;
|
|
ts->sigactive = 0;
|
|
|
|
if (E1 != ts->linemode) {
|
|
ts->span.channels = 24;
|
|
ts->span.deflaw = DAHDI_LAW_MULAW;
|
|
ts->span.linecompat = DAHDI_CONFIG_AMI |
|
|
DAHDI_CONFIG_B8ZS | DAHDI_CONFIG_D4 |
|
|
DAHDI_CONFIG_ESF;
|
|
} else {
|
|
ts->span.channels = 31;
|
|
ts->span.deflaw = DAHDI_LAW_ALAW;
|
|
ts->span.linecompat = DAHDI_CONFIG_AMI |
|
|
DAHDI_CONFIG_HDB3 | DAHDI_CONFIG_CCS |
|
|
DAHDI_CONFIG_CRC4;
|
|
}
|
|
ts->span.chans = ts->chans;
|
|
ts->span.flags = DAHDI_FLAG_RBS;
|
|
|
|
for (i = 0; i < ts->span.channels; i++) {
|
|
struct dahdi_chan *const chan = ts->chans[i];
|
|
chan->pvt = wc;
|
|
snprintf(chan->name, sizeof(chan->name) - 1,
|
|
"%s/%d", ts->span.name, i + 1);
|
|
t4_chan_set_sigcap(&ts->span, i);
|
|
chan->chanpos = i + 1;
|
|
}
|
|
|
|
/* Enable 1sec timer interrupt */
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
reg = __t4_framer_in(wc, ts->span.offset, FMR1_T);
|
|
__t4_framer_out(wc, ts->span.offset, FMR1_T, (reg | FMR1_ECM));
|
|
|
|
/* Enable Errored Second interrupt */
|
|
__t4_framer_out(wc, ts->span.offset, ESM, 0);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
t4_reset_counters(&ts->span);
|
|
}
|
|
|
|
/**
|
|
* t4_set_linemode - Allows user space to change the linemode before spans are assigned.
|
|
* @span: span on which to change the linemode.
|
|
* @linemode: A value from enumerated spantypes
|
|
*
|
|
* This callback is used to override the E1/T1 mode jumper settings and set
|
|
* the linemode on for each span. Called when the "spantype" attribute
|
|
* is written in sysfs under the dahdi_device.
|
|
*
|
|
*/
|
|
static int t4_set_linemode(struct dahdi_span *span, enum spantypes linemode)
|
|
{
|
|
struct t4_span *ts = container_of(span, struct t4_span, span);
|
|
struct t4 *wc = ts->owner;
|
|
int res = 0;
|
|
enum linemode mode;
|
|
const char *old_name;
|
|
static DEFINE_MUTEX(linemode_lock);
|
|
unsigned long flags;
|
|
|
|
dev_dbg(&wc->dev->dev, "Setting '%s' to '%s'\n", span->name,
|
|
dahdi_spantype2str(linemode));
|
|
|
|
if (span->spantype == linemode)
|
|
return 0;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
wc->reset_required = 1;
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
/* Do not allow the t1e1 member to be changed by multiple threads. */
|
|
mutex_lock(&linemode_lock);
|
|
old_name = dahdi_spantype2str(span->spantype);
|
|
switch (linemode) {
|
|
case SPANTYPE_DIGITAL_T1:
|
|
dev_info(&wc->dev->dev,
|
|
"Changing from %s to T1 line mode.\n", old_name);
|
|
mode = T1;
|
|
wc->t1e1 &= ~(1 << span->offset);
|
|
break;
|
|
case SPANTYPE_DIGITAL_E1:
|
|
dev_info(&wc->dev->dev,
|
|
"Changing from %s to E1 line mode.\n", old_name);
|
|
mode = E1;
|
|
wc->t1e1 |= (1 << span->offset);
|
|
break;
|
|
case SPANTYPE_DIGITAL_J1:
|
|
dev_info(&wc->dev->dev,
|
|
"Changing from %s to J1 line mode.\n", old_name);
|
|
mode = J1;
|
|
wc->t1e1 &= ~(1 << span->offset);
|
|
break;
|
|
default:
|
|
dev_err(&wc->dev->dev,
|
|
"Got invalid linemode %d from dahdi\n", linemode);
|
|
res = -EINVAL;
|
|
}
|
|
|
|
if (!res) {
|
|
t4_alloc_channels(wc, ts, mode);
|
|
t4_init_one_span(wc, ts);
|
|
dahdi_init_span(span);
|
|
}
|
|
|
|
mutex_unlock(&linemode_lock);
|
|
return res;
|
|
}
|
|
|
|
static const struct dahdi_span_ops t4_gen1_span_ops = {
|
|
.owner = THIS_MODULE,
|
|
.spanconfig = t4_spanconfig,
|
|
.chanconfig = t4_chanconfig,
|
|
.startup = t4_startup,
|
|
.shutdown = t4_shutdown,
|
|
.rbsbits = t4_rbsbits,
|
|
.maint = t4_maint,
|
|
.ioctl = t4_ioctl,
|
|
.hdlc_hard_xmit = t4_hdlc_hard_xmit,
|
|
.assigned = t4_span_assigned,
|
|
.set_spantype = t4_set_linemode,
|
|
};
|
|
|
|
static const struct dahdi_span_ops t4_gen2_span_ops = {
|
|
.owner = THIS_MODULE,
|
|
.spanconfig = t4_spanconfig,
|
|
.chanconfig = t4_chanconfig,
|
|
.startup = t4_startup,
|
|
.shutdown = t4_shutdown,
|
|
.rbsbits = t4_rbsbits,
|
|
.maint = t4_maint,
|
|
.ioctl = t4_ioctl,
|
|
.hdlc_hard_xmit = t4_hdlc_hard_xmit,
|
|
.dacs = t4_dacs,
|
|
.assigned = t4_span_assigned,
|
|
.set_spantype = t4_set_linemode,
|
|
#ifdef VPM_SUPPORT
|
|
.echocan_create = t4_echocan_create,
|
|
.echocan_name = t4_echocan_name,
|
|
#endif
|
|
};
|
|
|
|
/**
|
|
* init_spans - Do first initialization on all the spans
|
|
* @wc: Card to initialize the spans on.
|
|
*
|
|
* This function is called *before* the dahdi_device is first registered
|
|
* with the system. What happens in t4_init_one_span can happen between
|
|
* when the device is registered and when the spans are assigned via
|
|
* sysfs (or automatically).
|
|
*
|
|
*/
|
|
static void init_spans(struct t4 *wc)
|
|
{
|
|
int x, y;
|
|
int gen2;
|
|
struct t4_span *ts;
|
|
unsigned int reg;
|
|
unsigned long flags;
|
|
|
|
gen2 = (wc->tspans[0]->spanflags & FLAG_2NDGEN);
|
|
for (x = 0; x < wc->numspans; x++) {
|
|
ts = wc->tspans[x];
|
|
|
|
sprintf(ts->span.name, "TE%d/%d/%d", wc->numspans, wc->num, x + 1);
|
|
snprintf(ts->span.desc, sizeof(ts->span.desc) - 1,
|
|
"T%dXXP (PCI) Card %d Span %d", wc->numspans, wc->num, x+1);
|
|
switch (ts->linemode) {
|
|
case T1:
|
|
ts->span.spantype = SPANTYPE_DIGITAL_T1;
|
|
break;
|
|
case E1:
|
|
ts->span.spantype = SPANTYPE_DIGITAL_E1;
|
|
break;
|
|
case J1:
|
|
ts->span.spantype = SPANTYPE_DIGITAL_J1;
|
|
break;
|
|
}
|
|
|
|
/* HDLC Specific init */
|
|
ts->sigchan = NULL;
|
|
ts->sigmode = sigmode;
|
|
ts->sigactive = 0;
|
|
|
|
if (E1 != ts->linemode) {
|
|
ts->span.channels = 24;
|
|
ts->span.deflaw = DAHDI_LAW_MULAW;
|
|
ts->span.linecompat = DAHDI_CONFIG_AMI |
|
|
DAHDI_CONFIG_B8ZS | DAHDI_CONFIG_D4 |
|
|
DAHDI_CONFIG_ESF;
|
|
} else {
|
|
ts->span.channels = 31;
|
|
ts->span.deflaw = DAHDI_LAW_ALAW;
|
|
ts->span.linecompat = DAHDI_CONFIG_AMI |
|
|
DAHDI_CONFIG_HDB3 | DAHDI_CONFIG_CCS |
|
|
DAHDI_CONFIG_CRC4;
|
|
}
|
|
ts->span.chans = ts->chans;
|
|
ts->span.flags = DAHDI_FLAG_RBS;
|
|
|
|
ts->owner = wc;
|
|
ts->span.offset = x;
|
|
ts->writechunk = (void *)(wc->writechunk + x * 32 * 2);
|
|
ts->readchunk = (void *)(wc->readchunk + x * 32 * 2);
|
|
|
|
if (gen2) {
|
|
ts->span.ops = &t4_gen2_span_ops;
|
|
} else {
|
|
ts->span.ops = &t4_gen1_span_ops;
|
|
}
|
|
|
|
for (y=0;y<wc->tspans[x]->span.channels;y++) {
|
|
struct dahdi_chan *mychans = ts->chans[y];
|
|
sprintf(mychans->name, "TE%d/%d/%d/%d", wc->numspans, wc->num, x + 1, y + 1);
|
|
t4_chan_set_sigcap(&ts->span, x);
|
|
mychans->pvt = wc;
|
|
mychans->chanpos = y + 1;
|
|
}
|
|
|
|
/* Start checking for alarms in 250 ms */
|
|
ts->alarmcheck_time = jiffies + msecs_to_jiffies(250);
|
|
|
|
/* Enable 1sec timer interrupt */
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
reg = __t4_framer_in(wc, x, FMR1_T);
|
|
__t4_framer_out(wc, x, FMR1_T, (reg | FMR1_ECM));
|
|
|
|
/* Enable Errored Second interrupt */
|
|
__t4_framer_out(wc, x, ESM, 0);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
t4_reset_counters(&ts->span);
|
|
|
|
}
|
|
|
|
set_span_devicetype(wc);
|
|
setup_chunks(wc, 0);
|
|
wc->lastindex = 0;
|
|
}
|
|
|
|
static int syncsrc = 0;
|
|
static int syncnum = 0 /* -1 */;
|
|
static int syncspan = 0;
|
|
static DEFINE_SPINLOCK(synclock);
|
|
|
|
static void __t4_set_rclk_src(struct t4 *wc, int span)
|
|
{
|
|
if (is_octal(wc)) {
|
|
int cmr5 = 0x00 | (span << 5);
|
|
int cmr1 = 0x38; /* Clock Mode: RCLK sourced by DCO-R1
|
|
by default, Disable Clock-Switching */
|
|
|
|
__t4_framer_out(wc, 0, 0x44, cmr1);
|
|
__t4_framer_out(wc, 0, FRMR_CMR5, cmr5);
|
|
} else {
|
|
int cmr1 = 0x38; /* Clock Mode: RCLK sourced by DCO-R1
|
|
by default, Disable Clock-Switching */
|
|
cmr1 |= (span << 6);
|
|
__t4_framer_out(wc, 0, 0x44, cmr1);
|
|
}
|
|
|
|
dev_info(&wc->dev->dev, "RCLK source set to span %d\n", span+1);
|
|
}
|
|
|
|
static void __t4_set_sclk_src(struct t4 *wc, int mode, int master, int slave)
|
|
{
|
|
if (slave) {
|
|
wc->dmactrl |= (1 << 25);
|
|
dev_info(&wc->dev->dev, "SCLK is slaved to timing cable\n");
|
|
} else {
|
|
wc->dmactrl &= ~(1 << 25);
|
|
}
|
|
|
|
if (master) {
|
|
wc->dmactrl |= (1 << 24);
|
|
dev_info(&wc->dev->dev, "SCLK is master to timing cable\n");
|
|
} else {
|
|
wc->dmactrl &= ~(1 << 24);
|
|
}
|
|
|
|
if (mode == WC_RECOVER)
|
|
wc->dmactrl |= (1 << 29); /* Recover timing from RCLK */
|
|
|
|
if (mode == WC_SELF)
|
|
wc->dmactrl &= ~(1 << 29);/* Provide timing from MCLK */
|
|
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
}
|
|
|
|
static ssize_t t4_timing_master_show(struct device *dev,
|
|
struct device_attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct t4 *wc = dev_get_drvdata(dev);
|
|
if (wc->dmactrl & (1 << 29))
|
|
return sprintf(buf, "%d\n", wc->syncsrc);
|
|
else
|
|
return sprintf(buf, "%d\n", -1);
|
|
}
|
|
|
|
static DEVICE_ATTR(timing_master, 0400, t4_timing_master_show, NULL);
|
|
|
|
static void create_sysfs_files(struct t4 *wc)
|
|
{
|
|
int ret;
|
|
ret = device_create_file(&wc->dev->dev,
|
|
&dev_attr_timing_master);
|
|
if (ret) {
|
|
dev_info(&wc->dev->dev,
|
|
"Failed to create device attributes.\n");
|
|
}
|
|
}
|
|
|
|
static void remove_sysfs_files(struct t4 *wc)
|
|
{
|
|
device_remove_file(&wc->dev->dev,
|
|
&dev_attr_timing_master);
|
|
}
|
|
|
|
static inline void __t4_update_timing(struct t4 *wc)
|
|
{
|
|
int i;
|
|
/* update sync src info */
|
|
if (wc->syncsrc != syncsrc) {
|
|
dev_info(&wc->dev->dev, "Swapping card %d from %d to %d\n",
|
|
wc->num, wc->syncsrc, syncsrc);
|
|
wc->syncsrc = syncsrc;
|
|
/* Update sync sources */
|
|
for (i = 0; i < wc->numspans; i++) {
|
|
wc->tspans[i]->span.syncsrc = wc->syncsrc;
|
|
}
|
|
if (syncnum == wc->num) {
|
|
__t4_set_rclk_src(wc, syncspan-1);
|
|
__t4_set_sclk_src(wc, WC_RECOVER, 1, 0);
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "Card %d, using sync "
|
|
"span %d, master\n", wc->num, syncspan);
|
|
} else {
|
|
__t4_set_sclk_src(wc, WC_RECOVER, 0, 1);
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "Card %d, using "
|
|
"Timing Bus, NOT master\n", wc->num);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int __t4_findsync(struct t4 *wc)
|
|
{
|
|
int i;
|
|
int x;
|
|
unsigned long flags;
|
|
int p;
|
|
int nonzero;
|
|
int newsyncsrc = 0; /* DAHDI span number */
|
|
int newsyncnum = 0; /* wct4xxp card number */
|
|
int newsyncspan = 0; /* span on given wct4xxp card */
|
|
spin_lock_irqsave(&synclock, flags);
|
|
if (!wc->num) {
|
|
/* If we're the first card, go through all the motions, up to 8 levels
|
|
of sync source */
|
|
p = 1;
|
|
while (p < 8) {
|
|
nonzero = 0;
|
|
for (x=0;cards[x];x++) {
|
|
for (i = 0; i < cards[x]->numspans; i++) {
|
|
if (cards[x]->tspans[i]->syncpos) {
|
|
nonzero = 1;
|
|
if ((cards[x]->tspans[i]->syncpos == p) &&
|
|
!(cards[x]->tspans[i]->span.alarms & (DAHDI_ALARM_RED | DAHDI_ALARM_BLUE | DAHDI_ALARM_LOOPBACK)) &&
|
|
(cards[x]->tspans[i]->span.flags & DAHDI_FLAG_RUNNING)) {
|
|
/* This makes a good sync source */
|
|
newsyncsrc = cards[x]->tspans[i]->span.spanno;
|
|
newsyncnum = x;
|
|
newsyncspan = i + 1;
|
|
/* Jump out */
|
|
goto found;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if (nonzero)
|
|
p++;
|
|
else
|
|
break;
|
|
}
|
|
found:
|
|
if ((syncnum != newsyncnum) || (syncsrc != newsyncsrc) || (newsyncspan != syncspan)) {
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "New syncnum: %d "
|
|
"(was %d), syncsrc: %d (was %d), "
|
|
"syncspan: %d (was %d)\n", newsyncnum,
|
|
syncnum, newsyncsrc, syncsrc,
|
|
newsyncspan, syncspan);
|
|
syncnum = newsyncnum;
|
|
syncsrc = newsyncsrc;
|
|
syncspan = newsyncspan;
|
|
for (x=0;cards[x];x++) {
|
|
__t4_update_timing(cards[x]);
|
|
}
|
|
}
|
|
}
|
|
__t4_update_timing(wc);
|
|
spin_unlock_irqrestore(&synclock, flags);
|
|
return 0;
|
|
}
|
|
|
|
static void __t4_set_timing_source_auto(struct t4 *wc)
|
|
{
|
|
int x, i;
|
|
int firstprio, secondprio;
|
|
firstprio = secondprio = 4;
|
|
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "timing source auto\n");
|
|
clear_bit(T4_CHECK_TIMING, &wc->checkflag);
|
|
if (timingcable) {
|
|
__t4_findsync(wc);
|
|
} else {
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "Evaluating spans for timing "
|
|
"source\n");
|
|
for (x=0;x<wc->numspans;x++) {
|
|
if ((wc->tspans[x]->span.flags & DAHDI_FLAG_RUNNING) &&
|
|
!(wc->tspans[x]->span.alarms & (DAHDI_ALARM_RED |
|
|
DAHDI_ALARM_BLUE))) {
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "span %d is "
|
|
"green : syncpos %d\n", x+1,
|
|
wc->tspans[x]->syncpos);
|
|
if (wc->tspans[x]->syncpos) {
|
|
/* Valid rsync source in recovered
|
|
timing mode */
|
|
if (firstprio == 4)
|
|
firstprio = x;
|
|
else if (wc->tspans[x]->syncpos <
|
|
wc->tspans[firstprio]->syncpos)
|
|
firstprio = x;
|
|
} else {
|
|
/* Valid rsync source in system timing
|
|
mode */
|
|
if (secondprio == 4)
|
|
secondprio = x;
|
|
}
|
|
}
|
|
}
|
|
if (firstprio != 4) {
|
|
wc->syncsrc = firstprio;
|
|
__t4_set_rclk_src(wc, firstprio);
|
|
__t4_set_sclk_src(wc, WC_RECOVER, 0, 0);
|
|
dev_info(&wc->dev->dev, "Recovered timing mode, "\
|
|
"RCLK set to span %d\n",
|
|
firstprio+1);
|
|
} else if (secondprio != 4) {
|
|
wc->syncsrc = -1;
|
|
__t4_set_rclk_src(wc, secondprio);
|
|
__t4_set_sclk_src(wc, WC_SELF, 0, 0);
|
|
dev_info(&wc->dev->dev, "System timing mode, "\
|
|
"RCLK set to span %d\n",
|
|
secondprio+1);
|
|
} else {
|
|
wc->syncsrc = -1;
|
|
dev_info(&wc->dev->dev, "All spans in alarm : No valid"\
|
|
"span to source RCLK from\n");
|
|
/* Default rclk to lock with span 1 */
|
|
__t4_set_rclk_src(wc, 0);
|
|
__t4_set_sclk_src(wc, WC_SELF, 0, 0);
|
|
}
|
|
|
|
/* Propagate sync selection to dahdi_span struct
|
|
* this is read by dahdi_tool to display the span's
|
|
* master/slave sync information */
|
|
for (i = 0; i < wc->numspans; i++) {
|
|
wc->tspans[i]->span.syncsrc = wc->syncsrc + 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void __t4_configure_t1(struct t4 *wc, int unit, int lineconfig, int txlevel)
|
|
{
|
|
unsigned int fmr4, fmr2, fmr1, fmr0, lim2;
|
|
char *framing, *line;
|
|
int mytxlevel;
|
|
if ((txlevel > 7) || (txlevel < 4))
|
|
mytxlevel = 0;
|
|
else
|
|
mytxlevel = txlevel - 4;
|
|
|
|
if (is_octal(wc))
|
|
fmr1 = 0x9c | 0x02; /* FMR1: Mode 1, T1 mode, CRC on for ESF, 8.192 Mhz system data rate, no XAIS */
|
|
else
|
|
fmr1 = 0x9c; /* FMR1: Mode 1, T1 mode, CRC on for ESF, 8.192 Mhz system data rate, no XAIS */
|
|
|
|
fmr2 = 0x20; /* FMR2: no payload loopback, don't auto yellow */
|
|
fmr4 = 0x0c; /* FMR4: Lose sync on 2 out of 5 framing bits, auto resync */
|
|
lim2 = 0x21; /* LIM2: 50% peak is a "1", Advanced Loss recovery */
|
|
lim2 |= (mytxlevel << 6); /* LIM2: Add line buildout */
|
|
__t4_framer_out(wc, unit, 0x1d, fmr1);
|
|
__t4_framer_out(wc, unit, 0x1e, fmr2);
|
|
|
|
/* Configure line interface */
|
|
if (lineconfig & DAHDI_CONFIG_AMI) {
|
|
line = "AMI";
|
|
/* workaround for errata #2 in ES v3 09-10-16 */
|
|
fmr0 = (is_octal(wc) || wc->falc31) ? 0xb0 : 0xa0;
|
|
} else {
|
|
line = "B8ZS";
|
|
fmr0 = 0xf0;
|
|
}
|
|
if (lineconfig & DAHDI_CONFIG_D4) {
|
|
framing = "D4";
|
|
} else {
|
|
framing = "ESF";
|
|
fmr4 |= 0x2;
|
|
fmr2 |= 0xc0;
|
|
}
|
|
__t4_framer_out(wc, unit, 0x1c, fmr0);
|
|
__t4_framer_out(wc, unit, 0x20, fmr4);
|
|
__t4_framer_out(wc, unit, FMR5, FMR5_EIBR); /* FMR5: Enable RBS mode */
|
|
|
|
__t4_framer_out(wc, unit, 0x37, 0xf0 ); /* LIM1: Clear data in case of LOS, Set receiver threshold (0.5V), No remote loop, no DRS */
|
|
__t4_framer_out(wc, unit, 0x36, 0x08); /* LIM0: Enable auto long haul mode, no local loop (must be after LIM1) */
|
|
|
|
__t4_framer_out(wc, unit, 0x02, 0x50); /* CMDR: Reset the receiver and transmitter line interface */
|
|
__t4_framer_out(wc, unit, 0x02, 0x00); /* CMDR: Reset the receiver and transmitter line interface */
|
|
|
|
if (wc->falc31) {
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "card %d span %d: setting Rtx "
|
|
"to 0ohm for T1\n", wc->num, unit);
|
|
__t4_framer_out(wc, unit, 0x86, 0x00); /* PC6: set Rtx to 0ohm for T1 */
|
|
|
|
// Hitting the bugfix register to fix errata #3
|
|
__t4_framer_out(wc, unit, 0xbd, 0x05);
|
|
}
|
|
|
|
__t4_framer_out(wc, unit, 0x3a, lim2); /* LIM2: 50% peak amplitude is a "1" */
|
|
__t4_framer_out(wc, unit, 0x38, 0x0a); /* PCD: LOS after 176 consecutive "zeros" */
|
|
__t4_framer_out(wc, unit, 0x39, 0x15); /* PCR: 22 "ones" clear LOS */
|
|
|
|
/* Generate pulse mask for T1 */
|
|
switch(mytxlevel) {
|
|
case 3:
|
|
__t4_framer_out(wc, unit, 0x26, 0x07); /* XPM0 */
|
|
__t4_framer_out(wc, unit, 0x27, 0x01); /* XPM1 */
|
|
__t4_framer_out(wc, unit, 0x28, 0x00); /* XPM2 */
|
|
break;
|
|
case 2:
|
|
__t4_framer_out(wc, unit, 0x26, 0x8c); /* XPM0 */
|
|
__t4_framer_out(wc, unit, 0x27, 0x11); /* XPM1 */
|
|
__t4_framer_out(wc, unit, 0x28, 0x01); /* XPM2 */
|
|
break;
|
|
case 1:
|
|
__t4_framer_out(wc, unit, 0x26, 0x8c); /* XPM0 */
|
|
__t4_framer_out(wc, unit, 0x27, 0x01); /* XPM1 */
|
|
__t4_framer_out(wc, unit, 0x28, 0x00); /* XPM2 */
|
|
break;
|
|
case 0:
|
|
default:
|
|
__t4_framer_out(wc, unit, 0x26, 0xd7); /* XPM0 */
|
|
__t4_framer_out(wc, unit, 0x27, 0x22); /* XPM1 */
|
|
__t4_framer_out(wc, unit, 0x28, 0x01); /* XPM2 */
|
|
break;
|
|
}
|
|
|
|
/* Don't mask framer interrupts if hardware HDLC is in use */
|
|
__t4_framer_out(wc, unit, FRMR_IMR0, 0xff & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR0_MASK : 0)); /* IMR0: We care about CAS changes, etc */
|
|
__t4_framer_out(wc, unit, FRMR_IMR1, 0xff & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR1_MASK : 0)); /* IMR1: We care about nothing */
|
|
__t4_framer_out(wc, unit, 0x16, 0x00); /* IMR2: All the alarm stuff! */
|
|
__t4_framer_out(wc, unit, 0x17, 0x34); /* IMR3: AIS and friends */
|
|
__t4_framer_out(wc, unit, 0x18, 0x3f); /* IMR4: Slips on transmit */
|
|
|
|
dev_info(&wc->dev->dev, "Span %d configured for %s/%s\n", unit + 1,
|
|
framing, line);
|
|
}
|
|
|
|
static void __t4_configure_e1(struct t4 *wc, int unit, int lineconfig)
|
|
{
|
|
unsigned int fmr2, fmr1, fmr0;
|
|
unsigned int cas = 0;
|
|
unsigned int imr3extra=0;
|
|
char *crc4 = "";
|
|
char *framing, *line;
|
|
if (is_octal(wc)) {
|
|
/* 16 MHz */
|
|
fmr1 = 0x44 | 0x02; /* FMR1: E1 mode, Automatic force resync, PCM30 mode, 8.192 Mhz backplane, no XAIS */
|
|
} else {
|
|
/* 8 MHz */
|
|
fmr1 = 0x44; /* FMR1: E1 mode, Automatic force resync, PCM30 mode, 8.192 Mhz backplane, no XAIS */
|
|
}
|
|
fmr2 = 0x03; /* FMR2: Auto transmit remote alarm, auto loss of multiframe recovery, no payload loopback */
|
|
if (lineconfig & DAHDI_CONFIG_CRC4) {
|
|
fmr1 |= 0x08; /* CRC4 transmit */
|
|
fmr2 |= 0xc0; /* CRC4 receive */
|
|
crc4 = "/CRC4";
|
|
}
|
|
__t4_framer_out(wc, unit, 0x1d, fmr1);
|
|
__t4_framer_out(wc, unit, 0x1e, fmr2);
|
|
|
|
/* Configure line interface */
|
|
if (lineconfig & DAHDI_CONFIG_AMI) {
|
|
line = "AMI";
|
|
/* workaround for errata #2 in ES v3 09-10-16 */
|
|
fmr0 = (is_octal(wc) || wc->falc31) ? 0xb0 : 0xa0;
|
|
} else {
|
|
line = "HDB3";
|
|
fmr0 = 0xf0;
|
|
}
|
|
if (lineconfig & DAHDI_CONFIG_CCS) {
|
|
framing = "CCS";
|
|
imr3extra = 0x28;
|
|
} else {
|
|
framing = "CAS";
|
|
cas = 0x40;
|
|
}
|
|
__t4_framer_out(wc, unit, 0x1c, fmr0);
|
|
|
|
__t4_framer_out(wc, unit, 0x37, 0xf0 /*| 0x6 */ ); /* LIM1: Clear data in case of LOS, Set receiver threshold (0.5V), No remote loop, no DRS */
|
|
__t4_framer_out(wc, unit, 0x36, 0x08); /* LIM0: Enable auto long haul mode, no local loop (must be after LIM1) */
|
|
|
|
__t4_framer_out(wc, unit, 0x02, 0x50); /* CMDR: Reset the receiver and transmitter line interface */
|
|
__t4_framer_out(wc, unit, 0x02, 0x00); /* CMDR: Reset the receiver and transmitter line interface */
|
|
|
|
if (wc->falc31) {
|
|
if (debug)
|
|
dev_info(&wc->dev->dev,
|
|
"setting Rtx to 7.5ohm for E1\n");
|
|
__t4_framer_out(wc, unit, 0x86, 0x40); /* PC6: turn on 7.5ohm Rtx for E1 */
|
|
}
|
|
|
|
/* Condition receive line interface for E1 after reset */
|
|
__t4_framer_out(wc, unit, 0xbb, 0x17);
|
|
__t4_framer_out(wc, unit, 0xbc, 0x55);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x97);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x11);
|
|
__t4_framer_out(wc, unit, 0xbc, 0xaa);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x91);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x12);
|
|
__t4_framer_out(wc, unit, 0xbc, 0x55);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x92);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x0c);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x00);
|
|
__t4_framer_out(wc, unit, 0xbb, 0x8c);
|
|
|
|
__t4_framer_out(wc, unit, 0x3a, 0x20); /* LIM2: 50% peak amplitude is a "1" */
|
|
__t4_framer_out(wc, unit, 0x38, 0x0a); /* PCD: LOS after 176 consecutive "zeros" */
|
|
__t4_framer_out(wc, unit, 0x39, 0x15); /* PCR: 22 "ones" clear LOS */
|
|
|
|
__t4_framer_out(wc, unit, 0x20, 0x9f); /* XSW: Spare bits all to 1 */
|
|
__t4_framer_out(wc, unit, 0x21, 0x1c|cas); /* XSP: E-bit set when async. AXS auto, XSIF to 1 */
|
|
|
|
|
|
/* Generate pulse mask for E1 */
|
|
__t4_framer_out(wc, unit, 0x26, 0x54); /* XPM0 */
|
|
__t4_framer_out(wc, unit, 0x27, 0x02); /* XPM1 */
|
|
__t4_framer_out(wc, unit, 0x28, 0x00); /* XPM2 */
|
|
|
|
/* Don't mask framer interrupts if hardware HDLC is in use */
|
|
__t4_framer_out(wc, unit, FRMR_IMR0, 0xff & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR0_MASK : 0)); /* IMR0: We care about CRC errors, CAS changes, etc */
|
|
__t4_framer_out(wc, unit, FRMR_IMR1, 0x3f & ~((wc->tspans[unit]->sigchan) ? HDLC_IMR1_MASK : 0)); /* IMR1: We care about loopup / loopdown */
|
|
__t4_framer_out(wc, unit, 0x16, 0x00); /* IMR2: We care about all the alarm stuff! */
|
|
__t4_framer_out(wc, unit, 0x17, 0x04 | imr3extra); /* IMR3: AIS */
|
|
__t4_framer_out(wc, unit, 0x18, 0x3f); /* IMR4: We care about slips on transmit */
|
|
|
|
__t4_framer_out(wc, unit, 0x2f, 0x00);
|
|
__t4_framer_out(wc, unit, 0x30, 0x00);
|
|
__t4_framer_out(wc, unit, 0x31, 0x00);
|
|
|
|
dev_info(&wc->dev->dev, "TE%dXXP: Span %d configured for %s/%s%s\n",
|
|
wc->numspans, unit + 1, framing, line, crc4);
|
|
}
|
|
|
|
/**
|
|
* t4_check_for_interrupts - Return 0 if the card is generating interrupts.
|
|
* @wc: The card to check.
|
|
*
|
|
* If the card is not generating interrupts, this function will also place all
|
|
* the spans on the card into red alarm.
|
|
*
|
|
*/
|
|
static int t4_check_for_interrupts(struct t4 *wc)
|
|
{
|
|
unsigned int starting_intcount = wc->intcount;
|
|
unsigned long stop_time = jiffies + HZ*2;
|
|
unsigned long flags;
|
|
int x;
|
|
|
|
msleep(20);
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
while (starting_intcount == wc->intcount) {
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
if (time_after(jiffies, stop_time)) {
|
|
for (x = 0; x < wc->numspans; x++)
|
|
wc->tspans[x]->span.alarms = DAHDI_ALARM_RED;
|
|
dev_err(&wc->dev->dev, "Interrupts not detected.\n");
|
|
return -EIO;
|
|
}
|
|
msleep(100);
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
}
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int _t4_startup(struct file *file, struct dahdi_span *span)
|
|
{
|
|
#ifdef SUPPORT_GEN1
|
|
int i;
|
|
#endif
|
|
int tspan;
|
|
unsigned long flags;
|
|
int alreadyrunning;
|
|
struct t4_span *ts = container_of(span, struct t4_span, span);
|
|
struct t4 *wc = ts->owner;
|
|
|
|
set_bit(T4_IGNORE_LATENCY, &wc->checkflag);
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "About to enter startup!\n");
|
|
|
|
tspan = span->offset + 1;
|
|
if (tspan < 0) {
|
|
dev_info(&wc->dev->dev, "TE%dXXP: Span '%d' isn't us?\n",
|
|
wc->numspans, span->spanno);
|
|
return -1;
|
|
}
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
|
|
alreadyrunning = span->flags & DAHDI_FLAG_RUNNING;
|
|
|
|
#ifdef SUPPORT_GEN1
|
|
/* initialize the start value for the entire chunk of last ec buffer */
|
|
for(i = 0; i < span->channels; i++)
|
|
{
|
|
memset(ts->ec_chunk1[i],
|
|
DAHDI_LIN2X(0,span->chans[i]),DAHDI_CHUNKSIZE);
|
|
memset(ts->ec_chunk2[i],
|
|
DAHDI_LIN2X(0,span->chans[i]),DAHDI_CHUNKSIZE);
|
|
}
|
|
#endif
|
|
/* Force re-evaluation of timing source */
|
|
wc->syncsrc = -1;
|
|
set_bit(T4_CHECK_TIMING, &wc->checkflag);
|
|
|
|
if (E1 == ts->linemode)
|
|
__t4_configure_e1(wc, span->offset, span->lineconfig);
|
|
else
|
|
__t4_configure_t1(wc, span->offset, span->lineconfig, span->txlevel);
|
|
|
|
/* Note clear channel status */
|
|
wc->tspans[span->offset]->notclear = 0;
|
|
__set_clear(wc, span->offset);
|
|
|
|
if (!alreadyrunning) {
|
|
span->flags |= DAHDI_FLAG_RUNNING;
|
|
wc->spansstarted++;
|
|
|
|
if (wc->devtype->flags & FLAG_5THGEN)
|
|
__t4_pci_out(wc, 5, (ms_per_irq << 16) | wc->numbufs);
|
|
else
|
|
__t4_pci_out(wc, 5, (1 << 16) | 1);
|
|
/* enable interrupts */
|
|
/* Start DMA, enabling DMA interrupts on read only */
|
|
wc->dmactrl |= (ts->spanflags & FLAG_2NDGEN) ? 0xc0000000 : 0xc0000003;
|
|
#ifdef VPM_SUPPORT
|
|
wc->dmactrl |= (wc->vpm) ? T4_VPM_PRESENT : 0;
|
|
#endif
|
|
/* Seed interrupt register */
|
|
__t4_pci_out(wc, WC_INTR, 0x0c);
|
|
if (noburst || !(ts->spanflags & FLAG_BURST))
|
|
wc->dmactrl |= (1 << 26);
|
|
else
|
|
wc->dmactrl &= ~(1 << 26);
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
|
|
/* Startup HDLC controller too */
|
|
}
|
|
|
|
if (ts->sigchan) {
|
|
struct dahdi_chan *sigchan = ts->sigchan;
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
if (hdlc_start(wc, span->offset, sigchan, ts->sigmode)) {
|
|
dev_notice(&wc->dev->dev, "Error initializing "
|
|
"signalling controller\n");
|
|
return -1;
|
|
}
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
}
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
local_irq_save(flags);
|
|
t4_check_alarms(wc, span->offset);
|
|
t4_check_sigbits(wc, span->offset);
|
|
local_irq_restore(flags);
|
|
|
|
if (wc->tspans[0]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Primary Sync Source\n",
|
|
span->spanno);
|
|
if (wc->tspans[1]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Secondary Sync Source\n",
|
|
span->spanno);
|
|
if (wc->numspans >= 4) {
|
|
if (wc->tspans[2]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Tertiary Sync Source"
|
|
"\n", span->spanno);
|
|
if (wc->tspans[3]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Quaternary Sync "
|
|
"Source\n", span->spanno);
|
|
}
|
|
if (wc->numspans == 8) {
|
|
if (wc->tspans[4]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Quinary Sync "
|
|
"Source\n", span->spanno);
|
|
if (wc->tspans[5]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Senary Sync "
|
|
"Source\n", span->spanno);
|
|
if (wc->tspans[6]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Septenary Sync "
|
|
"Source\n", span->spanno);
|
|
if (wc->tspans[7]->sync == span->spanno)
|
|
dev_info(&wc->dev->dev, "SPAN %d: Octonary Sync "
|
|
"Source\n", span->spanno);
|
|
}
|
|
|
|
if (!alreadyrunning) {
|
|
if (t4_check_for_interrupts(wc))
|
|
return -EIO;
|
|
}
|
|
|
|
if (debug)
|
|
dev_info(&wc->dev->dev, "Completed startup!\n");
|
|
clear_bit(T4_IGNORE_LATENCY, &wc->checkflag);
|
|
return 0;
|
|
}
|
|
|
|
static int t4_startup(struct file *file, struct dahdi_span *span)
|
|
{
|
|
int ret;
|
|
struct dahdi_device *const ddev = span->parent;
|
|
struct dahdi_span *s;
|
|
|
|
ret = _t4_startup(file, span);
|
|
list_for_each_entry(s, &ddev->spans, device_node) {
|
|
if (!test_bit(DAHDI_FLAGBIT_RUNNING, &s->flags)) {
|
|
_t4_startup(file, s);
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
#ifdef SUPPORT_GEN1
|
|
static inline void e1_check(struct t4 *wc, int span, int val)
|
|
{
|
|
struct t4_span *ts = wc->tspans[span];
|
|
if ((ts->span.channels > 24) &&
|
|
(ts->span.flags & DAHDI_FLAG_RUNNING) &&
|
|
!(ts->span.alarms) &&
|
|
(!wc->e1recover)) {
|
|
if (val != 0x1b) {
|
|
ts->e1check++;
|
|
} else
|
|
ts->e1check = 0;
|
|
if (ts->e1check > 100) {
|
|
/* Wait 1000 ms */
|
|
wc->e1recover = 1000 * 8;
|
|
wc->tspans[0]->e1check = wc->tspans[1]->e1check = 0;
|
|
if (wc->numspans == 4)
|
|
wc->tspans[2]->e1check = wc->tspans[3]->e1check = 0;
|
|
if (debug & DEBUG_MAIN)
|
|
dev_notice(&wc->dev->dev, "Detected loss of "
|
|
"E1 alignment on span %d!\n", span);
|
|
t4_reset_dma(wc);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void t4_receiveprep(struct t4 *wc, int irq)
|
|
{
|
|
unsigned int *readchunk;
|
|
int dbl = 0;
|
|
int x,y,z;
|
|
unsigned int tmp;
|
|
int offset=0;
|
|
if (!has_e1_span(wc))
|
|
offset = 4;
|
|
if (irq & 1) {
|
|
/* First part */
|
|
readchunk = wc->readchunk;
|
|
if (!wc->last0)
|
|
dbl = 1;
|
|
wc->last0 = 0;
|
|
} else {
|
|
readchunk = wc->readchunk + DAHDI_CHUNKSIZE * 32;
|
|
if (wc->last0)
|
|
dbl = 1;
|
|
wc->last0 = 1;
|
|
}
|
|
if (unlikely(dbl && (debug & DEBUG_MAIN)))
|
|
dev_notice(&wc->dev->dev, "Double/missed interrupt detected\n");
|
|
|
|
for (x=0;x<DAHDI_CHUNKSIZE;x++) {
|
|
for (z=0;z<24;z++) {
|
|
/* All T1/E1 channels */
|
|
tmp = readchunk[z+1+offset];
|
|
if (wc->numspans == 4) {
|
|
wc->tspans[3]->span.chans[z]->readchunk[x] = tmp & 0xff;
|
|
wc->tspans[2]->span.chans[z]->readchunk[x] = (tmp & 0xff00) >> 8;
|
|
}
|
|
wc->tspans[1]->span.chans[z]->readchunk[x] = (tmp & 0xff0000) >> 16;
|
|
wc->tspans[0]->span.chans[z]->readchunk[x] = tmp >> 24;
|
|
}
|
|
if (has_e1_span(wc)) {
|
|
if (wc->e1recover > 0)
|
|
wc->e1recover--;
|
|
tmp = readchunk[0];
|
|
if (wc->numspans == 4) {
|
|
e1_check(wc, 3, (tmp & 0x7f));
|
|
e1_check(wc, 2, (tmp & 0x7f00) >> 8);
|
|
}
|
|
e1_check(wc, 1, (tmp & 0x7f0000) >> 16);
|
|
e1_check(wc, 0, (tmp & 0x7f000000) >> 24);
|
|
for (z=24;z<31;z++) {
|
|
/* Only E1 channels now */
|
|
tmp = readchunk[z+1];
|
|
if (wc->numspans == 4) {
|
|
if (wc->tspans[3]->span.channels > 24)
|
|
wc->tspans[3]->span.chans[z]->readchunk[x] = tmp & 0xff;
|
|
if (wc->tspans[2]->span.channels > 24)
|
|
wc->tspans[2]->span.chans[z]->readchunk[x] = (tmp & 0xff00) >> 8;
|
|
}
|
|
if (wc->tspans[1]->span.channels > 24)
|
|
wc->tspans[1]->span.chans[z]->readchunk[x] = (tmp & 0xff0000) >> 16;
|
|
if (wc->tspans[0]->span.channels > 24)
|
|
wc->tspans[0]->span.chans[z]->readchunk[x] = tmp >> 24;
|
|
}
|
|
}
|
|
/* Advance pointer by 4 TDM frame lengths */
|
|
readchunk += 32;
|
|
}
|
|
for (x=0;x<wc->numspans;x++) {
|
|
if (wc->tspans[x]->span.flags & DAHDI_FLAG_RUNNING) {
|
|
for (y=0;y<wc->tspans[x]->span.channels;y++) {
|
|
/* Echo cancel double buffered data */
|
|
dahdi_ec_chunk(wc->tspans[x]->span.chans[y],
|
|
wc->tspans[x]->span.chans[y]->readchunk,
|
|
wc->tspans[x]->ec_chunk2[y]);
|
|
memcpy(wc->tspans[x]->ec_chunk2[y],wc->tspans[x]->ec_chunk1[y],
|
|
DAHDI_CHUNKSIZE);
|
|
memcpy(wc->tspans[x]->ec_chunk1[y],
|
|
wc->tspans[x]->span.chans[y]->writechunk,
|
|
DAHDI_CHUNKSIZE);
|
|
}
|
|
_dahdi_receive(&wc->tspans[x]->span);
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
#if (DAHDI_CHUNKSIZE != 8)
|
|
#error Sorry, nextgen does not support chunksize != 8
|
|
#endif
|
|
|
|
static void __receive_span(struct t4_span *ts)
|
|
{
|
|
#ifdef VPM_SUPPORT
|
|
int y;
|
|
unsigned long merged;
|
|
merged = ts->dtmfactive & ts->dtmfmutemask;
|
|
if (merged) {
|
|
for (y=0;y<ts->span.channels;y++) {
|
|
/* Mute any DTMFs which are supposed to be muted */
|
|
if (test_bit(y, &merged)) {
|
|
memset(ts->span.chans[y]->readchunk, DAHDI_XLAW(0, ts->span.chans[y]), DAHDI_CHUNKSIZE);
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
_dahdi_ec_span(&ts->span);
|
|
_dahdi_receive(&ts->span);
|
|
}
|
|
|
|
static inline void __transmit_span(struct t4_span *ts)
|
|
{
|
|
_dahdi_transmit(&ts->span);
|
|
}
|
|
|
|
#ifdef ENABLE_WORKQUEUES
|
|
static void workq_handlespan(void *data)
|
|
{
|
|
struct t4_span *ts = data;
|
|
struct t4 *wc = ts->owner;
|
|
|
|
__receive_span(ts);
|
|
__transmit_span(ts);
|
|
atomic_dec(&wc->worklist);
|
|
if (!atomic_read(&wc->worklist))
|
|
t4_pci_out(wc, WC_INTR, 0);
|
|
}
|
|
#else
|
|
static void t4_prep_gen2(struct t4 *wc)
|
|
{
|
|
int x;
|
|
for (x=0;x<wc->numspans;x++) {
|
|
if (wc->tspans[x]->span.flags & DAHDI_FLAG_RUNNING) {
|
|
__receive_span(wc->tspans[x]);
|
|
__transmit_span(wc->tspans[x]);
|
|
}
|
|
}
|
|
}
|
|
|
|
#endif
|
|
#ifdef SUPPORT_GEN1
|
|
static void t4_transmitprep(struct t4 *wc, int irq)
|
|
{
|
|
u32 *writechunk;
|
|
int x, y, z;
|
|
unsigned int tmp;
|
|
int offset = 0;
|
|
if (!has_e1_span(wc))
|
|
offset = 4;
|
|
if (irq & 1) {
|
|
/* First part */
|
|
writechunk = wc->writechunk + 1;
|
|
} else {
|
|
writechunk = wc->writechunk + DAHDI_CHUNKSIZE * 32 + 1;
|
|
}
|
|
for (y=0;y<wc->numspans;y++) {
|
|
if (wc->tspans[y]->span.flags & DAHDI_FLAG_RUNNING)
|
|
_dahdi_transmit(&wc->tspans[y]->span);
|
|
}
|
|
|
|
for (x=0;x<DAHDI_CHUNKSIZE;x++) {
|
|
/* Once per chunk */
|
|
for (z=0;z<24;z++) {
|
|
/* All T1/E1 channels */
|
|
tmp = (wc->tspans[3]->span.chans[z]->writechunk[x]) |
|
|
(wc->tspans[2]->span.chans[z]->writechunk[x] << 8) |
|
|
(wc->tspans[1]->span.chans[z]->writechunk[x] << 16) |
|
|
(wc->tspans[0]->span.chans[z]->writechunk[x] << 24);
|
|
writechunk[z+offset] = tmp;
|
|
}
|
|
if (has_e1_span(wc)) {
|
|
for (z=24;z<31;z++) {
|
|
/* Only E1 channels now */
|
|
tmp = 0;
|
|
if (wc->numspans == 4) {
|
|
if (wc->tspans[3]->span.channels > 24)
|
|
tmp |= wc->tspans[3]->span.chans[z]->writechunk[x];
|
|
if (wc->tspans[2]->span.channels > 24)
|
|
tmp |= (wc->tspans[2]->span.chans[z]->writechunk[x] << 8);
|
|
}
|
|
if (wc->tspans[1]->span.channels > 24)
|
|
tmp |= (wc->tspans[1]->span.chans[z]->writechunk[x] << 16);
|
|
if (wc->tspans[0]->span.channels > 24)
|
|
tmp |= (wc->tspans[0]->span.chans[z]->writechunk[x] << 24);
|
|
writechunk[z] = tmp;
|
|
}
|
|
}
|
|
/* Advance pointer by 4 TDM frame lengths */
|
|
writechunk += 32;
|
|
}
|
|
|
|
}
|
|
#endif
|
|
|
|
static void t4_dahdi_rbsbits(struct dahdi_chan *const chan, int rxs)
|
|
{
|
|
if ((debug & DEBUG_RBS) && printk_ratelimit()) {
|
|
const struct t4_span *tspan = container_of(chan->span,
|
|
struct t4_span,
|
|
span);
|
|
const struct t4 *const wc = tspan->owner;
|
|
dev_notice(&wc->dev->dev, "Detected sigbits change on " \
|
|
"channel %s to %04x\n", chan->name, rxs);
|
|
}
|
|
dahdi_rbsbits(chan, rxs);
|
|
}
|
|
|
|
static void t4_check_sigbits(struct t4 *wc, int span)
|
|
{
|
|
int a,i,rxs;
|
|
struct t4_span *ts = wc->tspans[span];
|
|
|
|
if (debug & DEBUG_RBS)
|
|
dev_notice(&wc->dev->dev, "Checking sigbits on span %d\n",
|
|
span + 1);
|
|
|
|
if (E1 == ts->linemode) {
|
|
for (i = 0; i < 15; i++) {
|
|
a = t4_framer_in(wc, span, 0x71 + i);
|
|
/* Get high channel in low bits */
|
|
rxs = (a & 0xf);
|
|
if (!(ts->span.chans[i+16]->sig & DAHDI_SIG_CLEAR)) {
|
|
if (ts->span.chans[i+16]->rxsig != rxs)
|
|
t4_dahdi_rbsbits(ts->span.chans[i+16], rxs);
|
|
}
|
|
rxs = (a >> 4) & 0xf;
|
|
if (!(ts->span.chans[i]->sig & DAHDI_SIG_CLEAR)) {
|
|
if (ts->span.chans[i]->rxsig != rxs)
|
|
t4_dahdi_rbsbits(ts->span.chans[i], rxs);
|
|
}
|
|
}
|
|
} else if (ts->span.lineconfig & DAHDI_CONFIG_D4) {
|
|
for (i = 0; i < 24; i+=4) {
|
|
a = t4_framer_in(wc, span, 0x70 + (i>>2));
|
|
/* Get high channel in low bits */
|
|
rxs = (a & 0x3) << 2;
|
|
if (!(ts->span.chans[i+3]->sig & DAHDI_SIG_CLEAR)) {
|
|
if (ts->span.chans[i+3]->rxsig != rxs)
|
|
t4_dahdi_rbsbits(ts->span.chans[i+3], rxs);
|
|
}
|
|
rxs = (a & 0xc);
|
|
if (!(ts->span.chans[i+2]->sig & DAHDI_SIG_CLEAR)) {
|
|
if (ts->span.chans[i+2]->rxsig != rxs)
|
|
t4_dahdi_rbsbits(ts->span.chans[i+2], rxs);
|
|
}
|
|
rxs = (a >> 2) & 0xc;
|
|
if (!(ts->span.chans[i+1]->sig & DAHDI_SIG_CLEAR)) {
|
|
if (ts->span.chans[i+1]->rxsig != rxs)
|
|
t4_dahdi_rbsbits(ts->span.chans[i+1], rxs);
|
|
}
|
|
rxs = (a >> 4) & 0xc;
|
|
if (!(ts->span.chans[i]->sig & DAHDI_SIG_CLEAR)) {
|
|
if (ts->span.chans[i]->rxsig != rxs)
|
|
t4_dahdi_rbsbits(ts->span.chans[i], rxs);
|
|
}
|
|
}
|
|
} else {
|
|
for (i = 0; i < 24; i+=2) {
|
|
a = t4_framer_in(wc, span, 0x70 + (i>>1));
|
|
/* Get high channel in low bits */
|
|
rxs = (a & 0xf);
|
|
if (!(ts->span.chans[i+1]->sig & DAHDI_SIG_CLEAR)) {
|
|
/* XXX Not really reset on every trans! XXX */
|
|
if (ts->span.chans[i+1]->rxsig != rxs) {
|
|
t4_dahdi_rbsbits(ts->span.chans[i+1], rxs);
|
|
}
|
|
}
|
|
rxs = (a >> 4) & 0xf;
|
|
if (!(ts->span.chans[i]->sig & DAHDI_SIG_CLEAR)) {
|
|
/* XXX Not really reset on every trans! XXX */
|
|
if (ts->span.chans[i]->rxsig != rxs) {
|
|
t4_dahdi_rbsbits(ts->span.chans[i], rxs);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Must be called from within hardirq context. */
|
|
static void t4_check_alarms(struct t4 *wc, int span)
|
|
{
|
|
unsigned char c, d, e;
|
|
int alarms;
|
|
int x,j;
|
|
struct t4_span *ts = wc->tspans[span];
|
|
|
|
if (time_before(jiffies, ts->alarmcheck_time))
|
|
return;
|
|
|
|
if (!(ts->span.flags & DAHDI_FLAG_RUNNING))
|
|
return;
|
|
|
|
spin_lock(&wc->reglock);
|
|
|
|
c = __t4_framer_in(wc, span, 0x4c);
|
|
d = __t4_framer_in(wc, span, 0x4d);
|
|
|
|
/* Assume no alarms */
|
|
alarms = 0;
|
|
|
|
/* And consider only carrier alarms */
|
|
ts->span.alarms &= (DAHDI_ALARM_RED | DAHDI_ALARM_BLUE | DAHDI_ALARM_NOTOPEN);
|
|
|
|
if (E1 == ts->linemode) {
|
|
if (c & 0x04) {
|
|
/* No multiframe found, force RAI high after 400ms only if
|
|
we haven't found a multiframe since last loss
|
|
of frame */
|
|
if (!(ts->spanflags & FLAG_NMF)) {
|
|
__t4_framer_out(wc, span, 0x20, 0x9f | 0x20); /* LIM0: Force RAI High */
|
|
ts->spanflags |= FLAG_NMF;
|
|
dev_notice(&wc->dev->dev,
|
|
"Lost crc4-multiframe alignment\n");
|
|
}
|
|
__t4_framer_out(wc, span, 0x1e, 0xc3); /* Reset to CRC4 mode */
|
|
__t4_framer_out(wc, span, 0x1c, 0xf2); /* Force Resync */
|
|
__t4_framer_out(wc, span, 0x1c, 0xf0); /* Force Resync */
|
|
} else if (!(c & 0x02)) {
|
|
if ((ts->spanflags & FLAG_NMF)) {
|
|
__t4_framer_out(wc, span, 0x20, 0x9f); /* LIM0: Clear forced RAI */
|
|
ts->spanflags &= ~FLAG_NMF;
|
|
dev_notice(&wc->dev->dev,
|
|
"Obtained crc4-multiframe alignment\n");
|
|
}
|
|
}
|
|
} else {
|
|
/* Detect loopup code if we're not sending one */
|
|
if ((!ts->span.mainttimer) && (d & 0x08)) {
|
|
/* Loop-up code detected */
|
|
if ((ts->loopupcnt++ > 80) && (ts->span.maintstat != DAHDI_MAINT_REMOTELOOP)) {
|
|
dev_notice(&wc->dev->dev,
|
|
"span %d: Loopup detected,"\
|
|
" enabling remote loop\n",
|
|
span+1);
|
|
__t4_framer_out(wc, span, 0x36, 0x08); /* LIM0: Disable any local loop */
|
|
__t4_framer_out(wc, span, 0x37, 0xf6 ); /* LIM1: Enable remote loop */
|
|
ts->span.maintstat = DAHDI_MAINT_REMOTELOOP;
|
|
}
|
|
} else
|
|
ts->loopupcnt = 0;
|
|
/* Same for loopdown code */
|
|
if ((!ts->span.mainttimer) && (d & 0x10)) {
|
|
/* Loop-down code detected */
|
|
if ((ts->loopdowncnt++ > 80) && (ts->span.maintstat == DAHDI_MAINT_REMOTELOOP)) {
|
|
dev_notice(&wc->dev->dev,
|
|
"span %d: Loopdown detected,"\
|
|
" disabling remote loop\n",
|
|
span+1);
|
|
__t4_framer_out(wc, span, 0x36, 0x08); /* LIM0: Disable any local loop */
|
|
__t4_framer_out(wc, span, 0x37, 0xf0 ); /* LIM1: Disable remote loop */
|
|
ts->span.maintstat = DAHDI_MAINT_NONE;
|
|
}
|
|
} else
|
|
ts->loopdowncnt = 0;
|
|
}
|
|
|
|
if (ts->span.lineconfig & DAHDI_CONFIG_NOTOPEN) {
|
|
for (x=0,j=0;x < ts->span.channels;x++)
|
|
if ((ts->span.chans[x]->flags & DAHDI_FLAG_OPEN) ||
|
|
dahdi_have_netdev(ts->span.chans[x]))
|
|
j++;
|
|
if (!j)
|
|
alarms |= DAHDI_ALARM_NOTOPEN;
|
|
}
|
|
|
|
/* Loss of Frame Alignment */
|
|
if (c & 0x20) {
|
|
if (!ts->alarm_time) {
|
|
if (unlikely(debug)) {
|
|
/* starting to debounce LOF/LFA */
|
|
dev_info(&wc->dev->dev, "wct%dxxp: LOF/LFA "
|
|
"detected on span %d but debouncing "
|
|
"for %d ms\n", wc->numspans, span + 1,
|
|
alarmdebounce);
|
|
}
|
|
ts->alarm_time = jiffies +
|
|
msecs_to_jiffies(alarmdebounce);
|
|
} else if (time_after(jiffies, ts->alarm_time)) {
|
|
/* Disable Slip Interrupts */
|
|
e = __t4_framer_in(wc, span, 0x17);
|
|
__t4_framer_out(wc, span, 0x17, (e|0x03));
|
|
|
|
alarms |= DAHDI_ALARM_RED;
|
|
}
|
|
} else {
|
|
ts->alarm_time = 0;
|
|
}
|
|
|
|
/* Loss of Signal */
|
|
if (c & 0x80) {
|
|
if (!ts->losalarm_time) {
|
|
if (unlikely(debug)) {
|
|
/* starting to debounce LOS */
|
|
dev_info(&wc->dev->dev, "wct%dxxp: LOS "
|
|
"detected on span %d but debouncing "
|
|
"for %d ms\n", wc->numspans,
|
|
span + 1, losalarmdebounce);
|
|
}
|
|
ts->losalarm_time = jiffies +
|
|
msecs_to_jiffies(losalarmdebounce);
|
|
} else if (time_after(jiffies, ts->losalarm_time)) {
|
|
/* Disable Slip Interrupts */
|
|
e = __t4_framer_in(wc, span, 0x17);
|
|
__t4_framer_out(wc, span, 0x17, (e|0x03));
|
|
|
|
alarms |= DAHDI_ALARM_RED;
|
|
}
|
|
} else {
|
|
ts->losalarm_time = 0;
|
|
}
|
|
|
|
/* Alarm Indication Signal */
|
|
if (c & 0x40) {
|
|
if (!ts->aisalarm_time) {
|
|
if (unlikely(debug)) {
|
|
/* starting to debounce AIS */
|
|
dev_info(&wc->dev->dev, "wct%dxxp: AIS "
|
|
"detected on span %d but debouncing "
|
|
"for %d ms\n", wc->numspans,
|
|
span + 1, aisalarmdebounce);
|
|
}
|
|
ts->aisalarm_time = jiffies +
|
|
msecs_to_jiffies(aisalarmdebounce);
|
|
} else if (time_after(jiffies, ts->aisalarm_time)) {
|
|
alarms |= DAHDI_ALARM_BLUE;
|
|
}
|
|
} else {
|
|
ts->aisalarm_time = 0;
|
|
}
|
|
|
|
/* Add detailed alarm status information to a red alarm state */
|
|
if (alarms & DAHDI_ALARM_RED) {
|
|
if (c & FRS0_LOS)
|
|
alarms |= DAHDI_ALARM_LOS;
|
|
if (c & FRS0_LFA)
|
|
alarms |= DAHDI_ALARM_LFA;
|
|
if (c & FRS0_LMFA)
|
|
alarms |= DAHDI_ALARM_LMFA;
|
|
}
|
|
|
|
if (unlikely(debug)) {
|
|
/* Check to ensure the xmit line isn't shorted */
|
|
if (unlikely(d & FRS1_XLS)) {
|
|
dev_info(&wc->dev->dev,
|
|
"Detected a possible hardware malfunction"\
|
|
" this card may need servicing\n");
|
|
}
|
|
}
|
|
|
|
if (((!ts->span.alarms) && alarms) ||
|
|
(ts->span.alarms && (!alarms)))
|
|
set_bit(T4_CHECK_TIMING, &wc->checkflag);
|
|
|
|
/* Keep track of recovering */
|
|
if ((!alarms) && ts->span.alarms)
|
|
ts->alarmtimer = DAHDI_ALARMSETTLE_TIME;
|
|
if (ts->alarmtimer)
|
|
alarms |= DAHDI_ALARM_RECOVER;
|
|
|
|
/* If receiving alarms, go into Yellow alarm state */
|
|
if (alarms && !(ts->spanflags & FLAG_SENDINGYELLOW)) {
|
|
/* We manually do yellow alarm to handle RECOVER and NOTOPEN, otherwise it's auto anyway */
|
|
unsigned char fmr4;
|
|
fmr4 = __t4_framer_in(wc, span, 0x20);
|
|
__t4_framer_out(wc, span, 0x20, fmr4 | 0x20);
|
|
dev_info(&wc->dev->dev, "Setting yellow alarm span %d\n",
|
|
span+1);
|
|
ts->spanflags |= FLAG_SENDINGYELLOW;
|
|
} else if ((!alarms) && (ts->spanflags & FLAG_SENDINGYELLOW)) {
|
|
unsigned char fmr4;
|
|
/* We manually do yellow alarm to handle RECOVER */
|
|
fmr4 = __t4_framer_in(wc, span, 0x20);
|
|
__t4_framer_out(wc, span, 0x20, fmr4 & ~0x20);
|
|
dev_info(&wc->dev->dev, "Clearing yellow alarm span %d\n",
|
|
span+1);
|
|
|
|
/* Re-enable timing slip interrupts */
|
|
e = __t4_framer_in(wc, span, 0x17);
|
|
|
|
__t4_framer_out(wc, span, 0x17, (e & ~(0x03)));
|
|
|
|
ts->spanflags &= ~FLAG_SENDINGYELLOW;
|
|
}
|
|
|
|
/* Re-check the timing source when we enter/leave alarm, not withstanding
|
|
yellow alarm */
|
|
if (c & 0x10) { /* receiving yellow (RAI) */
|
|
if (!ts->yelalarm_time) {
|
|
if (unlikely(debug)) {
|
|
/* starting to debounce AIS */
|
|
dev_info(&wc->dev->dev, "wct%dxxp: yellow "
|
|
"(RAI) detected on span %d but "
|
|
"debouncing for %d ms\n",
|
|
wc->numspans, span + 1,
|
|
yelalarmdebounce);
|
|
}
|
|
ts->yelalarm_time = jiffies +
|
|
msecs_to_jiffies(yelalarmdebounce);
|
|
} else if (time_after(jiffies, ts->yelalarm_time)) {
|
|
alarms |= DAHDI_ALARM_YELLOW;
|
|
}
|
|
} else {
|
|
ts->yelalarm_time = 0;
|
|
}
|
|
|
|
if (alarms)
|
|
ts->alarmcheck_time = jiffies + msecs_to_jiffies(100);
|
|
else
|
|
ts->alarmcheck_time = jiffies + msecs_to_jiffies(50);
|
|
|
|
if (ts->span.mainttimer || ts->span.maintstat)
|
|
alarms |= DAHDI_ALARM_LOOPBACK;
|
|
ts->span.alarms = alarms;
|
|
|
|
spin_unlock(&wc->reglock);
|
|
dahdi_alarm_notify(&ts->span);
|
|
}
|
|
|
|
static void t4_do_counters(struct t4 *wc)
|
|
{
|
|
int span;
|
|
for (span = 0; span < wc->numspans; span++) {
|
|
struct t4_span *ts = wc->tspans[span];
|
|
|
|
spin_lock(&wc->reglock);
|
|
if (ts->alarmtimer && (0 == (--ts->alarmtimer)))
|
|
ts->span.alarms &= ~(DAHDI_ALARM_RECOVER);
|
|
spin_unlock(&wc->reglock);
|
|
|
|
t4_check_alarms(wc, span);
|
|
}
|
|
}
|
|
|
|
static inline void __handle_leds(struct t4 *wc)
|
|
{
|
|
int x;
|
|
|
|
wc->blinktimer++;
|
|
for (x=0;x<wc->numspans;x++) {
|
|
struct t4_span *ts = wc->tspans[x];
|
|
if (ts->span.flags & DAHDI_FLAG_RUNNING) {
|
|
if ((ts->span.alarms & (DAHDI_ALARM_RED |
|
|
DAHDI_ALARM_BLUE)) ||
|
|
ts->losalarm_time) {
|
|
#ifdef FANCY_ALARM
|
|
if (wc->blinktimer == (altab[wc->alarmpos] >> 1)) {
|
|
__t4_set_led(wc, x, WC_RED);
|
|
}
|
|
if (wc->blinktimer == 0xf) {
|
|
__t4_set_led(wc, x, WC_OFF);
|
|
}
|
|
#else
|
|
if (wc->blinktimer == 160) {
|
|
__t4_set_led(wc, x, WC_RED);
|
|
} else if (wc->blinktimer == 480) {
|
|
__t4_set_led(wc, x, WC_OFF);
|
|
}
|
|
#endif
|
|
} else if (ts->span.alarms & DAHDI_ALARM_YELLOW) {
|
|
/* Yellow Alarm */
|
|
__t4_set_led(wc, x, WC_YELLOW);
|
|
} else if (ts->span.mainttimer || ts->span.maintstat) {
|
|
#ifdef FANCY_ALARM
|
|
if (wc->blinktimer == (altab[wc->alarmpos] >> 1)) {
|
|
__t4_set_led(wc, x, WC_GREEN);
|
|
}
|
|
if (wc->blinktimer == 0xf) {
|
|
__t4_set_led(wc, x, WC_OFF);
|
|
}
|
|
#else
|
|
if (wc->blinktimer == 160) {
|
|
__t4_set_led(wc, x, WC_GREEN);
|
|
} else if (wc->blinktimer == 480) {
|
|
__t4_set_led(wc, x, WC_OFF);
|
|
}
|
|
#endif
|
|
} else {
|
|
/* No Alarm */
|
|
__t4_set_led(wc, x, WC_GREEN);
|
|
}
|
|
} else
|
|
__t4_set_led(wc, x, WC_OFF);
|
|
|
|
}
|
|
#ifdef FANCY_ALARM
|
|
if (wc->blinktimer == 0xf) {
|
|
wc->blinktimer = -1;
|
|
wc->alarmpos++;
|
|
if (wc->alarmpos >= ARRAY_SIZE(altab))
|
|
wc->alarmpos = 0;
|
|
}
|
|
#else
|
|
if (wc->blinktimer == 480)
|
|
wc->blinktimer = 0;
|
|
#endif
|
|
}
|
|
|
|
static inline void t4_framer_interrupt(struct t4 *wc, int span)
|
|
{
|
|
/* Check interrupts for a given span */
|
|
unsigned char gis, isr0, isr1, isr2, isr3, isr4;
|
|
int readsize = -1;
|
|
struct t4_span *ts = wc->tspans[span];
|
|
struct dahdi_chan *sigchan;
|
|
unsigned long flags;
|
|
bool recheck_sigbits = false;
|
|
|
|
|
|
/* 1st gen cards isn't used interrupts */
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
gis = __t4_framer_in(wc, span, FRMR_GIS);
|
|
isr0 = (gis & FRMR_GIS_ISR0) ? __t4_framer_in(wc, span, FRMR_ISR0) : 0;
|
|
isr1 = (gis & FRMR_GIS_ISR1) ? __t4_framer_in(wc, span, FRMR_ISR1) : 0;
|
|
isr2 = (gis & FRMR_GIS_ISR2) ? __t4_framer_in(wc, span, FRMR_ISR2) : 0;
|
|
isr3 = (gis & FRMR_GIS_ISR3) ? __t4_framer_in(wc, span, FRMR_ISR3) : 0;
|
|
isr4 = (gis & FRMR_GIS_ISR4) ? __t4_framer_in(wc, span, FRMR_ISR4) : 0;
|
|
|
|
if ((debug & DEBUG_FRAMER) && !(isr3 & ISR3_SEC)) {
|
|
dev_info(&wc->dev->dev, "gis: %02x, isr0: %02x, isr1: %02x, "\
|
|
"isr2: %02x, isr3: %08x, isr4: %02x, intcount=%u\n",
|
|
gis, isr0, isr1, isr2, isr3, isr4, wc->intcount);
|
|
}
|
|
|
|
/* Collect performance counters once per second */
|
|
if (isr3 & ISR3_SEC) {
|
|
ts->span.count.fe += __t4_framer_in(wc, span, FECL_T);
|
|
ts->span.count.crc4 += __t4_framer_in(wc, span, CEC1L_T);
|
|
ts->span.count.cv += __t4_framer_in(wc, span, CVCL_T);
|
|
ts->span.count.ebit += __t4_framer_in(wc, span, EBCL_T);
|
|
ts->span.count.be += __t4_framer_in(wc, span, BECL_T);
|
|
ts->span.count.prbs = __t4_framer_in(wc, span, FRS1_T);
|
|
if (DAHDI_RXSIG_INITIAL == ts->span.chans[0]->rxhooksig)
|
|
recheck_sigbits = true;
|
|
}
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
/* Collect errored second counter once per second */
|
|
if (isr3 & ISR3_ES) {
|
|
ts->span.count.errsec += 1;
|
|
}
|
|
|
|
if (isr0 & 0x08 || recheck_sigbits)
|
|
t4_check_sigbits(wc, span);
|
|
|
|
if (E1 == ts->linemode) {
|
|
/* E1 checks */
|
|
if ((isr3 & 0x38) || isr2 || isr1)
|
|
t4_check_alarms(wc, span);
|
|
} else {
|
|
/* T1 checks */
|
|
if (isr2 || (isr3 & 0x08))
|
|
t4_check_alarms(wc, span);
|
|
}
|
|
if (!ts->span.alarms) {
|
|
if ((isr3 & 0x3) || (isr4 & 0xc0))
|
|
ts->span.count.timingslips++;
|
|
|
|
if (debug & DEBUG_MAIN) {
|
|
if (isr3 & 0x02)
|
|
dev_notice(&wc->dev->dev, "TE%d10P: RECEIVE "
|
|
"slip NEGATIVE on span %d\n",
|
|
wc->numspans, span + 1);
|
|
if (isr3 & 0x01)
|
|
dev_notice(&wc->dev->dev, "TE%d10P: RECEIVE "
|
|
"slip POSITIVE on span %d\n",
|
|
wc->numspans, span + 1);
|
|
if (isr4 & 0x80)
|
|
dev_notice(&wc->dev->dev, "TE%dXXP: TRANSMIT "
|
|
"slip POSITIVE on span %d\n",
|
|
wc->numspans, span + 1);
|
|
if (isr4 & 0x40)
|
|
dev_notice(&wc->dev->dev, "TE%d10P: TRANSMIT "
|
|
"slip NEGATIVE on span %d\n",
|
|
wc->numspans, span + 1);
|
|
}
|
|
} else
|
|
ts->span.count.timingslips = 0;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
/* HDLC controller checks - receive side */
|
|
if (!ts->sigchan) {
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
return;
|
|
}
|
|
|
|
sigchan = ts->sigchan;
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
if (isr0 & FRMR_ISR0_RME) {
|
|
readsize = (t4_framer_in(wc, span, FRMR_RBCH) << 8) | t4_framer_in(wc, span, FRMR_RBCL);
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Received data length is %d "
|
|
"(%d)\n", readsize,
|
|
readsize & FRMR_RBCL_MAX_SIZE);
|
|
/* RPF isn't set on last part of frame */
|
|
if ((readsize > 0) && ((readsize &= FRMR_RBCL_MAX_SIZE) == 0))
|
|
readsize = FRMR_RBCL_MAX_SIZE + 1;
|
|
} else if (isr0 & FRMR_ISR0_RPF)
|
|
readsize = FRMR_RBCL_MAX_SIZE + 1;
|
|
|
|
if (readsize > 0) {
|
|
int i;
|
|
unsigned char readbuf[FRMR_RBCL_MAX_SIZE + 1];
|
|
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Framer %d: Got RPF/RME! "
|
|
"readsize is %d\n", sigchan->span->offset,
|
|
readsize);
|
|
|
|
for (i = 0; i < readsize; i++)
|
|
readbuf[i] = t4_framer_in(wc, span, FRMR_RXFIFO);
|
|
|
|
/* Tell the framer to clear the RFIFO */
|
|
t4_framer_cmd_wait(wc, span, FRMR_CMDR_RMC);
|
|
|
|
if (debug & DEBUG_FRAMER) {
|
|
dev_notice(&wc->dev->dev, "RX(");
|
|
for (i = 0; i < readsize; i++)
|
|
dev_notice(&wc->dev->dev, "%s%02x",
|
|
(i ? " " : ""), readbuf[i]);
|
|
dev_notice(&wc->dev->dev, ")\n");
|
|
}
|
|
|
|
if (isr0 & FRMR_ISR0_RME) {
|
|
/* Do checks for HDLC problems */
|
|
unsigned char rsis = readbuf[readsize-1];
|
|
unsigned char rsis_reg = t4_framer_in(wc, span, FRMR_RSIS);
|
|
|
|
++ts->frames_in;
|
|
if ((debug & DEBUG_FRAMER) && !(ts->frames_in & 0x0f))
|
|
dev_notice(&wc->dev->dev, "Received %d frames "
|
|
"on span %d\n", ts->frames_in, span);
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Received HDLC frame"
|
|
" %d. RSIS = 0x%x (%x)\n",
|
|
ts->frames_in, rsis, rsis_reg);
|
|
if (!(rsis & FRMR_RSIS_CRC16)) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "CRC check "
|
|
"failed %d\n", span);
|
|
dahdi_hdlc_abort(sigchan, DAHDI_EVENT_BADFCS);
|
|
} else if (rsis & FRMR_RSIS_RAB) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "ABORT of "
|
|
"current frame due to "
|
|
"overflow %d\n", span);
|
|
dahdi_hdlc_abort(sigchan, DAHDI_EVENT_ABORT);
|
|
} else if (rsis & FRMR_RSIS_RDO) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "HDLC "
|
|
"overflow occured %d\n",
|
|
span);
|
|
dahdi_hdlc_abort(sigchan, DAHDI_EVENT_OVERRUN);
|
|
} else if (!(rsis & FRMR_RSIS_VFR)) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Valid Frame"
|
|
" check failed on span %d\n",
|
|
span);
|
|
dahdi_hdlc_abort(sigchan, DAHDI_EVENT_ABORT);
|
|
} else {
|
|
dahdi_hdlc_putbuf(sigchan, readbuf, readsize - 1);
|
|
dahdi_hdlc_finish(sigchan);
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Received "
|
|
"valid HDLC frame on span %d"
|
|
"\n", span);
|
|
}
|
|
} else if (isr0 & FRMR_ISR0_RPF)
|
|
dahdi_hdlc_putbuf(sigchan, readbuf, readsize);
|
|
}
|
|
|
|
/* Transmit side */
|
|
if (isr1 & FRMR_ISR1_XDU) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "XDU: Resetting signal "
|
|
"controller!\n");
|
|
t4_framer_cmd_wait(wc, span, FRMR_CMDR_SRES);
|
|
} else if (isr1 & FRMR_ISR1_XPR) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Sigchan %d is %p\n",
|
|
sigchan->chanpos, sigchan);
|
|
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "Framer %d: Got XPR!\n",
|
|
sigchan->span->offset);
|
|
t4_hdlc_xmit_fifo(wc, span, ts);
|
|
}
|
|
|
|
if (isr1 & FRMR_ISR1_ALLS) {
|
|
if (debug & DEBUG_FRAMER)
|
|
dev_notice(&wc->dev->dev, "ALLS received\n");
|
|
}
|
|
}
|
|
|
|
#ifdef SUPPORT_GEN1
|
|
static irqreturn_t _t4_interrupt(int irq, void *dev_id)
|
|
{
|
|
struct t4 *wc = dev_id;
|
|
unsigned long flags;
|
|
int x;
|
|
|
|
unsigned int status;
|
|
unsigned int status2;
|
|
|
|
/* Make sure it's really for us */
|
|
status = __t4_pci_in(wc, WC_INTR);
|
|
|
|
/* Process framer interrupts */
|
|
status2 = t4_framer_in(wc, 0, FRMR_CIS);
|
|
if (status2 & 0x0f) {
|
|
for (x = 0; x < wc->numspans; ++x) {
|
|
if (status2 & (1 << x))
|
|
t4_framer_interrupt(wc, x);
|
|
}
|
|
}
|
|
|
|
/* Ignore if it's not for us */
|
|
if (!status)
|
|
return IRQ_NONE;
|
|
|
|
__t4_pci_out(wc, WC_INTR, 0);
|
|
|
|
if (!wc->spansstarted) {
|
|
dev_notice(&wc->dev->dev, "Not prepped yet!\n");
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
wc->intcount++;
|
|
|
|
if (status & 0x3) {
|
|
t4_receiveprep(wc, status);
|
|
t4_transmitprep(wc, status);
|
|
}
|
|
|
|
t4_do_counters(wc);
|
|
|
|
x = wc->intcount & 15 /* 63 */;
|
|
switch(x) {
|
|
case 0:
|
|
case 1:
|
|
case 2:
|
|
case 3:
|
|
t4_check_sigbits(wc, x);
|
|
break;
|
|
case 4:
|
|
case 5:
|
|
case 6:
|
|
case 7:
|
|
t4_check_alarms(wc, x - 4);
|
|
break;
|
|
}
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
|
|
__handle_leds(wc);
|
|
|
|
if (test_bit(T4_CHECK_TIMING, &wc->checkflag))
|
|
__t4_set_timing_source_auto(wc);
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
return IRQ_RETVAL(1);
|
|
}
|
|
|
|
DAHDI_IRQ_HANDLER(t4_interrupt)
|
|
{
|
|
irqreturn_t ret;
|
|
unsigned long flags;
|
|
local_irq_save(flags);
|
|
ret = _t4_interrupt(irq, dev_id);
|
|
local_irq_restore(flags);
|
|
return ret;
|
|
}
|
|
#endif
|
|
|
|
static int t4_allocate_buffers(struct t4 *wc, int numbufs,
|
|
void **oldalloc, dma_addr_t *oldwritedma)
|
|
{
|
|
void *alloc;
|
|
dma_addr_t writedma;
|
|
|
|
/* 32 channels, Double-buffer, Read/Write, 4 spans */
|
|
alloc = pci_alloc_consistent(wc->dev, numbufs * T4_BASE_SIZE(wc) * 2,
|
|
&writedma);
|
|
|
|
if (!alloc) {
|
|
dev_notice(&wc->dev->dev, "wct%dxxp: Unable to allocate "
|
|
"DMA-able memory\n", wc->numspans);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (oldwritedma)
|
|
*oldwritedma = wc->writedma;
|
|
if (oldalloc)
|
|
*oldalloc = wc->writechunk;
|
|
|
|
wc->writechunk = alloc;
|
|
wc->writedma = writedma;
|
|
|
|
/* Read is after the whole write piece (in words) */
|
|
wc->readchunk = wc->writechunk + (T4_BASE_SIZE(wc) * numbufs) / 4;
|
|
|
|
/* Same thing but in bytes... */
|
|
wc->readdma = wc->writedma + (T4_BASE_SIZE(wc) * numbufs);
|
|
|
|
wc->numbufs = numbufs;
|
|
|
|
/* Initialize Write/Buffers to all blank data */
|
|
memset(wc->writechunk, 0x00, T4_BASE_SIZE(wc) * numbufs);
|
|
memset(wc->readchunk, 0xff, T4_BASE_SIZE(wc) * numbufs);
|
|
|
|
if (debug) {
|
|
dev_notice(&wc->dev->dev, "DMA memory base of size %d at " \
|
|
"%p. Read: %p and Write %p\n",
|
|
numbufs * T4_BASE_SIZE(wc) * 2, wc->writechunk,
|
|
wc->readchunk, wc->writechunk);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void t4_increase_latency(struct t4 *wc, int newlatency)
|
|
{
|
|
unsigned long flags;
|
|
void *oldalloc;
|
|
dma_addr_t oldaddr;
|
|
int oldbufs;
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
|
|
__t4_pci_out(wc, WC_DMACTRL, 0x00000000);
|
|
/* Acknowledge any pending interrupts */
|
|
__t4_pci_out(wc, WC_INTR, 0x00000000);
|
|
|
|
__t4_pci_in(wc, WC_VERSION);
|
|
|
|
oldbufs = wc->numbufs;
|
|
|
|
if (t4_allocate_buffers(wc, newlatency, &oldalloc, &oldaddr)) {
|
|
dev_info(&wc->dev->dev, "Error allocating latency buffers for "
|
|
"latency of %d\n", newlatency);
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
return;
|
|
}
|
|
|
|
__t4_pci_out(wc, WC_RDADDR, wc->readdma);
|
|
__t4_pci_out(wc, WC_WRADDR, wc->writedma);
|
|
|
|
__t4_pci_in(wc, WC_VERSION);
|
|
|
|
__t4_pci_out(wc, 5, (ms_per_irq << 16) | newlatency);
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
|
|
__t4_pci_in(wc, WC_VERSION);
|
|
|
|
wc->rxident = 0;
|
|
wc->lastindex = 0;
|
|
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
pci_free_consistent(wc->dev, T4_BASE_SIZE(wc) * oldbufs * 2,
|
|
oldalloc, oldaddr);
|
|
|
|
dev_info(&wc->dev->dev, "Increased latency to %d\n", newlatency);
|
|
|
|
}
|
|
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 20)
|
|
static void t4_work_func(void *data)
|
|
{
|
|
struct t4 *wc = data;
|
|
#else
|
|
static void t4_work_func(struct work_struct *work)
|
|
{
|
|
struct t4 *wc = container_of(work, struct t4, bh_work);
|
|
#endif
|
|
|
|
if (test_bit(T4_CHANGE_LATENCY, &wc->checkflag)) {
|
|
if (wc->needed_latency != wc->numbufs) {
|
|
t4_increase_latency(wc, wc->needed_latency);
|
|
clear_bit(T4_CHANGE_LATENCY, &wc->checkflag);
|
|
}
|
|
}
|
|
#ifdef VPM_SUPPORT
|
|
if (wc->vpm) {
|
|
if (test_and_clear_bit(T4_CHECK_VPM, &wc->checkflag)) {
|
|
/* How stupid is it that the octasic can't generate an
|
|
* interrupt when there's a tone, in spite of what
|
|
* their documentation says? */
|
|
t4_check_vpm(wc);
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
static irqreturn_t _t4_interrupt_gen2(int irq, void *dev_id)
|
|
{
|
|
struct t4 *wc = dev_id;
|
|
unsigned int status;
|
|
unsigned char rxident, expected;
|
|
|
|
/* Check this first in case we get a spurious interrupt */
|
|
if (unlikely(test_bit(T4_STOP_DMA, &wc->checkflag))) {
|
|
/* Stop DMA cleanly if requested */
|
|
wc->dmactrl = 0x0;
|
|
t4_pci_out(wc, WC_DMACTRL, 0x00000000);
|
|
/* Acknowledge any pending interrupts */
|
|
t4_pci_out(wc, WC_INTR, 0x00000000);
|
|
spin_lock(&wc->reglock);
|
|
__t4_set_sclk_src(wc, WC_SELF, 0, 0);
|
|
spin_unlock(&wc->reglock);
|
|
return IRQ_RETVAL(1);
|
|
}
|
|
|
|
/* Make sure it's really for us */
|
|
status = __t4_pci_in(wc, WC_INTR);
|
|
|
|
/* Ignore if it's not for us */
|
|
if (!(status & 0x7)) {
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
#ifdef ENABLE_WORKQUEUES
|
|
__t4_pci_out(wc, WC_INTR, status & 0x00000008);
|
|
#endif
|
|
|
|
if (unlikely(!wc->spansstarted)) {
|
|
dev_info(&wc->dev->dev, "Not prepped yet!\n");
|
|
return IRQ_NONE;
|
|
}
|
|
|
|
wc->intcount++;
|
|
if ((wc->devtype->flags & FLAG_5THGEN) && (status & 0x2)) {
|
|
rxident = (status >> 16) & 0x7f;
|
|
expected = (wc->rxident + ms_per_irq) % 128;
|
|
|
|
if ((rxident != expected) && !test_bit(T4_IGNORE_LATENCY, &wc->checkflag)) {
|
|
int needed_latency;
|
|
int smallest_max;
|
|
|
|
if (debug & DEBUG_MAIN)
|
|
dev_warn(&wc->dev->dev, "Missed interrupt. "
|
|
"Expected ident of %d and got ident "
|
|
"of %d\n", expected, rxident);
|
|
|
|
if (test_bit(T4_IGNORE_LATENCY, &wc->checkflag)) {
|
|
dev_info(&wc->dev->dev,
|
|
"Should have ignored latency\n");
|
|
}
|
|
if (rxident > wc->rxident) {
|
|
needed_latency = rxident - wc->rxident;
|
|
} else {
|
|
needed_latency = (128 - wc->rxident) + rxident;
|
|
}
|
|
|
|
needed_latency += 1;
|
|
|
|
smallest_max = (max_latency >= GEN5_MAX_LATENCY) ? GEN5_MAX_LATENCY : max_latency;
|
|
|
|
if (needed_latency > smallest_max) {
|
|
dev_info(&wc->dev->dev, "Truncating latency "
|
|
"request to %d instead of %d\n",
|
|
smallest_max, needed_latency);
|
|
needed_latency = smallest_max;
|
|
}
|
|
|
|
if (needed_latency > wc->numbufs) {
|
|
dev_info(&wc->dev->dev, "Need to increase "
|
|
"latency. Estimated latency should "
|
|
"be %d\n", needed_latency);
|
|
wc->ddev->irqmisses++;
|
|
wc->needed_latency = needed_latency;
|
|
__t4_pci_out(wc, WC_DMACTRL, 0x00000000);
|
|
set_bit(T4_CHANGE_LATENCY, &wc->checkflag);
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
wc->rxident = rxident;
|
|
}
|
|
|
|
#ifdef DEBUG
|
|
if (unlikely((wc->intcount < 20)))
|
|
dev_dbg(&wc->dev->dev, "2G: Got interrupt, status = %08x, "
|
|
"CIS = %04x\n", status, t4_framer_in(wc, 0, FRMR_CIS));
|
|
#endif
|
|
|
|
if (likely(status & 0x2)) {
|
|
#ifdef ENABLE_WORKQUEUES
|
|
int cpus = num_online_cpus();
|
|
atomic_set(&wc->worklist, wc->numspans);
|
|
if (wc->tspans[0]->span.flags & DAHDI_FLAG_RUNNING)
|
|
t4_queue_work(wc->workq, &wc->tspans[0]->swork, 0);
|
|
else
|
|
atomic_dec(&wc->worklist);
|
|
if (wc->tspans[1]->span.flags & DAHDI_FLAG_RUNNING)
|
|
t4_queue_work(wc->workq, &wc->tspans[1]->swork, 1 % cpus);
|
|
else
|
|
atomic_dec(&wc->worklist);
|
|
if (wc->numspans == 4) {
|
|
if (wc->tspans[2]->span.flags & DAHDI_FLAG_RUNNING)
|
|
t4_queue_work(wc->workq, &wc->tspans[2]->swork, 2 % cpus);
|
|
else
|
|
atomic_dec(&wc->worklist);
|
|
if (wc->tspans[3]->span.flags & DAHDI_FLAG_RUNNING)
|
|
t4_queue_work(wc->workq, &wc->tspans[3]->swork, 3 % cpus);
|
|
else
|
|
atomic_dec(&wc->worklist);
|
|
}
|
|
#else
|
|
unsigned int reg5 = __t4_pci_in(wc, 5);
|
|
|
|
#ifdef DEBUG
|
|
if (wc->intcount < 20)
|
|
dev_info(&wc->dev->dev, "Reg 5 is %08x\n", reg5);
|
|
#endif
|
|
|
|
if (wc->devtype->flags & FLAG_5THGEN) {
|
|
unsigned int current_index = (reg5 >> 8) & 0x7f;
|
|
|
|
while (((wc->lastindex + 1) % wc->numbufs) != current_index) {
|
|
wc->lastindex = (wc->lastindex + 1) % wc->numbufs;
|
|
setup_chunks(wc, wc->lastindex);
|
|
t4_prep_gen2(wc);
|
|
}
|
|
} else {
|
|
t4_prep_gen2(wc);
|
|
}
|
|
|
|
#endif
|
|
t4_do_counters(wc);
|
|
spin_lock(&wc->reglock);
|
|
__handle_leds(wc);
|
|
spin_unlock(&wc->reglock);
|
|
|
|
}
|
|
|
|
if (unlikely(status & 0x1)) {
|
|
unsigned char cis;
|
|
|
|
cis = t4_framer_in(wc, 0, FRMR_CIS);
|
|
if (cis & FRMR_CIS_GIS1)
|
|
t4_framer_interrupt(wc, 0);
|
|
if (cis & FRMR_CIS_GIS2)
|
|
t4_framer_interrupt(wc, 1);
|
|
if (cis & FRMR_CIS_GIS3)
|
|
t4_framer_interrupt(wc, 2);
|
|
if (cis & FRMR_CIS_GIS4)
|
|
t4_framer_interrupt(wc, 3);
|
|
|
|
if (is_octal(wc)) {
|
|
if (cis & FRMR_CIS_GIS5)
|
|
t4_framer_interrupt(wc, 4);
|
|
if (cis & FRMR_CIS_GIS6)
|
|
t4_framer_interrupt(wc, 5);
|
|
if (cis & FRMR_CIS_GIS7)
|
|
t4_framer_interrupt(wc, 6);
|
|
if (cis & FRMR_CIS_GIS8)
|
|
t4_framer_interrupt(wc, 7);
|
|
}
|
|
}
|
|
|
|
#ifdef VPM_SUPPORT
|
|
if (wc->vpm && vpmdtmfsupport) {
|
|
/* How stupid is it that the octasic can't generate an
|
|
* interrupt when there's a tone, in spite of what their
|
|
* documentation says? */
|
|
if (!(wc->intcount & 0xf))
|
|
set_bit(T4_CHECK_VPM, &wc->checkflag);
|
|
}
|
|
#endif
|
|
|
|
spin_lock(&wc->reglock);
|
|
|
|
if (unlikely(test_bit(T4_CHECK_TIMING, &wc->checkflag))) {
|
|
__t4_set_timing_source_auto(wc);
|
|
}
|
|
|
|
spin_unlock(&wc->reglock);
|
|
|
|
out:
|
|
if (unlikely(test_bit(T4_CHANGE_LATENCY, &wc->checkflag) || test_bit(T4_CHECK_VPM, &wc->checkflag)))
|
|
schedule_work(&wc->bh_work);
|
|
|
|
#ifndef ENABLE_WORKQUEUES
|
|
__t4_pci_out(wc, WC_INTR, 0);
|
|
#endif
|
|
|
|
return IRQ_RETVAL(1);
|
|
}
|
|
|
|
DAHDI_IRQ_HANDLER(t4_interrupt_gen2)
|
|
{
|
|
irqreturn_t ret;
|
|
unsigned long flags;
|
|
local_irq_save(flags);
|
|
ret = _t4_interrupt_gen2(irq, dev_id);
|
|
local_irq_restore(flags);
|
|
return ret;
|
|
}
|
|
|
|
#ifdef SUPPORT_GEN1
|
|
static int t4_reset_dma(struct t4 *wc)
|
|
{
|
|
/* Turn off DMA and such */
|
|
wc->dmactrl = 0x0;
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
t4_pci_out(wc, WC_COUNT, 0);
|
|
t4_pci_out(wc, WC_RDADDR, 0);
|
|
t4_pci_out(wc, WC_WRADDR, 0);
|
|
t4_pci_out(wc, WC_INTR, 0);
|
|
/* Turn it all back on */
|
|
t4_pci_out(wc, WC_RDADDR, wc->readdma);
|
|
t4_pci_out(wc, WC_WRADDR, wc->writedma);
|
|
t4_pci_out(wc, WC_COUNT, ((DAHDI_MAX_CHUNKSIZE * 2 * 32 - 1) << 18) | ((DAHDI_MAX_CHUNKSIZE * 2 * 32 - 1) << 2));
|
|
t4_pci_out(wc, WC_INTR, 0);
|
|
#ifdef VPM_SUPPORT
|
|
wc->dmactrl = 0xc0000000 | (1 << 29) |
|
|
((wc->vpm) ? T4_VPM_PRESENT : 0);
|
|
#else
|
|
wc->dmactrl = 0xc0000000 | (1 << 29);
|
|
#endif
|
|
if (noburst)
|
|
wc->dmactrl |= (1 << 26);
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
#ifdef VPM_SUPPORT
|
|
static void t4_vpm_init(struct t4 *wc)
|
|
{
|
|
int laws[8] = { 0, };
|
|
int x;
|
|
unsigned int vpm_capacity;
|
|
struct firmware embedded_firmware;
|
|
const struct firmware *firmware = &embedded_firmware;
|
|
#if !defined(HOTPLUG_FIRMWARE)
|
|
extern void _binary_dahdi_fw_oct6114_064_bin_size;
|
|
extern void _binary_dahdi_fw_oct6114_128_bin_size;
|
|
extern void _binary_dahdi_fw_oct6114_256_bin_size;
|
|
extern u8 _binary_dahdi_fw_oct6114_064_bin_start[];
|
|
extern u8 _binary_dahdi_fw_oct6114_128_bin_start[];
|
|
extern u8 _binary_dahdi_fw_oct6114_256_bin_start[];
|
|
#else
|
|
static const char oct064_firmware[] = "dahdi-fw-oct6114-064.bin";
|
|
static const char oct128_firmware[] = "dahdi-fw-oct6114-128.bin";
|
|
static const char oct256_firmware[] = "dahdi-fw-oct6114-256.bin";
|
|
#endif
|
|
|
|
if (!vpmsupport) {
|
|
dev_info(&wc->dev->dev, "VPM450: Support Disabled\n");
|
|
return;
|
|
}
|
|
|
|
/* Turn on GPIO/DATA mux if supported */
|
|
t4_gpio_setdir(wc, (1 << 24), (1 << 24));
|
|
__t4_raw_oct_out(wc, 0x000a, 0x5678);
|
|
__t4_raw_oct_out(wc, 0x0004, 0x1234);
|
|
__t4_raw_oct_in(wc, 0x0004);
|
|
__t4_raw_oct_in(wc, 0x000a);
|
|
if (debug)
|
|
dev_notice(&wc->dev->dev, "OCT Result: %04x/%04x\n",
|
|
__t4_raw_oct_in(wc, 0x0004),
|
|
__t4_raw_oct_in(wc, 0x000a));
|
|
if (__t4_raw_oct_in(wc, 0x0004) != 0x1234) {
|
|
dev_notice(&wc->dev->dev, "VPM450: Not Present\n");
|
|
return;
|
|
}
|
|
|
|
/* Setup alaw vs ulaw rules */
|
|
for (x = 0;x < wc->numspans; x++) {
|
|
if (wc->tspans[x]->span.channels > 24)
|
|
laws[x] = 1;
|
|
}
|
|
|
|
vpm_capacity = get_vpm450m_capacity(&wc->dev->dev);
|
|
if (vpm_capacity != wc->numspans * 32) {
|
|
dev_info(&wc->dev->dev, "Disabling VPMOCT%03d. TE%dXXP"\
|
|
" requires a VPMOCT%03d", vpm_capacity,
|
|
wc->numspans, wc->numspans*32);
|
|
return;
|
|
}
|
|
|
|
switch (vpm_capacity) {
|
|
case 64:
|
|
#if defined(HOTPLUG_FIRMWARE)
|
|
if ((request_firmware(&firmware, oct064_firmware, &wc->dev->dev) != 0) ||
|
|
!firmware) {
|
|
dev_notice(&wc->dev->dev, "VPM450: firmware %s not "
|
|
"available from userspace\n", oct064_firmware);
|
|
return;
|
|
}
|
|
#else
|
|
embedded_firmware.data = _binary_dahdi_fw_oct6114_064_bin_start;
|
|
/* Yes... this is weird. objcopy gives us a symbol containing
|
|
the size of the firmware, not a pointer a variable containing
|
|
the size. The only way we can get the value of the symbol
|
|
is to take its address, so we define it as a pointer and
|
|
then cast that value to the proper type.
|
|
*/
|
|
embedded_firmware.size = (size_t) &_binary_dahdi_fw_oct6114_064_bin_size;
|
|
#endif
|
|
break;
|
|
case 128:
|
|
#if defined(HOTPLUG_FIRMWARE)
|
|
if ((request_firmware(&firmware, oct128_firmware, &wc->dev->dev) != 0) ||
|
|
!firmware) {
|
|
dev_notice(&wc->dev->dev, "VPM450: firmware %s not "
|
|
"available from userspace\n", oct128_firmware);
|
|
return;
|
|
}
|
|
#else
|
|
embedded_firmware.data = _binary_dahdi_fw_oct6114_128_bin_start;
|
|
/* Yes... this is weird. objcopy gives us a symbol containing
|
|
the size of the firmware, not a pointer a variable containing
|
|
the size. The only way we can get the value of the symbol
|
|
is to take its address, so we define it as a pointer and
|
|
then cast that value to the proper type.
|
|
*/
|
|
embedded_firmware.size = (size_t) &_binary_dahdi_fw_oct6114_128_bin_size;
|
|
#endif
|
|
break;
|
|
case 256:
|
|
#if defined(HOTPLUG_FIRMWARE)
|
|
if ((request_firmware(&firmware, oct256_firmware, &wc->dev->dev) != 0) ||
|
|
!firmware) {
|
|
dev_notice(&wc->dev->dev, "VPM450: firmware %s not "
|
|
"available from userspace\n", oct256_firmware);
|
|
return;
|
|
}
|
|
#else
|
|
embedded_firmware.data = _binary_dahdi_fw_oct6114_256_bin_start;
|
|
/* Yes... this is weird. objcopy gives us a symbol containing
|
|
the size of the firmware, not a pointer a variable containing
|
|
the size. The only way we can get the value of the symbol
|
|
is to take its address, so we define it as a pointer and
|
|
then cast that value to the proper type.
|
|
*/
|
|
embedded_firmware.size = (size_t) &_binary_dahdi_fw_oct6114_256_bin_size;
|
|
#endif
|
|
break;
|
|
default:
|
|
dev_notice(&wc->dev->dev, "Unsupported channel capacity found "
|
|
"on VPM module (%d).\n", vpm_capacity);
|
|
return;
|
|
}
|
|
|
|
wc->vpm = init_vpm450m(&wc->dev->dev, laws, wc->numspans, firmware);
|
|
if (!wc->vpm) {
|
|
dev_notice(&wc->dev->dev, "VPM450: Failed to initialize\n");
|
|
if (firmware != &embedded_firmware)
|
|
release_firmware(firmware);
|
|
return;
|
|
}
|
|
|
|
if (firmware != &embedded_firmware)
|
|
release_firmware(firmware);
|
|
|
|
if (vpmdtmfsupport == -1) {
|
|
dev_info(&wc->dev->dev, "VPM450: hardware DTMF disabled.\n");
|
|
vpmdtmfsupport = 0;
|
|
}
|
|
|
|
dev_info(&wc->dev->dev, "VPM450: Present and operational servicing %d "
|
|
"span(s)\n", wc->numspans);
|
|
|
|
}
|
|
#endif /* VPM_SUPPORT */
|
|
|
|
static void t4_tsi_reset(struct t4 *wc)
|
|
{
|
|
int x;
|
|
if (is_octal(wc)) {
|
|
for (x = 0; x < 256; x++) {
|
|
wc->dmactrl &= ~0x0001ffff;
|
|
wc->dmactrl |= (0x00004000 | ((x & 0x7f) << 7) | ((x >> 7) << 15));
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
}
|
|
wc->dmactrl &= ~0x0001ffff;
|
|
} else {
|
|
for (x = 0; x < 128; x++) {
|
|
wc->dmactrl &= ~0x00007fff;
|
|
wc->dmactrl |= (0x00004000 | (x << 7));
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
}
|
|
wc->dmactrl &= ~0x00007fff;
|
|
}
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
}
|
|
|
|
/* Note that channels here start from 1 */
|
|
static void t4_tsi_assign(struct t4 *wc, int fromspan, int fromchan, int tospan, int tochan)
|
|
{
|
|
unsigned long flags;
|
|
int fromts, tots;
|
|
|
|
fromts = (fromspan << 5) |(fromchan);
|
|
tots = (tospan << 5) | (tochan);
|
|
|
|
if (!has_e1_span(wc)) {
|
|
fromts += 4;
|
|
tots += 4;
|
|
}
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
if (is_octal(wc)) {
|
|
int fromts_b = fromts & 0x7f;
|
|
int fromts_t = fromts >> 7;
|
|
int tots_b = tots & 0x7f;
|
|
int tots_t = tots >> 7;
|
|
|
|
wc->dmactrl &= ~0x0001ffff;
|
|
wc->dmactrl |= ((fromts_t << 16) | (tots_t << 15) | 0x00004000 | (tots_b << 7) | (fromts_b));
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
wc->dmactrl &= ~0x0001ffff;
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
} else {
|
|
wc->dmactrl &= ~0x00007fff;
|
|
wc->dmactrl |= (0x00004000 | (tots << 7) | (fromts));
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
wc->dmactrl &= ~0x00007fff;
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
}
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
static void t4_tsi_unassign(struct t4 *wc, int tospan, int tochan)
|
|
{
|
|
unsigned long flags;
|
|
int tots;
|
|
|
|
tots = (tospan << 5) | (tochan);
|
|
|
|
if (!has_e1_span(wc))
|
|
tots += 4;
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
if (is_octal(wc)) {
|
|
int tots_b = tots & 0x7f;
|
|
int tots_t = tots >> 7;
|
|
|
|
wc->dmactrl &= ~0x0001ffff;
|
|
wc->dmactrl |= ((tots_t << 15) | 0x00004000 | (tots_b << 7));
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
if (debug & DEBUG_TSI)
|
|
dev_notice(&wc->dev->dev, "Sending '%08x\n", wc->dmactrl);
|
|
wc->dmactrl &= ~0x0001ffff;
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
} else {
|
|
wc->dmactrl &= ~0x00007fff;
|
|
wc->dmactrl |= (0x00004000 | (tots << 7));
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
if (debug & DEBUG_TSI)
|
|
dev_notice(&wc->dev->dev, "Sending '%08x\n", wc->dmactrl);
|
|
wc->dmactrl &= ~0x00007fff;
|
|
__t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
}
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
}
|
|
|
|
#ifndef CONFIG_NOEXTENDED_RESET
|
|
static void t4_extended_reset(struct t4 *wc)
|
|
{
|
|
unsigned int oldreg = t4_pci_in(wc, 0x4);
|
|
|
|
udelay(1000);
|
|
|
|
t4_pci_out(wc, 0x4, 0x42000000);
|
|
t4_pci_out(wc, 0xa, 0x42000000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00180000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00180000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00180000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00180000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00180000);
|
|
t4_pci_out(wc, 0xa, 0x00080000);
|
|
t4_pci_out(wc, 0xa, 0x00180000);
|
|
t4_pci_out(wc, 0x4, oldreg);
|
|
|
|
udelay(1000);
|
|
}
|
|
#endif
|
|
|
|
#define SPI_CS (0)
|
|
#define SPI_CLK (1)
|
|
#define SPI_IO0 (2)
|
|
#define SPI_IO1 (3)
|
|
#define SPI_IO3 (4)
|
|
#define SPI_IO2 (5)
|
|
#define SPI_IN (0)
|
|
#define SPI_OUT (1)
|
|
#define ESPI_REG 13
|
|
|
|
static void t8_clear_bit(struct t4 *wc, int whichb)
|
|
{
|
|
wc->st.wrreg &= ~(1 << whichb);
|
|
}
|
|
|
|
static void t8_set_bit(struct t4 *wc, int whichb, int val)
|
|
{
|
|
t8_clear_bit(wc, whichb);
|
|
wc->st.wrreg |= (val << whichb);
|
|
}
|
|
|
|
static int t8_get_bit(struct t4 *wc, int whichb)
|
|
{
|
|
return (wc->st.rdreg >> whichb) & 1;
|
|
}
|
|
|
|
static void set_iodir(struct t4 *wc, int whichb, int dir)
|
|
{
|
|
whichb += 16;
|
|
t8_clear_bit(wc, whichb);
|
|
t8_set_bit(wc, whichb, dir);
|
|
}
|
|
|
|
static void write_hwreg(struct t4 *wc)
|
|
{
|
|
t4_pci_out(wc, ESPI_REG, wc->st.wrreg);
|
|
}
|
|
|
|
static void read_hwreg(struct t4 *wc)
|
|
{
|
|
wc->st.rdreg = t4_pci_in(wc, ESPI_REG);
|
|
}
|
|
|
|
static void set_cs(struct t4 *wc, int state)
|
|
{
|
|
t8_set_bit(wc, SPI_CS, state);
|
|
write_hwreg(wc);
|
|
}
|
|
|
|
static void set_clk(struct t4 *wc, int clk)
|
|
{
|
|
t8_set_bit(wc, SPI_CLK, clk);
|
|
write_hwreg(wc);
|
|
}
|
|
|
|
static void clk_bit_out(struct t4 *wc, int val)
|
|
{
|
|
t8_set_bit(wc, SPI_IO0, val & 1);
|
|
set_clk(wc, 0);
|
|
set_clk(wc, 1);
|
|
}
|
|
|
|
static void shift_out(struct t4 *wc, int val)
|
|
{
|
|
int i;
|
|
for (i = 7; i >= 0; i--)
|
|
clk_bit_out(wc, (val >> i) & 1);
|
|
}
|
|
|
|
static int clk_bit_in(struct t4 *wc)
|
|
{
|
|
int ret;
|
|
set_clk(wc, 0);
|
|
read_hwreg(wc);
|
|
ret = t8_get_bit(wc, SPI_IO1);
|
|
set_clk(wc, 1);
|
|
return ret;
|
|
}
|
|
|
|
static int shift_in(struct t4 *wc)
|
|
{
|
|
int ret = 0;
|
|
int i;
|
|
int bit;
|
|
|
|
for (i = 7; i >= 0; i--) {
|
|
bit = clk_bit_in(wc);
|
|
ret |= ((bit & 1) << i);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static void write_enable(struct t4 *wc)
|
|
{
|
|
int cmd = 0x06;
|
|
set_cs(wc, 0);
|
|
shift_out(wc, cmd);
|
|
set_cs(wc, 1);
|
|
}
|
|
|
|
static int read_sr1(struct t4 *wc)
|
|
{
|
|
int cmd = 0x05;
|
|
int ret;
|
|
set_cs(wc, 0);
|
|
shift_out(wc, cmd);
|
|
ret = shift_in(wc);
|
|
set_cs(wc, 1);
|
|
return ret;
|
|
}
|
|
|
|
static void clear_busy(struct t4 *wc)
|
|
{
|
|
static const int SR1_BUSY = (1 << 0);
|
|
unsigned long stop;
|
|
|
|
stop = jiffies + 2*HZ;
|
|
while (read_sr1(wc) & SR1_BUSY) {
|
|
if (time_after(jiffies, stop)) {
|
|
if (printk_ratelimit()) {
|
|
dev_err(&wc->dev->dev,
|
|
"Lockup in %s\n", __func__);
|
|
}
|
|
break;
|
|
}
|
|
cond_resched();
|
|
}
|
|
}
|
|
|
|
static void sector_erase(struct t4 *wc, uint32_t addr)
|
|
{
|
|
int cmd = 0x20;
|
|
write_enable(wc);
|
|
set_cs(wc, 0);
|
|
shift_out(wc, cmd);
|
|
shift_out(wc, (addr >> 16) & 0xff);
|
|
shift_out(wc, (addr >> 8) & 0xff);
|
|
shift_out(wc, (addr >> 0) & 0xff);
|
|
set_cs(wc, 1);
|
|
clear_busy(wc);
|
|
}
|
|
|
|
static void erase_half(struct t4 *wc)
|
|
{
|
|
uint32_t addr = 0x00080000;
|
|
uint32_t i;
|
|
|
|
dev_info(&wc->dev->dev, "Erasing octal firmware\n");
|
|
|
|
for (i = addr; i < (addr + 0x80000); i += 4096)
|
|
sector_erase(wc, i);
|
|
}
|
|
|
|
|
|
#define T8_FLASH_PAGE_SIZE 256UL
|
|
|
|
static void t8_update_firmware_page(struct t4 *wc, u32 address,
|
|
const u8 *page_data, size_t size)
|
|
{
|
|
int i;
|
|
|
|
write_enable(wc);
|
|
set_cs(wc, 0);
|
|
shift_out(wc, 0x02);
|
|
shift_out(wc, (address >> 16) & 0xff);
|
|
shift_out(wc, (address >> 8) & 0xff);
|
|
shift_out(wc, (address >> 0) & 0xff);
|
|
|
|
for (i = 0; i < size; ++i)
|
|
shift_out(wc, page_data[i]);
|
|
|
|
set_cs(wc, 1);
|
|
clear_busy(wc);
|
|
}
|
|
|
|
static int t8_update_firmware(struct t4 *wc, const struct firmware *fw,
|
|
const char *t8_firmware)
|
|
{
|
|
int res;
|
|
size_t offset = 0;
|
|
const u32 BASE_ADDRESS = 0x00080000;
|
|
const u8 *data, *end;
|
|
size_t size = 0;
|
|
|
|
/* Erase flash */
|
|
erase_half(wc);
|
|
|
|
dev_info(&wc->dev->dev,
|
|
"Uploading %s. This can take up to 30 seconds.\n", t8_firmware);
|
|
|
|
data = &fw->data[sizeof(struct t8_firm_header)];
|
|
end = &fw->data[fw->size];
|
|
|
|
while (data < end) {
|
|
/* Calculate the tail end of data that's shorter than a page */
|
|
size = min(T8_FLASH_PAGE_SIZE, (unsigned long)(end - data));
|
|
|
|
t8_update_firmware_page(wc, BASE_ADDRESS + offset,
|
|
data, size);
|
|
data += T8_FLASH_PAGE_SIZE;
|
|
offset += T8_FLASH_PAGE_SIZE;
|
|
|
|
cond_resched();
|
|
}
|
|
|
|
/* Reset te820 fpga after loading firmware */
|
|
dev_info(&wc->dev->dev, "Firmware load complete. Reseting device.\n");
|
|
res = pci_save_state(wc->dev);
|
|
if (res)
|
|
goto error_exit;
|
|
/* Set the fpga reset bits and clobber the remainder of the
|
|
* register, device will be reset anyway */
|
|
t4_pci_out(wc, WC_LEDS, 0xe0000000);
|
|
msleep(1000);
|
|
|
|
pci_restore_state(wc->dev);
|
|
|
|
/* Signal the driver to restart initialization.
|
|
* This will back out all initialization so far and
|
|
* restart the driver load process */
|
|
return -EAGAIN;
|
|
|
|
error_exit:
|
|
return res;
|
|
}
|
|
|
|
static char read_flash_byte(struct t4 *wc, unsigned int addr)
|
|
{
|
|
int cmd = 0x03;
|
|
char data;
|
|
|
|
set_cs(wc, 0);
|
|
|
|
shift_out(wc, cmd);
|
|
|
|
shift_out(wc, (addr >> 16) & 0xff);
|
|
shift_out(wc, (addr >> 8) & 0xff);
|
|
shift_out(wc, (addr >> 0) & 0xff);
|
|
|
|
data = shift_in(wc) & 0xff;
|
|
|
|
set_cs(wc, 1);
|
|
|
|
return data;
|
|
}
|
|
|
|
/**
|
|
* t8_read_serial - Returns the serial number of the board.
|
|
* @wc: The board whos serial number we are reading.
|
|
*
|
|
* The buffer returned is dynamically allocated and must be kfree'd by the
|
|
* caller. If memory could not be allocated, NULL is returned.
|
|
*
|
|
* Must be called in process context.
|
|
*
|
|
*/
|
|
static char *t8_read_serial(struct t4 *wc)
|
|
{
|
|
int i;
|
|
static const int MAX_SERIAL = 14;
|
|
static const u32 base_addr = 0x00080000-256;
|
|
unsigned char c;
|
|
unsigned char *serial = kzalloc(MAX_SERIAL + 1, GFP_KERNEL);
|
|
|
|
if (!serial)
|
|
return NULL;
|
|
|
|
for (i = 0; i < MAX_SERIAL; ++i) {
|
|
c = read_flash_byte(wc, base_addr+i);
|
|
if ((c >= 'a' && c <= 'z') ||
|
|
(c >= 'A' && c <= 'Z') ||
|
|
(c >= '0' && c <= '9'))
|
|
serial[i] = c;
|
|
else
|
|
break;
|
|
|
|
}
|
|
|
|
if (!i) {
|
|
kfree(serial);
|
|
serial = NULL;
|
|
}
|
|
|
|
return serial;
|
|
}
|
|
|
|
static void setup_spi(struct t4 *wc)
|
|
{
|
|
wc->st.rdreg = wc->st.wrreg = 0;
|
|
|
|
set_iodir(wc, SPI_IO0, SPI_OUT);
|
|
set_iodir(wc, SPI_IO1, SPI_IN);
|
|
set_iodir(wc, SPI_CS, SPI_OUT);
|
|
set_iodir(wc, SPI_CLK, SPI_OUT);
|
|
|
|
t8_set_bit(wc, SPI_CS, 1);
|
|
t8_set_bit(wc, SPI_CLK, 1);
|
|
|
|
write_hwreg(wc);
|
|
}
|
|
|
|
static int t8_check_firmware(struct t4 *wc, unsigned int version)
|
|
{
|
|
const struct firmware *fw;
|
|
static const char t8_firmware[] = "dahdi-fw-te820.bin";
|
|
const struct t8_firm_header *header;
|
|
int res = 0;
|
|
u32 crc;
|
|
|
|
res = request_firmware(&fw, t8_firmware, &wc->dev->dev);
|
|
if (res) {
|
|
dev_info(&wc->dev->dev, "firmware %s not "
|
|
"available from userspace\n", t8_firmware);
|
|
goto cleanup;
|
|
}
|
|
|
|
header = (const struct t8_firm_header *)fw->data;
|
|
|
|
/* Check the crc before anything else */
|
|
crc = crc32(~0, &fw->data[10], fw->size - 10) ^ ~0;
|
|
if (memcmp("DIGIUM", header->header, sizeof(header->header)) ||
|
|
(le32_to_cpu(header->chksum) != crc)) {
|
|
dev_info(&wc->dev->dev,
|
|
"%s is invalid. Please reinstall.\n", t8_firmware);
|
|
goto cleanup;
|
|
}
|
|
|
|
/* Spi struct must be setup before any access to flash memory */
|
|
setup_spi(wc);
|
|
|
|
/* Check the two firmware versions */
|
|
if (le32_to_cpu(header->version) == version)
|
|
goto cleanup;
|
|
|
|
dev_info(&wc->dev->dev, "%s Version: %08x available for flash\n",
|
|
t8_firmware, header->version);
|
|
|
|
res = t8_update_firmware(wc, fw, t8_firmware);
|
|
if (res && res != -EAGAIN) {
|
|
dev_info(&wc->dev->dev, "Failed to load firmware %s\n",
|
|
t8_firmware);
|
|
}
|
|
|
|
cleanup:
|
|
release_firmware(fw);
|
|
return res;
|
|
}
|
|
|
|
static int
|
|
__t4_hardware_init_1(struct t4 *wc, unsigned int cardflags, bool first_time)
|
|
{
|
|
unsigned int version;
|
|
int res;
|
|
|
|
version = t4_pci_in(wc, WC_VERSION);
|
|
if (is_octal(wc) && first_time) {
|
|
dev_info(&wc->dev->dev, "Firmware Version: %01x.%02x\n",
|
|
(version & 0xf00) >> 8,
|
|
version & 0xff);
|
|
} else if (first_time) {
|
|
dev_info(&wc->dev->dev, "Firmware Version: %08x\n", version);
|
|
}
|
|
if (debug) {
|
|
dev_info(&wc->dev->dev, "Burst Mode: %s\n",
|
|
(!(cardflags & FLAG_BURST) && noburst) ? "Off" : "On");
|
|
#ifdef ENABLE_WORKQUEUES
|
|
dev_info(&wc->dev->dev, "Work Queues: Enabled\n");
|
|
#endif
|
|
}
|
|
|
|
/* Check the field updatable firmware for the wcte820 */
|
|
if (is_octal(wc)) {
|
|
res = t8_check_firmware(wc, version);
|
|
if (res)
|
|
return res;
|
|
|
|
wc->ddev->hardware_id = t8_read_serial(wc);
|
|
}
|
|
|
|
#if defined(CONFIG_FORCE_EXTENDED_RESET)
|
|
t4_extended_reset(wc);
|
|
#elif !defined(CONFIG_NOEXTENDED_RESET)
|
|
if (wc->devtype->flags & FLAG_EXPRESS)
|
|
t4_extended_reset(wc);
|
|
#endif
|
|
|
|
/* Make sure DMA engine is not running and interrupts are acknowledged */
|
|
wc->dmactrl = 0x0;
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
/* Reset Framer and friends */
|
|
t4_pci_out(wc, WC_LEDS, 0x00000000);
|
|
|
|
/* Set DMA addresses */
|
|
t4_pci_out(wc, WC_RDADDR, wc->readdma);
|
|
t4_pci_out(wc, WC_WRADDR, wc->writedma);
|
|
|
|
/* Setup counters, interrupt flags (ignored in Gen2) */
|
|
if (cardflags & FLAG_2NDGEN) {
|
|
t4_tsi_reset(wc);
|
|
} else {
|
|
t4_pci_out(wc, WC_COUNT, ((DAHDI_MAX_CHUNKSIZE * 2 * 32 - 1) << 18) | ((DAHDI_MAX_CHUNKSIZE * 2 * 32 - 1) << 2));
|
|
}
|
|
|
|
/* Reset pending interrupts */
|
|
t4_pci_out(wc, WC_INTR, 0x00000000);
|
|
|
|
/* Read T1/E1 status */
|
|
if (first_time) {
|
|
if (!strcasecmp("auto", default_linemode)) {
|
|
if (-1 == t1e1override) {
|
|
wc->t1e1 = (((t4_pci_in(wc, WC_LEDS)) &
|
|
0x0f00) >> 8);
|
|
wc->t1e1 &= 0xf;
|
|
if (is_octal(wc)) {
|
|
wc->t1e1 |= ((t4_pci_in(wc, WC_LEDS2)) &
|
|
0x0f00) >> 4;
|
|
}
|
|
} else {
|
|
dev_warn(&wc->dev->dev,
|
|
"'t1e1override' is deprecated. Please use 'default_linemode'.\n");
|
|
wc->t1e1 = t1e1override & 0xf;
|
|
}
|
|
} else if (!strcasecmp("t1", default_linemode)) {
|
|
wc->t1e1 = 0;
|
|
} else if (!strcasecmp("e1", default_linemode)) {
|
|
wc->t1e1 = 0xff;
|
|
} else if (!strcasecmp("j1", default_linemode)) {
|
|
wc->t1e1 = 0;
|
|
j1mode = 1;
|
|
} else {
|
|
dev_err(&wc->dev->dev, "'%s' is an unknown linemode.\n",
|
|
default_linemode);
|
|
wc->t1e1 = 0;
|
|
}
|
|
}
|
|
|
|
wc->order = ((t4_pci_in(wc, WC_LEDS)) & 0xf0000000) >> 28;
|
|
order_index[wc->order]++;
|
|
|
|
/* TE820 Auth Check */
|
|
if (is_octal(wc)) {
|
|
unsigned long stop = jiffies + HZ;
|
|
uint32_t donebit;
|
|
|
|
t4_pci_out(wc, WC_LEDS2, WC_SET_AUTH);
|
|
donebit = t4_pci_in(wc, WC_LEDS2);
|
|
while (!(donebit & WC_GET_AUTH)) {
|
|
if (time_after(jiffies, stop)) {
|
|
/* Encryption check failed, stop operation */
|
|
dev_info(&wc->dev->dev,
|
|
"Failed encryption check. "
|
|
"Unloading driver.\n");
|
|
return -EIO;
|
|
}
|
|
msleep(20);
|
|
donebit = t4_pci_in(wc, WC_LEDS2);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int t4_hardware_init_1(struct t4 *wc, unsigned int cardflags)
|
|
{
|
|
return __t4_hardware_init_1(wc, cardflags, true);
|
|
}
|
|
|
|
static int __t4_hardware_init_2(struct t4 *wc, bool first_time)
|
|
{
|
|
int x;
|
|
unsigned int regval;
|
|
unsigned long flags;
|
|
|
|
if (t4_pci_in(wc, WC_VERSION) >= 0xc01a0165) {
|
|
wc->tspans[0]->spanflags |= FLAG_OCTOPT;
|
|
}
|
|
/* Setup LEDS, take out of reset */
|
|
t4_pci_out(wc, WC_LEDS, 0x000000ff);
|
|
udelay(100);
|
|
t4_activate(wc);
|
|
udelay(100);
|
|
|
|
/* In order to find out the QFALC framer version, we have to
|
|
* temporarily turn off compat mode and take a peak at VSTR. We turn
|
|
* compat back on when we are done.
|
|
*
|
|
*/
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
regval = __t4_framer_in(wc, 0, 0xd6);
|
|
if (is_octal(wc))
|
|
regval |= (1 << 4); /* SSI16 - For 16 MHz multiplex mode with comp = 1 */
|
|
else
|
|
regval |= (1 << 5); /* set COMP_DIS*/
|
|
|
|
__t4_framer_out(wc, 0, 0xd6, regval);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
if (!is_octal(wc)) {
|
|
regval = t4_framer_in(wc, 0, 0x4a);
|
|
if (first_time && regval == 0x05) {
|
|
dev_info(&wc->dev->dev, "FALC Framer Version: 2.1 or "
|
|
"earlier\n");
|
|
} else if (regval == 0x20) {
|
|
if (first_time)
|
|
dev_info(&wc->dev->dev, "FALC Framer Version: 3.1\n");
|
|
wc->falc31 = 1;
|
|
} else if (first_time) {
|
|
dev_info(&wc->dev->dev, "FALC Framer Version: Unknown "
|
|
"(VSTR = 0x%02x)\n", regval);
|
|
}
|
|
}
|
|
|
|
spin_lock_irqsave(&wc->reglock, flags);
|
|
regval = __t4_framer_in(wc, 0, 0xd6);
|
|
regval &= ~(1 << 5); /* clear COMP_DIS*/
|
|
__t4_framer_out(wc, 0, 0xd6, regval);
|
|
__t4_framer_out(wc, 0, 0x4a, 0xaa);
|
|
spin_unlock_irqrestore(&wc->reglock, flags);
|
|
|
|
if (debug) {
|
|
dev_info(&wc->dev->dev, "Board ID: %02x\n", wc->order);
|
|
for (x = 0; x < 11; x++) {
|
|
dev_info(&wc->dev->dev, "Reg %d: 0x%08x\n", x,
|
|
t4_pci_in(wc, x));
|
|
}
|
|
}
|
|
|
|
wc->gpio = 0x00000000;
|
|
t4_pci_out(wc, WC_GPIO, wc->gpio);
|
|
t4_gpio_setdir(wc, (1 << 17), (1 << 17));
|
|
t4_gpio_setdir(wc, (0xff), (0xff));
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int t4_hardware_init_2(struct t4 *wc)
|
|
{
|
|
return __t4_hardware_init_2(wc, true);
|
|
}
|
|
|
|
static int __devinit t4_launch(struct t4 *wc)
|
|
{
|
|
int x;
|
|
int res;
|
|
|
|
if (test_bit(DAHDI_FLAGBIT_REGISTERED, &wc->tspans[0]->span.flags))
|
|
return 0;
|
|
|
|
if (debug) {
|
|
dev_info(&wc->dev->dev,
|
|
"TE%dXXP: Launching card: %d\n", wc->numspans,
|
|
wc->order);
|
|
}
|
|
|
|
wc->ddev->manufacturer = "Digium";
|
|
if (!ignore_rotary && (1 == order_index[wc->order])) {
|
|
wc->ddev->location = kasprintf(GFP_KERNEL,
|
|
"Board ID Switch %d", wc->order);
|
|
} else {
|
|
bool express = ((wc->tspans[0]->spanflags & FLAG_EXPRESS) > 0);
|
|
wc->ddev->location = kasprintf(GFP_KERNEL,
|
|
"PCI%s Bus %02d Slot %02d",
|
|
(express) ? " Express" : "",
|
|
wc->dev->bus->number,
|
|
PCI_SLOT(wc->dev->devfn) + 1);
|
|
}
|
|
|
|
if (!wc->ddev->location)
|
|
return -ENOMEM;
|
|
|
|
for (x = 0; x < wc->numspans; ++x) {
|
|
list_add_tail(&wc->tspans[x]->span.device_node,
|
|
&wc->ddev->spans);
|
|
}
|
|
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 20)
|
|
INIT_WORK(&wc->bh_work, t4_work_func, wc);
|
|
#else
|
|
INIT_WORK(&wc->bh_work, t4_work_func);
|
|
#endif
|
|
|
|
res = dahdi_register_device(wc->ddev, &wc->dev->dev);
|
|
if (res) {
|
|
dev_err(&wc->dev->dev, "Failed to register with DAHDI.\n");
|
|
return res;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* wct4xxp_sort_cards - Sort the cards in card array by rotary switch settings.
|
|
*
|
|
*/
|
|
static void wct4xxp_sort_cards(void)
|
|
{
|
|
int x;
|
|
|
|
/* get the current number of probed cards and run a slice of a tail
|
|
* insertion sort */
|
|
for (x = 0; x < MAX_T4_CARDS; x++) {
|
|
if (!cards[x+1])
|
|
break;
|
|
}
|
|
for ( ; x > 0; x--) {
|
|
if (cards[x]->order < cards[x-1]->order) {
|
|
struct t4 *tmp = cards[x];
|
|
cards[x] = cards[x-1];
|
|
cards[x-1] = tmp;
|
|
} else {
|
|
/* if we're not moving it, we won't move any more
|
|
* since all cards are sorted on addition */
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
static int __devinit
|
|
t4_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|
{
|
|
int res;
|
|
struct t4 *wc;
|
|
unsigned int x;
|
|
int init_latency;
|
|
|
|
if (pci_enable_device(pdev)) {
|
|
return -EIO;
|
|
}
|
|
|
|
wc = kzalloc(sizeof(*wc), GFP_KERNEL);
|
|
if (!wc)
|
|
return -ENOMEM;
|
|
|
|
wc->ddev = dahdi_create_device();
|
|
if (!wc->ddev) {
|
|
kfree(wc);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
spin_lock_init(&wc->reglock);
|
|
wc->devtype = (const struct devtype *)(ent->driver_data);
|
|
|
|
#ifdef CONFIG_WCT4XXP_DISABLE_ASPM
|
|
if (is_pcie(wc)) {
|
|
pci_disable_link_state(pdev->bus->self, PCIE_LINK_STATE_L0S |
|
|
PCIE_LINK_STATE_L1 | PCIE_LINK_STATE_CLKPM);
|
|
};
|
|
#endif
|
|
|
|
if (is_octal(wc))
|
|
wc->numspans = 8;
|
|
else if (wc->devtype->flags & FLAG_2PORT)
|
|
wc->numspans = 2;
|
|
else
|
|
wc->numspans = 4;
|
|
|
|
wc->membase = pci_iomap(pdev, 0, 0);
|
|
/* This rids of the Double missed interrupt message after loading */
|
|
wc->last0 = 1;
|
|
if (pci_request_regions(pdev, wc->devtype->desc))
|
|
dev_info(&pdev->dev, "wct%dxxp: Unable to request regions\n",
|
|
wc->numspans);
|
|
|
|
if (debug)
|
|
dev_info(&pdev->dev, "Found TE%dXXP\n", wc->numspans);
|
|
|
|
wc->dev = pdev;
|
|
|
|
/* Enable bus mastering */
|
|
pci_set_master(pdev);
|
|
|
|
/* Keep track of which device we are */
|
|
pci_set_drvdata(pdev, wc);
|
|
|
|
if (wc->devtype->flags & FLAG_5THGEN) {
|
|
if ((ms_per_irq > 1) && (latency <= ((ms_per_irq) << 1))) {
|
|
init_latency = ms_per_irq << 1;
|
|
} else {
|
|
if (latency > 2)
|
|
init_latency = latency;
|
|
else
|
|
init_latency = 2;
|
|
}
|
|
dev_info(&wc->dev->dev, "5th gen card with initial latency of "
|
|
"%d and %d ms per IRQ\n", init_latency, ms_per_irq);
|
|
} else {
|
|
if (wc->devtype->flags & FLAG_2NDGEN)
|
|
init_latency = 1;
|
|
else
|
|
init_latency = 2;
|
|
}
|
|
|
|
if (max_latency < init_latency) {
|
|
printk(KERN_INFO "maxlatency must be set to something greater than %d ms, increasing it to %d\n", init_latency, init_latency);
|
|
max_latency = init_latency;
|
|
}
|
|
|
|
if (t4_allocate_buffers(wc, init_latency, NULL, NULL)) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Initialize hardware */
|
|
res = t4_hardware_init_1(wc, wc->devtype->flags);
|
|
if (res) {
|
|
/* If this function returns -EAGAIN, we expect
|
|
* to attempt another driver load. Clean everything
|
|
* up first */
|
|
pci_iounmap(wc->dev, wc->membase);
|
|
pci_release_regions(wc->dev);
|
|
pci_free_consistent(wc->dev, T4_BASE_SIZE(wc) * wc->numbufs * 2,
|
|
wc->writechunk, wc->writedma);
|
|
pci_set_drvdata(wc->dev, NULL);
|
|
free_wc(wc);
|
|
return res;
|
|
}
|
|
|
|
for(x = 0; x < MAX_T4_CARDS; x++) {
|
|
if (!cards[x])
|
|
break;
|
|
}
|
|
|
|
if (x >= MAX_T4_CARDS) {
|
|
dev_notice(&wc->dev->dev, "No cards[] slot available!!\n");
|
|
kfree(wc);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
wc->num = x;
|
|
cards[x] = wc;
|
|
|
|
#ifdef ENABLE_WORKQUEUES
|
|
if (wc->devtype->flags & FLAG_2NDGEN) {
|
|
char tmp[20];
|
|
|
|
sprintf(tmp, "te%dxxp[%d]", wc->numspans, wc->num);
|
|
wc->workq = create_workqueue(tmp);
|
|
}
|
|
#endif
|
|
|
|
/* Allocate pieces we need here */
|
|
for (x = 0; x < ports_on_framer(wc); x++) {
|
|
struct t4_span *ts;
|
|
enum linemode linemode;
|
|
|
|
ts = kzalloc(sizeof(*ts), GFP_KERNEL);
|
|
if (!ts) {
|
|
free_wc(wc);
|
|
return -ENOMEM;
|
|
}
|
|
wc->tspans[x] = ts;
|
|
|
|
#ifdef ENABLE_WORKQUEUES
|
|
INIT_WORK(&ts->swork, workq_handlespan, ts);
|
|
#endif
|
|
ts->spanflags |= wc->devtype->flags;
|
|
linemode = (wc->t1e1 & (1 << x)) ? E1 : ((j1mode) ? J1 : T1);
|
|
t4_alloc_channels(wc, wc->tspans[x], linemode);
|
|
}
|
|
|
|
/* Continue hardware intiialization */
|
|
t4_hardware_init_2(wc);
|
|
|
|
#ifdef SUPPORT_GEN1
|
|
if (request_irq(pdev->irq, (wc->devtype->flags & FLAG_2NDGEN) ?
|
|
t4_interrupt_gen2 : t4_interrupt,
|
|
IRQF_SHARED,
|
|
(wc->numspans == 8) ? "wct8xxp" :
|
|
(wc->numspans == 2) ? "wct2xxp" :
|
|
"wct4xxp",
|
|
wc)) {
|
|
#else
|
|
if (!(wc->tspans[0]->spanflags & FLAG_2NDGEN)) {
|
|
dev_notice(&wc->dev->dev, "This driver does not "
|
|
"support 1st gen modules\n");
|
|
free_wc(wc);
|
|
return -ENODEV;
|
|
}
|
|
|
|
if (request_irq(pdev->irq, t4_interrupt_gen2,
|
|
IRQF_SHARED, "t4xxp", wc)) {
|
|
#endif
|
|
dev_notice(&wc->dev->dev, "Unable to request IRQ %d\n",
|
|
pdev->irq);
|
|
free_wc(wc);
|
|
return -EIO;
|
|
}
|
|
|
|
init_spans(wc);
|
|
|
|
if (!ignore_rotary)
|
|
wct4xxp_sort_cards();
|
|
|
|
if (wc->ddev->hardware_id) {
|
|
dev_info(&wc->dev->dev,
|
|
"Found a Wildcard: %s (SN: %s)\n", wc->devtype->desc,
|
|
wc->ddev->hardware_id);
|
|
} else {
|
|
dev_info(&wc->dev->dev,
|
|
"Found a Wildcard: %s\n", wc->devtype->desc);
|
|
}
|
|
|
|
#ifdef VPM_SUPPORT
|
|
if (!wc->vpm) {
|
|
t4_vpm_init(wc);
|
|
wc->dmactrl |= (wc->vpm) ? T4_VPM_PRESENT : 0;
|
|
t4_pci_out(wc, WC_DMACTRL, wc->dmactrl);
|
|
if (wc->vpm)
|
|
set_span_devicetype(wc);
|
|
}
|
|
#endif
|
|
|
|
create_sysfs_files(wc);
|
|
|
|
res = 0;
|
|
if (ignore_rotary)
|
|
res = t4_launch(wc);
|
|
|
|
return res;
|
|
}
|
|
|
|
static int t4_hardware_stop(struct t4 *wc)
|
|
{
|
|
|
|
/* Turn off DMA, leave interrupts enabled */
|
|
set_bit(T4_STOP_DMA, &wc->checkflag);
|
|
|
|
/* Wait for interrupts to stop */
|
|
msleep(25);
|
|
|
|
/* Turn off counter, address, etc */
|
|
if (wc->tspans[0]->spanflags & FLAG_2NDGEN) {
|
|
t4_tsi_reset(wc);
|
|
} else {
|
|
t4_pci_out(wc, WC_COUNT, 0x000000);
|
|
}
|
|
t4_pci_out(wc, WC_RDADDR, 0x0000000);
|
|
t4_pci_out(wc, WC_WRADDR, 0x0000000);
|
|
wc->gpio = 0x00000000;
|
|
t4_pci_out(wc, WC_GPIO, wc->gpio);
|
|
t4_pci_out(wc, WC_LEDS, 0x00000000);
|
|
|
|
if (debug) {
|
|
dev_notice(&wc->dev->dev, "Stopped TE%dXXP, Turned off DMA\n",
|
|
wc->numspans);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int __devinit
|
|
t4_init_one_retry(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|
{
|
|
int res;
|
|
res = t4_init_one(pdev, ent);
|
|
|
|
/* If the driver was reset by a firmware load,
|
|
* try to load once again */
|
|
if (-EAGAIN == res) {
|
|
res = t4_init_one(pdev, ent);
|
|
if (-EAGAIN == res) {
|
|
dev_err(&pdev->dev, "Failed to update firmware.\n");
|
|
res = -EIO;
|
|
}
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
static void _t4_remove_one(struct t4 *wc)
|
|
{
|
|
int basesize;
|
|
|
|
if (!wc)
|
|
return;
|
|
|
|
dahdi_unregister_device(wc->ddev);
|
|
|
|
remove_sysfs_files(wc);
|
|
|
|
/* Stop hardware */
|
|
t4_hardware_stop(wc);
|
|
|
|
#ifdef VPM_SUPPORT
|
|
/* Release vpm */
|
|
if (wc->vpm)
|
|
release_vpm450m(wc->vpm);
|
|
wc->vpm = NULL;
|
|
#endif
|
|
/* Unregister spans */
|
|
|
|
basesize = DAHDI_MAX_CHUNKSIZE * 32 * 4;
|
|
if (!(wc->tspans[0]->spanflags & FLAG_2NDGEN))
|
|
basesize = basesize * 2;
|
|
|
|
#ifdef ENABLE_WORKQUEUES
|
|
if (wc->workq) {
|
|
flush_workqueue(wc->workq);
|
|
destroy_workqueue(wc->workq);
|
|
}
|
|
#endif
|
|
|
|
free_irq(wc->dev->irq, wc);
|
|
|
|
if (wc->membase)
|
|
pci_iounmap(wc->dev, wc->membase);
|
|
|
|
pci_release_regions(wc->dev);
|
|
|
|
/* Immediately free resources */
|
|
pci_free_consistent(wc->dev, T4_BASE_SIZE(wc) * wc->numbufs * 2,
|
|
wc->writechunk, wc->writedma);
|
|
|
|
order_index[wc->order]--;
|
|
|
|
cards[wc->num] = NULL;
|
|
pci_set_drvdata(wc->dev, NULL);
|
|
free_wc(wc);
|
|
}
|
|
|
|
static void __devexit t4_remove_one(struct pci_dev *pdev)
|
|
{
|
|
struct t4 *wc = pci_get_drvdata(pdev);
|
|
if (!wc)
|
|
return;
|
|
|
|
_t4_remove_one(wc);
|
|
}
|
|
|
|
|
|
static DEFINE_PCI_DEVICE_TABLE(t4_pci_tbl) =
|
|
{
|
|
{ 0x10ee, 0x0314, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct4xxp },
|
|
|
|
{ 0xd161, 0x1820, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct820p5 },
|
|
{ 0xd161, 0x1420, 0x0005, PCI_ANY_ID, 0, 0, (unsigned long)&wct420p5 },
|
|
{ 0xd161, 0x1410, 0x0005, PCI_ANY_ID, 0, 0, (unsigned long)&wct410p5 },
|
|
{ 0xd161, 0x1405, 0x0005, PCI_ANY_ID, 0, 0, (unsigned long)&wct405p5 },
|
|
{ 0xd161, 0x0420, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct420p4 },
|
|
{ 0xd161, 0x0410, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct410p4 },
|
|
{ 0xd161, 0x0405, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct405p4 },
|
|
{ 0xd161, 0x0410, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct410p3 },
|
|
{ 0xd161, 0x0405, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct405p3 },
|
|
{ 0xd161, 0x0410, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct410p2 },
|
|
{ 0xd161, 0x0405, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct405p2 },
|
|
|
|
{ 0xd161, 0x1220, 0x0005, PCI_ANY_ID, 0, 0, (unsigned long)&wct220p5 },
|
|
{ 0xd161, 0x1205, 0x0005, PCI_ANY_ID, 0, 0, (unsigned long)&wct205p5 },
|
|
{ 0xd161, 0x1210, 0x0005, PCI_ANY_ID, 0, 0, (unsigned long)&wct210p5 },
|
|
{ 0xd161, 0x0220, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct220p4 },
|
|
{ 0xd161, 0x0205, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct205p4 },
|
|
{ 0xd161, 0x0210, 0x0004, PCI_ANY_ID, 0, 0, (unsigned long)&wct210p4 },
|
|
{ 0xd161, 0x0205, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct205p3 },
|
|
{ 0xd161, 0x0210, 0x0003, PCI_ANY_ID, 0, 0, (unsigned long)&wct210p3 },
|
|
{ 0xd161, 0x0205, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct205 },
|
|
{ 0xd161, 0x0210, PCI_ANY_ID, PCI_ANY_ID, 0, 0, (unsigned long)&wct210 },
|
|
{ 0, }
|
|
};
|
|
|
|
static void _t4_shutdown(struct pci_dev *pdev)
|
|
{
|
|
struct t4 *wc = pci_get_drvdata(pdev);
|
|
t4_hardware_stop(wc);
|
|
}
|
|
|
|
static int t4_suspend(struct pci_dev *pdev, pm_message_t state)
|
|
{
|
|
return -ENOSYS;
|
|
}
|
|
|
|
static struct pci_driver t4_driver = {
|
|
.name = "wct4xxp",
|
|
.probe = t4_init_one_retry,
|
|
.remove = __devexit_p(t4_remove_one),
|
|
.shutdown = _t4_shutdown,
|
|
.suspend = t4_suspend,
|
|
.id_table = t4_pci_tbl,
|
|
};
|
|
|
|
static int __init t4_init(void)
|
|
{
|
|
int i;
|
|
int res;
|
|
|
|
if (-1 != t1e1override) {
|
|
pr_info("'t1e1override' module parameter is deprecated. "
|
|
"Please use 'default_linemode' instead.\n");
|
|
}
|
|
|
|
res = dahdi_pci_module(&t4_driver);
|
|
if (res)
|
|
return -ENODEV;
|
|
|
|
/* If we're ignoring the rotary switch settings, then we've already
|
|
* registered in the context of .probe */
|
|
if (!ignore_rotary) {
|
|
|
|
/* Initialize cards since we have all of them. Warn for
|
|
* missing zero and duplicate numbers. */
|
|
|
|
if (cards[0] && cards[0]->order != 0) {
|
|
printk(KERN_NOTICE "wct4xxp: Ident of first card is not zero (%d)\n",
|
|
cards[0]->order);
|
|
}
|
|
|
|
for (i = 0; cards[i]; i++) {
|
|
/* warn the user of duplicate ident values it is
|
|
* probably unintended */
|
|
if (debug && res < 15 && cards[i+1] &&
|
|
cards[res]->order == cards[i+1]->order) {
|
|
printk(KERN_NOTICE "wct4xxp: Duplicate ident "
|
|
"value found (%d)\n", cards[i]->order);
|
|
}
|
|
res = t4_launch(cards[i]);
|
|
if (res) {
|
|
int j;
|
|
for (j = 0; j < i; ++j)
|
|
_t4_remove_one(cards[j]);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
return res;
|
|
}
|
|
|
|
static void __exit t4_cleanup(void)
|
|
{
|
|
pci_unregister_driver(&t4_driver);
|
|
}
|
|
|
|
MODULE_AUTHOR("Digium Incorporated <support@digium.com>");
|
|
MODULE_DESCRIPTION("Wildcard Dual/Quad/Octal-port Digital Card Driver");
|
|
MODULE_ALIAS("wct2xxp");
|
|
MODULE_LICENSE("GPL v2");
|
|
|
|
module_param(debug, int, 0600);
|
|
module_param(noburst, int, 0600);
|
|
module_param(timingcable, int, 0600);
|
|
module_param(t1e1override, int, 0400);
|
|
module_param(default_linemode, charp, S_IRUGO);
|
|
MODULE_PARM_DESC(default_linemode, "\"auto\"(default), \"e1\", \"t1\", "
|
|
"or \"j1\". \"auto\" will use the value from any hardware "
|
|
"jumpers.");
|
|
module_param(alarmdebounce, int, 0600);
|
|
module_param(losalarmdebounce, int, 0600);
|
|
module_param(aisalarmdebounce, int, 0600);
|
|
module_param(yelalarmdebounce, int, 0600);
|
|
module_param(max_latency, int, 0600);
|
|
module_param(j1mode, int, 0600);
|
|
module_param(sigmode, int, 0600);
|
|
module_param(latency, int, 0600);
|
|
module_param(ms_per_irq, int, 0600);
|
|
module_param(ignore_rotary, int, 0400);
|
|
MODULE_PARM_DESC(ignore_rotary, "Set to > 0 to ignore the rotary switch when " \
|
|
"registering with DAHDI.");
|
|
|
|
#ifdef VPM_SUPPORT
|
|
module_param(vpmsupport, int, 0600);
|
|
module_param(vpmdtmfsupport, int, 0600);
|
|
#endif
|
|
|
|
MODULE_DEVICE_TABLE(pci, t4_pci_tbl);
|
|
|
|
module_init(t4_init);
|
|
module_exit(t4_cleanup);
|