dahdi-linux/drivers/dahdi/wctc4xxp/base.c

4262 lines
111 KiB
C
Raw Normal View History

/* Wildcard TC400B Driver
*
* Copyright (C) 2006-2012, Digium, Inc.
*
* All rights reserved.
*
*/
/*
* See http://www.asterisk.org for more information about
* the Asterisk project. Please do not directly contact
* any of the maintainers of this project for assistance;
* the project provides a web site, mailing lists and IRC
* channels for your use.
*
* This program is free software, distributed under the terms of
* the GNU General Public License Version 2 as published by the
* Free Software Foundation. See the LICENSE file included with
* this program for more details.
*/
#include <linux/kernel.h>
#include <linux/errno.h>
#include <linux/module.h>
#include <linux/init.h>
#include <linux/slab.h>
#include <linux/kmod.h>
#include <linux/sched.h>
#include <linux/pci.h>
#include <linux/interrupt.h>
#include <linux/delay.h>
#include <linux/jiffies.h>
#include <linux/moduleparam.h>
#include <linux/firmware.h>
#include <linux/if_ether.h>
#include <linux/ip.h>
#include <linux/udp.h>
#include <linux/etherdevice.h>
#include <linux/timer.h>
#include <stdbool.h>
#include <dahdi/kernel.h>
#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 11, 0)
#include <linux/sched/signal.h>
#endif /* 4.11.0 */
#include <linux/io.h>
/* COMPILE TIME OPTIONS =================================================== */
#define INTERRUPT 0
#define WORKQUEUE 1
#define TASKLET 2
/* Define if you want a debug attribute to test alert processing. */
#undef EXPORT_FOR_ALERT_ATTRIBUTE
#ifndef DEFERRED_PROCESSING
# define DEFERRED_PROCESSING WORKQUEUE
#endif
#if DEFERRED_PROCESSING == INTERRUPT
# define ALLOC_FLAGS GFP_ATOMIC
#elif DEFERRED_PROCESSING == TASKLET
# define ALLOC_FLAGS GFP_ATOMIC
#else
# define ALLOC_FLAGS GFP_KERNEL
#endif
#define WARN_ALWAYS() WARN_ON(1)
#define DTE_DEBUG(_dbgmask, _fmt, _args...) \
if ((debug & _dbgmask) == (_dbgmask)) { \
dev_info(&(wc)->pdev->dev, _fmt, ## _args); \
} \
/* define CONFIG_WCTC4XXP_POLLING to operate in a pure polling mode. This is
* was placed in as a debugging tool for a particluar system that wasn't
* routing the interrupt properly. Therefore it is off by default and the
* driver must be recompiled to enable it. */
#undef CONFIG_WCTC4XXP_POLLING
/* The total number of active channels over which the driver will start polling
* the card every 10 ms. */
#define POLLING_CALL_THRESHOLD 40
#define INVALID 999 /* Used to mark invalid channels, commands, etc.. */
#define MAX_CHANNEL_PACKETS 5
#define G729_LENGTH 20
#define G723_LENGTH 30
#define G729_SAMPLES 160 /* G.729 */
#define G723_SAMPLES 240 /* G.723.1 */
#define G729_BYTES 20 /* G.729 */
#define G723_6K_BYTES 24 /* G.723.1 at 6.3kb/s */
#define G723_5K_BYTES 20 /* G.723.1 at 5.3kb/s */
#define G723_SID_BYTES 4 /* G.723.1 SID frame */
#define MAX_CAPTURED_PACKETS 5000
/* The following bit fields are used to set the various debug levels. */
#define DTE_DEBUG_GENERAL (1 << 0) /* 1 */
#define DTE_DEBUG_CHANNEL_SETUP (1 << 1) /* 2 */
#define DTE_DEBUG_RTP_TX (1 << 2) /* 4 */
#define DTE_DEBUG_RTP_RX (1 << 3) /* 8 */
#define DTE_DEBUG_RX_TIMEOUT (1 << 4) /* 16 */
#define DTE_DEBUG_NETWORK_IF (1 << 5) /* 32 */
#define DTE_DEBUG_NETWORK_EARLY (1 << 6) /* 64 */
#define DTE_DEBUG_ETH_STATS (1 << 7) /* 128 */
static int debug;
static char *mode;
static spinlock_t wctc4xxp_list_lock;
static struct list_head wctc4xxp_list;
#define ETH_P_CSM_ENCAPS 0x889B
struct rtphdr {
#if defined(__LITTLE_ENDIAN_BITFIELD)
__u8 csrc_count:4;
__u8 extension:1;
__u8 padding:1;
__u8 ver:2;
__u8 type:7;
__u8 marker:1;
#elif defined(__BIG_ENDIAN_BITFIELD)
__u8 ver:2;
__u8 padding:1;
__u8 extension:1;
__u8 csrc_count:4;
__u8 marker:1;
__u8 type:7;
#else
#error "Please fix <asm/byteorder.h>"
#endif
__be16 seqno;
__be32 timestamp;
__be32 ssrc;
} __attribute__((packed));
struct rtp_packet {
struct ethhdr ethhdr;
struct iphdr iphdr;
struct udphdr udphdr;
struct rtphdr rtphdr;
__u8 payload[0];
} __attribute__((packed));
struct csm_encaps_cmd {
/* COMMON PART OF PAYLOAD HEADER */
__u8 length;
__u8 index;
__u8 type;
__u8 class;
__le16 function;
__le16 reserved;
__le16 params[0];
} __attribute__((packed));
/* Ethernet packet type for communication control information to the DTE. */
struct csm_encaps_hdr {
struct ethhdr ethhdr;
/* CSM_ENCAPS HEADER */
__be16 op_code;
__u8 seq_num;
__u8 control;
__be16 channel;
/* There is always at least one command. */
struct csm_encaps_cmd cmd;
} __attribute__((packed));
#define CONTROL_PACKET_OPCODE 0x0001
/* Control bits */
#define LITTLE_ENDIAN 0x01
#define SUPPRESS_ACK 0x40
#define MESSAGE_PACKET 0x80
#define SUPERVISOR_CHANNEL 0xffff
/* Supervisor function codes */
#define SUPVSR_CREATE_CHANNEL 0x0010
#define MONITOR_LIVE_INDICATION_TYPE 0x75
#define VOIP_VCEINFO_TYPE 0x0e
#define CONFIG_CHANGE_TYPE 0x00
#define CONFIG_CHANNEL_CLASS 0x02
#define CONFIG_DEVICE_CLASS 0x06
/* Individual channel config commands */
#define MAX_FRAME_SIZE 1518
#define SFRAME_SIZE MAX_FRAME_SIZE
#define DEFAULT_RX_DRING_SIZE (1 << 6) /* Must be a power of two */
/* Keep the TX ring shorter in order to reduce the amount of time needed to
* bring up channels when sending high priority csm_encaps packets. */
#define DEFAULT_TX_DRING_SIZE (1 << 4) /* Must be a power of two */
#define MIN_PACKET_LEN 64
/* Transcoder buffer (tcb) */
struct tcb {
void *data;
struct list_head node;
unsigned long timeout;
unsigned long retries;
#define TX_COMPLETE (1 << 1)
#define DO_NOT_CAPTURE (1 << 2)
#define WAIT_FOR_ACK (1 << 3)
#define WAIT_FOR_RESPONSE (1 << 4)
#define DTE_CMD_TIMEOUT (1 << 5)
u16 flags;
u16 next_index;
struct completion *complete;
struct tcb *response;
struct channel_pvt *cpvt;
/* The number of bytes available in data. */
int data_len;
};
static inline const struct csm_encaps_hdr *
response_header(struct tcb *cmd)
{
BUG_ON(!cmd->response);
return (const struct csm_encaps_hdr *)(cmd)->response->data;
}
static inline void
initialize_cmd(struct tcb *cmd, unsigned long cmd_flags)
{
INIT_LIST_HEAD(&cmd->node);
cmd->flags = cmd_flags;
}
/*! Used to allocate commands to submit to the dte. */
static struct kmem_cache *cmd_cache;
static inline struct tcb *
__alloc_cmd(size_t size, gfp_t alloc_flags, unsigned long cmd_flags)
{
struct tcb *cmd;
if (unlikely(size > SFRAME_SIZE))
return NULL;
if (size < MIN_PACKET_LEN)
size = MIN_PACKET_LEN;
cmd = kmem_cache_alloc(cmd_cache, alloc_flags);
if (likely(cmd)) {
memset(cmd, 0, sizeof(*cmd));
cmd->data = kzalloc(size, alloc_flags);
if (unlikely(!cmd->data)) {
kmem_cache_free(cmd_cache, cmd);
return NULL;
}
cmd->data_len = size;
initialize_cmd(cmd, cmd_flags);
}
return cmd;
}
static struct tcb *
alloc_cmd(size_t size)
{
return __alloc_cmd(size, GFP_KERNEL, 0);
}
static void
__free_cmd(struct tcb *cmd)
{
if (cmd)
kfree(cmd->data);
kmem_cache_free(cmd_cache, cmd);
return;
}
static void
free_cmd(struct tcb *cmd)
{
if (cmd->response)
__free_cmd(cmd->response);
__free_cmd(cmd);
}
struct channel_stats {
atomic_t packets_sent;
atomic_t packets_received;
};
struct channel_pvt {
spinlock_t lock; /* Lock for this structure */
struct wcdte *wc;
u16 seqno;
u8 cmd_seqno;
u8 ssrc;
u8 last_rx_seq_num;
u16 timeslot_in_num; /* DTE timeslot to receive from */
u16 timeslot_out_num; /* DTE timeslot to send data to */
u16 chan_in_num; /* DTE channel to receive from */
u16 chan_out_num; /* DTE channel to send data to */
u32 last_timestamp;
struct {
u8 encoder:1; /* If we're an encoder */
};
struct channel_stats stats;
struct list_head rx_queue; /* Transcoded packets for this channel. */
/* Used to prevent user space from flooding the firmware. */
long samples_in_flight;
unsigned long send_time;
};
struct wcdte {
char board_name[40];
const char *variety;
int pos;
struct list_head node;
spinlock_t reglock;
wait_queue_head_t waitq;
struct mutex chanlock;
#define DTE_READY 1
#define DTE_SHUTDOWN 2
#define DTE_POLLING 3
#define DTE_RELOAD 4
unsigned long flags;
/* This is a device-global list of commands that are waiting to be
* transmited (and did not fit on the transmit descriptor ring) */
spinlock_t cmd_list_lock;
struct list_head cmd_list;
struct list_head waiting_for_response_list;
spinlock_t rx_list_lock;
struct list_head rx_list;
spinlock_t rx_lock;
unsigned int seq_num;
int last_rx_seq_num;
unsigned char numchannels;
unsigned char complexname[40];
/* This section contains the members necessary to communicate with the
* physical interface to the transcoding engine. */
struct pci_dev *pdev;
unsigned int intmask;
void __iomem *iobase;
struct wctc4xxp_descriptor_ring *txd;
struct wctc4xxp_descriptor_ring *rxd;
struct dahdi_transcoder *uencode;
struct dahdi_transcoder *udecode;
struct channel_pvt *encoders;
struct channel_pvt *decoders;
#if DEFERRED_PROCESSING == WORKQUEUE
struct work_struct deferred_work;
#endif
/*
* This section contains the members necessary for exporting the
* network interface to the host system. This is only used for
* debugging purposes.
*
*/
struct sk_buff_head captured_packets;
struct net_device *netdev;
struct net_device_stats net_stats;
struct napi_struct napi;
struct timer_list watchdog;
u16 open_channels;
unsigned long reported_packet_errors;
};
struct wcdte_netdev_priv {
struct wcdte *wc;
};
static inline struct wcdte *
wcdte_from_netdev(struct net_device *netdev)
{
struct wcdte_netdev_priv *priv;
priv = netdev_priv(netdev);
return priv->wc;
}
static inline void wctc4xxp_set_ready(struct wcdte *wc)
{
set_bit(DTE_READY, &wc->flags);
}
static inline int wctc4xxp_is_ready(struct wcdte *wc)
{
return test_bit(DTE_READY, &wc->flags);
}
#define DTE_FORMAT_ULAW 0x00
#define DTE_FORMAT_G723_1 0x04
#define DTE_FORMAT_ALAW 0x08
#define DTE_FORMAT_G729A 0x12
#define DTE_FORMAT_UNDEF 0xFF
static inline u8 wctc4xxp_dahdifmt_to_dtefmt(unsigned int fmt)
{
u8 pt;
switch (fmt) {
case DAHDI_FORMAT_G723_1:
pt = DTE_FORMAT_G723_1;
break;
case DAHDI_FORMAT_ULAW:
pt = DTE_FORMAT_ULAW;
break;
case DAHDI_FORMAT_ALAW:
pt = DTE_FORMAT_ALAW;
break;
case DAHDI_FORMAT_G729A:
pt = DTE_FORMAT_G729A;
break;
default:
pt = DTE_FORMAT_UNDEF;
break;
}
return pt;
}
static struct sk_buff *
tcb_to_skb(struct net_device *netdev, const struct tcb *cmd)
{
struct sk_buff *skb;
skb = alloc_skb(cmd->data_len, in_atomic() ? GFP_ATOMIC : GFP_KERNEL);
if (skb) {
skb->dev = netdev;
skb_put(skb, cmd->data_len);
memcpy(skb->data, cmd->data, cmd->data_len);
skb->protocol = eth_type_trans(skb, netdev);
}
return skb;
}
/**
* wctc4xxp_skb_to_cmd - Convert a socket buffer (skb) to a tcb
* @wc: The transcoder that we're going to send this command to.
* @skb: socket buffer to convert.
*
*/
static struct tcb *
wctc4xxp_skb_to_cmd(struct wcdte *wc, const struct sk_buff *skb)
{
const gfp_t alloc_flags = in_interrupt() ? GFP_ATOMIC : GFP_KERNEL;
struct tcb *cmd;
cmd = __alloc_cmd(skb->len, alloc_flags, 0);
if (cmd) {
int res;
cmd->data_len = skb->len;
res = skb_copy_bits(skb, 0, cmd->data, cmd->data_len);
if (res) {
dev_warn(&wc->pdev->dev,
"Failed call to skb_copy_bits.\n");
free_cmd(cmd);
cmd = NULL;
}
}
return cmd;
}
#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 2, 0)
static void
wctc4xxp_net_set_multi(struct net_device *netdev)
{
struct wcdte *wc = wcdte_from_netdev(netdev);
DTE_DEBUG(DTE_DEBUG_GENERAL, "%s promiscuity:%d\n",
__func__, netdev->promiscuity);
}
#else
static void
wctc4xxp_set_rx_mode(struct net_device *netdev)
{
struct wcdte *wc = wcdte_from_netdev(netdev);
DTE_DEBUG(DTE_DEBUG_GENERAL, "%s promiscuity:%d\n",
__func__, netdev->promiscuity);
}
#endif
static int
wctc4xxp_net_up(struct net_device *netdev)
{
struct wcdte *wc = wcdte_from_netdev(netdev);
DTE_DEBUG(DTE_DEBUG_GENERAL, "%s\n", __func__);
napi_enable(&wc->napi);
return 0;
}
static int
wctc4xxp_net_down(struct net_device *netdev)
{
struct wcdte *wc = wcdte_from_netdev(netdev);
DTE_DEBUG(DTE_DEBUG_GENERAL, "%s\n", __func__);
napi_disable(&wc->napi);
return 0;
}
static void wctc4xxp_transmit_cmd(struct wcdte *, struct tcb *);
static int
wctc4xxp_net_hard_start_xmit(struct sk_buff *skb, struct net_device *netdev)
{
struct wcdte *wc = wcdte_from_netdev(netdev);
struct tcb *cmd;
/* We set DO_NOT_CAPTURE because this packet was already captured by
* in code higher up in the networking stack. We don't want to
* capture it twice.
*/
cmd = wctc4xxp_skb_to_cmd(wc, skb);
if (cmd) {
cmd->flags |= DO_NOT_CAPTURE;
wctc4xxp_transmit_cmd(wc, cmd);
}
dev_kfree_skb_any(skb);
return NETDEV_TX_OK;
}
static int
wctc4xxp_net_receive(struct wcdte *wc, int max)
{
int count = 0;
struct sk_buff *skb;
WARN_ON(0 == max);
while ((skb = skb_dequeue(&wc->captured_packets))) {
netif_receive_skb(skb);
if (++count >= max)
break;
}
return count;
}
static int
wctc4xxp_poll(struct napi_struct *napi, int budget)
{
struct wcdte *wc = container_of(napi, struct wcdte, napi);
int count;
count = wctc4xxp_net_receive(wc, budget);
if (!skb_queue_len(&wc->captured_packets)) {
#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 29)
netif_rx_complete(wc->netdev, &wc->napi);
#elif LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 30)
netif_rx_complete(&wc->napi);
#else
napi_complete(&wc->napi);
#endif
}
return count;
}
static struct net_device_stats *
wctc4xxp_net_get_stats(struct net_device *netdev)
{
struct wcdte *wc = wcdte_from_netdev(netdev);
return &wc->net_stats;
}
/* Wait until this device is put into promiscuous mode, or we timeout. */
static void
wctc4xxp_net_waitfor_promiscuous(struct wcdte *wc)
{
unsigned int seconds = 15;
unsigned long start = jiffies;
struct net_device *netdev = wc->netdev;
dev_info(&wc->pdev->dev,
"Waiting %d seconds for adapter to be placed in " \
"promiscuous mode for early trace.\n", seconds);
while (!netdev->promiscuity) {
if (signal_pending(current)) {
dev_info(&wc->pdev->dev,
"Aborting wait due to signal.\n");
break;
}
msleep(100);
if (time_after(jiffies, start + (seconds * HZ))) {
dev_info(&wc->pdev->dev,
"Aborting wait due to timeout.\n");
break;
}
}
}
#ifdef HAVE_NET_DEVICE_OPS
static const struct net_device_ops wctc4xxp_netdev_ops = {
#if LINUX_VERSION_CODE < KERNEL_VERSION(3, 2, 0)
.ndo_set_multicast_list = &wctc4xxp_net_set_multi,
#else
.ndo_set_rx_mode = &wctc4xxp_set_rx_mode,
#endif
.ndo_open = &wctc4xxp_net_up,
.ndo_stop = &wctc4xxp_net_down,
.ndo_start_xmit = &wctc4xxp_net_hard_start_xmit,
.ndo_get_stats = &wctc4xxp_net_get_stats,
};
#endif
/**
* wctc4xxp_net_register - Register a new network interface.
* @wc: transcoder card to register the interface for.
*
* The network interface is primarily used for debugging in order to watch the
* traffic between the transcoder and the host.
*
*/
static int
wctc4xxp_net_register(struct wcdte *wc)
{
int res;
struct net_device *netdev;
struct wcdte_netdev_priv *priv;
const char our_mac[] = { 0xaa, 0xbb, 0xcc, 0xdd, 0xee, 0xff};
#if LINUX_VERSION_CODE >= KERNEL_VERSION(3, 17, 0)
netdev = alloc_netdev(sizeof(*priv), wc->board_name,
NET_NAME_UNKNOWN, ether_setup);
#else
netdev = alloc_netdev(sizeof(*priv), wc->board_name, ether_setup);
#endif
if (!netdev)
return -ENOMEM;
priv = netdev_priv(netdev);
priv->wc = wc;
memcpy(netdev->dev_addr, our_mac, sizeof(our_mac));
# ifdef HAVE_NET_DEVICE_OPS
netdev->netdev_ops = &wctc4xxp_netdev_ops;
# else
netdev->set_multicast_list = &wctc4xxp_net_set_multi;
netdev->open = &wctc4xxp_net_up;
netdev->stop = &wctc4xxp_net_down;
netdev->hard_start_xmit = &wctc4xxp_net_hard_start_xmit;
netdev->get_stats = &wctc4xxp_net_get_stats;
# endif
netdev->promiscuity = 0;
netdev->flags |= IFF_NOARP;
netif_napi_add(netdev, &wc->napi, &wctc4xxp_poll, 64);
res = register_netdev(netdev);
if (res) {
dev_warn(&wc->pdev->dev,
"Failed to register network device %s.\n",
wc->board_name);
goto error_sw;
}
wc->netdev = netdev;
skb_queue_head_init(&wc->captured_packets);
if (debug & DTE_DEBUG_NETWORK_EARLY)
wctc4xxp_net_waitfor_promiscuous(wc);
dev_info(&wc->pdev->dev,
"Created network device %s for debug.\n", wc->board_name);
return 0;
error_sw:
if (netdev)
free_netdev(netdev);
return res;
}
static void
wctc4xxp_net_unregister(struct wcdte *wc)
{
struct sk_buff *skb;
if (!wc->netdev)
return;
unregister_netdev(wc->netdev);
while ((skb = skb_dequeue(&wc->captured_packets)))
kfree_skb(skb);
free_netdev(wc->netdev);
wc->netdev = NULL;
}
/**
* wctc4xxp_net_capture_cmd - Send a tcb to the network stack.
* @wc: transcoder that received the command.
* @cmd: command to send to network stack.
*
*/
static void
wctc4xxp_net_capture_cmd(struct wcdte *wc, const struct tcb *cmd)
{
struct sk_buff *skb;
struct net_device *netdev = wc->netdev;
if (!netdev)
return;
/* No need to capture if there isn't anyone listening. */
if (!(netdev->flags & IFF_UP))
return;
if (skb_queue_len(&wc->captured_packets) > MAX_CAPTURED_PACKETS) {
WARN_ON_ONCE(1);
return;
}
skb = tcb_to_skb(netdev, cmd);
if (!skb)
return;
skb_queue_tail(&wc->captured_packets, skb);
# if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 29)
netif_rx_schedule(netdev, &wc->napi);
# elif LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 30)
netif_rx_schedule(&wc->napi);
# else
napi_schedule(&wc->napi);
# endif
return;
}
/*! In-memory structure shared by the host and the adapter. */
struct wctc4xxp_descriptor {
__le32 des0;
__le32 des1;
__le32 buffer1;
__le32 container; /* Unused */
} __attribute__((packed));
struct wctc4xxp_descriptor_ring {
/* Pointer to an array of descriptors to give to hardware. */
struct wctc4xxp_descriptor *desc;
/* Read completed buffers from the head. */
unsigned int head;
/* Write ready buffers to the tail. */
unsigned int tail;
/* Array to save the kernel virtual address of pending commands. */
struct tcb **pending;
/* PCI Bus address of the descriptor list. */
dma_addr_t desc_dma;
/*! either DMA_FROM_DEVICE or DMA_TO_DEVICE */
unsigned int direction;
/*! The number of buffers currently submitted to the hardware. */
unsigned int count;
/*! The number of bytes to pad each descriptor for cache alignment. */
unsigned int padding;
/*! Protects this structure from concurrent access. */
spinlock_t lock;
/*! PCI device for the card associated with this ring. */
struct pci_dev *pdev;
/*! The size of the dring. */
unsigned long size;
/*! Total number of packets completed. */
unsigned long packet_count;
/*! Total number of packets with errors. */
unsigned long packet_errors;
};
/**
* wctc4xxp_descriptor - Returns the desriptor at index.
* @dr: The descriptor ring we're using.
* @index: index of the descriptor we want.
*
* We need this function because we do not know what the padding on the
* descriptors will be. Otherwise, we would just use an array.
*/
static inline struct wctc4xxp_descriptor *
wctc4xxp_descriptor(struct wctc4xxp_descriptor_ring *dr, int index)
{
return (struct wctc4xxp_descriptor *)((u8 *)dr->desc +
((sizeof(*dr->desc) + dr->padding) * index));
}
static int
wctc4xxp_initialize_descriptor_ring(struct pci_dev *pdev,
struct wctc4xxp_descriptor_ring *dr, u32 des1, unsigned int direction,
unsigned long size)
{
int i;
const u32 END_OF_RING = 0x02000000;
u8 cache_line_size = 0;
int add_padding;
struct wctc4xxp_descriptor *d = NULL;
BUG_ON(!pdev);
BUG_ON(!dr);
if (pci_read_config_byte(pdev, 0x0c, &cache_line_size))
return -EIO;
memset(dr, 0, sizeof(*dr));
dr->size = size;
/*
* Add some padding to each descriptor to ensure that they are
* aligned on host system cache-line boundaries, but only for the
* cache-line sizes that we support.
*
*/
add_padding = (0x08 == cache_line_size) ||
(0x10 == cache_line_size) ||
(0x20 == cache_line_size);
if (add_padding)
dr->padding = (cache_line_size*sizeof(u32)) - sizeof(*d);
dr->pending = kmalloc(sizeof(struct tcb *) * dr->size, GFP_KERNEL);
if (!dr->pending)
return -ENOMEM;
dr->desc = pci_alloc_consistent(pdev,
(sizeof(*d)+dr->padding)*dr->size, &dr->desc_dma);
if (!dr->desc) {
kfree(dr->pending);
return -ENOMEM;
}
memset(dr->desc, 0, (sizeof(*d) + dr->padding) * dr->size);
for (i = 0; i < dr->size; ++i) {
d = wctc4xxp_descriptor(dr, i);
memset(d, 0, sizeof(*d));
d->des1 = cpu_to_le32(des1);
}
d->des1 |= cpu_to_le32(END_OF_RING);
dr->direction = direction;
spin_lock_init(&dr->lock);
dr->pdev = pdev;
return 0;
}
#define OWN_BIT cpu_to_le32(0x80000000)
#define OWNED(_d_) (((_d_)->des0)&OWN_BIT)
#define SET_OWNED(_d_) do { wmb(); (_d_)->des0 |= OWN_BIT; wmb(); } while (0)
static const unsigned int BUFFER1_SIZE_MASK = 0x7ff;
static int
wctc4xxp_submit(struct wctc4xxp_descriptor_ring *dr, struct tcb *c)
{
volatile struct wctc4xxp_descriptor *d;
unsigned int len;
unsigned long flags;
WARN_ON(!c);
len = (c->data_len < MIN_PACKET_LEN) ? MIN_PACKET_LEN : c->data_len;
if (c->data_len > MAX_FRAME_SIZE) {
WARN_ON_ONCE(!"Invalid command length passed\n");
c->data_len = MAX_FRAME_SIZE;
}
spin_lock_irqsave(&dr->lock, flags);
d = wctc4xxp_descriptor(dr, dr->tail);
WARN_ON(!d);
if (d->buffer1) {
spin_unlock_irqrestore(&dr->lock, flags);
/* Do not overwrite a buffer that is still in progress. */
return -EBUSY;
}
d->des1 &= cpu_to_le32(~(BUFFER1_SIZE_MASK));
d->des1 |= cpu_to_le32(len & BUFFER1_SIZE_MASK);
d->buffer1 = cpu_to_le32(pci_map_single(dr->pdev, c->data,
SFRAME_SIZE, dr->direction));
SET_OWNED(d); /* That's it until the hardware is done with it. */
dr->pending[dr->tail] = c;
dr->tail = (dr->tail + 1) & (dr->size-1);
++dr->count;
spin_unlock_irqrestore(&dr->lock, flags);
return 0;
}
static inline struct tcb*
wctc4xxp_retrieve(struct wctc4xxp_descriptor_ring *dr)
{
volatile struct wctc4xxp_descriptor *d;
struct tcb *c;
unsigned int head = dr->head;
unsigned long flags;
u32 des0;
spin_lock_irqsave(&dr->lock, flags);
d = wctc4xxp_descriptor(dr, head);
if (d->buffer1 && !OWNED(d)) {
pci_unmap_single(dr->pdev, le32_to_cpu(d->buffer1),
SFRAME_SIZE, dr->direction);
c = dr->pending[head];
WARN_ON(!c);
dr->head = (++head) & (dr->size-1);
d->buffer1 = 0;
--dr->count;
WARN_ON(!c);
des0 = le32_to_cpu(d->des0);
c->data_len = (des0 >> 16) & BUFFER1_SIZE_MASK;
if (des0 & (1<<15)) {
++dr->packet_errors;
/* The upper layers won't be able to do anything with
* this packet. Free it up and log the error. */
free_cmd(c);
c = NULL;
} else {
++dr->packet_count;
WARN_ON(c->data_len > SFRAME_SIZE);
}
} else {
c = NULL;
}
spin_unlock_irqrestore(&dr->lock, flags);
return c;
}
static inline int wctc4xxp_getcount(struct wctc4xxp_descriptor_ring *dr)
{
int count;
unsigned long flags;
spin_lock_irqsave(&dr->lock, flags);
count = dr->count;
spin_unlock_irqrestore(&dr->lock, flags);
return count;
}
static inline int wctc4xxp_get_packet_count(struct wctc4xxp_descriptor_ring *dr)
{
unsigned long count;
unsigned long flags;
spin_lock_irqsave(&dr->lock, flags);
count = dr->packet_count;
spin_unlock_irqrestore(&dr->lock, flags);
return count;
}
static inline int
wctc4xxp_get_packet_errors(struct wctc4xxp_descriptor_ring *dr)
{
unsigned long count;
unsigned long flags;
spin_lock_irqsave(&dr->lock, flags);
count = dr->packet_errors;
spin_unlock_irqrestore(&dr->lock, flags);
return count;
}
static inline void
wctc4xxp_set_packet_count(struct wctc4xxp_descriptor_ring *dr,
unsigned long count)
{
unsigned long flags;
spin_lock_irqsave(&dr->lock, flags);
dr->packet_count = count;
spin_unlock_irqrestore(&dr->lock, flags);
}
static inline void
__wctc4xxp_setctl(struct wcdte *wc, unsigned int addr, unsigned int val)
{
writel(val, wc->iobase + addr);
readl(wc->iobase + addr);
}
static inline unsigned int
__wctc4xxp_getctl(struct wcdte *wc, unsigned int addr)
{
return readl(wc->iobase + addr);
}
static inline void
wctc4xxp_setctl(struct wcdte *wc, unsigned int addr, unsigned int val)
{
unsigned long flags;
spin_lock_irqsave(&wc->reglock, flags);
__wctc4xxp_setctl(wc, addr, val);
spin_unlock_irqrestore(&wc->reglock, flags);
}
static inline void
wctc4xxp_receive_demand_poll(struct wcdte *wc)
{
__wctc4xxp_setctl(wc, 0x0010, 0x00000000);
}
static inline void
wctc4xxp_transmit_demand_poll(struct wcdte *wc)
{
return;
# if 0
__wctc4xxp_setctl(wc, 0x0008, 0x00000000);
/* \todo Investigate why this register needs to be written twice in
* order to get it to poll reliably. So far, most of the problems
* I've seen with timeouts had more to do with an untransmitted
* packet sitting in the outbound descriptor list as opposed to any
* problem with the dte firmware.
*/
__wctc4xxp_setctl(wc, 0x0008, 0x00000000);
#endif
}
#define LENGTH_WITH_N_PARAMETERS(__n) (((__n) * sizeof(u16)) + \
sizeof(struct csm_encaps_cmd))
static const u8 dst_mac[6] = {0x00, 0x11, 0x22, 0x33, 0x44, 0x55};
static const u8 src_mac[6] = {0xAA, 0xBB, 0xCC, 0xDD, 0xEE, 0xFF};
static int wctc4xxp_transmit_cmd_and_wait(struct wcdte *wc, struct tcb *cmd);
static void
setup_common_header(struct wcdte *wc, struct csm_encaps_hdr *hdr)
{
memcpy(hdr->ethhdr.h_dest, dst_mac, sizeof(dst_mac));
memcpy(hdr->ethhdr.h_source, src_mac, sizeof(src_mac));
hdr->ethhdr.h_proto = cpu_to_be16(ETH_P_CSM_ENCAPS);
}
static void
setup_supervisor_header(struct wcdte *wc, struct csm_encaps_hdr *hdr)
{
setup_common_header(wc, hdr);
hdr->op_code = cpu_to_be16(CONTROL_PACKET_OPCODE);
hdr->control = LITTLE_ENDIAN;
hdr->seq_num = (wc->seq_num++)&0xf;
hdr->channel = cpu_to_be16(SUPERVISOR_CHANNEL);
}
static void
create_supervisor_cmd(struct wcdte *wc, struct tcb *cmd, u8 type, u8 class,
u16 function, const u16 *parameters, const int num_parameters)
{
struct csm_encaps_hdr *hdr = cmd->data;
int i;
if (cmd->response) {
free_cmd(cmd->response);
cmd->response = NULL;
}
setup_supervisor_header(wc, hdr);
hdr->cmd.length = LENGTH_WITH_N_PARAMETERS(num_parameters);
hdr->cmd.index = 0;
hdr->cmd.type = type;
hdr->cmd.class = class;
hdr->cmd.function = cpu_to_le16(function);
hdr->cmd.reserved = 0;
for (i = 0; i < num_parameters; ++i)
hdr->cmd.params[i] = cpu_to_le16(parameters[i]);
cmd->flags = WAIT_FOR_RESPONSE;
cmd->data_len = sizeof(struct csm_encaps_hdr) -
sizeof(struct csm_encaps_cmd) +
hdr->cmd.length;
cmd->cpvt = NULL;
}
static void
setup_channel_header(struct channel_pvt *pvt, struct tcb *cmd)
{
struct csm_encaps_hdr *hdr = cmd->data;
if (cmd->response) {
free_cmd(cmd->response);
cmd->response = NULL;
}
setup_common_header(pvt->wc, hdr);
hdr->op_code = cpu_to_be16(CONTROL_PACKET_OPCODE);
hdr->seq_num = (pvt->cmd_seqno++)&0xf;
hdr->channel = cpu_to_be16(pvt->chan_in_num);
cmd->flags = WAIT_FOR_RESPONSE;
cmd->data_len = sizeof(struct csm_encaps_hdr) -
sizeof(struct csm_encaps_cmd);
cmd->cpvt = pvt;
cmd->next_index = 0;
}
static void
append_channel_cmd(struct tcb *cmd, u8 type, u8 class, u16 function,
const u16 *parameters, int num_parameters)
{
int i;
struct csm_encaps_cmd *csm_cmd = cmd->data + cmd->data_len;
csm_cmd->length = LENGTH_WITH_N_PARAMETERS(num_parameters);
csm_cmd->index = cmd->next_index++;
csm_cmd->type = type;
csm_cmd->class = class;
csm_cmd->function = cpu_to_le16(function);
csm_cmd->reserved = 0;
for (i = 0; i < num_parameters; ++i)
csm_cmd->params[i] = cpu_to_le16(parameters[i]);
cmd->data_len += csm_cmd->length;
/* Pad it out to a DW boundary */
if (cmd->data_len % 4)
cmd->data_len += 4 - (cmd->data_len % 4);
WARN_ON(cmd->data_len >= SFRAME_SIZE);
}
static void
create_channel_cmd(struct channel_pvt *pvt, struct tcb *cmd, u8 type, u8 class,
u16 function, const u16 *parameters, int num_parameters)
{
setup_channel_header(pvt, cmd);
append_channel_cmd(cmd, type, class, function, parameters,
num_parameters);
}
static int
send_create_channel_cmd(struct wcdte *wc, struct tcb *cmd, u16 timeslot,
u16 *channel_number)
{
int res;
const u16 parameters[] = {0x0002, timeslot};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, SUPVSR_CREATE_CHANNEL,
parameters, ARRAY_SIZE(parameters));
res = wctc4xxp_transmit_cmd_and_wait(wc, cmd);
if (res)
return res;
if (0x0000 != response_header(cmd)->cmd.params[0]) {
if (printk_ratelimit()) {
dev_warn(&wc->pdev->dev,
"Failed to create channel in timeslot " \
"%d. Response from DTE was (%04x).\n",
timeslot, response_header(cmd)->cmd.params[0]);
}
free_cmd(cmd->response);
cmd->response = NULL;
return -EIO;
}
*channel_number = le16_to_cpu(response_header(cmd)->cmd.params[1]);
free_cmd(cmd->response);
cmd->response = NULL;
return 0;
}
static int
send_set_arm_clk_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x012c, 0x0000};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0411, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_set_spu_clk_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x012c, 0x0000};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0412, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_tdm_select_bus_mode_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x0004};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0417, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_set_eth_header_cmd(struct wcdte *wc, struct tcb *cmd,
const u8 *host_mac, const u8 *assigned_mac)
{
u16 parameters[8];
u16 *part;
parameters[0] = 0x0001;
part = (u16 *)host_mac;
parameters[1] = part[0];
parameters[2] = part[1];
parameters[3] = part[2];
part = (u16 *)assigned_mac;
parameters[4] = part[0];
parameters[5] = part[1];
parameters[6] = part[2];
parameters[7] = 0x0008;
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0100, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_supvsr_setup_tdm_parms(struct wcdte *wc, struct tcb *cmd,
u8 bus_number)
{
const u16 parameters[] = {0x8380, 0x0c00, 0, (bus_number << 2)&0xc};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0407, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_ip_service_config_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x0200};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0302, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_arp_service_config_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x0001};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0105, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_icmp_service_config_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0xff01};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0304, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_device_set_country_code_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x0000};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x041b, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_spu_features_control_cmd(struct wcdte *wc, struct tcb *cmd, u16 options)
{
const u16 parameters[] = {options};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0013, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
/* Allows sending more than one CSM_ENCAPS packet in a single ethernet frame. */
static int send_csme_multi_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x1};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x010a, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_tdm_opt_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x0000};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0435, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
send_destroy_channel_cmd(struct wcdte *wc, struct tcb *cmd, u16 channel)
{
int res;
u16 result;
const u16 parameters[] = {channel};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0011, parameters,
ARRAY_SIZE(parameters));
res = wctc4xxp_transmit_cmd_and_wait(wc, cmd);
if (res)
return res;
/* Let's check the response for any error codes.... */
result = le16_to_cpu(response_header(cmd)->cmd.params[0]);
if (0x0000 != result) {
dev_err(&wc->pdev->dev,
"Failed to destroy channel %04d (%04x)\n",
channel, result);
return -EIO;
}
return 0;
}
static void
append_set_ip_hdr_channel_cmd(struct tcb *cmd)
{
const u16 parameters[] = {0, 0x0045, 0, 0, 0x0040, 0x1180, 0,
0xa8c0, 0x0309, 0xa8c0, 0x0309,
swab16(cmd->cpvt->timeslot_out_num + 0x5000),
swab16(cmd->cpvt->timeslot_in_num + 0x5000),
0, 0};
append_channel_cmd(cmd, CONFIG_CHANGE_TYPE, CONFIG_CHANNEL_CLASS,
0x9000, parameters, ARRAY_SIZE(parameters));
}
static void
append_voip_vceopt_cmd(struct tcb *cmd, u16 length)
{
const u16 parameters[] = {((length << 8)|0x21), 0x1c00,
0x0004, 0, 0};
append_channel_cmd(cmd, CONFIG_CHANGE_TYPE, CONFIG_CHANNEL_CLASS,
0x8001, parameters, ARRAY_SIZE(parameters));
}
static void
append_voip_tonectl_cmd(struct tcb *cmd)
{
const u16 parameters[] = {0};
append_channel_cmd(cmd, CONFIG_CHANGE_TYPE, CONFIG_CHANNEL_CLASS,
0x805b, parameters, ARRAY_SIZE(parameters));
}
static void
append_voip_dtmfopt_cmd(struct tcb *cmd)
{
const u16 parameters[] = {0x0008};
append_channel_cmd(cmd, CONFIG_CHANGE_TYPE, CONFIG_CHANNEL_CLASS,
0x8002, parameters, ARRAY_SIZE(parameters));
}
static void
append_voip_indctrl_cmd(struct tcb *cmd)
{
const u16 parameters[] = {0x0007};
append_channel_cmd(cmd, CONFIG_CHANGE_TYPE, CONFIG_CHANNEL_CLASS,
0x8084, parameters, ARRAY_SIZE(parameters));
}
static void
send_voip_vopena_cmd(struct channel_pvt *pvt, struct tcb *cmd, u8 format)
{
const u16 parameters[] = {1, ((format<<8)|0x80), 0, 0, 0,
0x3412, 0x7856};
create_channel_cmd(pvt, cmd, CONFIG_CHANGE_TYPE, CONFIG_CHANNEL_CLASS,
0x8000, parameters, ARRAY_SIZE(parameters));
wctc4xxp_transmit_cmd(pvt->wc, cmd);
}
static int
send_voip_vopena_close_cmd(struct channel_pvt *pvt, struct tcb *cmd)
{
int res;
const u16 parameters[] = {0};
create_channel_cmd(pvt, cmd, CONFIG_CHANGE_TYPE, CONFIG_CHANNEL_CLASS,
0x8000, parameters, ARRAY_SIZE(parameters));
res = wctc4xxp_transmit_cmd_and_wait(pvt->wc, cmd);
if (res)
return res;
/* Let's check the response for any error codes.... */
if (0x0000 != response_header(cmd)->cmd.params[0]) {
WARN_ON(1);
return -EIO;
}
return 0;
}
static int
send_ip_options_cmd(struct wcdte *wc, struct tcb *cmd)
{
const u16 parameters[] = {0x0002};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0306, parameters,
ARRAY_SIZE(parameters));
return wctc4xxp_transmit_cmd_and_wait(wc, cmd);
}
static int
_send_trans_connect_cmd(struct wcdte *wc, struct tcb *cmd, u16 enable, u16
encoder_channel, u16 decoder_channel, u16 encoder_format,
u16 decoder_format)
{
int res;
const u16 parameters[] = {enable, encoder_channel, encoder_format,
decoder_channel, decoder_format};
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x9322, parameters,
ARRAY_SIZE(parameters));
res = wctc4xxp_transmit_cmd_and_wait(wc, cmd);
if (res)
return res;
/* Let's check the response for any error codes.... */
if (0x0000 != response_header(cmd)->cmd.params[0]) {
WARN_ON(1);
return -EIO;
}
return 0;
}
static int
send_trans_connect_cmd(struct wcdte *wc, struct tcb *cmd, const u16
encoder_channel, const u16 decoder_channel, const u16 encoder_format,
const u16 decoder_format)
{
return _send_trans_connect_cmd(wc, cmd, 1, encoder_channel,
decoder_channel, encoder_format, decoder_format);
}
static int
send_trans_disconnect_cmd(struct wcdte *wc, struct tcb *cmd, const u16
encoder_channel, const u16 decoder_channel, const u16 encoder_format,
const u16 decoder_format)
{
return _send_trans_connect_cmd(wc, cmd, 0, encoder_channel,
decoder_channel, encoder_format, decoder_format);
}
static int
send_voip_vceinfo_cmd(struct channel_pvt *pvt, struct tcb *cmd)
{
int res;
const u16 parameters[] = {0};
static const int CONFIG_CHANNEL_STATS_CLASS = 1;
create_channel_cmd(pvt, cmd,
VOIP_VCEINFO_TYPE, CONFIG_CHANNEL_STATS_CLASS,
0x0000, parameters, 0);
res = wctc4xxp_transmit_cmd_and_wait(pvt->wc, cmd);
return res;
}
static int
send_eth_statistics_cmd(struct wcdte *wc, struct tcb *cmd)
{
int res;
const u16 parameters[] = {0};
create_supervisor_cmd(wc, cmd, 0x00, 0x05, 0x0000,
parameters, ARRAY_SIZE(parameters));
res = wctc4xxp_transmit_cmd_and_wait(wc, cmd);
if (res)
return -EIO;
if (0x0000 != response_header(cmd)->cmd.params[0]) {
dev_info(&wc->pdev->dev,
"Failed to get ethernet stats: 0x%04x\n",
response_header(cmd)->cmd.params[0]);
res = -EIO;
}
return res;
}
static void wctc4xxp_match_packet_counts(struct wcdte *wc)
{
struct tcb *cmd = alloc_cmd(SFRAME_SIZE);
int res;
u32 *parms;
res = send_eth_statistics_cmd(wc, cmd);
if (0 == res) {
parms = (u32 *)(&response_header(cmd)->cmd.params[0]);
wctc4xxp_set_packet_count(wc->rxd, parms[1]);
wctc4xxp_set_packet_count(wc->txd, parms[2]);
}
free_cmd(cmd);
}
static inline u32 wctc4xxp_bytes_to_samples(u32 fmt, size_t count)
{
switch (fmt) {
case DAHDI_FORMAT_G723_1:
return count * (G723_SAMPLES/G723_5K_BYTES);
case DAHDI_FORMAT_ULAW:
case DAHDI_FORMAT_ALAW:
return count;
case DAHDI_FORMAT_G729A:
return count * (G729_SAMPLES/G729_BYTES);
default:
WARN_ON(1);
return 0;
}
}
static struct tcb *
wctc4xxp_create_rtp_cmd(struct wcdte *wc, struct dahdi_transcoder_channel *dtc,
size_t inbytes)
{
struct channel_pvt *cpvt = dtc->pvt;
struct rtp_packet *packet;
struct tcb *cmd;
cmd = alloc_cmd(sizeof(*packet) + inbytes);
if (!cmd)
return NULL;
cmd->cpvt = cpvt;
packet = cmd->data;
BUG_ON(cmd->data_len < sizeof(*packet));
/* setup the ethernet header */
memcpy(packet->ethhdr.h_dest, dst_mac, sizeof(dst_mac));
memcpy(packet->ethhdr.h_source, src_mac, sizeof(src_mac));
packet->ethhdr.h_proto = cpu_to_be16(ETH_P_IP);
/* setup the IP header */
packet->iphdr.ihl = 5;
packet->iphdr.version = 4;
packet->iphdr.tos = 0;
packet->iphdr.tot_len = cpu_to_be16(inbytes+40);
packet->iphdr.id = 0;
packet->iphdr.frag_off = cpu_to_be16(0x4000);
packet->iphdr.ttl = 64;
packet->iphdr.protocol = 0x11; /* UDP */
packet->iphdr.check = 0;
packet->iphdr.saddr = cpu_to_be32(0xc0a80903);
packet->iphdr.daddr = cpu_to_be32(0xc0a80903);
packet->iphdr.check = ip_fast_csum((void *)&packet->iphdr,
packet->iphdr.ihl);
/* setup the UDP header */
packet->udphdr.source = cpu_to_be16(cpvt->timeslot_out_num + 0x5000);
packet->udphdr.dest = cpu_to_be16(cpvt->timeslot_in_num + 0x5000);
packet->udphdr.len = cpu_to_be16(inbytes + sizeof(struct rtphdr) +
sizeof(struct udphdr));
packet->udphdr.check = 0;
/* Setup the RTP header */
packet->rtphdr.ver = 2;
packet->rtphdr.padding = 0;
packet->rtphdr.extension = 0;
packet->rtphdr.csrc_count = 0;
packet->rtphdr.marker = 0;
packet->rtphdr.type = wctc4xxp_dahdifmt_to_dtefmt(dtc->srcfmt);
packet->rtphdr.seqno = cpu_to_be16(cpvt->seqno);
packet->rtphdr.timestamp = cpu_to_be32(cpvt->last_timestamp);
packet->rtphdr.ssrc = cpu_to_be32(cpvt->ssrc);
cpvt->last_timestamp += wctc4xxp_bytes_to_samples(dtc->srcfmt,
inbytes);
WARN_ON(cmd->data_len > SFRAME_SIZE);
return cmd;
}
static void
wctc4xxp_cleanup_descriptor_ring(struct wctc4xxp_descriptor_ring *dr)
{
int i;
struct wctc4xxp_descriptor *d;
if (!dr || !dr->desc)
return;
for (i = 0; i < dr->size; ++i) {
d = wctc4xxp_descriptor(dr, i);
if (d->buffer1) {
pci_unmap_single(dr->pdev, d->buffer1,
SFRAME_SIZE, dr->direction);
d->buffer1 = 0;
/* Commands will also be sitting on the waiting for
* response list, so we want to make sure to delete
* them from that list as well. */
list_del_init(&(dr->pending[i])->node);
free_cmd(dr->pending[i]);
dr->pending[i] = NULL;
}
}
dr->head = 0;
dr->tail = 0;
dr->count = 0;
pci_free_consistent(dr->pdev, (sizeof(*d)+dr->padding) * dr->size,
dr->desc, dr->desc_dma);
kfree(dr->pending);
}
static void wctc4xxp_timeout_all_commands(struct wcdte *wc)
{
struct tcb *cmd;
struct tcb *temp;
unsigned long flags;
LIST_HEAD(local_list);
spin_lock_irqsave(&wc->cmd_list_lock, flags);
list_splice_init(&wc->waiting_for_response_list, &local_list);
list_splice_init(&wc->cmd_list, &local_list);
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
list_for_each_entry_safe(cmd, temp, &local_list, node) {
list_del_init(&cmd->node);
if (cmd->complete) {
cmd->flags |= DTE_CMD_TIMEOUT;
complete(cmd->complete);
} else {
free_cmd(cmd);
}
}
}
static void wctc4xxp_cleanup_command_list(struct wcdte *wc)
{
struct tcb *cmd;
unsigned long flags;
LIST_HEAD(local_list);
spin_lock_irqsave(&wc->cmd_list_lock, flags);
list_splice_init(&wc->cmd_list, &local_list);
list_splice_init(&wc->waiting_for_response_list, &local_list);
list_splice_init(&wc->rx_list, &local_list);
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
while (!list_empty(&local_list)) {
cmd = list_entry(local_list.next, struct tcb, node);
list_del_init(&cmd->node);
free_cmd(cmd);
}
}
static inline bool is_rtp_packet(const struct tcb *cmd)
{
const struct ethhdr *ethhdr = cmd->data;
return (cpu_to_be16(ETH_P_IP) == ethhdr->h_proto);
}
static void
wctc4xxp_transmit_cmd(struct wcdte *wc, struct tcb *cmd)
{
int res;
unsigned long flags;
/* If we're shutdown all commands will timeout. Just complete the
* command here with the timeout flag */
if (unlikely(test_bit(DTE_SHUTDOWN, &wc->flags))) {
if (cmd->complete) {
cmd->flags |= DTE_CMD_TIMEOUT;
list_del_init(&cmd->node);
complete(cmd->complete);
} else {
list_del(&cmd->node);
free_cmd(cmd);
}
return;
}
if (cmd->data_len < MIN_PACKET_LEN) {
memset((u8 *)(cmd->data) + cmd->data_len, 0,
MIN_PACKET_LEN-cmd->data_len);
cmd->data_len = MIN_PACKET_LEN;
}
WARN_ON(cmd->response);
WARN_ON(cmd->flags & TX_COMPLETE);
cmd->timeout = jiffies + HZ/4;
spin_lock_irqsave(&wc->cmd_list_lock, flags);
if (cmd->flags & (WAIT_FOR_ACK | WAIT_FOR_RESPONSE)) {
if (cmd->flags & WAIT_FOR_RESPONSE) {
/* We don't need both an ACK and a response. Let's
* tell the DTE not to generate an ACK, and we'll just
* retry if we do not get the response within the
* timeout period. */
struct csm_encaps_hdr *hdr = cmd->data;
hdr->control |= SUPPRESS_ACK;
}
WARN_ON(!list_empty(&cmd->node));
list_add_tail(&cmd->node, &wc->waiting_for_response_list);
mod_timer(&wc->watchdog, jiffies + HZ/2);
}
if (!list_empty(&wc->cmd_list)) {
if (is_rtp_packet(cmd))
list_add_tail(&cmd->node, &wc->cmd_list);
else
list_move(&cmd->node, &wc->cmd_list);
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
return;
}
res = wctc4xxp_submit(wc->txd, cmd);
if (-EBUSY == res) {
/* Looks like we're out of room in the descriptor
* ring. We'll add this command to the pending list
* and the interrupt service routine will pull from
* this list as it clears up room in the descriptor
* ring. */
list_move_tail(&cmd->node, &wc->cmd_list);
} else if (0 == res) {
if (!(cmd->flags & DO_NOT_CAPTURE))
wctc4xxp_net_capture_cmd(wc, cmd);
wctc4xxp_transmit_demand_poll(wc);
} else {
/* Unknown return value... */
WARN_ON(1);
}
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
}
static int
wctc4xxp_transmit_cmd_and_wait(struct wcdte *wc, struct tcb *cmd)
{
DECLARE_COMPLETION_ONSTACK(done);
cmd->complete = &done;
wctc4xxp_transmit_cmd(wc, cmd);
wait_for_completion(&done);
cmd->complete = NULL;
if (cmd->flags & DTE_CMD_TIMEOUT) {
DTE_DEBUG(DTE_DEBUG_GENERAL, "Timeout waiting for command.\n");
return -EIO;
}
return 0;
}
static int wctc4xxp_create_channel_pair(struct wcdte *wc,
struct channel_pvt *cpvt, u8 simple, u8 complicated);
static int wctc4xxp_destroy_channel_pair(struct wcdte *wc,
struct channel_pvt *cpvt);
static void
wctc4xxp_init_state(struct channel_pvt *cpvt, int encoder,
unsigned int channel, struct wcdte *wc)
{
memset(cpvt, 0, sizeof(*cpvt));
cpvt->encoder = encoder;
cpvt->wc = wc;
cpvt->chan_in_num = INVALID;
cpvt->chan_out_num = INVALID;
cpvt->ssrc = 0x78;
cpvt->timeslot_in_num = channel*2;
cpvt->timeslot_out_num = channel*2;
cpvt->last_rx_seq_num = 0xff;
if (encoder)
++cpvt->timeslot_out_num;
else
++cpvt->timeslot_in_num;
spin_lock_init(&cpvt->lock);
INIT_LIST_HEAD(&cpvt->rx_queue);
}
static unsigned int
wctc4xxp_getctl(struct wcdte *wc, unsigned int addr)
{
unsigned int val;
unsigned long flags;
spin_lock_irqsave(&wc->reglock, flags);
val = __wctc4xxp_getctl(wc, addr);
spin_unlock_irqrestore(&wc->reglock, flags);
return val;
}
static void
wctc4xxp_cleanup_channel_private(struct wcdte *wc,
struct dahdi_transcoder_channel *dtc)
{
struct tcb *cmd, *temp;
struct channel_pvt *cpvt = dtc->pvt;
unsigned long flags;
LIST_HEAD(local_list);
/* Once we cleanup this channel, we do not want any queued packets
* waiting to be transmitted. Anything on the hardware descriptor ring
* will be flushed by the csm_encaps command to shutdown the channel. */
spin_lock_irqsave(&wc->cmd_list_lock, flags);
list_for_each_entry_safe(cmd, temp, &wc->cmd_list, node) {
if (cmd->cpvt == cpvt)
list_move(&cmd->node, &local_list);
}
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
spin_lock_irqsave(&cpvt->lock, flags);
list_splice_init(&cpvt->rx_queue, &local_list);
dahdi_tc_clear_data_waiting(dtc);
cpvt->samples_in_flight = 0;
spin_unlock_irqrestore(&cpvt->lock, flags);
memset(&cpvt->stats, 0, sizeof(cpvt->stats));
list_for_each_entry_safe(cmd, temp, &local_list, node) {
list_del(&cmd->node);
free_cmd(cmd);
}
}
static void
wctc4xxp_mark_channel_complement_built(struct wcdte *wc,
struct dahdi_transcoder_channel *dtc)
{
int index;
struct channel_pvt *cpvt = dtc->pvt;
struct dahdi_transcoder_channel *compl_dtc;
struct channel_pvt *compl_cpvt;
BUG_ON(!cpvt);
index = cpvt->timeslot_in_num/2;
BUG_ON(index >= wc->numchannels);
if (cpvt->encoder)
compl_dtc = &(wc->udecode->channels[index]);
else
compl_dtc = &(wc->uencode->channels[index]);
/* It shouldn't already have been built... */
WARN_ON(dahdi_tc_is_built(compl_dtc));
compl_dtc->built_fmts = dtc->dstfmt | dtc->srcfmt;
compl_cpvt = compl_dtc->pvt;
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"dtc: %p is the complement to %p\n", compl_dtc, dtc);
compl_cpvt->chan_in_num = cpvt->chan_out_num;
compl_cpvt->chan_out_num = cpvt->chan_in_num;
dahdi_tc_set_built(compl_dtc);
wctc4xxp_cleanup_channel_private(wc, dtc);
}
static int
do_channel_allocate(struct dahdi_transcoder_channel *dtc)
{
struct channel_pvt *cpvt = dtc->pvt;
struct wcdte *wc = cpvt->wc;
u8 wctc4xxp_srcfmt; /* Digium Transcoder Engine Source Format */
u8 wctc4xxp_dstfmt; /* Digium Transcoder Engine Dest Format */
int res;
/* Check again to see if the channel was built after grabbing the
* channel lock, in case the previous holder of the lock
* built this channel as a complement to itself. */
if (dahdi_tc_is_built(dtc)) {
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"Allocating channel %p which is already built.\n", dtc);
return 0;
}
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"Entering %s for channel %p.\n", __func__, dtc);
/* Anything on the rx queue now is old news... */
wctc4xxp_cleanup_channel_private(wc, dtc);
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"Allocating a new channel: %p.\n", dtc);
wctc4xxp_srcfmt = wctc4xxp_dahdifmt_to_dtefmt(dtc->srcfmt);
wctc4xxp_dstfmt = wctc4xxp_dahdifmt_to_dtefmt(dtc->dstfmt);
res = wctc4xxp_create_channel_pair(wc, cpvt, wctc4xxp_srcfmt,
wctc4xxp_dstfmt);
if (res) {
dev_err(&wc->pdev->dev, "Failed to create channel pair.\n");
/* A failure to create a channel pair is normally a critical
* error in the firmware state. Reload the firmware when this
* handle is closed. */
set_bit(DTE_RELOAD, &wc->flags);
set_bit(DTE_SHUTDOWN, &wc->flags);
return res;
}
/* Mark this channel as built */
dahdi_tc_set_built(dtc);
dtc->built_fmts = dtc->dstfmt | dtc->srcfmt;
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"Channel %p has dstfmt=%x and srcfmt=%x\n", dtc, dtc->dstfmt,
dtc->srcfmt);
/* Mark the channel complement (other half of encoder/decoder pair) as
* built */
wctc4xxp_mark_channel_complement_built(wc, dtc);
dahdi_transcoder_alert(dtc);
return 0;
}
static void
wctc4xxp_setintmask(struct wcdte *wc, unsigned int intmask)
{
wc->intmask = intmask;
wctc4xxp_setctl(wc, 0x0038, intmask);
}
static const u32 DEFAULT_INTERRUPTS = 0x0001a0c0;
static void
wctc4xxp_enable_interrupts(struct wcdte *wc)
{
wctc4xxp_setintmask(wc, DEFAULT_INTERRUPTS);
}
static void
wctc4xxp_disable_interrupts(struct wcdte *wc)
{
/* Disable interrupts */
wctc4xxp_setintmask(wc, 0x00000000);
wctc4xxp_setctl(wc, 0x0084, 0x00000000);
}
static void
wctc4xxp_enable_polling(struct wcdte *wc)
{
set_bit(DTE_POLLING, &wc->flags);
wctc4xxp_setctl(wc, 0x0058, 0x10003);
/* Enable the general purpose timer interrupt. */
wctc4xxp_setintmask(wc, (DEFAULT_INTERRUPTS | (1 << 11)) & ~0x41);
}
static int wctc4xxp_reset_driver_state(struct wcdte *wc);
static bool wctc4xxp_need_firmware_reload(struct wcdte *wc)
{
return !!test_bit(DTE_RELOAD, &wc->flags) &&
(1 == wc->open_channels);
}
static int wctc4xxp_reload_firmware(struct wcdte *wc)
{
int res;
clear_bit(DTE_SHUTDOWN, &wc->flags);
res = wctc4xxp_reset_driver_state(wc);
if (res)
set_bit(DTE_SHUTDOWN, &wc->flags);
else
clear_bit(DTE_RELOAD, &wc->flags);
return res;
}
static int
wctc4xxp_operation_allocate(struct dahdi_transcoder_channel *dtc)
{
int res = 0;
struct wcdte *wc = ((struct channel_pvt *)(dtc->pvt))->wc;
res = mutex_lock_killable(&wc->chanlock);
if (res)
return res;
++wc->open_channels;
if (test_bit(DTE_SHUTDOWN, &wc->flags)) {
res = -EIO;
if (wctc4xxp_need_firmware_reload(wc))
res = wctc4xxp_reload_firmware(wc);
} else if (wctc4xxp_need_firmware_reload(wc)) {
res = wctc4xxp_reload_firmware(wc);
}
if (res)
goto error_exit;
if (wc->open_channels > POLLING_CALL_THRESHOLD) {
if (!test_bit(DTE_POLLING, &wc->flags))
wctc4xxp_enable_polling(wc);
}
if (dahdi_tc_is_built(dtc)) {
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"Allocating channel %p which is already built.\n", dtc);
res = 0;
} else {
res = do_channel_allocate(dtc);
}
error_exit:
mutex_unlock(&wc->chanlock);
return res;
}
static void
wctc4xxp_disable_polling(struct wcdte *wc)
{
clear_bit(DTE_POLLING, &wc->flags);
wctc4xxp_setctl(wc, 0x0058, 0x0);
wctc4xxp_enable_interrupts(wc);
}
static void wctc4xxp_check_for_rx_errors(struct wcdte *wc)
{
/* get_packet_errors() returns the accumulated total errors */
unsigned long errors = wctc4xxp_get_packet_errors(wc->rxd);
/* Print warning when the number of errors changes */
if (wc->reported_packet_errors != errors) {
if (printk_ratelimit()) {
dev_err(&wc->pdev->dev,
"%lu errored receive packets.\n",
errors - wc->reported_packet_errors);
wc->reported_packet_errors = errors;
}
}
}
static int
wctc4xxp_operation_release(struct dahdi_transcoder_channel *dtc)
{
int res = 0;
int index;
/* This is the 'complimentary channel' to dtc. I.e., if dtc is an
* encoder, compl_dtc is the decoder and vice-versa */
struct dahdi_transcoder_channel *compl_dtc;
struct channel_pvt *compl_cpvt;
struct channel_pvt *cpvt = dtc->pvt;
struct wcdte *wc = cpvt->wc;
int packets_received, packets_sent;
BUG_ON(!cpvt);
BUG_ON(!wc);
res = mutex_lock_killable(&wc->chanlock);
if (res)
return res;
if (test_bit(DTE_SHUTDOWN, &wc->flags)) {
/* On shutdown, if we reload the firmware we will reset the
* state of all the channels. Therefore we do not want to
* process any of the channel release logic even if the firmware
* was reloaded successfully. */
if (wctc4xxp_need_firmware_reload(wc))
wctc4xxp_reload_firmware(wc);
res = -EIO;
} else if (wctc4xxp_need_firmware_reload(wc)) {
wctc4xxp_reload_firmware(wc);
res = -EIO;
}
if (wc->open_channels) {
--wc->open_channels;
#if !defined(CONFIG_WCTC4XXP_POLLING)
if (wc->open_channels < POLLING_CALL_THRESHOLD) {
if (test_bit(DTE_POLLING, &wc->flags))
wctc4xxp_disable_polling(wc);
}
#endif
}
if (res)
goto error_exit;
packets_received = atomic_read(&cpvt->stats.packets_received);
packets_sent = atomic_read(&cpvt->stats.packets_sent);
DTE_DEBUG(DTE_DEBUG_ETH_STATS,
"%s channel %d sent %d packets and received %d packets.\n",
(cpvt->encoder) ? "encoder" : "decoder", cpvt->chan_out_num,
packets_sent, packets_received);
/* Remove any packets that are waiting on the outbound queue. */
dahdi_tc_clear_busy(dtc);
wctc4xxp_cleanup_channel_private(wc, dtc);
index = cpvt->timeslot_in_num/2;
BUG_ON(index >= wc->numchannels);
if (cpvt->encoder)
compl_dtc = &(wc->udecode->channels[index]);
else
compl_dtc = &(wc->uencode->channels[index]);
BUG_ON(!compl_dtc);
if (!dahdi_tc_is_built(compl_dtc)) {
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"Releasing a channel that was never built.\n");
res = 0;
goto error_exit;
}
/* If the channel complement (other half of the encoder/decoder pair) is
* being used. */
if (dahdi_tc_is_busy(compl_dtc)) {
res = 0;
goto error_exit;
}
res = wctc4xxp_destroy_channel_pair(wc, cpvt);
if (res)
goto error_exit;
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP, "Releasing channel: %p\n", dtc);
/* Mark this channel as not built */
dahdi_tc_clear_built(dtc);
dtc->built_fmts = 0;
cpvt->chan_in_num = INVALID;
cpvt->chan_out_num = INVALID;
/* Mark the channel complement as not built */
dahdi_tc_clear_built(compl_dtc);
compl_dtc->built_fmts = 0;
compl_cpvt = compl_dtc->pvt;
compl_cpvt->chan_in_num = INVALID;
compl_cpvt->chan_out_num = INVALID;
wctc4xxp_check_for_rx_errors(wc);
error_exit:
mutex_unlock(&wc->chanlock);
return res;
}
static inline struct tcb*
get_ready_cmd(struct dahdi_transcoder_channel *dtc)
{
struct channel_pvt *cpvt = dtc->pvt;
struct tcb *cmd;
unsigned long flags;
spin_lock_irqsave(&cpvt->lock, flags);
if (!list_empty(&cpvt->rx_queue)) {
WARN_ON(!dahdi_tc_is_data_waiting(dtc));
cmd = list_entry(cpvt->rx_queue.next, struct tcb, node);
list_del_init(&cmd->node);
} else {
cmd = NULL;
}
if (list_empty(&cpvt->rx_queue))
dahdi_tc_clear_data_waiting(dtc);
spin_unlock_irqrestore(&cpvt->lock, flags);
return cmd;
}
static int
wctc4xxp_handle_receive_ring(struct wcdte *wc)
{
struct tcb *cmd;
unsigned long flags;
unsigned int count = 0;
/* If we can't grab this lock, another thread must already be checking
* the receive ring...so we should just finish up, and we'll try again
* later. */
#if defined(spin_trylock_irqsave)
if (!spin_trylock_irqsave(&wc->rx_lock, flags))
return 0;
#else
if (spin_is_locked(&wc->rx_lock))
return 0;
spin_lock_irqsave(&wc->rx_lock, flags);
#endif
while ((cmd = wctc4xxp_retrieve(wc->rxd))) {
++count;
spin_lock(&wc->rx_list_lock);
list_add_tail(&cmd->node, &wc->rx_list);
spin_unlock(&wc->rx_list_lock);
cmd = __alloc_cmd(SFRAME_SIZE, GFP_ATOMIC, 0);
if (!cmd) {
dev_err(&wc->pdev->dev,
"Out of memory in %s.\n", __func__);
} else {
if (wctc4xxp_submit(wc->rxd, cmd)) {
dev_err(&wc->pdev->dev, "Failed submit in %s\n",
__func__);
free_cmd(cmd);
}
}
}
spin_unlock_irqrestore(&wc->rx_lock, flags);
return count;
}
/* Called with a buffer in which to copy a transcoded frame. */
static ssize_t
wctc4xxp_read(struct file *file, char __user *frame, size_t count, loff_t *ppos)
{
ssize_t ret;
struct dahdi_transcoder_channel *dtc = file->private_data;
struct channel_pvt *cpvt = dtc->pvt;
struct wcdte *wc = cpvt->wc;
struct tcb *cmd;
struct rtp_packet *packet;
ssize_t payload_bytes;
ssize_t returned_bytes = 0;
unsigned long flags;
BUG_ON(!dtc);
BUG_ON(!cpvt);
if (unlikely(test_bit(DTE_SHUTDOWN, &wc->flags))) {
/* The shudown flags can also be set if there is a
* catastrophic failure. */
return -EIO;
}
cmd = get_ready_cmd(dtc);
if (!cmd) {
if (file->f_flags & O_NONBLOCK)
return -EAGAIN;
ret = wait_event_interruptible(dtc->ready,
dahdi_tc_is_data_waiting(dtc));
if (-ERESTARTSYS == ret)
return -EINTR;
/* List went not empty. */
cmd = get_ready_cmd(dtc);
}
do {
BUG_ON(!cmd);
packet = cmd->data;
payload_bytes = be16_to_cpu(packet->udphdr.len) -
sizeof(struct rtphdr) -
sizeof(struct udphdr);
if (count < (payload_bytes + returned_bytes)) {
if (returned_bytes) {
/* If we have already returned at least one
* packets worth of data, we'll add this next
* packet to the head of the receive queue so
* it will be picked up next time. */
spin_lock_irqsave(&cpvt->lock, flags);
list_add(&cmd->node, &cpvt->rx_queue);
dahdi_tc_set_data_waiting(dtc);
spin_unlock_irqrestore(&cpvt->lock, flags);
return returned_bytes;
}
if (printk_ratelimit()) {
dev_err(&wc->pdev->dev,
"Cannot copy %zd bytes into %zd byte user " \
"buffer.\n", payload_bytes, count);
}
free_cmd(cmd);
return -EFBIG;
}
atomic_inc(&cpvt->stats.packets_received);
ret = copy_to_user(&frame[returned_bytes],
&packet->payload[0], payload_bytes);
if (unlikely(ret)) {
dev_err(&wc->pdev->dev, "Failed to copy data in %s\n",
__func__);
free_cmd(cmd);
return -EFAULT;
}
returned_bytes += payload_bytes;
free_cmd(cmd);
} while ((cmd = get_ready_cmd(dtc)));
return returned_bytes;
}
/* Called with a frame in the srcfmt to be transcoded into the dstfmt. */
static ssize_t
wctc4xxp_write(struct file *file, const char __user *frame,
size_t count, loff_t *ppos)
{
struct dahdi_transcoder_channel *dtc = file->private_data;
struct channel_pvt *cpvt = dtc->pvt;
struct wcdte *wc = cpvt->wc;
struct tcb *cmd;
u32 samples;
unsigned long flags;
const unsigned long MAX_SAMPLES_IN_FLIGHT = 640;
const unsigned long MAX_RTP_PAYLOAD = 500;
BUG_ON(!cpvt);
BUG_ON(!wc);
if (unlikely(test_bit(DTE_SHUTDOWN, &wc->flags)))
return -EIO;
if (!test_bit(DAHDI_TC_FLAG_CHAN_BUILT, &dtc->flags))
return -EAGAIN;
if (count < 2) {
DTE_DEBUG(DTE_DEBUG_GENERAL,
"Cannot request to transcode a packet that is less than " \
"2 bytes.\n");
return -EINVAL;
}
if (count > MAX_RTP_PAYLOAD) {
DTE_DEBUG(DTE_DEBUG_GENERAL,
"Cannot transcode packet of %Zu bytes. This exceeds the maximum size of %lu bytes.\n",
count, MAX_RTP_PAYLOAD);
return -EINVAL;
}
if (DAHDI_FORMAT_G723_1 == dtc->srcfmt) {
if ((G723_5K_BYTES != count) && (G723_6K_BYTES != count) &&
(G723_SID_BYTES != count)) {
DTE_DEBUG(DTE_DEBUG_GENERAL,
"Trying to transcode packet into G723 format " \
"that is %Zu bytes instead of the expected " \
"%d/%d/%d bytes.\n", count, G723_5K_BYTES,
G723_6K_BYTES, G723_SID_BYTES);
return -EINVAL;
}
}
/* Do not flood the firmware with packets. This can result in out of
* memory conditions in the firmware. */
spin_lock_irqsave(&cpvt->lock, flags);
if (time_after(jiffies, cpvt->send_time)) {
cpvt->samples_in_flight = max(0L,
cpvt->samples_in_flight - 160L);
}
samples = wctc4xxp_bytes_to_samples(dtc->srcfmt, count);
if ((cpvt->samples_in_flight + samples) > MAX_SAMPLES_IN_FLIGHT) {
spin_unlock_irqrestore(&cpvt->lock, flags);
/* This should most likely be an error, but it results in
* codec_dahdi spamming when it's not set to wait for new
* packets. Instead we will silently drop the bytes. */
return count;
}
cpvt->send_time = jiffies + msecs_to_jiffies(20);
spin_unlock_irqrestore(&cpvt->lock, flags);
cmd = wctc4xxp_create_rtp_cmd(wc, dtc, count);
if (!cmd)
return -ENOMEM;
/* Copy the data directly from user space into the command buffer. */
if (copy_from_user(&((struct rtp_packet *)(cmd->data))->payload[0],
frame, count)) {
dev_err(&wc->pdev->dev,
"Failed to copy packet from userspace.\n");
free_cmd(cmd);
return -EFAULT;
}
cpvt->seqno += 1;
DTE_DEBUG(DTE_DEBUG_RTP_TX,
"Sending packet of %Zu byte on channel (%p).\n", count, dtc);
atomic_inc(&cpvt->stats.packets_sent);
wctc4xxp_transmit_cmd(wc, cmd);
return count;
}
static void
wctc4xxp_send_ack(struct wcdte *wc, u8 seqno, __be16 channel, __le16 function)
{
struct tcb *cmd;
struct csm_encaps_hdr *hdr;
cmd = __alloc_cmd(sizeof(*hdr), ALLOC_FLAGS, 0);
if (!cmd) {
WARN_ON(1);
return;
}
hdr = cmd->data;
BUG_ON(sizeof(*hdr) > cmd->data_len);
setup_common_header(wc, hdr);
hdr->op_code = cpu_to_be16(0x0001);
hdr->seq_num = seqno;
hdr->control = 0xe0;
hdr->channel = channel;
hdr->cmd.function = function;
wctc4xxp_transmit_cmd(wc, cmd);
}
static void do_rx_response_packet(struct wcdte *wc, struct tcb *cmd)
{
struct csm_encaps_hdr *rxhdr;
const struct csm_encaps_hdr *listhdr;
struct tcb *pos, *temp;
unsigned long flags;
bool handled = false;
rxhdr = cmd->data;
/* Check if duplicated response on the supervisor channel. */
if (SUPERVISOR_CHANNEL == rxhdr->channel) {
if (rxhdr->seq_num == wc->last_rx_seq_num) {
free_cmd(cmd);
return;
}
wc->last_rx_seq_num = rxhdr->seq_num;
}
spin_lock_irqsave(&wc->cmd_list_lock, flags);
list_for_each_entry_safe(pos, temp,
&wc->waiting_for_response_list, node) {
listhdr = pos->data;
if ((listhdr->cmd.function == rxhdr->cmd.function) &&
(listhdr->channel == rxhdr->channel)) {
/* If this is a channel command, do not complete it if
* the seq_num is the same as previous. */
if (pos->cpvt) {
if (rxhdr->seq_num ==
pos->cpvt->last_rx_seq_num) {
break;
}
pos->cpvt->last_rx_seq_num = rxhdr->seq_num;
}
list_del_init(&pos->node);
pos->flags &= ~(WAIT_FOR_RESPONSE);
pos->response = cmd;
/* If this isn't TX_COMPLETE yet, then this packet will
* be completed in service_tx_ring. */
if (pos->flags & TX_COMPLETE && pos->complete)
complete(pos->complete);
handled = true;
break;
}
}
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
if (!handled) {
DTE_DEBUG(DTE_DEBUG_GENERAL,
"Freeing unhandled response ch:(%04x)\n",
be16_to_cpu(rxhdr->channel));
free_cmd(cmd);
}
}
static void
do_rx_ack_packet(struct wcdte *wc, struct tcb *cmd)
{
const struct csm_encaps_hdr *listhdr, *rxhdr;
struct tcb *pos, *temp;
unsigned long flags;
rxhdr = cmd->data;
spin_lock_irqsave(&wc->cmd_list_lock, flags);
list_for_each_entry_safe(pos, temp,
&wc->waiting_for_response_list, node) {
listhdr = pos->data;
if (cpu_to_be16(0xefed) == listhdr->ethhdr.h_proto) {
wc->seq_num = (rxhdr->seq_num + 1) & 0xff;
WARN_ON(!(pos->complete));
WARN_ON(!(pos->flags & TX_COMPLETE));
list_del_init(&pos->node);
if (pos->complete)
complete(pos->complete);
} else if ((listhdr->seq_num == rxhdr->seq_num) &&
(listhdr->channel == rxhdr->channel)) {
if (pos->flags & WAIT_FOR_RESPONSE) {
pos->flags &= ~(WAIT_FOR_ACK);
} else {
list_del_init(&pos->node);
if (pos->complete) {
WARN_ON(!(pos->flags & TX_COMPLETE));
complete(pos->complete);
} else {
free_cmd(pos);
}
}
break;
}
}
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
/* There is never a reason to store up the ack packets. */
free_cmd(cmd);
}
static inline int
is_response(const struct csm_encaps_hdr *hdr)
{
return ((0x02 == hdr->cmd.type) ||
(0x04 == hdr->cmd.type) ||
(0x0e == hdr->cmd.type) ||
(0x00 == hdr->cmd.type)) ? 1 : 0;
}
static void
print_command(struct wcdte *wc, const struct csm_encaps_hdr *hdr)
{
int i, curlength;
char *buffer;
const int BUFFER_SIZE = 1024;
int parameters = ((hdr->cmd.length - 8)/sizeof(__le16));
buffer = kzalloc(BUFFER_SIZE + 1, GFP_ATOMIC);
if (!buffer) {
dev_info(&wc->pdev->dev, "Failed print_command\n");
return;
}
curlength = snprintf(buffer, BUFFER_SIZE,
"opcode: %04x seq: %02x control: %02x "
"channel: %04x ", be16_to_cpu(hdr->op_code),
hdr->seq_num, hdr->control, be16_to_cpu(hdr->channel));
curlength += snprintf(buffer + curlength, BUFFER_SIZE - curlength,
"length: %02x index: %02x type: %02x "
"class: %02x function: %04x",
hdr->cmd.length, hdr->cmd.index, hdr->cmd.type, hdr->cmd.class,
le16_to_cpu(hdr->cmd.function));
for (i = 0; i < parameters; ++i) {
curlength += snprintf(buffer + curlength,
BUFFER_SIZE - curlength, " %04x",
le16_to_cpu(hdr->cmd.params[i]));
}
dev_info(&wc->pdev->dev, "%s\n", buffer);
kfree(buffer);
}
static inline void wctc4xxp_reset_processor(struct wcdte *wc)
{
wctc4xxp_setctl(wc, 0x00A0, 0x04000000);
}
static void handle_csm_alert(struct wcdte *wc,
const struct csm_encaps_hdr *hdr)
{
const struct csm_encaps_cmd *c = &hdr->cmd;
if (c->function == 0x0000) {
u16 alert_type = le16_to_cpu(c->params[0]);
u16 action_required = le16_to_cpu(c->params[1]) >> 8;
const bool fatal_error = action_required != 0;
dev_err(&wc->pdev->dev,
"Received alert (0x%04x) from dsp. Firmware will be reloaded when possible.\n",
alert_type);
if (fatal_error) {
/* If any fatal errors are reported we'll just shut
* everything down so that we do not hang up any user
* process trying to wait for commands to complete. */
wctc4xxp_reset_processor(wc);
set_bit(DTE_RELOAD, &wc->flags);
set_bit(DTE_SHUTDOWN, &wc->flags);
wctc4xxp_timeout_all_commands(wc);
} else {
/* For non-fatal errors we'll try to proceed and reload
* the firmware when all open channels are closed. This
* will prevent impacting any normal calls in progress.
*
*/
set_bit(DTE_RELOAD, &wc->flags);
}
} else {
if (debug) {
dev_warn(&wc->pdev->dev,
"Received diagnostic message:\n");
}
}
if (debug) {
print_command(wc, hdr);
}
}
static void
receive_csm_encaps_packet(struct wcdte *wc, struct tcb *cmd)
{
const struct csm_encaps_hdr *hdr = cmd->data;
const struct csm_encaps_cmd *c = &hdr->cmd;
if (!(hdr->control & MESSAGE_PACKET)) {
const bool suppress_ack = ((hdr->control & SUPPRESS_ACK) > 0);
if (!suppress_ack) {
wctc4xxp_send_ack(wc, hdr->seq_num, hdr->channel,
c->function);
}
if (is_response(hdr)) {
do_rx_response_packet(wc, cmd);
} else if (0xc1 == c->type) {
if (0x75 == c->class) {
dev_warn(&wc->pdev->dev,
"Received alert (0x%04x) from dsp\n",
le16_to_cpu(c->params[0]));
}
free_cmd(cmd);
} else if (0xd4 == c->type) {
if (c->params[0] != le16_to_cpu(0xffff)) {
dev_warn(&wc->pdev->dev,
"DTE Failed self test (%04x).\n",
le16_to_cpu(c->params[0]));
} else if ((c->params[1] != le16_to_cpu(0x000c)) &&
(c->params[1] != le16_to_cpu(0x010c))) {
dev_warn(&wc->pdev->dev,
"Unexpected ERAM status (%04x).\n",
le16_to_cpu(c->params[1]));
} else {
wctc4xxp_set_ready(wc);
wake_up(&wc->waitq);
}
free_cmd(cmd);
} else if (MONITOR_LIVE_INDICATION_TYPE == c->type) {
handle_csm_alert(wc, hdr);
free_cmd(cmd);
} else {
dev_warn(&wc->pdev->dev,
"Unknown command type received. %02x\n",
c->type);
free_cmd(cmd);
}
} else {
do_rx_ack_packet(wc, cmd);
}
}
static void
queue_rtp_packet(struct wcdte *wc, struct tcb *cmd)
{
unsigned index;
struct dahdi_transcoder_channel *dtc;
struct channel_pvt *cpvt;
struct rtp_packet *packet = cmd->data;
unsigned long flags;
long samples;
if (unlikely(ip_fast_csum((void *)(&packet->iphdr),
packet->iphdr.ihl))) {
DTE_DEBUG(DTE_DEBUG_GENERAL,
"Invalid checksum in RTP packet %04x\n",
ip_fast_csum((void *)(&packet->iphdr),
packet->iphdr.ihl));
free_cmd(cmd);
return;
}
index = (be16_to_cpu(packet->udphdr.dest) - 0x5000) / 2;
if (unlikely(!(index < wc->numchannels))) {
dev_err(&wc->pdev->dev,
"Invalid channel number in response from DTE.\n");
free_cmd(cmd);
return;
}
switch (packet->rtphdr.type) {
case 0x00:
case 0x08:
dtc = &(wc->udecode->channels[index]);
break;
case 0x04:
case 0x12:
dtc = &(wc->uencode->channels[index]);
break;
default:
dev_err(&wc->pdev->dev, "Unknown codec in packet (0x%02x).\n",\
packet->rtphdr.type);
free_cmd(cmd);
return;
}
cpvt = dtc->pvt;
if (!dahdi_tc_is_busy(dtc)) {
free_cmd(cmd);
return;
}
spin_lock_irqsave(&cpvt->lock, flags);
samples = wctc4xxp_bytes_to_samples(dtc->dstfmt,
be16_to_cpu(packet->udphdr.len) -
sizeof(struct rtphdr) - sizeof(struct udphdr));
cpvt->samples_in_flight = max(cpvt->samples_in_flight - samples, 0L);
list_add_tail(&cmd->node, &cpvt->rx_queue);
dahdi_tc_set_data_waiting(dtc);
spin_unlock_irqrestore(&cpvt->lock, flags);
dahdi_transcoder_alert(dtc);
return;
}
static void service_tx_ring(struct wcdte *wc)
{
struct tcb *cmd;
unsigned long flags;
spin_lock_irqsave(&wc->cmd_list_lock, flags);
while ((cmd = wctc4xxp_retrieve(wc->txd))) {
cmd->flags |= TX_COMPLETE;
if (!(cmd->flags & (WAIT_FOR_ACK | WAIT_FOR_RESPONSE))) {
/* If we're not waiting for an ACK or Response from
* the DTE, this message should not be sitting on any
* lists. */
WARN_ON(!list_empty(&cmd->node));
if (cmd->complete) {
WARN_ON(!(cmd->flags & TX_COMPLETE));
complete(cmd->complete);
} else {
free_cmd(cmd);
}
}
/* We've freed up a spot in the hardware ring buffer. If
* another packet is queued up, let's submit it to the
* hardware. */
if (!list_empty(&wc->cmd_list)) {
cmd = list_entry(wc->cmd_list.next, struct tcb, node);
list_del_init(&cmd->node);
if (cmd->flags & (WAIT_FOR_ACK | WAIT_FOR_RESPONSE)) {
list_add_tail(&cmd->node,
&wc->waiting_for_response_list);
}
wctc4xxp_submit(wc->txd, cmd);
}
}
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
}
static void service_rx_ring(struct wcdte *wc)
{
struct tcb *cmd;
unsigned long flags;
LIST_HEAD(local_list);
spin_lock_irqsave(&wc->rx_list_lock, flags);
list_splice_init(&wc->rx_list, &local_list);
spin_unlock_irqrestore(&wc->rx_list_lock, flags);
/*
* Process the received packets
*/
while (!list_empty(&local_list)) {
const struct ethhdr *ethhdr;
cmd = container_of(local_list.next, struct tcb, node);
ethhdr = (const struct ethhdr *)(cmd->data);
list_del_init(&cmd->node);
wctc4xxp_net_capture_cmd(wc, cmd);
if (cpu_to_be16(ETH_P_IP) == ethhdr->h_proto) {
queue_rtp_packet(wc, cmd);
} else if (cpu_to_be16(ETH_P_CSM_ENCAPS) == ethhdr->h_proto) {
receive_csm_encaps_packet(wc, cmd);
} else {
DTE_DEBUG(DTE_DEBUG_GENERAL,
"Unknown packet protocol received: %04x.\n",
be16_to_cpu(ethhdr->h_proto));
free_cmd(cmd);
}
}
wctc4xxp_receive_demand_poll(wc);
}
static void deferred_work_func(struct work_struct *work)
{
struct wcdte *wc = container_of(work, struct wcdte, deferred_work);
service_rx_ring(wc);
}
static irqreturn_t wctc4xxp_interrupt(int irq, void *dev_id)
{
struct wcdte *wc = dev_id;
bool packets_to_process = false;
u32 ints;
#define NORMAL_INTERRUPT_SUMMARY (1<<16)
#define ABNORMAL_INTERRUPT_SUMMARY (1<<15)
#define TX_COMPLETE_INTERRUPT 0x00000001
#define RX_COMPLETE_INTERRUPT 0x00000040
#define TIMER_INTERRUPT (1<<11)
#define NORMAL_INTERRUPTS (TX_COMPLETE_INTERRUPT | RX_COMPLETE_INTERRUPT | \
TIMER_INTERRUPT)
/* Read and clear interrupts */
ints = __wctc4xxp_getctl(wc, 0x0028);
if (!(ints & (NORMAL_INTERRUPT_SUMMARY|ABNORMAL_INTERRUPT_SUMMARY)))
return IRQ_NONE;
/* Clear all the pending interrupts. */
__wctc4xxp_setctl(wc, 0x0028, ints);
if (ints & (RX_COMPLETE_INTERRUPT | TIMER_INTERRUPT)) {
packets_to_process = wctc4xxp_handle_receive_ring(wc) > 0;
service_tx_ring(wc);
#if DEFERRED_PROCESSING == WORKQUEUE
if (packets_to_process)
schedule_work(&wc->deferred_work);
#elif DEFERRED_PROCESSING == INTERRUPT
#error "You will need to change the locks if you want to run the processing " \
"in the interrupt handler."
#else
#error "Define a deferred processing function in kernel/wctc4xxp/wctc4xxp.h"
#endif
} else {
if ((ints & 0x00008000) && debug)
dev_info(&wc->pdev->dev, "Abnormal Interrupt.\n");
if (ints & 0x00002000)
dev_err(&wc->pdev->dev, "Fatal Bus Error Detected.\n");
if ((ints & 0x00000100) && debug)
dev_info(&wc->pdev->dev, "Receive Stopped INT\n");
if ((ints & 0x00000080) && debug) {
dev_info(&wc->pdev->dev,
"Receive Desciptor Unavailable INT " \
"(%d)\n", wctc4xxp_getcount(wc->rxd));
}
if ((ints & 0x00000020) && debug)
dev_info(&wc->pdev->dev, "Transmit Under-flow INT\n");
if ((ints & 0x00000008) && debug)
dev_info(&wc->pdev->dev, "Jabber Timer Time-out INT\n");
if ((ints & 0x00000002) && debug) {
dev_info(&wc->pdev->dev,
"Transmit Processor Stopped INT\n");
}
}
return IRQ_HANDLED;
}
static int
wctc4xxp_hardware_init(struct wcdte *wc)
{
/* Hardware stuff */
enum {
/* Software Reset */
SWR = (1 << 0),
/* Bus Arbitration (1 for priority transmit) */
BAR = (1 << 1),
/* Memory Write Invalidate */
MWI = (1 << 24),
/* Memory Read Line */
MRL = (1 << 23),
/* Descriptor Skip Length */
DSLShift = 2,
/* Cache Alignment */
CALShift = 14,
/* Transmit Auto Pollling */
TAPShift = 17,
};
u32 reg;
unsigned long newjiffies;
u8 cache_line_size;
const u32 DEFAULT_PCI_ACCESS = (MWI | (11 << TAPShift));
if (pci_read_config_byte(wc->pdev, 0x0c, &cache_line_size))
return -EIO;
switch (cache_line_size) {
case 0x08:
reg = DEFAULT_PCI_ACCESS | (0x1 << CALShift);
break;
case 0x10:
reg = DEFAULT_PCI_ACCESS | (0x2 << CALShift);
break;
case 0x20:
reg = DEFAULT_PCI_ACCESS | (0x3 << CALShift);
break;
default:
reg = (11 << TAPShift);
break;
}
reg |= ((wc->txd->padding / sizeof(u32)) << 2) & 0x7c;
/* Reset the DTE... */
wctc4xxp_setctl(wc, 0x0000, reg | 1);
newjiffies = jiffies + HZ; /* One second timeout */
/* ...and wait for it to come out of reset. */
while (((wctc4xxp_getctl(wc, 0x0000)) & 0x00000001) &&
(newjiffies > jiffies))
msleep(1);
wctc4xxp_setctl(wc, 0x0000, reg | 0x60000);
/* Configure watchdogs, access, etc */
wctc4xxp_setctl(wc, 0x0030, 0x00280048);
wctc4xxp_setctl(wc, 0x0078, 0x00000013);
reg = wctc4xxp_getctl(wc, 0x00fc);
wctc4xxp_setctl(wc, 0x00fc, (reg & ~0x7) | 0x7);
reg = wctc4xxp_getctl(wc, 0x00fc);
return 0;
}
static void
wctc4xxp_start_dma(struct wcdte *wc)
{
int res;
int i;
u32 reg;
struct tcb *cmd;
for (i = 0; i < wc->rxd->size; ++i) {
cmd = alloc_cmd(SFRAME_SIZE);
if (!cmd) {
WARN_ALWAYS();
return;
}
WARN_ON(SFRAME_SIZE != cmd->data_len);
res = wctc4xxp_submit(wc->rxd, cmd);
if (res) {
/* When we're starting the DMA, we should always be
* able to fill the ring....so something is wrong
* here. */
WARN_ALWAYS();
free_cmd(cmd);
break;
}
}
wmb();
wctc4xxp_setctl(wc, 0x0020, wc->txd->desc_dma);
wctc4xxp_setctl(wc, 0x0018, wc->rxd->desc_dma);
/* Start receiver/transmitter */
reg = wctc4xxp_getctl(wc, 0x0030);
wctc4xxp_setctl(wc, 0x0030, reg | 0x00002002);
wctc4xxp_receive_demand_poll(wc);
reg = wctc4xxp_getctl(wc, 0x0028);
wctc4xxp_setctl(wc, 0x0028, reg);
}
static void
_wctc4xxp_stop_dma(struct wcdte *wc)
{
/* Disable interrupts and reset */
unsigned int reg;
/* Disable interrupts */
wctc4xxp_setintmask(wc, 0x00000000);
wctc4xxp_setctl(wc, 0x0084, 0x00000000);
wctc4xxp_setctl(wc, 0x0048, 0x00000000);
/* Reset the part to be on the safe side */
reg = wctc4xxp_getctl(wc, 0x0000);
reg |= 0x00000001;
wctc4xxp_setctl(wc, 0x0000, reg);
}
static void
wctc4xxp_stop_dma(struct wcdte *wc)
{
unsigned long newjiffies;
_wctc4xxp_stop_dma(wc);
newjiffies = jiffies + HZ; /* One second timeout */
/* We'll wait here for the part to come out of reset */
while (((wctc4xxp_getctl(wc, 0x0000)) & 0x00000001) &&
(newjiffies > jiffies))
msleep(1);
}
#define MDIO_SHIFT_CLK 0x10000
#define MDIO_DATA_WRITE1 0x20000
#define MDIO_ENB 0x00000
#define MDIO_ENB_IN 0x40000
#define MDIO_DATA_READ 0x80000
static int
wctc4xxp_read_phy(struct wcdte *wc, int location)
{
int i;
long mdio_addr = 0x0048;
int read_cmd = (0xf6 << 10) | (1 << 5) | location;
int retval = 0;
/* Establish sync by sending at least 32 logic ones. */
for (i = 32; i >= 0; --i) {
wctc4xxp_setctl(wc, mdio_addr,
MDIO_ENB | MDIO_DATA_WRITE1);
wctc4xxp_getctl(wc, mdio_addr);
wctc4xxp_setctl(wc, mdio_addr,
MDIO_ENB | MDIO_DATA_WRITE1 | MDIO_SHIFT_CLK);
wctc4xxp_getctl(wc, mdio_addr);
}
/* Shift the read command bits out. */
for (i = 17; i >= 0; --i) {
int dataval = (read_cmd & (1 << i)) ? MDIO_DATA_WRITE1 : 0;
wctc4xxp_setctl(wc, mdio_addr, MDIO_ENB | dataval);
wctc4xxp_getctl(wc, mdio_addr);
wctc4xxp_setctl(wc, mdio_addr,
MDIO_ENB | dataval | MDIO_SHIFT_CLK);
wctc4xxp_getctl(wc, mdio_addr);
}
/* Read the two transition, 16 data, and wire-idle bits. */
for (i = 19; i > 0; --i) {
wctc4xxp_setctl(wc, mdio_addr, MDIO_ENB_IN);
wctc4xxp_getctl(wc, mdio_addr);
retval = (retval << 1) |
((wctc4xxp_getctl(wc, mdio_addr) & MDIO_DATA_READ) ?
1 : 0);
wctc4xxp_setctl(wc, mdio_addr, MDIO_ENB_IN | MDIO_SHIFT_CLK);
wctc4xxp_getctl(wc, mdio_addr);
}
retval = (retval>>1) & 0xffff;
return retval;
}
static void
wctc4xxp_write_phy(struct wcdte *wc, int location, int value)
{
int i;
int cmd = (0x5002 << 16) | (1 << 23) | (location<<18) | value;
long mdio_addr = 0x0048;
/* Establish sync by sending 32 logic ones. */
for (i = 32; i >= 0; --i) {
wctc4xxp_setctl(wc, mdio_addr, MDIO_ENB | MDIO_DATA_WRITE1);
wctc4xxp_getctl(wc, mdio_addr);
wctc4xxp_setctl(wc, mdio_addr,
MDIO_ENB | MDIO_DATA_WRITE1 | MDIO_SHIFT_CLK);
wctc4xxp_getctl(wc, mdio_addr);
}
/* Shift the command bits out. */
for (i = 31; i >= 0; --i) {
int dataval = (cmd & (1 << i)) ? MDIO_DATA_WRITE1 : 0;
wctc4xxp_setctl(wc, mdio_addr, MDIO_ENB | dataval);
wctc4xxp_getctl(wc, mdio_addr);
wctc4xxp_setctl(wc, mdio_addr,
MDIO_ENB | dataval | MDIO_SHIFT_CLK);
wctc4xxp_getctl(wc, mdio_addr);
}
/* Clear out extra bits. */
for (i = 2; i > 0; --i) {
wctc4xxp_setctl(wc, mdio_addr, MDIO_ENB_IN);
wctc4xxp_getctl(wc, mdio_addr);
wctc4xxp_setctl(wc, mdio_addr, MDIO_ENB_IN | MDIO_SHIFT_CLK);
wctc4xxp_getctl(wc, mdio_addr);
}
return;
}
static int
wctc4xxp_wait_for_link(struct wcdte *wc)
{
int reg;
unsigned int delay_count = 0;
do {
reg = wctc4xxp_getctl(wc, 0x00fc);
msleep(2);
delay_count++;
if (delay_count >= 5000) {
dev_err(&wc->pdev->dev,
"Failed to link to DTE processor!\n");
return -EIO;
}
} while ((reg & 0xE0000000) != 0xE0000000);
return 0;
}
static int
wctc4xxp_load_firmware(struct wcdte *wc, const struct firmware *firmware)
{
unsigned int byteloc;
unsigned int length;
struct tcb *cmd;
DECLARE_COMPLETION_ONSTACK(done);
byteloc = 17;
cmd = alloc_cmd(SFRAME_SIZE);
if (!cmd)
return -ENOMEM;
#if defined(CONFIG_WCTC4XXP_POLLING)
wctc4xxp_enable_polling(wc);
#endif
clear_bit(DTE_READY, &wc->flags);
while (byteloc < (firmware->size-20)) {
length = (firmware->data[byteloc] << 8) |
firmware->data[byteloc+1];
byteloc += 2;
cmd->data_len = length;
BUG_ON(length > cmd->data_len);
memcpy(cmd->data, &firmware->data[byteloc], length);
byteloc += length;
cmd->flags = WAIT_FOR_ACK;
cmd->complete = &done;
wctc4xxp_transmit_cmd(wc, cmd);
wait_for_completion(&done);
if (cmd->flags & DTE_CMD_TIMEOUT) {
free_cmd(cmd);
dev_err(&wc->pdev->dev, "Failed to load firmware.\n");
#if defined(CONFIG_WCTC4XXP_POLLING)
wctc4xxp_disable_polling(wc);
#endif
return -EIO;
}
}
free_cmd(cmd);
if (!wait_event_timeout(wc->waitq, wctc4xxp_is_ready(wc), 15*HZ)) {
dev_err(&wc->pdev->dev, "Failed to boot firmware.\n");
#if defined(CONFIG_WCTC4XXP_POLLING)
wctc4xxp_disable_polling(wc);
#endif
return -EIO;
}
#if defined(CONFIG_WCTC4XXP_POLLING)
wctc4xxp_disable_polling(wc);
#endif
return 0;
}
static int
wctc4xxp_turn_off_booted_led(struct wcdte *wc)
{
int ret = 0;
int reg;
/* Turn off auto negotiation */
wctc4xxp_write_phy(wc, 0, 0x2100);
DTE_DEBUG(DTE_DEBUG_GENERAL, "PHY register 0 = %X\n",
wctc4xxp_read_phy(wc, 0));
wctc4xxp_reset_processor(wc);
/* Wait 4 ms to ensure processor reset */
msleep(4);
/* Clear reset */
wctc4xxp_setctl(wc, 0x00A0, 0x04080000);
/* Wait for the ethernet link */
ret = wctc4xxp_wait_for_link(wc);
if (ret)
return ret;
/* Turn off booted LED */
wctc4xxp_setctl(wc, 0x00A0, 0x04084000);
reg = wctc4xxp_getctl(wc, 0x00fc);
DTE_DEBUG(DTE_DEBUG_GENERAL, "LINK STATUS: reg(0xfc) = %X\n", reg);
reg = wctc4xxp_getctl(wc, 0x00A0);
return ret;
}
static void
wctc4xxp_turn_on_booted_led(struct wcdte *wc)
{
wctc4xxp_setctl(wc, 0x00A0, 0x04080000);
}
static int
wctc4xxp_boot_processor(struct wcdte *wc, const struct firmware *firmware)
{
int ret;
wctc4xxp_turn_off_booted_led(wc);
ret = wctc4xxp_load_firmware(wc, firmware);
if (ret)
return ret;
wctc4xxp_turn_on_booted_led(wc);
DTE_DEBUG(DTE_DEBUG_GENERAL, "Successfully booted DTE processor.\n");
return 0;
}
static void
setup_half_channel(struct channel_pvt *pvt, struct tcb *cmd, u16 length)
{
setup_channel_header(pvt, cmd);
append_set_ip_hdr_channel_cmd(cmd);
append_voip_vceopt_cmd(cmd, length);
append_voip_tonectl_cmd(cmd);
append_voip_dtmfopt_cmd(cmd);
append_voip_indctrl_cmd(cmd);
/* To indicate the end of multiple messages. */
cmd->data_len += 4;
WARN_ON(cmd->data_len >= SFRAME_SIZE);
wctc4xxp_transmit_cmd(pvt->wc, cmd);
}
static int wctc4xxp_setup_channels(struct wcdte *wc,
struct channel_pvt *encoder_pvt,
struct channel_pvt *decoder_pvt,
u16 length)
{
int res = 0;
struct tcb *encoder_cmd;
struct tcb *decoder_cmd;
DECLARE_COMPLETION_ONSTACK(encoder_done);
DECLARE_COMPLETION_ONSTACK(decoder_done);
encoder_cmd = alloc_cmd(SFRAME_SIZE);
decoder_cmd = alloc_cmd(SFRAME_SIZE);
if (!encoder_cmd || !decoder_cmd) {
res = -ENOMEM;
goto error_exit;
}
encoder_cmd->complete = &encoder_done;
decoder_cmd->complete = &decoder_done;
setup_half_channel(encoder_pvt, encoder_cmd, length);
setup_half_channel(decoder_pvt, decoder_cmd, length);
wait_for_completion(&decoder_done);
wait_for_completion(&encoder_done);
if (encoder_cmd->flags & DTE_CMD_TIMEOUT ||
decoder_cmd->flags & DTE_CMD_TIMEOUT) {
DTE_DEBUG(DTE_DEBUG_GENERAL, "Timeout waiting for command.\n");
res = -EIO;
}
if ((0x0000 != response_header(encoder_cmd)->cmd.params[0]) ||
(0x0000 != response_header(encoder_cmd)->cmd.params[0]))
res = -EIO;
error_exit:
free_cmd(encoder_cmd);
free_cmd(decoder_cmd);
return res;
}
static int wctc4xxp_enable_channels(struct wcdte *wc,
struct channel_pvt *encoder_pvt,
struct channel_pvt *decoder_pvt,
u8 complicated, u8 simple)
{
int res = 0;
struct tcb *encoder_cmd;
struct tcb *decoder_cmd;
DECLARE_COMPLETION_ONSTACK(encoder_done);
DECLARE_COMPLETION_ONSTACK(decoder_done);
encoder_cmd = alloc_cmd(SFRAME_SIZE);
decoder_cmd = alloc_cmd(SFRAME_SIZE);
if (!encoder_cmd || !decoder_cmd) {
res = -ENOMEM;
goto error_exit;
}
encoder_cmd->complete = &encoder_done;
decoder_cmd->complete = &decoder_done;
send_voip_vopena_cmd(encoder_pvt, encoder_cmd, complicated);
send_voip_vopena_cmd(decoder_pvt, decoder_cmd, simple);
wait_for_completion(&decoder_done);
wait_for_completion(&encoder_done);
if ((0x0000 != response_header(encoder_cmd)->cmd.params[0]) ||
(0x0000 != response_header(decoder_cmd)->cmd.params[0]))
res = -EIO;
error_exit:
free_cmd(encoder_cmd);
free_cmd(decoder_cmd);
return res;
}
static int
wctc4xxp_create_channel_pair(struct wcdte *wc, struct channel_pvt *cpvt,
u8 simple, u8 complicated)
{
struct channel_pvt *encoder_pvt, *decoder_pvt;
u16 encoder_timeslot, decoder_timeslot;
u16 encoder_channel, decoder_channel;
struct tcb *cmd;
u16 length;
cmd = alloc_cmd(SFRAME_SIZE);
if (!cmd)
return -ENOMEM;
BUG_ON(!wc || !cpvt);
if (cpvt->encoder) {
encoder_timeslot = cpvt->timeslot_in_num;
decoder_timeslot = cpvt->timeslot_out_num;
} else {
u8 temp;
encoder_timeslot = cpvt->timeslot_out_num;
decoder_timeslot = cpvt->timeslot_in_num;
temp = simple;
simple = complicated;
complicated = temp;
}
length = (DTE_FORMAT_G729A == complicated) ? G729_LENGTH :
(DTE_FORMAT_G723_1 == complicated) ? G723_LENGTH : 0;
BUG_ON(encoder_timeslot/2 >= wc->numchannels);
BUG_ON(decoder_timeslot/2 >= wc->numchannels);
encoder_pvt = wc->uencode->channels[encoder_timeslot/2].pvt;
decoder_pvt = wc->udecode->channels[decoder_timeslot/2].pvt;
BUG_ON(!encoder_pvt);
BUG_ON(!decoder_pvt);
encoder_pvt->last_rx_seq_num = 0xff;
decoder_pvt->last_rx_seq_num = 0xff;
WARN_ON(encoder_timeslot == decoder_timeslot);
/* First, let's create two channels, one for the simple -> complex
* encoder and another for the complex->simple decoder. */
if (send_create_channel_cmd(wc, cmd, encoder_timeslot,
&encoder_channel))
goto error_exit;
if (send_create_channel_cmd(wc, cmd, decoder_timeslot,
&decoder_channel))
goto error_exit;
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"DTE is using the following channels encoder_channel: " \
"%d decoder_channel: %d\n", encoder_channel, decoder_channel);
WARN_ON(encoder_channel == decoder_channel);
/* Now set all the default parameters for the encoder. */
encoder_pvt->chan_in_num = encoder_channel;
encoder_pvt->chan_out_num = decoder_channel;
decoder_pvt->chan_in_num = decoder_channel;
decoder_pvt->chan_out_num = encoder_channel;
if (wctc4xxp_setup_channels(wc, encoder_pvt, decoder_pvt, length))
goto error_exit;
if (send_trans_connect_cmd(wc, cmd, encoder_channel,
decoder_channel, complicated, simple))
goto error_exit;
if (wctc4xxp_enable_channels(wc, encoder_pvt, decoder_pvt,
complicated, simple))
goto error_exit;
DTE_DEBUG(DTE_DEBUG_CHANNEL_SETUP,
"DTE has completed setup and connected the " \
"two channels together.\n");
free_cmd(cmd);
return 0;
error_exit:
free_cmd(cmd);
return -EIO;
}
static void print_vceinfo_packet(struct wcdte *wc, struct tcb *cmd)
{
int i;
struct device *const dev = &wc->pdev->dev;
static const struct {
const char *name;
bool show;
} PARAMETERS[] = {
{ "Format Revision ", false},
{ "Reserved ", false},
{ "Call Timer (seconds) ", false},
{ "Current Playout Delay [to PCM] ", false},
{ "Minimum Playout Delay [to PCM] ", false},
{ "Maximum Playout Delay [to PCM] ", false},
{ "Clock Offset ", false},
{ "PeakJitter (ms) ", true},
{ "Interpolative Concealment [to PCM] ", false},
{ "Silence Concealment [to PCM] ", false},
{ "Jitter Buffer Overflow Discard [from IP] ", true},
{ "End-point Detection Errors ", true},
{ "Number of Tx Voice Packets [to IP] ", true},
{ "Number of Tx Signalling Packets [to IP] ", true},
{ "Number of Tx Comfort Noise Packets [to IP] ", true},
{ "Total Transmit Duration [to IP] ", true},
{ "Voice Transmit Duration [to IP] ", true},
{ "Number of Rx Voice Packets [from IP] ", true},
{ "Number of Rx Signalling Packets [from IP] ", true},
{ "Number of Rx Comfort Noise Packets [from IP] ", true},
{ "Total Receive Duration [from IP] ", true},
{ "Voice Receive Duration [from IP] ", true},
{ "Packets Out of Sequence [from IP] ", true},
{ "Bad Protocol Headers [from IP] ", true},
{ "Late Packets [from IP] ", true},
{ "Reserved (Early Packets) always zero ", false},
{ "Number of Rx Voice bytes ", true},
{ "Number of Lost Packets [from IP] ", true},
{ "Current Transmit Power [from PCM] ", false},
{ "Mean Transmit Power [from PCM] ", false},
{ "Current Receive Power [to PCM] ", false},
{ "Mean Receive Power [to PCM] ", false},
{ "Background Noise [to PCM] ", false},
{ "ERL Level [to PCM] ", false},
{ "ACOM Level [from PCM] ", false},
{ "Current Transmit Activity [from PCM] ", false},
{ "Current Receive Activity [to PCM] ", false},
{ "Discarded Unexpected Packets ", true},
{ "Discard Packets Due to Rx Disabled ", true},
{ "Discarded Duplicate Packets ", true},
{ "Discarded Packets Due to Incorrect Payload Length ", true},
{ "Discarded Packets Due to Channel Inactive ", true},
{ "Discarded Packets Due to Insufficient Memory ", true}
};
u32 *parms = (u32 *)(&response_header(cmd)->cmd.params[0]);
for (i = 0; i < 43; ++i) {
if (PARAMETERS[i].show)
dev_info(dev, "%s%d\n", PARAMETERS[i].name, parms[i]);
}
}
static void print_eth_statistics_packet(struct wcdte *wc, struct tcb *cmd)
{
int i;
struct device *const dev = &wc->pdev->dev;
static const struct {
const char *name;
bool show;
} PARAMETERS[] = {
{ "Format Revision ", true},
{ "Emitted Frames ", true},
{ "Received Frames ", true},
{ "Unknown Packet Type ", true},
{ "Received Broadcast Packets ", true},
{ "Unknown Broadcast ", true},
{ "Emitted VLAN frames ", true},
{ "Received VLAN frames ", true},
{ "Received VLAN frames with E-RIF ", true}
};
u32 *parms = (u32 *)(&response_header(cmd)->cmd.params[0]);
for (i = 0; i < sizeof(PARAMETERS)/sizeof(PARAMETERS[0]); ++i) {
if (PARAMETERS[i].show)
dev_info(dev, "%s%d\n", PARAMETERS[i].name, parms[i]);
}
}
static int
wctc4xxp_destroy_channel_pair(struct wcdte *wc, struct channel_pvt *cpvt)
{
struct dahdi_transcoder_channel *dtc1, *dtc2;
struct channel_pvt *encoder_pvt, *decoder_pvt;
int chan1, chan2, timeslot1, timeslot2;
struct tcb *cmd;
cmd = alloc_cmd(SFRAME_SIZE);
if (!cmd)
return -ENOMEM;
if (cpvt->encoder) {
chan1 = cpvt->chan_in_num;
timeslot1 = cpvt->timeslot_in_num;
chan2 = cpvt->chan_out_num;
timeslot2 = cpvt->timeslot_out_num;
} else {
chan1 = cpvt->chan_out_num;
timeslot1 = cpvt->timeslot_out_num;
chan2 = cpvt->chan_in_num;
timeslot2 = cpvt->timeslot_in_num;
}
if (timeslot1/2 >= wc->numchannels || timeslot2/2 >= wc->numchannels) {
dev_warn(&wc->pdev->dev,
"Invalid channel numbers in %s. chan1:%d chan2: %d\n",
__func__, timeslot1/2, timeslot2/2);
return 0;
}
dtc1 = &(wc->uencode->channels[timeslot1/2]);
dtc2 = &(wc->udecode->channels[timeslot2/2]);
encoder_pvt = dtc1->pvt;
decoder_pvt = dtc2->pvt;
if (debug & DTE_DEBUG_ETH_STATS) {
if (send_voip_vceinfo_cmd(encoder_pvt, cmd))
goto error_exit;
dev_warn(&wc->pdev->dev,
"****************************************\n");
dev_warn(&wc->pdev->dev,
"Encoder stats (ch: %d):\n",
encoder_pvt->timeslot_in_num);
print_vceinfo_packet(wc, cmd);
if (send_voip_vceinfo_cmd(decoder_pvt, cmd))
goto error_exit;
dev_warn(&wc->pdev->dev,
"****************************************\n");
dev_warn(&wc->pdev->dev,
"Decoder stats (ch: %d):\n",
decoder_pvt->timeslot_in_num);
print_vceinfo_packet(wc, cmd);
}
if (send_voip_vopena_close_cmd(encoder_pvt, cmd))
goto error_exit;
if (send_voip_vopena_close_cmd(decoder_pvt, cmd))
goto error_exit;
if (send_trans_disconnect_cmd(wc, cmd, chan1, chan2, 0, 0))
goto error_exit;
if (send_destroy_channel_cmd(wc, cmd, chan1))
goto error_exit;
if (send_destroy_channel_cmd(wc, cmd, chan2))
goto error_exit;
if (debug & DTE_DEBUG_ETH_STATS) {
if (send_eth_statistics_cmd(wc, cmd))
goto error_exit;
print_eth_statistics_packet(wc, cmd);
dev_info(&wc->pdev->dev, "AN983 tx packets: %d rx packets: %d\n",
wctc4xxp_get_packet_count(wc->txd),
wctc4xxp_get_packet_count(wc->rxd));
}
free_cmd(cmd);
return 0;
error_exit:
free_cmd(cmd);
return -1;
}
static int wctc4xxp_setup_device(struct wcdte *wc)
{
struct tcb *cmd;
int tdm_bus;
cmd = alloc_cmd(SFRAME_SIZE);
if (!cmd)
return -ENOMEM;
if (send_set_arm_clk_cmd(wc, cmd))
goto error_exit;
if (send_set_spu_clk_cmd(wc, cmd))
goto error_exit;
if (send_tdm_select_bus_mode_cmd(wc, cmd))
goto error_exit;
for (tdm_bus = 0; tdm_bus < 4; ++tdm_bus) {
if (send_supvsr_setup_tdm_parms(wc, cmd, tdm_bus))
goto error_exit;
}
if (send_set_eth_header_cmd(wc, cmd, src_mac, dst_mac))
goto error_exit;
if (send_ip_service_config_cmd(wc, cmd))
goto error_exit;
if (send_arp_service_config_cmd(wc, cmd))
goto error_exit;
if (send_icmp_service_config_cmd(wc, cmd))
goto error_exit;
if (send_device_set_country_code_cmd(wc, cmd))
goto error_exit;
if (send_spu_features_control_cmd(wc, cmd, 0x02))
goto error_exit;
if (send_ip_options_cmd(wc, cmd))
goto error_exit;
if (send_spu_features_control_cmd(wc, cmd, 0x04))
goto error_exit;
if (send_csme_multi_cmd(wc, cmd))
goto error_exit;
if (send_tdm_opt_cmd(wc, cmd))
goto error_exit;
free_cmd(cmd);
return 0;
error_exit:
free_cmd(cmd);
return -1;
}
static void wctc4xxp_setup_file_operations(struct file_operations *fops)
{
fops->owner = THIS_MODULE;
fops->read = wctc4xxp_read;
fops->write = wctc4xxp_write;
}
static int
initialize_channel_pvt(struct wcdte *wc, int encoder,
struct channel_pvt **cpvt)
{
int chan;
*cpvt = kmalloc(sizeof(struct channel_pvt) * wc->numchannels,
GFP_KERNEL);
if (!(*cpvt))
return -ENOMEM;
for (chan = 0; chan < wc->numchannels; ++chan)
wctc4xxp_init_state((*cpvt) + chan, encoder, chan, wc);
return 0;
}
static int
initialize_transcoder(struct wcdte *wc, unsigned int srcfmts,
unsigned int dstfmts, struct channel_pvt *pvts,
struct dahdi_transcoder **zt)
{
int chan;
*zt = dahdi_transcoder_alloc(wc->numchannels);
if (!(*zt))
return -ENOMEM;
(*zt)->srcfmts = srcfmts;
(*zt)->dstfmts = dstfmts;
(*zt)->allocate = wctc4xxp_operation_allocate;
(*zt)->release = wctc4xxp_operation_release;
wctc4xxp_setup_file_operations(&((*zt)->fops));
for (chan = 0; chan < wc->numchannels; ++chan)
(*zt)->channels[chan].pvt = &pvts[chan];
return 0;
}
static int initialize_encoders(struct wcdte *wc, unsigned int complexfmts)
{
int res;
res = initialize_channel_pvt(wc, 1, &wc->encoders);
if (res)
return res;
res = initialize_transcoder(wc, DAHDI_FORMAT_ULAW | DAHDI_FORMAT_ALAW,
complexfmts, wc->encoders, &wc->uencode);
if (res)
return res;
sprintf(wc->uencode->name, "DTE Encoder");
return res;
}
static int
initialize_decoders(struct wcdte *wc, unsigned int complexfmts)
{
int res;
res = initialize_channel_pvt(wc, 0, &wc->decoders);
if (res)
return res;
res = initialize_transcoder(wc, complexfmts,
DAHDI_FORMAT_ULAW | DAHDI_FORMAT_ALAW,
wc->decoders, &wc->udecode);
if (res)
return res;
sprintf(wc->udecode->name, "DTE Decoder");
return res;
}
static void
wctc4xxp_send_commands(struct wcdte *wc, struct list_head *to_send)
{
struct tcb *cmd;
while (!list_empty(to_send)) {
cmd = container_of(to_send->next, struct tcb, node);
list_del_init(&cmd->node);
wctc4xxp_transmit_cmd(wc, cmd);
}
}
static void
wctc4xxp_watchdog(TIMER_DATA_TYPE timer)
{
struct wcdte *wc = from_timer(wc, timer, watchdog);
struct tcb *cmd, *temp;
LIST_HEAD(cmds_to_retry);
const int MAX_RETRIES = 5;
int reschedule_timer = 0;
unsigned long flags;
service_tx_ring(wc);
spin_lock_irqsave(&wc->cmd_list_lock, flags);
/* Go through the list of messages that are waiting for responses from
* the DTE, and complete or retry any that have timed out. */
list_for_each_entry_safe(cmd, temp,
&wc->waiting_for_response_list, node) {
if (!time_after(jiffies, cmd->timeout))
continue;
if (++cmd->retries > MAX_RETRIES) {
if (!(cmd->flags & TX_COMPLETE)) {
cmd->flags |= DTE_CMD_TIMEOUT;
list_del_init(&cmd->node);
if (cmd->complete)
complete(cmd->complete);
wctc4xxp_reset_processor(wc);
set_bit(DTE_SHUTDOWN, &wc->flags);
spin_unlock_irqrestore(&wc->cmd_list_lock,
flags);
_wctc4xxp_stop_dma(wc);
dev_err(&wc->pdev->dev,
"Board malfunctioning. Halting operation.\n");
reschedule_timer = 0;
spin_lock_irqsave(&wc->cmd_list_lock, flags);
break;
}
/* ERROR: We've retried the command and
* haven't received the ACK or the response.
*/
cmd->flags |= DTE_CMD_TIMEOUT;
list_del_init(&cmd->node);
if (cmd->complete)
complete(cmd->complete);
} else if (cmd->flags & TX_COMPLETE) {
/* Move this to the local list because we're
* going to resend it once we free the locks
*/
list_move_tail(&cmd->node, &cmds_to_retry);
cmd->flags &= ~(TX_COMPLETE);
} else {
/* The command is still sitting on the tx
* descriptor ring. We don't want to move it
* off any lists, lets just reset the timeout
* and tell the hardware to look for another
* command . */
dev_warn(&wc->pdev->dev,
"Retrying command that was still on descriptor list.\n");
cmd->timeout = jiffies + HZ/4;
wctc4xxp_transmit_demand_poll(wc);
reschedule_timer = 1;
}
}
spin_unlock_irqrestore(&wc->cmd_list_lock, flags);
if (list_empty(&cmds_to_retry) && reschedule_timer)
mod_timer(&wc->watchdog, jiffies + HZ/2);
else if (!list_empty(&cmds_to_retry))
wctc4xxp_send_commands(wc, &cmds_to_retry);
}
/**
* Insert an struct wcdte on the global list in sorted order
*
*/
static int __devinit
wctc4xxp_add_to_device_list(struct wcdte *wc)
{
struct wcdte *cur;
int pos = 0;
INIT_LIST_HEAD(&wc->node);
spin_lock(&wctc4xxp_list_lock);
list_for_each_entry(cur, &wctc4xxp_list, node) {
if (cur->pos != pos) {
/* Add the new entry before the one here */
list_add_tail(&wc->node, &cur->node);
break;
} else {
++pos;
}
}
/* If we didn't already add the new entry to the list, add it now */
if (list_empty(&wc->node))
list_add_tail(&wc->node, &wctc4xxp_list);
spin_unlock(&wctc4xxp_list_lock);
return pos;
}
static void wctc4xxp_remove_from_device_list(struct wcdte *wc)
{
spin_lock(&wctc4xxp_list_lock);
list_del(&wc->node);
spin_unlock(&wctc4xxp_list_lock);
}
struct wctc4xxp_desc {
const char *short_name;
const char *long_name;
};
static struct wctc4xxp_desc wctc400p = {
.short_name = "tc400b",
.long_name = "Wildcard TC400P+TC400M",
};
static struct wctc4xxp_desc wctce400 = {
.short_name = "tce400",
.long_name = "Wildcard TCE400+TC400M",
};
static void wctc4xxp_cleanup_channels(struct wcdte *wc);
static int wctc4xxp_reset_and_reload_firmware(struct wcdte *wc,
const struct firmware *firmware)
{
int res;
wctc4xxp_cleanup_command_list(wc);
wctc4xxp_cleanup_channels(wc);
res = wctc4xxp_boot_processor(wc, firmware);
if (res)
return res;
#if defined(CONFIG_WCTC4XXP_POLLING)
wctc4xxp_enable_polling(wc);
#endif
res = wctc4xxp_setup_device(wc);
if (res) {
dev_err(&wc->pdev->dev, "Failed to setup DTE\n");
return res;
}
return 0;
}
static int wctc4xxp_reset_driver_state(struct wcdte *wc)
{
int res;
struct firmware embedded_firmware;
unsigned long flags;
const struct firmware *firmware = &embedded_firmware;
#if !defined(HOTPLUG_FIRMWARE)
extern void _binary_dahdi_fw_tc400m_bin_size;
extern u8 _binary_dahdi_fw_tc400m_bin_start[];
embedded_firmware.data = _binary_dahdi_fw_tc400m_bin_start;
embedded_firmware.size = (size_t) &_binary_dahdi_fw_tc400m_bin_size;
#else
static const char tc400m_firmware[] = "dahdi-fw-tc400m.bin";
res = request_firmware(&firmware, tc400m_firmware, &wc->pdev->dev);
if (res || !firmware) {
dev_err(&wc->pdev->dev,
"Firmware %s not available from userspace. (%d)\n",
tc400m_firmware, res);
return res;
}
#endif
res = wctc4xxp_reset_and_reload_firmware(wc, firmware);
if (firmware != &embedded_firmware)
release_firmware(firmware);
spin_lock_irqsave(&wc->rxd->lock, flags);
wc->rxd->packet_errors = 0;
wc->reported_packet_errors = 0;
spin_unlock_irqrestore(&wc->rxd->lock, flags);
return res;
}
#ifdef EXPORT_FOR_ALERT_ATTRIBUTE
static ssize_t wctc4xxp_force_alert_store(struct device *dev,
struct device_attribute *attr,
const char *buf, size_t count)
{
int res;
unsigned int alert_type;
struct wcdte *wc = dev_get_drvdata(dev);
struct tcb *cmd = alloc_cmd(SFRAME_SIZE);
u16 parameters[] = {0};
if (!cmd)
return -ENOMEM;
res = sscanf(buf, "%x", &alert_type);
if (1 != res) {
free_cmd(cmd);
return -EINVAL;
}
dev_info(&wc->pdev->dev, "Forcing alert type: 0x%x\n", alert_type);
res = mutex_lock_killable(&wc->chanlock);
if (res) {
free_cmd(cmd);
return -EAGAIN;
}
parameters[0] = alert_type;
create_supervisor_cmd(wc, cmd, CONFIG_CHANGE_TYPE,
CONFIG_DEVICE_CLASS, 0x0409, parameters,
ARRAY_SIZE(parameters));
wctc4xxp_transmit_cmd(wc, cmd);
mutex_unlock(&wc->chanlock);
return count;
}
static DEVICE_ATTR(force_alert, 0200, NULL, wctc4xxp_force_alert_store);
static void wctc4xxp_create_sysfs_files(struct wcdte *wc)
{
int ret;
ret = device_create_file(&wc->pdev->dev, &dev_attr_force_alert);
if (ret) {
dev_info(&wc->pdev->dev,
"Failed to create device attributes.\n");
}
}
static void wctc4xxp_remove_sysfs_files(struct wcdte *wc)
{
device_remove_file(&wc->pdev->dev, &dev_attr_force_alert);
}
#else
static inline void wctc4xxp_create_sysfs_files(struct wcdte *wc) { return; }
static inline void wctc4xxp_remove_sysfs_files(struct wcdte *wc) { return; }
#endif
static int __devinit
wctc4xxp_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
{
int res, reg, position_on_list;
struct wcdte *wc = NULL;
struct wctc4xxp_desc *d = (struct wctc4xxp_desc *)ent->driver_data;
unsigned char g729_numchannels, g723_numchannels, min_numchannels;
unsigned char wctc4xxp_firmware_ver, wctc4xxp_firmware_ver_minor;
unsigned int complexfmts;
struct firmware embedded_firmware;
const struct firmware *firmware = &embedded_firmware;
#if !defined(HOTPLUG_FIRMWARE)
extern void _binary_dahdi_fw_tc400m_bin_size;
extern u8 _binary_dahdi_fw_tc400m_bin_start[];
#else
static const char tc400m_firmware[] = "dahdi-fw-tc400m.bin";
#endif
/* ------------------------------------------------------------------
* Setup the pure software constructs internal to this driver.
* --------------------------------------------------------------- */
wc = kzalloc(sizeof(*wc), GFP_KERNEL);
if (!wc)
return -ENOMEM;
position_on_list = wctc4xxp_add_to_device_list(wc);
snprintf(wc->board_name, sizeof(wc->board_name)-1, "%s%d",
d->short_name, position_on_list);
wc->iobase = pci_iomap(pdev, 1, 0);
wc->pdev = pdev;
wc->pos = position_on_list;
wc->variety = d->long_name;
wc->last_rx_seq_num = -1;
if (!request_mem_region(pci_resource_start(pdev, 1),
pci_resource_len(pdev, 1), wc->board_name)) {
dev_err(&pdev->dev, "IO Registers are in use by another "
"module.\n");
wctc4xxp_remove_from_device_list(wc);
kfree(wc);
return -EIO;
}
mutex_init(&wc->chanlock);
spin_lock_init(&wc->reglock);
spin_lock_init(&wc->cmd_list_lock);
spin_lock_init(&wc->rx_list_lock);
spin_lock_init(&wc->rx_lock);
INIT_LIST_HEAD(&wc->cmd_list);
INIT_LIST_HEAD(&wc->waiting_for_response_list);
INIT_LIST_HEAD(&wc->rx_list);
INIT_WORK(&wc->deferred_work, deferred_work_func);
init_waitqueue_head(&wc->waitq);
if (pci_set_dma_mask(wc->pdev, DMA_BIT_MASK(32))) {
release_mem_region(pci_resource_start(wc->pdev, 1),
pci_resource_len(wc->pdev, 1));
if (wc->iobase)
pci_iounmap(wc->pdev, wc->iobase);
pci_clear_mwi(wc->pdev);
dev_warn(&wc->pdev->dev, "No suitable DMA available.\n");
return -EIO;
}
wc->txd = kmalloc(sizeof(*wc->txd), GFP_KERNEL);
if (!wc->txd) {
res = -ENOMEM;
goto error_exit_swinit;
}
res = wctc4xxp_initialize_descriptor_ring(wc->pdev, wc->txd,
0xe0800000, DMA_TO_DEVICE, DEFAULT_TX_DRING_SIZE);
if (res)
goto error_exit_swinit;
wc->rxd = kmalloc(sizeof(*wc->rxd), GFP_KERNEL);
if (!wc->rxd) {
res = -ENOMEM;
goto error_exit_swinit;
}
res = wctc4xxp_initialize_descriptor_ring(wc->pdev, wc->rxd, 0,
DMA_FROM_DEVICE, DEFAULT_RX_DRING_SIZE);
if (res)
goto error_exit_swinit;
#if defined(HOTPLUG_FIRMWARE)
res = request_firmware(&firmware, tc400m_firmware, &wc->pdev->dev);
if (res || !firmware) {
dev_err(&wc->pdev->dev,
"Firmware %s not available from userspace. (%d)\n",
tc400m_firmware, res);
goto error_exit_swinit;
}
#else
embedded_firmware.data = _binary_dahdi_fw_tc400m_bin_start;
embedded_firmware.size = (size_t) &_binary_dahdi_fw_tc400m_bin_size;
#endif
wctc4xxp_firmware_ver = firmware->data[0];
wctc4xxp_firmware_ver_minor = firmware->data[16];
g729_numchannels = firmware->data[1];
g723_numchannels = firmware->data[2];
min_numchannels = min(g723_numchannels, g729_numchannels);
if (!mode || strlen(mode) < 4) {
sprintf(wc->complexname, "G.729a / G.723.1");
complexfmts = DAHDI_FORMAT_G729A | DAHDI_FORMAT_G723_1;
wc->numchannels = min_numchannels;
} else if (mode[3] == '9') { /* "G.729" */
sprintf(wc->complexname, "G.729a");
complexfmts = DAHDI_FORMAT_G729A;
wc->numchannels = g729_numchannels;
} else if (mode[3] == '3') { /* "G.723.1" */
sprintf(wc->complexname, "G.723.1");
complexfmts = DAHDI_FORMAT_G723_1;
wc->numchannels = g723_numchannels;
} else {
sprintf(wc->complexname, "G.729a / G.723.1");
complexfmts = DAHDI_FORMAT_G729A | DAHDI_FORMAT_G723_1;
wc->numchannels = min_numchannels;
}
res = initialize_encoders(wc, complexfmts);
if (res)
goto error_exit_swinit;
res = initialize_decoders(wc, complexfmts);
if (res)
goto error_exit_swinit;
if (DTE_DEBUG_NETWORK_IF & debug) {
res = wctc4xxp_net_register(wc);
if (res)
goto error_exit_swinit;
}
timer_setup(&wc->watchdog, wctc4xxp_watchdog, 0);
/* ------------------------------------------------------------------
* Load the firmware and start the DTE.
* --------------------------------------------------------------- */
res = pci_enable_device(pdev);
if (res) {
dev_err(&wc->pdev->dev, "Failed to enable device.\n");
goto error_exit_swinit;;
}
res = pci_set_mwi(wc->pdev);
if (res) {
dev_warn(&wc->pdev->dev,
"Failed to set Memory-Write Invalidate Command Bit..\n");
}
pci_set_master(pdev);
pci_set_drvdata(pdev, wc);
res = request_irq(pdev->irq, wctc4xxp_interrupt,
IRQF_SHARED, wc->board_name, wc);
if (res) {
dev_err(&wc->pdev->dev,
"Unable to request IRQ %d\n", pdev->irq);
if (firmware != &embedded_firmware)
release_firmware(firmware);
goto error_exit_hwinit;
}
res = wctc4xxp_hardware_init(wc);
if (res) {
if (firmware != &embedded_firmware)
release_firmware(firmware);
goto error_exit_hwinit;
}
wctc4xxp_enable_interrupts(wc);
wctc4xxp_start_dma(wc);
res = wctc4xxp_reset_and_reload_firmware(wc, firmware);
if (firmware != &embedded_firmware)
release_firmware(firmware);
if (res)
goto error_exit_hwinit;
/* \todo Read firmware version directly from tc400b.*/
dev_info(&wc->pdev->dev, "(%s) Transcoder support LOADED " \
"(firm ver = %d.%d)\n", wc->complexname, wctc4xxp_firmware_ver,
wctc4xxp_firmware_ver_minor);
reg = wctc4xxp_getctl(wc, 0x00fc);
DTE_DEBUG(DTE_DEBUG_GENERAL,
"debug: (post-boot) Reg fc is %08x\n", reg);
dev_info(&wc->pdev->dev, "Installed a Wildcard TC: %s\n", wc->variety);
DTE_DEBUG(DTE_DEBUG_GENERAL, "Operating in DEBUG mode.\n");
dahdi_transcoder_register(wc->uencode);
dahdi_transcoder_register(wc->udecode);
wctc4xxp_match_packet_counts(wc);
wctc4xxp_create_sysfs_files(wc);
return 0;
error_exit_hwinit:
#if defined(CONFIG_WCTC4XXP_POLLING)
wctc4xxp_disable_polling(wc);
#endif
wctc4xxp_stop_dma(wc);
wctc4xxp_cleanup_command_list(wc);
free_irq(pdev->irq, wc);
pci_set_drvdata(pdev, NULL);
error_exit_swinit:
wctc4xxp_net_unregister(wc);
kfree(wc->encoders);
kfree(wc->decoders);
dahdi_transcoder_free(wc->uencode);
dahdi_transcoder_free(wc->udecode);
wctc4xxp_cleanup_descriptor_ring(wc->txd);
kfree(wc->txd);
wctc4xxp_cleanup_descriptor_ring(wc->rxd);
kfree(wc->rxd);
release_mem_region(pci_resource_start(wc->pdev, 1),
pci_resource_len(wc->pdev, 1));
if (wc->iobase)
pci_iounmap(wc->pdev, wc->iobase);
pci_clear_mwi(wc->pdev);
wctc4xxp_remove_from_device_list(wc);
kfree(wc);
return res;
}
static void wctc4xxp_cleanup_channels(struct wcdte *wc)
{
int i;
struct dahdi_transcoder_channel *dtc_en, *dtc_de;
struct channel_pvt *cpvt;
for (i = 0; i < wc->numchannels; ++i) {
dtc_en = &(wc->uencode->channels[i]);
wctc4xxp_cleanup_channel_private(wc, dtc_en);
dahdi_tc_clear_busy(dtc_en);
dahdi_tc_clear_built(dtc_en);
dtc_en->built_fmts = 0;
cpvt = dtc_en->pvt;
cpvt->chan_in_num = INVALID;
cpvt->chan_out_num = INVALID;
dtc_de = &(wc->udecode->channels[i]);
wctc4xxp_cleanup_channel_private(wc, dtc_de);
dahdi_tc_clear_busy(dtc_de);
dahdi_tc_clear_built(dtc_de);
dtc_de->built_fmts = 0;
cpvt = dtc_de->pvt;
cpvt->chan_in_num = INVALID;
cpvt->chan_out_num = INVALID;
}
}
static void __devexit wctc4xxp_remove_one(struct pci_dev *pdev)
{
struct wcdte *wc = pci_get_drvdata(pdev);
if (!wc)
return;
wctc4xxp_remove_sysfs_files(wc);
wctc4xxp_remove_from_device_list(wc);
set_bit(DTE_SHUTDOWN, &wc->flags);
if (del_timer_sync(&wc->watchdog))
del_timer_sync(&wc->watchdog);
/* This should already be stopped, but it doesn't hurt to make sure. */
clear_bit(DTE_POLLING, &wc->flags);
wctc4xxp_net_unregister(wc);
/* Stop any DMA */
wctc4xxp_stop_dma(wc);
/* In case hardware is still there */
wctc4xxp_disable_interrupts(wc);
free_irq(pdev->irq, wc);
/* There isn't anything that would run in the workqueue that will wait
* on an interrupt. */
dahdi_transcoder_unregister(wc->udecode);
dahdi_transcoder_unregister(wc->uencode);
/* Free Resources */
release_mem_region(pci_resource_start(wc->pdev, 1),
pci_resource_len(wc->pdev, 1));
if (wc->iobase)
pci_iounmap(wc->pdev, wc->iobase);
pci_clear_mwi(wc->pdev);
wctc4xxp_cleanup_descriptor_ring(wc->txd);
kfree(wc->txd);
wctc4xxp_cleanup_descriptor_ring(wc->rxd);
kfree(wc->rxd);
wctc4xxp_cleanup_command_list(wc);
wctc4xxp_cleanup_channels(wc);
pci_set_drvdata(pdev, NULL);
dahdi_transcoder_free(wc->uencode);
dahdi_transcoder_free(wc->udecode);
kfree(wc->encoders);
kfree(wc->decoders);
kfree(wc);
}
static DEFINE_PCI_DEVICE_TABLE(wctc4xxp_pci_tbl) = {
{ 0xd161, 0x3400, PCI_ANY_ID, PCI_ANY_ID,
0, 0, (unsigned long) &wctc400p }, /* Digium board */
{ 0xd161, 0x8004, PCI_ANY_ID, PCI_ANY_ID,
0, 0, (unsigned long) &wctce400 }, /* Digium board */
{ 0 }
};
MODULE_DEVICE_TABLE(pci, wctc4xxp_pci_tbl);
static int wctc4xxp_suspend(struct pci_dev *pdev, pm_message_t state)
{
return -ENOSYS;
}
static struct pci_driver wctc4xxp_driver = {
.name = "wctc4xxp",
.probe = wctc4xxp_init_one,
.remove = __devexit_p(wctc4xxp_remove_one),
.id_table = wctc4xxp_pci_tbl,
.suspend = wctc4xxp_suspend,
};
static int __init wctc4xxp_init(void)
{
int res;
unsigned long cache_flags;
cache_flags = SLAB_HWCACHE_ALIGN;
cmd_cache = kmem_cache_create(THIS_MODULE->name, sizeof(struct tcb),
0, cache_flags, NULL);
if (!cmd_cache)
return -ENOMEM;
spin_lock_init(&wctc4xxp_list_lock);
INIT_LIST_HEAD(&wctc4xxp_list);
res = pci_register_driver(&wctc4xxp_driver);
if (res) {
kmem_cache_destroy(cmd_cache);
return -ENODEV;
}
return 0;
}
static void __exit wctc4xxp_cleanup(void)
{
pci_unregister_driver(&wctc4xxp_driver);
kmem_cache_destroy(cmd_cache);
}
module_param(debug, int, S_IRUGO | S_IWUSR);
module_param(mode, charp, S_IRUGO);
MODULE_PARM_DESC(mode, "'g729', 'g723.1', or 'any'. Default 'any'.");
MODULE_DESCRIPTION("Wildcard TC400P+TC400M Driver");
MODULE_AUTHOR("Digium Incorporated <support@digium.com>");
MODULE_LICENSE("GPL");
module_init(wctc4xxp_init);
module_exit(wctc4xxp_cleanup);