wctc4xxp: Allow the tx and rx descriptor rings to be different sizes
Keeping the transmit descriptor ring shorter reduces the time it takes to send CSM_ENCAP commands to the transcoding engine when the card is otherwise busy. Signed-off-by: Shaun Ruffell <sruffell@digium.com> Signed-off-by: Russ Meyerriecks <rmeyerriecks@digium.com>
This commit is contained in:
parent
b6a8623203
commit
ba05e31c8a
@ -182,8 +182,11 @@ struct csm_encaps_hdr {
|
|||||||
#define MAX_FRAME_SIZE 1518
|
#define MAX_FRAME_SIZE 1518
|
||||||
#define SFRAME_SIZE MAX_FRAME_SIZE
|
#define SFRAME_SIZE MAX_FRAME_SIZE
|
||||||
|
|
||||||
#define DRING_SIZE (1 << 7) /* Must be a power of two */
|
#define DEFAULT_RX_DRING_SIZE (1 << 6) /* Must be a power of two */
|
||||||
#define DRING_MASK (DRING_SIZE-1)
|
|
||||||
|
/* Keep the TX ring shorter in order to reduce the amount of time needed to
|
||||||
|
* bring up channels when sending high priority csm_encaps packets. */
|
||||||
|
#define DEFAULT_TX_DRING_SIZE (1 << 4) /* Must be a power of two */
|
||||||
#define MIN_PACKET_LEN 64
|
#define MIN_PACKET_LEN 64
|
||||||
|
|
||||||
/* Transcoder buffer (tcb) */
|
/* Transcoder buffer (tcb) */
|
||||||
@ -765,7 +768,7 @@ struct wctc4xxp_descriptor_ring {
|
|||||||
/* Write ready buffers to the tail. */
|
/* Write ready buffers to the tail. */
|
||||||
unsigned int tail;
|
unsigned int tail;
|
||||||
/* Array to save the kernel virtual address of pending commands. */
|
/* Array to save the kernel virtual address of pending commands. */
|
||||||
struct tcb *pending[DRING_SIZE];
|
struct tcb **pending;
|
||||||
/* PCI Bus address of the descriptor list. */
|
/* PCI Bus address of the descriptor list. */
|
||||||
dma_addr_t desc_dma;
|
dma_addr_t desc_dma;
|
||||||
/*! either DMA_FROM_DEVICE or DMA_TO_DEVICE */
|
/*! either DMA_FROM_DEVICE or DMA_TO_DEVICE */
|
||||||
@ -778,6 +781,8 @@ struct wctc4xxp_descriptor_ring {
|
|||||||
spinlock_t lock;
|
spinlock_t lock;
|
||||||
/*! PCI device for the card associated with this ring. */
|
/*! PCI device for the card associated with this ring. */
|
||||||
struct pci_dev *pdev;
|
struct pci_dev *pdev;
|
||||||
|
/*! The size of the dring. */
|
||||||
|
unsigned long size;
|
||||||
};
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@ -797,13 +802,14 @@ wctc4xxp_descriptor(struct wctc4xxp_descriptor_ring *dr, int index)
|
|||||||
|
|
||||||
static int
|
static int
|
||||||
wctc4xxp_initialize_descriptor_ring(struct pci_dev *pdev,
|
wctc4xxp_initialize_descriptor_ring(struct pci_dev *pdev,
|
||||||
struct wctc4xxp_descriptor_ring *dr, u32 des1, unsigned int direction)
|
struct wctc4xxp_descriptor_ring *dr, u32 des1, unsigned int direction,
|
||||||
|
unsigned long size)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
const u32 END_OF_RING = 0x02000000;
|
const u32 END_OF_RING = 0x02000000;
|
||||||
u8 cache_line_size = 0;
|
u8 cache_line_size = 0;
|
||||||
struct wctc4xxp_descriptor *d;
|
|
||||||
int add_padding;
|
int add_padding;
|
||||||
|
struct wctc4xxp_descriptor *d = NULL;
|
||||||
|
|
||||||
BUG_ON(!pdev);
|
BUG_ON(!pdev);
|
||||||
BUG_ON(!dr);
|
BUG_ON(!dr);
|
||||||
@ -812,6 +818,7 @@ wctc4xxp_initialize_descriptor_ring(struct pci_dev *pdev,
|
|||||||
return -EIO;
|
return -EIO;
|
||||||
|
|
||||||
memset(dr, 0, sizeof(*dr));
|
memset(dr, 0, sizeof(*dr));
|
||||||
|
dr->size = size;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Add some padding to each descriptor to ensure that they are
|
* Add some padding to each descriptor to ensure that they are
|
||||||
@ -825,14 +832,19 @@ wctc4xxp_initialize_descriptor_ring(struct pci_dev *pdev,
|
|||||||
if (add_padding)
|
if (add_padding)
|
||||||
dr->padding = (cache_line_size*sizeof(u32)) - sizeof(*d);
|
dr->padding = (cache_line_size*sizeof(u32)) - sizeof(*d);
|
||||||
|
|
||||||
dr->desc = pci_alloc_consistent(pdev,
|
dr->pending = kmalloc(sizeof(struct tcb *) * dr->size, GFP_KERNEL);
|
||||||
(sizeof(*d)+dr->padding)*DRING_SIZE, &dr->desc_dma);
|
if (!dr->pending)
|
||||||
|
|
||||||
if (!dr->desc)
|
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
memset(dr->desc, 0, (sizeof(*d) + dr->padding) * DRING_SIZE);
|
dr->desc = pci_alloc_consistent(pdev,
|
||||||
for (i = 0; i < DRING_SIZE; ++i) {
|
(sizeof(*d)+dr->padding)*dr->size, &dr->desc_dma);
|
||||||
|
if (!dr->desc) {
|
||||||
|
kfree(dr->pending);
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
|
memset(dr->desc, 0, (sizeof(*d) + dr->padding) * dr->size);
|
||||||
|
for (i = 0; i < dr->size; ++i) {
|
||||||
d = wctc4xxp_descriptor(dr, i);
|
d = wctc4xxp_descriptor(dr, i);
|
||||||
memset(d, 0, sizeof(*d));
|
memset(d, 0, sizeof(*d));
|
||||||
d->des1 = cpu_to_le32(des1);
|
d->des1 = cpu_to_le32(des1);
|
||||||
@ -880,7 +892,7 @@ wctc4xxp_submit(struct wctc4xxp_descriptor_ring *dr, struct tcb *c)
|
|||||||
|
|
||||||
SET_OWNED(d); /* That's it until the hardware is done with it. */
|
SET_OWNED(d); /* That's it until the hardware is done with it. */
|
||||||
dr->pending[dr->tail] = c;
|
dr->pending[dr->tail] = c;
|
||||||
dr->tail = (dr->tail + 1) & DRING_MASK;
|
dr->tail = (dr->tail + 1) & (dr->size-1);
|
||||||
++dr->count;
|
++dr->count;
|
||||||
spin_unlock_irqrestore(&dr->lock, flags);
|
spin_unlock_irqrestore(&dr->lock, flags);
|
||||||
return 0;
|
return 0;
|
||||||
@ -900,7 +912,7 @@ wctc4xxp_retrieve(struct wctc4xxp_descriptor_ring *dr)
|
|||||||
SFRAME_SIZE, dr->direction);
|
SFRAME_SIZE, dr->direction);
|
||||||
c = dr->pending[head];
|
c = dr->pending[head];
|
||||||
WARN_ON(!c);
|
WARN_ON(!c);
|
||||||
dr->head = (++head) & DRING_MASK;
|
dr->head = (++head) & (dr->size-1);
|
||||||
d->buffer1 = 0;
|
d->buffer1 = 0;
|
||||||
--dr->count;
|
--dr->count;
|
||||||
WARN_ON(!c);
|
WARN_ON(!c);
|
||||||
@ -1484,7 +1496,7 @@ wctc4xxp_cleanup_descriptor_ring(struct wctc4xxp_descriptor_ring *dr)
|
|||||||
if (!dr || !dr->desc)
|
if (!dr || !dr->desc)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
for (i = 0; i < DRING_SIZE; ++i) {
|
for (i = 0; i < dr->size; ++i) {
|
||||||
d = wctc4xxp_descriptor(dr, i);
|
d = wctc4xxp_descriptor(dr, i);
|
||||||
if (d->buffer1) {
|
if (d->buffer1) {
|
||||||
pci_unmap_single(dr->pdev, d->buffer1,
|
pci_unmap_single(dr->pdev, d->buffer1,
|
||||||
@ -1501,8 +1513,9 @@ wctc4xxp_cleanup_descriptor_ring(struct wctc4xxp_descriptor_ring *dr)
|
|||||||
dr->head = 0;
|
dr->head = 0;
|
||||||
dr->tail = 0;
|
dr->tail = 0;
|
||||||
dr->count = 0;
|
dr->count = 0;
|
||||||
pci_free_consistent(dr->pdev, (sizeof(*d)+dr->padding) * DRING_SIZE,
|
pci_free_consistent(dr->pdev, (sizeof(*d)+dr->padding) * dr->size,
|
||||||
dr->desc, dr->desc_dma);
|
dr->desc, dr->desc_dma);
|
||||||
|
kfree(dr->pending);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void wctc4xxp_timeout_all_commands(struct wcdte *wc)
|
static void wctc4xxp_timeout_all_commands(struct wcdte *wc)
|
||||||
@ -2697,7 +2710,7 @@ wctc4xxp_start_dma(struct wcdte *wc)
|
|||||||
u32 reg;
|
u32 reg;
|
||||||
struct tcb *cmd;
|
struct tcb *cmd;
|
||||||
|
|
||||||
for (i = 0; i < DRING_SIZE; ++i) {
|
for (i = 0; i < wc->rxd->size; ++i) {
|
||||||
cmd = alloc_cmd(SFRAME_SIZE);
|
cmd = alloc_cmd(SFRAME_SIZE);
|
||||||
if (!cmd) {
|
if (!cmd) {
|
||||||
WARN_ALWAYS();
|
WARN_ALWAYS();
|
||||||
@ -3463,7 +3476,7 @@ wctc4xxp_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||||||
}
|
}
|
||||||
|
|
||||||
res = wctc4xxp_initialize_descriptor_ring(wc->pdev, wc->txd,
|
res = wctc4xxp_initialize_descriptor_ring(wc->pdev, wc->txd,
|
||||||
0xe0800000, DMA_TO_DEVICE);
|
0xe0800000, DMA_TO_DEVICE, DEFAULT_TX_DRING_SIZE);
|
||||||
if (res)
|
if (res)
|
||||||
goto error_exit_swinit;
|
goto error_exit_swinit;
|
||||||
|
|
||||||
@ -3474,7 +3487,7 @@ wctc4xxp_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||||||
}
|
}
|
||||||
|
|
||||||
res = wctc4xxp_initialize_descriptor_ring(wc->pdev, wc->rxd, 0,
|
res = wctc4xxp_initialize_descriptor_ring(wc->pdev, wc->rxd, 0,
|
||||||
DMA_FROM_DEVICE);
|
DMA_FROM_DEVICE, DEFAULT_RX_DRING_SIZE);
|
||||||
if (res)
|
if (res)
|
||||||
goto error_exit_swinit;
|
goto error_exit_swinit;
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user