avf: deal willth tx ring full properly

Change-Id: I0b38e129ee01d212463253e5526bfc2d60fd88af
Signed-off-by: Damjan Marion <damarion@cisco.com>
This commit is contained in:
Damjan Marion
2018-08-28 19:40:30 +02:00
committed by Damjan Marion
parent f2de90d724
commit fcf9497d3b
3 changed files with 52 additions and 11 deletions
+11
View File
@@ -288,6 +288,17 @@ typedef struct
avf_rx_vector_entry_t rxve;
} avf_input_trace_t;
#define foreach_avf_tx_func_error \
_(NO_FREE_SLOTS, "no free tx slots")
typedef enum
{
#define _(f,s) AVF_TX_ERROR_##f,
foreach_avf_tx_func_error
#undef _
AVF_TX_N_ERROR,
} avf_tx_func_error_t;
#endif /* AVF_H */
/*
+8
View File
@@ -1314,6 +1314,12 @@ avf_set_interface_next_node (vnet_main_t * vnm, u32 hw_if_index,
vlib_node_add_next (vlib_get_main (), avf_input_node.index, node_index);
}
static char *avf_tx_func_error_strings[] = {
#define _(n,s) s,
foreach_avf_tx_func_error
#undef _
};
/* *INDENT-OFF* */
VNET_DEVICE_CLASS (avf_device_class,) =
{
@@ -1323,6 +1329,8 @@ VNET_DEVICE_CLASS (avf_device_class,) =
.admin_up_down_function = avf_interface_admin_up_down,
.rx_mode_change_function = avf_interface_rx_mode_change,
.rx_redirect_to_node = avf_set_interface_next_node,
.tx_function_n_errors = AVF_TX_N_ERROR,
.tx_function_error_strings = avf_tx_func_error_strings,
};
/* *INDENT-ON* */
+33 -11
View File
@@ -47,19 +47,25 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm,
avf_tx_desc_t *d0, *d1, *d2, *d3;
u32 *buffers = vlib_frame_args (frame);
u32 bi0, bi1, bi2, bi3;
u16 n_left = frame->n_vectors;
u16 n_left, n_left_to_send, n_in_batch;
vlib_buffer_t *b0, *b1, *b2, *b3;
u16 next;
u16 n_retry = 5;
u16 mask = txq->size - 1;
u64 bits = (AVF_TXQ_DESC_CMD_EOP | AVF_TXQ_DESC_CMD_RS |
AVF_TXQ_DESC_CMD_RSV);
clib_spinlock_lock_if_init (&txq->lock);
n_left_to_send = frame->n_vectors;
next = txq->next;
retry:
/* release cosumed bufs */
if (txq->n_enqueued)
{
u16 first, slot, n_free = 0;
first = slot = (txq->next - txq->n_enqueued) & mask;
first = slot = (next - txq->n_enqueued) & mask;
d0 = txq->descs + slot;
while (n_free < txq->n_enqueued && avf_tx_desc_get_dtyp (d0) == 0x0F)
{
@@ -76,6 +82,9 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm,
}
}
n_in_batch = clib_min (n_left_to_send, txq->size - txq->n_enqueued - 8);
n_left = n_in_batch;
while (n_left >= 8)
{
u16 slot0, slot1, slot2, slot3;
@@ -85,10 +94,10 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm,
vlib_prefetch_buffer_with_index (vm, buffers[6], LOAD);
vlib_prefetch_buffer_with_index (vm, buffers[7], LOAD);
slot0 = txq->next;
slot1 = (txq->next + 1) & mask;
slot2 = (txq->next + 2) & mask;
slot3 = (txq->next + 3) & mask;
slot0 = next;
slot1 = (next + 1) & mask;
slot2 = (next + 2) & mask;
slot3 = (next + 3) & mask;
d0 = txq->descs + slot0;
d1 = txq->descs + slot1;
@@ -124,7 +133,7 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm,
d2->qword[1] = ((u64) b2->current_length) << 34 | bits;
d3->qword[1] = ((u64) b3->current_length) << 34 | bits;
txq->next = (txq->next + 4) & mask;
next = (next + 4) & mask;
txq->n_enqueued += 4;
buffers += 4;
n_left -= 4;
@@ -132,9 +141,9 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm,
while (n_left)
{
d0 = txq->descs + txq->next;
d0 = txq->descs + next;
bi0 = buffers[0];
txq->bufs[txq->next] = bi0;
txq->bufs[next] = bi0;
b0 = vlib_get_buffer (vm, bi0);
#if 0
@@ -145,13 +154,26 @@ VNET_DEVICE_CLASS_TX_FN (avf_device_class) (vlib_main_t * vm,
#endif
d0->qword[1] = (((u64) b0->current_length) << 34) | bits;
txq->next = (txq->next + 1) & mask;
next = (next + 1) & mask;
txq->n_enqueued++;
buffers++;
n_left--;
}
CLIB_MEMORY_BARRIER ();
*(txq->qtx_tail) = txq->next;
*(txq->qtx_tail) = txq->next = next;
n_left_to_send -= n_in_batch;
if (n_left_to_send)
{
if (n_retry--)
goto retry;
vlib_buffer_free (vm, buffers, n_left_to_send);
vlib_error_count (vm, node->node_index,
AVF_TX_ERROR_NO_FREE_SLOTS, n_left_to_send);
}
clib_spinlock_unlock_if_init (&txq->lock);