MPLS Nodes Dual Loop

Change-Id: Ic54d4cb9dec8e91446b9b4d2b40ed69a14bd4355
Signed-off-by: Neale Ranns <nranns@cisco.com>
This commit is contained in:
Neale Ranns
2016-12-10 21:08:09 +00:00
committed by Dave Barach
parent fa0fb5821a
commit 9ca18c604a
4 changed files with 396 additions and 17 deletions

View File

@ -54,6 +54,7 @@ mpls_label_dpo_create (mpls_label_t *label_stack,
mld = mpls_label_dpo_alloc();
mld->mld_n_labels = vec_len(label_stack);
mld->mld_n_hdr_bytes = mld->mld_n_labels * sizeof(mld->mld_hdr[0]);
mld->mld_payload_proto = payload_proto;
/*
@ -179,6 +180,163 @@ mpls_label_imposition_inline (vlib_main_t * vm,
vlib_get_next_frame(vm, node, next_index, to_next, n_left_to_next);
while (n_left_from >= 4 && n_left_to_next >= 2)
{
mpls_unicast_header_t *hdr0, *hdr1;
mpls_label_dpo_t *mld0, *mld1;
u32 bi0, mldi0, bi1, mldi1;
vlib_buffer_t * b0, *b1;
u32 next0, next1;
u8 ttl0, ttl1;
bi0 = to_next[0] = from[0];
bi1 = to_next[1] = from[1];
/* Prefetch next iteration. */
{
vlib_buffer_t * p2, * p3;
p2 = vlib_get_buffer (vm, from[2]);
p3 = vlib_get_buffer (vm, from[3]);
vlib_prefetch_buffer_header (p2, STORE);
vlib_prefetch_buffer_header (p3, STORE);
CLIB_PREFETCH (p2->data, sizeof (hdr0[0]), STORE);
CLIB_PREFETCH (p3->data, sizeof (hdr0[0]), STORE);
}
from += 2;
to_next += 2;
n_left_from -= 2;
n_left_to_next -= 2;
b0 = vlib_get_buffer (vm, bi0);
b1 = vlib_get_buffer (vm, bi1);
/* dst lookup was done by ip4 lookup */
mldi0 = vnet_buffer(b0)->ip.adj_index[VLIB_TX];
mldi1 = vnet_buffer(b1)->ip.adj_index[VLIB_TX];
mld0 = mpls_label_dpo_get(mldi0);
mld1 = mpls_label_dpo_get(mldi1);
if (payload_is_ip4)
{
/*
* decrement the TTL on ingress to the LSP
*/
ip4_header_t * ip0 = vlib_buffer_get_current(b0);
ip4_header_t * ip1 = vlib_buffer_get_current(b1);
u32 checksum0;
u32 checksum1;
checksum0 = ip0->checksum + clib_host_to_net_u16 (0x0100);
checksum1 = ip1->checksum + clib_host_to_net_u16 (0x0100);
checksum0 += checksum0 >= 0xffff;
checksum1 += checksum1 >= 0xffff;
ip0->checksum = checksum0;
ip1->checksum = checksum1;
ip0->ttl -= 1;
ip1->ttl -= 1;
ttl1 = ip1->ttl;
ttl0 = ip0->ttl;
}
else if (payload_is_ip6)
{
/*
* decrement the TTL on ingress to the LSP
*/
ip6_header_t * ip0 = vlib_buffer_get_current(b0);
ip6_header_t * ip1 = vlib_buffer_get_current(b1);
ip0->hop_limit -= 1;
ip1->hop_limit -= 1;
ttl0 = ip0->hop_limit;
ttl1 = ip1->hop_limit;
}
else
{
/*
* else, the packet to be encapped is an MPLS packet
*/
if (PREDICT_TRUE(vnet_buffer(b0)->mpls.first))
{
/*
* The first label to be imposed on the packet. this is a label swap.
* in which case we stashed the TTL and EXP bits in the
* packet in the lookup node
*/
ASSERT(0 != vnet_buffer (b0)->mpls.ttl);
ttl0 = vnet_buffer(b0)->mpls.ttl - 1;
}
else
{
/*
* not the first label. implying we are recusring down a chain of
* output labels.
* Each layer is considered a new LSP - hence the TTL is reset.
*/
ttl0 = 255;
}
if (PREDICT_TRUE(vnet_buffer(b1)->mpls.first))
{
ASSERT(1 != vnet_buffer (b1)->mpls.ttl);
ttl1 = vnet_buffer(b1)->mpls.ttl - 1;
}
else
{
ttl1 = 255;
}
}
vnet_buffer(b0)->mpls.first = 0;
vnet_buffer(b1)->mpls.first = 0;
/* Paint the MPLS header */
vlib_buffer_advance(b0, -(mld0->mld_n_hdr_bytes));
vlib_buffer_advance(b1, -(mld1->mld_n_hdr_bytes));
hdr0 = vlib_buffer_get_current(b0);
hdr1 = vlib_buffer_get_current(b1);
clib_memcpy(hdr0, mld0->mld_hdr, mld0->mld_n_hdr_bytes);
clib_memcpy(hdr1, mld1->mld_hdr, mld1->mld_n_hdr_bytes);
/* fixup the TTL for the inner most label */
hdr0 = hdr0 + (mld0->mld_n_labels - 1);
hdr1 = hdr1 + (mld1->mld_n_labels - 1);
((char*)hdr0)[3] = ttl0;
((char*)hdr1)[3] = ttl1;
next0 = mld0->mld_dpo.dpoi_next_node;
next1 = mld1->mld_dpo.dpoi_next_node;
vnet_buffer(b0)->ip.adj_index[VLIB_TX] = mld0->mld_dpo.dpoi_index;
vnet_buffer(b1)->ip.adj_index[VLIB_TX] = mld1->mld_dpo.dpoi_index;
if (PREDICT_FALSE(b0->flags & VLIB_BUFFER_IS_TRACED))
{
mpls_label_imposition_trace_t *tr =
vlib_add_trace (vm, node, b0, sizeof (*tr));
tr->hdr = *hdr0;
}
if (PREDICT_FALSE(b1->flags & VLIB_BUFFER_IS_TRACED))
{
mpls_label_imposition_trace_t *tr =
vlib_add_trace (vm, node, b1, sizeof (*tr));
tr->hdr = *hdr1;
}
vlib_validate_buffer_enqueue_x2(vm, node, next_index, to_next,
n_left_to_next,
bi0, bi1, next0, next1);
}
while (n_left_from > 0 && n_left_to_next > 0)
{
mpls_unicast_header_t *hdr0;
@ -255,11 +413,9 @@ mpls_label_imposition_inline (vlib_main_t * vm,
vnet_buffer(b0)->mpls.first = 0;
/* Paint the MPLS header */
vlib_buffer_advance(b0, -(sizeof(*hdr0) * mld0->mld_n_labels));
vlib_buffer_advance(b0, -(mld0->mld_n_hdr_bytes));
hdr0 = vlib_buffer_get_current(b0);
clib_memcpy(hdr0, mld0->mld_hdr,
sizeof(*hdr0) * mld0->mld_n_labels);
clib_memcpy(hdr0, mld0->mld_hdr, mld0->mld_n_hdr_bytes);
/* fixup the TTL for the inner most label */
hdr0 = hdr0 + (mld0->mld_n_labels - 1);

View File

@ -45,6 +45,11 @@ typedef struct mpls_label_dpo_t
*/
u16 mld_n_labels;
/**
* Cached amount of header bytes to paint
*/
u16 mld_n_hdr_bytes;
/**
* Number of locks/users of the label
*/

View File

@ -58,6 +58,7 @@ mpls_output_inline (vlib_main_t * vm,
{
u32 n_left_from, next_index, * from, * to_next, cpu_index;
vlib_node_runtime_t * error_node;
u32 n_left_to_next;
cpu_index = os_get_cpu_number();
error_node = vlib_node_get_runtime (vm, mpls_output_node.index);
@ -67,11 +68,146 @@ mpls_output_inline (vlib_main_t * vm,
while (n_left_from > 0)
{
u32 n_left_to_next;
vlib_get_next_frame (vm, node, next_index,
to_next, n_left_to_next);
while (n_left_from >= 4 && n_left_to_next >= 2)
{
ip_adjacency_t * adj0;
mpls_unicast_header_t *hdr0;
vlib_buffer_t * p0;
u32 pi0, rw_len0, adj_index0, next0, error0;
ip_adjacency_t * adj1;
mpls_unicast_header_t *hdr1;
vlib_buffer_t * p1;
u32 pi1, rw_len1, adj_index1, next1, error1;
/* Prefetch next iteration. */
{
vlib_buffer_t * p2, * p3;
p2 = vlib_get_buffer (vm, from[2]);
p3 = vlib_get_buffer (vm, from[3]);
vlib_prefetch_buffer_header (p2, STORE);
vlib_prefetch_buffer_header (p3, STORE);
CLIB_PREFETCH (p2->data, sizeof (hdr0[0]), STORE);
CLIB_PREFETCH (p3->data, sizeof (hdr1[0]), STORE);
}
pi0 = to_next[0] = from[0];
pi1 = to_next[1] = from[1];
from += 2;
n_left_from -= 2;
to_next += 2;
n_left_to_next -= 2;
p0 = vlib_get_buffer (vm, pi0);
p1 = vlib_get_buffer (vm, pi1);
adj_index0 = vnet_buffer (p0)->ip.adj_index[VLIB_TX];
adj_index1 = vnet_buffer (p1)->ip.adj_index[VLIB_TX];
/* We should never rewrite a pkt using the MISS adjacency */
ASSERT(adj_index0);
ASSERT(adj_index1);
adj0 = adj_get(adj_index0);
adj1 = adj_get(adj_index1);
hdr0 = vlib_buffer_get_current (p0);
hdr1 = vlib_buffer_get_current (p1);
/* Guess we are only writing on simple Ethernet header. */
vnet_rewrite_two_headers (adj0[0], adj1[0], hdr0, hdr1,
sizeof (ethernet_header_t));
/* Update packet buffer attributes/set output interface. */
rw_len0 = adj0[0].rewrite_header.data_bytes;
rw_len1 = adj1[0].rewrite_header.data_bytes;
if (PREDICT_FALSE (rw_len0 > sizeof(ethernet_header_t)))
vlib_increment_combined_counter
(&adjacency_counters,
cpu_index, adj_index0,
/* packet increment */ 0,
/* byte increment */ rw_len0-sizeof(ethernet_header_t));
if (PREDICT_FALSE (rw_len1 > sizeof(ethernet_header_t)))
vlib_increment_combined_counter
(&adjacency_counters,
cpu_index, adj_index1,
/* packet increment */ 0,
/* byte increment */ rw_len1-sizeof(ethernet_header_t));
/* Check MTU of outgoing interface. */
if (PREDICT_TRUE(vlib_buffer_length_in_chain (vm, p0) <=
adj0[0].rewrite_header.max_l3_packet_bytes))
{
p0->current_data -= rw_len0;
p0->current_length += rw_len0;
vnet_buffer (p0)->sw_if_index[VLIB_TX] =
adj0[0].rewrite_header.sw_if_index;
next0 = adj0[0].rewrite_header.next_index;
error0 = IP4_ERROR_NONE;
if (is_midchain)
{
adj0->sub_type.midchain.fixup_func(vm, adj0, p0);
}
}
else
{
error0 = IP4_ERROR_MTU_EXCEEDED;
next0 = MPLS_OUTPUT_NEXT_DROP;
}
if (PREDICT_TRUE(vlib_buffer_length_in_chain (vm, p1) <=
adj1[0].rewrite_header.max_l3_packet_bytes))
{
p1->current_data -= rw_len1;
p1->current_length += rw_len1;
vnet_buffer (p1)->sw_if_index[VLIB_TX] =
adj1[0].rewrite_header.sw_if_index;
next1 = adj1[0].rewrite_header.next_index;
error1 = IP4_ERROR_NONE;
if (is_midchain)
{
adj1->sub_type.midchain.fixup_func(vm, adj1, p1);
}
}
else
{
error1 = IP4_ERROR_MTU_EXCEEDED;
next1 = MPLS_OUTPUT_NEXT_DROP;
}
p0->error = error_node->errors[error0];
p1->error = error_node->errors[error1];
if (PREDICT_FALSE(p0->flags & VLIB_BUFFER_IS_TRACED))
{
mpls_output_trace_t *tr = vlib_add_trace (vm, node,
p0, sizeof (*tr));
tr->adj_index = vnet_buffer(p0)->ip.adj_index[VLIB_TX];
tr->flow_hash = vnet_buffer(p0)->ip.flow_hash;
}
if (PREDICT_FALSE(p1->flags & VLIB_BUFFER_IS_TRACED))
{
mpls_output_trace_t *tr = vlib_add_trace (vm, node,
p1, sizeof (*tr));
tr->adj_index = vnet_buffer(p1)->ip.adj_index[VLIB_TX];
tr->flow_hash = vnet_buffer(p1)->ip.flow_hash;
}
vlib_validate_buffer_enqueue_x2 (vm, node, next_index,
to_next, n_left_to_next,
pi0, pi1, next0, next1);
}
while (n_left_from > 0 && n_left_to_next > 0)
{
ip_adjacency_t * adj0;
@ -106,16 +242,8 @@ mpls_output_inline (vlib_main_t * vm,
/* byte increment */ rw_len0-sizeof(ethernet_header_t));
/* Check MTU of outgoing interface. */
error0 = (vlib_buffer_length_in_chain (vm, p0)
> adj0[0].rewrite_header.max_l3_packet_bytes
? IP4_ERROR_MTU_EXCEEDED
: IP4_ERROR_NONE);
p0->error = error_node->errors[error0];
/* Don't adjust the buffer for ttl issue; icmp-error node wants
* to see the IP headerr */
if (PREDICT_TRUE(error0 == IP4_ERROR_NONE))
if (PREDICT_TRUE(vlib_buffer_length_in_chain (vm, p0) <=
adj0[0].rewrite_header.max_l3_packet_bytes))
{
p0->current_data -= rw_len0;
p0->current_length += rw_len0;
@ -123,6 +251,7 @@ mpls_output_inline (vlib_main_t * vm,
vnet_buffer (p0)->sw_if_index[VLIB_TX] =
adj0[0].rewrite_header.sw_if_index;
next0 = adj0[0].rewrite_header.next_index;
error0 = IP4_ERROR_NONE;
if (is_midchain)
{
@ -131,8 +260,10 @@ mpls_output_inline (vlib_main_t * vm,
}
else
{
error0 = IP4_ERROR_MTU_EXCEEDED;
next0 = MPLS_OUTPUT_NEXT_DROP;
}
p0->error = error_node->errors[error0];
from += 1;
n_left_from -= 1;

View File

@ -90,6 +90,93 @@ mpls_input_inline (vlib_main_t * vm,
vlib_get_next_frame (vm, node, next_index,
to_next, n_left_to_next);
while (n_left_from >= 4 && n_left_to_next >= 2)
{
u32 label0, bi0, next0, sw_if_index0;
u32 label1, bi1, next1, sw_if_index1;
mpls_unicast_header_t *h0, *h1;
vlib_buffer_t *b0, *b1;
/* Prefetch next iteration. */
{
vlib_buffer_t * p2, * p3;
p2 = vlib_get_buffer (vm, from[2]);
p3 = vlib_get_buffer (vm, from[3]);
vlib_prefetch_buffer_header (p2, STORE);
vlib_prefetch_buffer_header (p3, STORE);
CLIB_PREFETCH (p2->data, sizeof (h0[0]), STORE);
CLIB_PREFETCH (p3->data, sizeof (h1[0]), STORE);
}
bi0 = to_next[0] = from[0];
bi1 = to_next[1] = from[1];
from += 2;
to_next += 2;
n_left_from -= 2;
n_left_to_next -= 2;
b0 = vlib_get_buffer (vm, bi0);
b1 = vlib_get_buffer (vm, bi1);
h0 = vlib_buffer_get_current (b0);
h1 = vlib_buffer_get_current (b1);
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
sw_if_index1 = vnet_buffer (b1)->sw_if_index[VLIB_RX];
label0 = clib_net_to_host_u32 (h0->label_exp_s_ttl);
label1 = clib_net_to_host_u32 (h1->label_exp_s_ttl);
/* TTL expired? */
if (PREDICT_FALSE(vnet_mpls_uc_get_ttl (label0) == 0))
{
next0 = MPLS_INPUT_NEXT_DROP;
b0->error = node->errors[MPLS_ERROR_TTL_EXPIRED];
}
else
{
next0 = MPLS_INPUT_NEXT_LOOKUP;
vnet_feature_arc_start(mm->input_feature_arc_index, sw_if_index0, &next0, b0);
vlib_increment_simple_counter (cm, cpu_index, sw_if_index0, 1);
}
if (PREDICT_FALSE(vnet_mpls_uc_get_ttl (label1) == 0))
{
next1 = MPLS_INPUT_NEXT_DROP;
b1->error = node->errors[MPLS_ERROR_TTL_EXPIRED];
}
else
{
next1 = MPLS_INPUT_NEXT_LOOKUP;
vnet_feature_arc_start(mm->input_feature_arc_index, sw_if_index1, &next1, b1);
vlib_increment_simple_counter (cm, cpu_index, sw_if_index1, 1);
}
if (PREDICT_FALSE(b0->flags & VLIB_BUFFER_IS_TRACED))
{
mpls_input_trace_t *tr = vlib_add_trace (vm, node,
b0, sizeof (*tr));
tr->next_index = next0;
tr->label_host_byte_order = label0;
}
if (PREDICT_FALSE(b1->flags & VLIB_BUFFER_IS_TRACED))
{
mpls_input_trace_t *tr = vlib_add_trace (vm, node,
b1, sizeof (*tr));
tr->next_index = next1;
tr->label_host_byte_order = label1;
}
vlib_validate_buffer_enqueue_x2 (vm, node, next_index,
to_next, n_left_to_next,
bi0, bi1, next0, next1);
}
while (n_left_from > 0 && n_left_to_next > 0)
{
u32 bi0;