353 lines
9.6 KiB
C
353 lines
9.6 KiB
C
|
/*
|
||
|
* Copyright (c) 2016 Cisco and/or its affiliates.
|
||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||
|
* you may not use this file except in compliance with the License.
|
||
|
* You may obtain a copy of the License at:
|
||
|
*
|
||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||
|
*
|
||
|
* Unless required by applicable law or agreed to in writing, software
|
||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
|
* See the License for the specific language governing permissions and
|
||
|
* limitations under the License.
|
||
|
*/
|
||
|
#include <vlib/vlib.h>
|
||
|
#include <vnet/vnet.h>
|
||
|
#include <vnet/pg/pg.h>
|
||
|
#include <vppinfra/error.h>
|
||
|
#include <vnet/ip/ip.h>
|
||
|
#include <ioam/export/ioam_export.h>
|
||
|
|
||
|
typedef struct
|
||
|
{
|
||
|
u32 next_index;
|
||
|
u32 flow_label;
|
||
|
} export_trace_t;
|
||
|
|
||
|
/* packet trace format function */
|
||
|
static u8 *
|
||
|
format_export_trace (u8 * s, va_list * args)
|
||
|
{
|
||
|
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
|
||
|
CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
|
||
|
export_trace_t *t = va_arg (*args, export_trace_t *);
|
||
|
|
||
|
s = format (s, "EXPORT: flow_label %d, next index %d",
|
||
|
t->flow_label, t->next_index);
|
||
|
return s;
|
||
|
}
|
||
|
|
||
|
vlib_node_registration_t export_node;
|
||
|
|
||
|
#define foreach_export_error \
|
||
|
_(RECORDED, "Packets recorded for export")
|
||
|
|
||
|
typedef enum
|
||
|
{
|
||
|
#define _(sym,str) EXPORT_ERROR_##sym,
|
||
|
foreach_export_error
|
||
|
#undef _
|
||
|
EXPORT_N_ERROR,
|
||
|
} export_error_t;
|
||
|
|
||
|
static char *export_error_strings[] = {
|
||
|
#define _(sym,string) string,
|
||
|
foreach_export_error
|
||
|
#undef _
|
||
|
};
|
||
|
|
||
|
typedef enum
|
||
|
{
|
||
|
EXPORT_NEXT_POP_HBYH,
|
||
|
EXPORT_N_NEXT,
|
||
|
} export_next_t;
|
||
|
|
||
|
always_inline void
|
||
|
copy3cachelines (void *dst, const void *src, size_t n)
|
||
|
{
|
||
|
#if 0
|
||
|
if (PREDICT_FALSE (n < DEFAULT_EXPORT_SIZE))
|
||
|
{
|
||
|
/* Copy only the first 1/2 cache lines whatever is available */
|
||
|
if (n >= 64)
|
||
|
clib_mov64 ((u8 *) dst, (const u8 *) src);
|
||
|
if (n >= 128)
|
||
|
clib_mov64 ((u8 *) dst + 64, (const u8 *) src + 64);
|
||
|
return;
|
||
|
}
|
||
|
clib_mov64 ((u8 *) dst, (const u8 *) src);
|
||
|
clib_mov64 ((u8 *) dst + 64, (const u8 *) src + 64);
|
||
|
clib_mov64 ((u8 *) dst + 128, (const u8 *) src + 128);
|
||
|
#endif
|
||
|
#if 1
|
||
|
|
||
|
u64 *copy_dst, *copy_src;
|
||
|
int i;
|
||
|
copy_dst = (u64 *) dst;
|
||
|
copy_src = (u64 *) src;
|
||
|
if (PREDICT_FALSE (n < DEFAULT_EXPORT_SIZE))
|
||
|
{
|
||
|
for (i = 0; i < n / 64; i++)
|
||
|
{
|
||
|
copy_dst[0] = copy_src[0];
|
||
|
copy_dst[1] = copy_src[1];
|
||
|
copy_dst[2] = copy_src[2];
|
||
|
copy_dst[3] = copy_src[3];
|
||
|
copy_dst[4] = copy_src[4];
|
||
|
copy_dst[5] = copy_src[5];
|
||
|
copy_dst[6] = copy_src[6];
|
||
|
copy_dst[7] = copy_src[7];
|
||
|
copy_dst += 8;
|
||
|
copy_src += 8;
|
||
|
}
|
||
|
return;
|
||
|
}
|
||
|
for (i = 0; i < 3; i++)
|
||
|
{
|
||
|
copy_dst[0] = copy_src[0];
|
||
|
copy_dst[1] = copy_src[1];
|
||
|
copy_dst[2] = copy_src[2];
|
||
|
copy_dst[3] = copy_src[3];
|
||
|
copy_dst[4] = copy_src[4];
|
||
|
copy_dst[5] = copy_src[5];
|
||
|
copy_dst[6] = copy_src[6];
|
||
|
copy_dst[7] = copy_src[7];
|
||
|
copy_dst += 8;
|
||
|
copy_src += 8;
|
||
|
}
|
||
|
#endif
|
||
|
}
|
||
|
|
||
|
static uword
|
||
|
ip6_export_node_fn (vlib_main_t * vm,
|
||
|
vlib_node_runtime_t * node, vlib_frame_t * frame)
|
||
|
{
|
||
|
ioam_export_main_t *em = &ioam_export_main;
|
||
|
u32 n_left_from, *from, *to_next;
|
||
|
export_next_t next_index;
|
||
|
u32 pkts_recorded = 0;
|
||
|
ioam_export_buffer_t *my_buf = 0;
|
||
|
vlib_buffer_t *eb0 = 0;
|
||
|
u32 ebi0 = 0;
|
||
|
from = vlib_frame_vector_args (frame);
|
||
|
n_left_from = frame->n_vectors;
|
||
|
next_index = node->cached_next_index;
|
||
|
|
||
|
while (__sync_lock_test_and_set (em->lockp[vm->cpu_index], 1))
|
||
|
;
|
||
|
my_buf = ioam_export_get_my_buffer (vm->cpu_index);
|
||
|
my_buf->touched_at = vlib_time_now (vm);
|
||
|
while (n_left_from > 0)
|
||
|
{
|
||
|
u32 n_left_to_next;
|
||
|
|
||
|
vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
|
||
|
while (n_left_from >= 4 && n_left_to_next >= 2)
|
||
|
{
|
||
|
u32 next0 = EXPORT_NEXT_POP_HBYH;
|
||
|
u32 next1 = EXPORT_NEXT_POP_HBYH;
|
||
|
u32 bi0, bi1;
|
||
|
ip6_header_t *ip60, *ip61;
|
||
|
vlib_buffer_t *p0, *p1;
|
||
|
u32 ip_len0, ip_len1;
|
||
|
|
||
|
/* Prefetch next iteration. */
|
||
|
{
|
||
|
vlib_buffer_t *p2, *p3;
|
||
|
|
||
|
p2 = vlib_get_buffer (vm, from[2]);
|
||
|
p3 = vlib_get_buffer (vm, from[3]);
|
||
|
|
||
|
vlib_prefetch_buffer_header (p2, LOAD);
|
||
|
vlib_prefetch_buffer_header (p3, LOAD);
|
||
|
|
||
|
/* IPv6 + HbyH header + Trace option */
|
||
|
/* 40 + 2 + [4 hdr] + [16]* no_of_nodes */
|
||
|
/* 3 cache lines can get v6 hdr + trace option with upto 9 node trace */
|
||
|
CLIB_PREFETCH (p2->data, 3 * CLIB_CACHE_LINE_BYTES, LOAD);
|
||
|
CLIB_PREFETCH (p3->data, 3 * CLIB_CACHE_LINE_BYTES, LOAD);
|
||
|
}
|
||
|
|
||
|
/* speculatively enqueue p0 and p1 to the current next frame */
|
||
|
to_next[0] = bi0 = from[0];
|
||
|
to_next[1] = bi1 = from[1];
|
||
|
from += 2;
|
||
|
to_next += 2;
|
||
|
n_left_from -= 2;
|
||
|
n_left_to_next -= 2;
|
||
|
|
||
|
p0 = vlib_get_buffer (vm, bi0);
|
||
|
p1 = vlib_get_buffer (vm, bi1);
|
||
|
|
||
|
ip60 = vlib_buffer_get_current (p0);
|
||
|
ip61 = vlib_buffer_get_current (p1);
|
||
|
|
||
|
ip_len0 =
|
||
|
clib_net_to_host_u16 (ip60->payload_length) +
|
||
|
sizeof (ip6_header_t);
|
||
|
ip_len1 =
|
||
|
clib_net_to_host_u16 (ip61->payload_length) +
|
||
|
sizeof (ip6_header_t);
|
||
|
|
||
|
ebi0 = my_buf->buffer_index;
|
||
|
eb0 = vlib_get_buffer (vm, ebi0);
|
||
|
if (PREDICT_FALSE (eb0 == 0))
|
||
|
goto NO_BUFFER1;
|
||
|
|
||
|
ip_len0 =
|
||
|
ip_len0 > DEFAULT_EXPORT_SIZE ? DEFAULT_EXPORT_SIZE : ip_len0;
|
||
|
ip_len1 =
|
||
|
ip_len1 > DEFAULT_EXPORT_SIZE ? DEFAULT_EXPORT_SIZE : ip_len1;
|
||
|
|
||
|
copy3cachelines (eb0->data + eb0->current_length, ip60, ip_len0);
|
||
|
eb0->current_length += DEFAULT_EXPORT_SIZE;
|
||
|
/* To maintain uniform size per export, each
|
||
|
* record is default size, ip6 hdr can be
|
||
|
* used to parse the record correctly
|
||
|
*/
|
||
|
my_buf->records_in_this_buffer++;
|
||
|
/* if number of buf exceeds max that fits in a MTU sized buffer
|
||
|
* ship it to the queue and pick new one
|
||
|
*/
|
||
|
if (my_buf->records_in_this_buffer >= DEFAULT_EXPORT_RECORDS)
|
||
|
{
|
||
|
ioam_export_send_buffer (vm, my_buf);
|
||
|
ioam_export_init_buffer (vm, my_buf);
|
||
|
}
|
||
|
|
||
|
ebi0 = my_buf->buffer_index;
|
||
|
eb0 = vlib_get_buffer (vm, ebi0);
|
||
|
if (PREDICT_FALSE (eb0 == 0))
|
||
|
goto NO_BUFFER1;
|
||
|
|
||
|
copy3cachelines (eb0->data + eb0->current_length, ip61, ip_len1);
|
||
|
eb0->current_length += DEFAULT_EXPORT_SIZE;
|
||
|
my_buf->records_in_this_buffer++;
|
||
|
if (my_buf->records_in_this_buffer >= DEFAULT_EXPORT_RECORDS)
|
||
|
{
|
||
|
ioam_export_send_buffer (vm, my_buf);
|
||
|
ioam_export_init_buffer (vm, my_buf);
|
||
|
}
|
||
|
|
||
|
pkts_recorded += 2;
|
||
|
|
||
|
if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)))
|
||
|
{
|
||
|
if (p0->flags & VLIB_BUFFER_IS_TRACED)
|
||
|
{
|
||
|
export_trace_t *t =
|
||
|
vlib_add_trace (vm, node, p0, sizeof (*t));
|
||
|
t->flow_label =
|
||
|
clib_net_to_host_u32 (ip60->
|
||
|
ip_version_traffic_class_and_flow_label);
|
||
|
t->next_index = next0;
|
||
|
}
|
||
|
if (p1->flags & VLIB_BUFFER_IS_TRACED)
|
||
|
{
|
||
|
export_trace_t *t =
|
||
|
vlib_add_trace (vm, node, p1, sizeof (*t));
|
||
|
t->flow_label =
|
||
|
clib_net_to_host_u32 (ip61->
|
||
|
ip_version_traffic_class_and_flow_label);
|
||
|
t->next_index = next1;
|
||
|
}
|
||
|
}
|
||
|
NO_BUFFER1:
|
||
|
/* verify speculative enqueues, maybe switch current next frame */
|
||
|
vlib_validate_buffer_enqueue_x2 (vm, node, next_index,
|
||
|
to_next, n_left_to_next,
|
||
|
bi0, bi1, next0, next1);
|
||
|
}
|
||
|
|
||
|
while (n_left_from > 0 && n_left_to_next > 0)
|
||
|
{
|
||
|
u32 bi0;
|
||
|
vlib_buffer_t *p0;
|
||
|
u32 next0 = EXPORT_NEXT_POP_HBYH;
|
||
|
ip6_header_t *ip60;
|
||
|
u32 ip_len0;
|
||
|
|
||
|
/* speculatively enqueue p0 to the current next frame */
|
||
|
bi0 = from[0];
|
||
|
to_next[0] = bi0;
|
||
|
from += 1;
|
||
|
to_next += 1;
|
||
|
n_left_from -= 1;
|
||
|
n_left_to_next -= 1;
|
||
|
|
||
|
p0 = vlib_get_buffer (vm, bi0);
|
||
|
ip60 = vlib_buffer_get_current (p0);
|
||
|
ip_len0 =
|
||
|
clib_net_to_host_u16 (ip60->payload_length) +
|
||
|
sizeof (ip6_header_t);
|
||
|
|
||
|
ebi0 = my_buf->buffer_index;
|
||
|
eb0 = vlib_get_buffer (vm, ebi0);
|
||
|
if (PREDICT_FALSE (eb0 == 0))
|
||
|
goto NO_BUFFER;
|
||
|
|
||
|
ip_len0 =
|
||
|
ip_len0 > DEFAULT_EXPORT_SIZE ? DEFAULT_EXPORT_SIZE : ip_len0;
|
||
|
copy3cachelines (eb0->data + eb0->current_length, ip60, ip_len0);
|
||
|
eb0->current_length += DEFAULT_EXPORT_SIZE;
|
||
|
/* To maintain uniform size per export, each
|
||
|
* record is default size, ip6 hdr can be
|
||
|
* used to parse the record correctly
|
||
|
*/
|
||
|
my_buf->records_in_this_buffer++;
|
||
|
/* if number of buf exceeds max that fits in a MTU sized buffer
|
||
|
* ship it to the queue and pick new one
|
||
|
*/
|
||
|
if (my_buf->records_in_this_buffer >= DEFAULT_EXPORT_RECORDS)
|
||
|
{
|
||
|
ioam_export_send_buffer (vm, my_buf);
|
||
|
ioam_export_init_buffer (vm, my_buf);
|
||
|
}
|
||
|
if (PREDICT_FALSE ((node->flags & VLIB_NODE_FLAG_TRACE)
|
||
|
&& (p0->flags & VLIB_BUFFER_IS_TRACED)))
|
||
|
{
|
||
|
export_trace_t *t = vlib_add_trace (vm, node, p0, sizeof (*t));
|
||
|
t->flow_label =
|
||
|
clib_net_to_host_u32 (ip60->
|
||
|
ip_version_traffic_class_and_flow_label);
|
||
|
t->next_index = next0;
|
||
|
}
|
||
|
|
||
|
pkts_recorded += 1;
|
||
|
NO_BUFFER:
|
||
|
/* verify speculative enqueue, maybe switch current next frame */
|
||
|
vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
|
||
|
to_next, n_left_to_next,
|
||
|
bi0, next0);
|
||
|
}
|
||
|
|
||
|
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
|
||
|
}
|
||
|
|
||
|
vlib_node_increment_counter (vm, export_node.index,
|
||
|
EXPORT_ERROR_RECORDED, pkts_recorded);
|
||
|
*em->lockp[vm->cpu_index] = 0;
|
||
|
return frame->n_vectors;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* Node for IP6 export
|
||
|
*/
|
||
|
VLIB_REGISTER_NODE (export_node) =
|
||
|
{
|
||
|
.function = ip6_export_node_fn,
|
||
|
.name = "ip6-export",
|
||
|
.vector_size = sizeof (u32),
|
||
|
.format_trace = format_export_trace,
|
||
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
||
|
.n_errors = ARRAY_LEN (export_error_strings),
|
||
|
.error_strings = export_error_strings,
|
||
|
.n_next_nodes = EXPORT_N_NEXT,
|
||
|
/* edit / add dispositions here */
|
||
|
.next_nodes =
|
||
|
{
|
||
|
[EXPORT_NEXT_POP_HBYH] = "ip6-pop-hop-by-hop"
|
||
|
},
|
||
|
};
|