Files
vpp/src/vlib/node.c
Christian Hopps 2e8b0618b1 vlib: fix for vlib_node_add_next_with_slot
- vlib_node_add_next_with_slot was not cleaning the old next node
references to the given slot when replacing it with new next node. This mostly
worked until one tried to set the slot to a previously (but not currently) used
next node for that slot.

Type: fix

Signed-off-by: Christian Hopps <chopps@labn.net>
Change-Id: I7ee607625da874e320158b80f12ddc16e377f8e9
2019-11-04 12:31:51 +00:00

803 lines
21 KiB
C

/*
* Copyright (c) 2015 Cisco and/or its affiliates.
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at:
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
/*
* node.c: VLIB processing nodes
*
* Copyright (c) 2008 Eliot Dresselhaus
*
* Permission is hereby granted, free of charge, to any person obtaining
* a copy of this software and associated documentation files (the
* "Software"), to deal in the Software without restriction, including
* without limitation the rights to use, copy, modify, merge, publish,
* distribute, sublicense, and/or sell copies of the Software, and to
* permit persons to whom the Software is furnished to do so, subject to
* the following conditions:
*
* The above copyright notice and this permission notice shall be
* included in all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
* LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
* OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
* WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
*/
#include <vlib/vlib.h>
#include <vlib/threads.h>
/* Query node given name. */
vlib_node_t *
vlib_get_node_by_name (vlib_main_t * vm, u8 * name)
{
vlib_node_main_t *nm = &vm->node_main;
uword *p;
u8 *key = name;
if (!clib_mem_is_heap_object (vec_header (key, 0)))
key = format (0, "%s", key);
p = hash_get (nm->node_by_name, key);
if (key != name)
vec_free (key);
return p ? vec_elt (nm->nodes, p[0]) : 0;
}
static void
node_set_elog_name (vlib_main_t * vm, uword node_index)
{
vlib_node_t *n = vlib_get_node (vm, node_index);
elog_event_type_t *t;
t = vec_elt_at_index (vm->node_call_elog_event_types, node_index);
vec_free (t->format);
t->format = (char *) format (0, "%v-call: %%d%c", n->name, 0);
t = vec_elt_at_index (vm->node_return_elog_event_types, node_index);
vec_free (t->format);
t->format = (char *) format (0, "%v-return: %%d%c", n->name, 0);
n->name_elog_string = elog_string (&vm->elog_main, "%v%c", n->name, 0);
}
static void
vlib_worker_thread_node_rename (u32 node_index)
{
int i;
vlib_main_t *vm;
vlib_node_t *n;
if (vec_len (vlib_mains) == 1)
return;
vm = vlib_mains[0];
n = vlib_get_node (vm, node_index);
ASSERT (vlib_get_thread_index () == 0);
ASSERT (*vlib_worker_threads->wait_at_barrier == 1);
for (i = 1; i < vec_len (vlib_mains); i++)
{
vlib_main_t *vm_worker = vlib_mains[i];
vlib_node_t *n_worker = vlib_get_node (vm_worker, node_index);
n_worker->name = n->name;
n_worker->name_elog_string = n->name_elog_string;
}
}
void
vlib_node_rename (vlib_main_t * vm, u32 node_index, char *fmt, ...)
{
va_list va;
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n = vlib_get_node (vm, node_index);
va_start (va, fmt);
hash_unset (nm->node_by_name, n->name);
vec_free (n->name);
n->name = va_format (0, fmt, &va);
va_end (va);
hash_set (nm->node_by_name, n->name, n->index);
node_set_elog_name (vm, node_index);
/* Propagate the change to all worker threads */
vlib_worker_thread_node_rename (node_index);
}
static void
vlib_node_runtime_update (vlib_main_t * vm, u32 node_index, u32 next_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_runtime_t *r, *s;
vlib_node_t *node, *next_node;
vlib_next_frame_t *nf;
vlib_pending_frame_t *pf;
i32 i, j, n_insert;
node = vec_elt (nm->nodes, node_index);
r = vlib_node_get_runtime (vm, node_index);
n_insert = vec_len (node->next_nodes) - r->n_next_nodes;
if (n_insert > 0)
{
i = r->next_frame_index + r->n_next_nodes;
vec_insert (nm->next_frames, n_insert, i);
/* Initialize newly inserted next frames. */
for (j = 0; j < n_insert; j++)
vlib_next_frame_init (nm->next_frames + i + j);
/* Relocate other next frames at higher indices. */
for (j = 0; j < vec_len (nm->nodes); j++)
{
s = vlib_node_get_runtime (vm, j);
if (j != node_index && s->next_frame_index >= i)
s->next_frame_index += n_insert;
}
/* Pending frames may need to be relocated also. */
vec_foreach (pf, nm->pending_frames)
{
if (pf->next_frame_index != VLIB_PENDING_FRAME_NO_NEXT_FRAME
&& pf->next_frame_index >= i)
pf->next_frame_index += n_insert;
}
/* *INDENT-OFF* */
pool_foreach (pf, nm->suspended_process_frames, ({
if (pf->next_frame_index != ~0 && pf->next_frame_index >= i)
pf->next_frame_index += n_insert;
}));
/* *INDENT-ON* */
r->n_next_nodes = vec_len (node->next_nodes);
}
/* Set frame's node runtime index. */
next_node = vlib_get_node (vm, node->next_nodes[next_index]);
nf = nm->next_frames + r->next_frame_index + next_index;
nf->node_runtime_index = next_node->runtime_index;
vlib_worker_thread_node_runtime_update ();
}
uword
vlib_node_get_next (vlib_main_t * vm, uword node_index, uword next_node_index)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *node;
uword *p;
node = vec_elt (nm->nodes, node_index);
/* Runtime has to be initialized. */
ASSERT (nm->flags & VLIB_NODE_MAIN_RUNTIME_STARTED);
if ((p = hash_get (node->next_slot_by_node, next_node_index)))
{
return p[0];
}
return (~0);
}
/* Add next node to given node in given slot. */
uword
vlib_node_add_next_with_slot (vlib_main_t * vm,
uword node_index,
uword next_node_index, uword slot)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *node, *next, *old_next;
u32 old_next_index;
uword *p;
ASSERT (vlib_get_thread_index () == 0);
node = vec_elt (nm->nodes, node_index);
next = vec_elt (nm->nodes, next_node_index);
/* Runtime has to be initialized. */
ASSERT (nm->flags & VLIB_NODE_MAIN_RUNTIME_STARTED);
if ((p = hash_get (node->next_slot_by_node, next_node_index)))
{
/* Next already exists: slot must match. */
if (slot != ~0)
ASSERT (slot == p[0]);
return p[0];
}
vlib_worker_thread_barrier_sync (vm);
if (slot == ~0)
slot = vec_len (node->next_nodes);
vec_validate_init_empty (node->next_nodes, slot, ~0);
vec_validate (node->n_vectors_by_next_node, slot);
if ((old_next_index = node->next_nodes[slot]) != ~0u)
{
hash_unset (node->next_slot_by_node, old_next_index);
old_next = vlib_get_node (vm, old_next_index);
old_next->prev_node_bitmap =
clib_bitmap_andnoti (old_next->prev_node_bitmap, node_index);
}
node->next_nodes[slot] = next_node_index;
hash_set (node->next_slot_by_node, next_node_index, slot);
vlib_node_runtime_update (vm, node_index, slot);
next->prev_node_bitmap = clib_bitmap_ori (next->prev_node_bitmap,
node_index);
/* Siblings all get same node structure. */
{
uword sib_node_index, sib_slot;
vlib_node_t *sib_node;
/* *INDENT-OFF* */
clib_bitmap_foreach (sib_node_index, node->sibling_bitmap, ({
sib_node = vec_elt (nm->nodes, sib_node_index);
if (sib_node != node)
{
sib_slot = vlib_node_add_next_with_slot (vm, sib_node_index, next_node_index, slot);
ASSERT (sib_slot == slot);
}
}));
/* *INDENT-ON* */
}
vlib_worker_thread_barrier_release (vm);
return slot;
}
/* Add named next node to given node in given slot. */
uword
vlib_node_add_named_next_with_slot (vlib_main_t * vm,
uword node, char *name, uword slot)
{
vlib_node_main_t *nm;
vlib_node_t *n, *n_next;
nm = &vm->node_main;
n = vlib_get_node (vm, node);
n_next = vlib_get_node_by_name (vm, (u8 *) name);
if (!n_next)
{
if (nm->flags & VLIB_NODE_MAIN_RUNTIME_STARTED)
return ~0;
if (slot == ~0)
slot = clib_max (vec_len (n->next_node_names),
vec_len (n->next_nodes));
vec_validate (n->next_node_names, slot);
n->next_node_names[slot] = name;
return slot;
}
return vlib_node_add_next_with_slot (vm, node, n_next->index, slot);
}
static void
node_elog_init (vlib_main_t * vm, uword ni)
{
elog_event_type_t t;
clib_memset (&t, 0, sizeof (t));
/* 2 event types for this node: one when node function is called.
One when it returns. */
vec_validate (vm->node_call_elog_event_types, ni);
vm->node_call_elog_event_types[ni] = t;
vec_validate (vm->node_return_elog_event_types, ni);
vm->node_return_elog_event_types[ni] = t;
node_set_elog_name (vm, ni);
}
#ifdef CLIB_UNIX
#define STACK_ALIGN (clib_mem_get_page_size())
#else
#define STACK_ALIGN CLIB_CACHE_LINE_BYTES
#endif
static void
register_node (vlib_main_t * vm, vlib_node_registration_t * r)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n;
u32 page_size = clib_mem_get_page_size ();
int i;
if (CLIB_DEBUG > 0)
{
/* Default (0) type should match INTERNAL. */
vlib_node_t zero = { 0 };
ASSERT (VLIB_NODE_TYPE_INTERNAL == zero.type);
}
if (r->node_fn_registrations)
{
vlib_node_fn_registration_t *fnr = r->node_fn_registrations;
int priority = -1;
/* to avoid confusion, please remove ".function " statiement from
CLIB_NODE_REGISTRATION() if using function function candidates */
ASSERT (r->function == 0);
while (fnr)
{
if (fnr->priority > priority)
{
priority = fnr->priority;
r->function = fnr->function;
}
fnr = fnr->next_registration;
}
}
ASSERT (r->function != 0);
n = clib_mem_alloc_no_fail (sizeof (n[0]));
clib_memset (n, 0, sizeof (n[0]));
n->index = vec_len (nm->nodes);
n->node_fn_registrations = r->node_fn_registrations;
n->protocol_hint = r->protocol_hint;
vec_add1 (nm->nodes, n);
/* Name is always a vector so it can be formatted with %v. */
if (clib_mem_is_heap_object (vec_header (r->name, 0)))
n->name = vec_dup ((u8 *) r->name);
else
n->name = format (0, "%s", r->name);
if (!nm->node_by_name)
nm->node_by_name = hash_create_vec ( /* size */ 32,
sizeof (n->name[0]), sizeof (uword));
/* Node names must be unique. */
{
vlib_node_t *o = vlib_get_node_by_name (vm, n->name);
if (o)
clib_error ("more than one node named `%v'", n->name);
}
hash_set (nm->node_by_name, n->name, n->index);
r->index = n->index; /* save index in registration */
n->function = r->function;
/* Node index of next sibling will be filled in by vlib_node_main_init. */
n->sibling_of = r->sibling_of;
if (r->sibling_of && r->n_next_nodes > 0)
clib_error ("sibling node should not have any next nodes `%v'", n->name);
if (r->type == VLIB_NODE_TYPE_INTERNAL)
ASSERT (r->vector_size > 0);
#define _(f) n->f = r->f
_(type);
_(flags);
_(state);
_(scalar_size);
_(vector_size);
_(format_buffer);
_(unformat_buffer);
_(format_trace);
_(validate_frame);
/* Register error counters. */
vlib_register_errors (vm, n->index, r->n_errors, r->error_strings);
node_elog_init (vm, n->index);
_(runtime_data_bytes);
if (r->runtime_data_bytes > 0)
{
vec_resize (n->runtime_data, r->runtime_data_bytes);
if (r->runtime_data)
clib_memcpy (n->runtime_data, r->runtime_data, r->runtime_data_bytes);
}
vec_resize (n->next_node_names, r->n_next_nodes);
for (i = 0; i < r->n_next_nodes; i++)
n->next_node_names[i] = r->next_nodes[i];
vec_validate_init_empty (n->next_nodes, r->n_next_nodes - 1, ~0);
vec_validate (n->n_vectors_by_next_node, r->n_next_nodes - 1);
n->owner_node_index = n->owner_next_index = ~0;
/* Initialize node runtime. */
{
vlib_node_runtime_t *rt;
u32 i;
if (n->type == VLIB_NODE_TYPE_PROCESS)
{
vlib_process_t *p;
uword log2_n_stack_bytes;
log2_n_stack_bytes = clib_max (r->process_log2_n_stack_bytes, 15);
#ifdef CLIB_UNIX
/*
* Bump the stack size if running over a kernel with a large page size,
* and the stack isn't any too big to begin with. Otherwise, we'll
* trip over the stack guard page for sure.
*/
if ((page_size > (4 << 10)) && log2_n_stack_bytes < 19)
{
if ((1 << log2_n_stack_bytes) <= page_size)
log2_n_stack_bytes = min_log2 (page_size) + 1;
else
log2_n_stack_bytes++;
}
#endif
p = clib_mem_alloc_aligned_at_offset
(sizeof (p[0]) + (1 << log2_n_stack_bytes),
STACK_ALIGN, STRUCT_OFFSET_OF (vlib_process_t, stack),
0 /* no, don't call os_out_of_memory */ );
if (p == 0)
clib_panic ("failed to allocate process stack (%d bytes)",
1 << log2_n_stack_bytes);
clib_memset (p, 0, sizeof (p[0]));
p->log2_n_stack_bytes = log2_n_stack_bytes;
/* Process node's runtime index is really index into process
pointer vector. */
n->runtime_index = vec_len (nm->processes);
vec_add1 (nm->processes, p);
/* Paint first stack word with magic number so we can at least
detect process stack overruns. */
p->stack[0] = VLIB_PROCESS_STACK_MAGIC;
/* Node runtime is stored inside of process. */
rt = &p->node_runtime;
#ifdef CLIB_UNIX
/*
* Disallow writes to the bottom page of the stack, to
* catch stack overflows.
*/
if (mprotect (p->stack, page_size, PROT_READ) < 0)
clib_unix_warning ("process stack");
#endif
}
else
{
vec_add2_aligned (nm->nodes_by_type[n->type], rt, 1,
/* align */ CLIB_CACHE_LINE_BYTES);
n->runtime_index = rt - nm->nodes_by_type[n->type];
}
if (n->type == VLIB_NODE_TYPE_INPUT)
nm->input_node_counts_by_state[n->state] += 1;
rt->function = n->function;
rt->flags = n->flags;
rt->state = n->state;
rt->node_index = n->index;
rt->n_next_nodes = r->n_next_nodes;
rt->next_frame_index = vec_len (nm->next_frames);
vec_resize (nm->next_frames, rt->n_next_nodes);
for (i = 0; i < rt->n_next_nodes; i++)
vlib_next_frame_init (nm->next_frames + rt->next_frame_index + i);
vec_resize (rt->errors, r->n_errors);
for (i = 0; i < vec_len (rt->errors); i++)
rt->errors[i] = n->error_heap_index + i;
STATIC_ASSERT_SIZEOF (vlib_node_runtime_t, 128);
ASSERT (vec_len (n->runtime_data) <= VLIB_NODE_RUNTIME_DATA_SIZE);
if (vec_len (n->runtime_data) > 0)
clib_memcpy (rt->runtime_data, n->runtime_data,
vec_len (n->runtime_data));
vec_free (n->runtime_data);
}
}
/* Register new packet processing node. */
u32
vlib_register_node (vlib_main_t * vm, vlib_node_registration_t * r)
{
register_node (vm, r);
return r->index;
}
static uword
null_node_fn (vlib_main_t * vm,
vlib_node_runtime_t * node, vlib_frame_t * frame)
{
u16 n_vectors = frame->n_vectors;
vlib_node_increment_counter (vm, node->node_index, 0, n_vectors);
vlib_buffer_free (vm, vlib_frame_vector_args (frame), n_vectors);
vlib_frame_free (vm, node, frame);
return n_vectors;
}
void
vlib_register_all_static_nodes (vlib_main_t * vm)
{
vlib_node_registration_t *r;
static char *null_node_error_strings[] = {
"blackholed packets",
};
static vlib_node_registration_t null_node_reg = {
.function = null_node_fn,
.vector_size = sizeof (u32),
.name = "null-node",
.n_errors = 1,
.error_strings = null_node_error_strings,
};
/* make sure that node index 0 is not used by
real node */
register_node (vm, &null_node_reg);
r = vm->node_main.node_registrations;
while (r)
{
register_node (vm, r);
r = r->next_registration;
}
}
void
vlib_node_get_nodes (vlib_main_t * vm, u32 max_threads, int include_stats,
int barrier_sync, vlib_node_t **** node_dupsp,
vlib_main_t *** stat_vmsp)
{
vlib_node_main_t *nm = &vm->node_main;
vlib_node_t *n;
vlib_node_t ***node_dups = *node_dupsp;
vlib_node_t **nodes;
vlib_main_t **stat_vms = *stat_vmsp;
vlib_main_t *stat_vm;
uword i, j;
u32 threads_to_serialize;
if (vec_len (stat_vms) == 0)
{
for (i = 0; i < vec_len (vlib_mains); i++)
{
stat_vm = vlib_mains[i];
if (stat_vm)
vec_add1 (stat_vms, stat_vm);
}
}
threads_to_serialize = clib_min (max_threads, vec_len (stat_vms));
vec_validate (node_dups, threads_to_serialize - 1);
/*
* Barrier sync across stats scraping.
* Otherwise, the counts will be grossly inaccurate.
*/
if (barrier_sync)
vlib_worker_thread_barrier_sync (vm);
for (j = 0; j < threads_to_serialize; j++)
{
stat_vm = stat_vms[j];
nm = &stat_vm->node_main;
if (include_stats)
{
for (i = 0; i < vec_len (nm->nodes); i++)
{
n = nm->nodes[i];
vlib_node_sync_stats (stat_vm, n);
}
}
nodes = node_dups[j];
vec_validate (nodes, vec_len (nm->nodes) - 1);
clib_memcpy (nodes, nm->nodes, vec_len (nm->nodes) * sizeof (nodes[0]));
node_dups[j] = nodes;
}
if (barrier_sync)
vlib_worker_thread_barrier_release (vm);
*node_dupsp = node_dups;
*stat_vmsp = stat_vms;
}
clib_error_t *
vlib_node_main_init (vlib_main_t * vm)
{
vlib_node_main_t *nm = &vm->node_main;
clib_error_t *error = 0;
vlib_node_t *n;
uword ni;
nm->frame_sizes = vec_new (vlib_frame_size_t, 1);
#ifdef VLIB_SUPPORTS_ARBITRARY_SCALAR_SIZES
nm->frame_size_hash = hash_create (0, sizeof (uword));
#endif
nm->flags |= VLIB_NODE_MAIN_RUNTIME_STARTED;
/* Generate sibling relationships */
{
vlib_node_t *n, *sib;
uword si;
for (ni = 0; ni < vec_len (nm->nodes); ni++)
{
n = vec_elt (nm->nodes, ni);
if (!n->sibling_of)
continue;
sib = vlib_get_node_by_name (vm, (u8 *) n->sibling_of);
if (!sib)
{
error = clib_error_create ("sibling `%s' not found for node `%v'",
n->sibling_of, n->name);
goto done;
}
/* *INDENT-OFF* */
clib_bitmap_foreach (si, sib->sibling_bitmap, ({
vlib_node_t * m = vec_elt (nm->nodes, si);
/* Connect all of sibling's siblings to us. */
m->sibling_bitmap = clib_bitmap_ori (m->sibling_bitmap, n->index);
/* Connect us to all of sibling's siblings. */
n->sibling_bitmap = clib_bitmap_ori (n->sibling_bitmap, si);
}));
/* *INDENT-ON* */
/* Connect sibling to us. */
sib->sibling_bitmap = clib_bitmap_ori (sib->sibling_bitmap, n->index);
/* Connect us to sibling. */
n->sibling_bitmap = clib_bitmap_ori (n->sibling_bitmap, sib->index);
}
}
/* Resolve next names into next indices. */
for (ni = 0; ni < vec_len (nm->nodes); ni++)
{
uword i;
n = vec_elt (nm->nodes, ni);
for (i = 0; i < vec_len (n->next_node_names); i++)
{
char *a = n->next_node_names[i];
if (!a)
continue;
if (~0 == vlib_node_add_named_next_with_slot (vm, n->index, a, i))
{
error = clib_error_create
("node `%v' refers to unknown node `%s'", n->name, a);
goto done;
}
}
vec_free (n->next_node_names);
}
/* Set previous node pointers. */
for (ni = 0; ni < vec_len (nm->nodes); ni++)
{
vlib_node_t *n_next;
uword i;
n = vec_elt (nm->nodes, ni);
for (i = 0; i < vec_len (n->next_nodes); i++)
{
if (n->next_nodes[i] >= vec_len (nm->nodes))
continue;
n_next = vec_elt (nm->nodes, n->next_nodes[i]);
n_next->prev_node_bitmap =
clib_bitmap_ori (n_next->prev_node_bitmap, n->index);
}
}
{
vlib_next_frame_t *nf;
vlib_node_runtime_t *r;
vlib_node_t *next;
uword i;
vec_foreach (r, nm->nodes_by_type[VLIB_NODE_TYPE_INTERNAL])
{
if (r->n_next_nodes == 0)
continue;
n = vlib_get_node (vm, r->node_index);
nf = vec_elt_at_index (nm->next_frames, r->next_frame_index);
for (i = 0; i < vec_len (n->next_nodes); i++)
{
next = vlib_get_node (vm, n->next_nodes[i]);
/* Validate node runtime indices are correctly initialized. */
ASSERT (nf[i].node_runtime_index == next->runtime_index);
nf[i].flags = 0;
if (next->flags & VLIB_NODE_FLAG_FRAME_NO_FREE_AFTER_DISPATCH)
nf[i].flags |= VLIB_FRAME_NO_FREE_AFTER_DISPATCH;
}
}
}
done:
return error;
}
u32
vlib_process_create (vlib_main_t * vm, char *name,
vlib_node_function_t * f, u32 log2_n_stack_bytes)
{
vlib_node_registration_t r;
vlib_node_t *n;
memset (&r, 0, sizeof (r));
r.name = (char *) format (0, "%s", name, 0);
r.function = f;
r.process_log2_n_stack_bytes = log2_n_stack_bytes;
r.type = VLIB_NODE_TYPE_PROCESS;
vlib_worker_thread_barrier_sync (vm);
vlib_register_node (vm, &r);
vec_free (r.name);
vlib_worker_thread_node_runtime_update ();
vlib_worker_thread_barrier_release (vm);
n = vlib_get_node (vm, r.index);
vlib_start_process (vm, n->runtime_index);
return (r.index);
}
/*
* fd.io coding-style-patch-verification: ON
*
* Local Variables:
* eval: (c-set-style "gnu")
* End:
*/