When a FIB entry has multiple sources installed by covering inheritance, the first walk installs the covering source, which can be worse than the already used one (hence the source is not activated). If a covering path is added and a new fib cover walk is done, the deactivated source is compared to itself. This operation calls for a reactivation of the source which was not activated in the first place. Type: fix Change-Id: I2639401bed14eb5a2cb003664c04d6c1f9130de9 Signed-off-by: Maxime Peim <mpeim@cisco.com>
1935 lines
53 KiB
C
1935 lines
53 KiB
C
/*
|
|
* Copyright (c) 2016 Cisco and/or its affiliates.
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <vnet/adj/adj.h>
|
|
#include <vnet/dpo/load_balance.h>
|
|
#include <vnet/dpo/mpls_label_dpo.h>
|
|
#include <vnet/dpo/drop_dpo.h>
|
|
#include <vnet/dpo/replicate_dpo.h>
|
|
|
|
#include <vnet/fib/fib_entry_src.h>
|
|
#include <vnet/fib/fib_table.h>
|
|
#include <vnet/fib/fib_path_ext.h>
|
|
#include <vnet/fib/fib_urpf_list.h>
|
|
#include <vnet/fib/fib_entry_delegate.h>
|
|
|
|
/*
|
|
* per-source type vft
|
|
*/
|
|
static fib_entry_src_vft_t fib_entry_src_bh_vft[FIB_SOURCE_BH_MAX];
|
|
|
|
/**
|
|
* Get the VFT for a given source. This is a combination of the source
|
|
* enum and the interposer flags
|
|
*/
|
|
const fib_entry_src_vft_t*
|
|
fib_entry_src_get_vft (const fib_entry_src_t *esrc)
|
|
{
|
|
fib_source_behaviour_t bh;
|
|
|
|
bh = fib_source_get_behaviour(esrc->fes_src);
|
|
|
|
if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_INTERPOSE)
|
|
{
|
|
return (&fib_entry_src_bh_vft[FIB_SOURCE_BH_INTERPOSE]);
|
|
}
|
|
|
|
ASSERT(bh < FIB_SOURCE_BH_MAX);
|
|
return (&fib_entry_src_bh_vft[bh]);
|
|
}
|
|
|
|
static void
|
|
fib_entry_src_copy_default (const fib_entry_src_t *orig_src,
|
|
const fib_entry_t *fib_entry,
|
|
fib_entry_src_t *copy_src)
|
|
{
|
|
clib_memcpy(©_src->u, &orig_src->u, sizeof(copy_src->u));
|
|
}
|
|
|
|
void
|
|
fib_entry_src_behaviour_register (fib_source_behaviour_t bh,
|
|
const fib_entry_src_vft_t *vft)
|
|
{
|
|
fib_entry_src_bh_vft[bh] = *vft;
|
|
|
|
if (NULL == fib_entry_src_bh_vft[bh].fesv_copy)
|
|
{
|
|
fib_entry_src_bh_vft[bh].fesv_copy = fib_entry_src_copy_default;
|
|
}
|
|
}
|
|
|
|
static int
|
|
fib_entry_src_cmp_for_sort (void * v1,
|
|
void * v2)
|
|
{
|
|
fib_entry_src_t *esrc1 = v1, *esrc2 = v2;
|
|
|
|
return (fib_source_get_prio(esrc1->fes_src) -
|
|
fib_source_get_prio(esrc2->fes_src));
|
|
}
|
|
|
|
static void
|
|
fib_entry_src_action_init (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
fib_entry_flag_t flags)
|
|
{
|
|
fib_entry_src_t esrc = {
|
|
.fes_pl = FIB_NODE_INDEX_INVALID,
|
|
.fes_flags = FIB_ENTRY_SRC_FLAG_NONE,
|
|
.fes_src = source,
|
|
.fes_entry_flags = flags,
|
|
};
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, &esrc, fesv_init, (&esrc));
|
|
|
|
vec_add1(fib_entry->fe_srcs, esrc);
|
|
vec_sort_with_function(fib_entry->fe_srcs,
|
|
fib_entry_src_cmp_for_sort);
|
|
}
|
|
|
|
static fib_entry_src_t *
|
|
fib_entry_src_find_i (const fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
u32 *index)
|
|
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
int ii;
|
|
|
|
ii = 0;
|
|
vec_foreach(esrc, fib_entry->fe_srcs)
|
|
{
|
|
if (esrc->fes_src == source)
|
|
{
|
|
if (NULL != index)
|
|
{
|
|
*index = ii;
|
|
}
|
|
return (esrc);
|
|
}
|
|
else
|
|
{
|
|
ii++;
|
|
}
|
|
}
|
|
|
|
return (NULL);
|
|
}
|
|
|
|
fib_entry_src_t *
|
|
fib_entry_src_find (const fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
|
|
{
|
|
return (fib_entry_src_find_i(fib_entry, source, NULL));
|
|
}
|
|
|
|
int
|
|
fib_entry_is_sourced (fib_node_index_t fib_entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
|
|
return (NULL != fib_entry_src_find(fib_entry, source));
|
|
}
|
|
|
|
int
|
|
fib_entry_is_marked (fib_node_index_t fib_entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL == esrc)
|
|
{
|
|
return (0);
|
|
}
|
|
else
|
|
{
|
|
return (!!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_STALE));
|
|
}
|
|
}
|
|
|
|
void
|
|
fib_entry_mark (fib_node_index_t fib_entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_STALE;
|
|
}
|
|
}
|
|
|
|
static fib_entry_src_t *
|
|
fib_entry_src_find_or_create (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
fib_entry_flag_t flags)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL == esrc)
|
|
{
|
|
fib_entry_src_action_init(fib_entry, source, flags);
|
|
}
|
|
|
|
return (fib_entry_src_find(fib_entry, source));
|
|
}
|
|
|
|
static void
|
|
fib_entry_src_action_deinit (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
u32 index = ~0;
|
|
|
|
esrc = fib_entry_src_find_i(fib_entry, source, &index);
|
|
|
|
ASSERT(NULL != esrc);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_deinit, (esrc));
|
|
|
|
fib_path_ext_list_flush(&esrc->fes_path_exts);
|
|
vec_del1(fib_entry->fe_srcs, index);
|
|
vec_sort_with_function(fib_entry->fe_srcs,
|
|
fib_entry_src_cmp_for_sort);
|
|
}
|
|
|
|
fib_entry_src_cover_res_t
|
|
fib_entry_src_action_cover_change (fib_entry_t *fib_entry,
|
|
fib_entry_src_t *esrc)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_cover_change,
|
|
(esrc, fib_entry));
|
|
|
|
fib_entry_src_cover_res_t res = {
|
|
.install = !0,
|
|
.bw_reason = FIB_NODE_BW_REASON_FLAG_NONE,
|
|
};
|
|
return (res);
|
|
}
|
|
|
|
fib_entry_src_cover_res_t
|
|
fib_entry_src_action_cover_update (fib_entry_t *fib_entry,
|
|
fib_entry_src_t *esrc)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_cover_update,
|
|
(esrc, fib_entry));
|
|
|
|
fib_entry_src_cover_res_t res = {
|
|
.install = !0,
|
|
.bw_reason = FIB_NODE_BW_REASON_FLAG_NONE,
|
|
};
|
|
return (res);
|
|
}
|
|
|
|
typedef struct fib_entry_src_collect_forwarding_ctx_t_
|
|
{
|
|
load_balance_path_t *next_hops;
|
|
const fib_entry_t *fib_entry;
|
|
i32 start_source_index, end_source_index;
|
|
fib_forward_chain_type_t fct;
|
|
int n_recursive_constrained;
|
|
u16 preference;
|
|
dpo_proto_t payload_proto;
|
|
} fib_entry_src_collect_forwarding_ctx_t;
|
|
|
|
/**
|
|
* @brief Determine whether this FIB entry should use a load-balance MAP
|
|
* to support PIC edge fast convergence
|
|
*/
|
|
static load_balance_flags_t
|
|
fib_entry_calc_lb_flags (fib_entry_src_collect_forwarding_ctx_t *ctx,
|
|
const fib_entry_src_t *esrc)
|
|
{
|
|
/**
|
|
* We'll use a LB map if the path-list has multiple recursive paths.
|
|
* recursive paths implies BGP, and hence scale.
|
|
*/
|
|
if (ctx->n_recursive_constrained > 1 &&
|
|
fib_path_list_is_popular(esrc->fes_pl))
|
|
{
|
|
return (LOAD_BALANCE_FLAG_USES_MAP);
|
|
}
|
|
return (LOAD_BALANCE_FLAG_NONE);
|
|
}
|
|
|
|
static int
|
|
fib_entry_src_valid_out_label (mpls_label_t label)
|
|
{
|
|
return ((MPLS_LABEL_IS_REAL(label) ||
|
|
MPLS_LABEL_POP == label ||
|
|
MPLS_IETF_IPV4_EXPLICIT_NULL_LABEL == label ||
|
|
MPLS_IETF_IPV6_EXPLICIT_NULL_LABEL == label ||
|
|
MPLS_IETF_IMPLICIT_NULL_LABEL == label));
|
|
}
|
|
|
|
static dpo_proto_t
|
|
fib_prefix_get_payload_proto (const fib_prefix_t *pfx)
|
|
{
|
|
switch (pfx->fp_proto)
|
|
{
|
|
case FIB_PROTOCOL_IP4:
|
|
return (DPO_PROTO_IP4);
|
|
case FIB_PROTOCOL_IP6:
|
|
return (DPO_PROTO_IP6);
|
|
case FIB_PROTOCOL_MPLS:
|
|
return (pfx->fp_payload_proto);
|
|
}
|
|
|
|
ASSERT(0);
|
|
return (DPO_PROTO_IP4);
|
|
}
|
|
|
|
static void
|
|
fib_entry_src_get_path_forwarding (fib_node_index_t path_index,
|
|
fib_entry_src_collect_forwarding_ctx_t *ctx)
|
|
{
|
|
load_balance_path_t *nh;
|
|
|
|
/*
|
|
* no extension => no out-going label for this path. that's OK
|
|
* in the case of an IP or EOS chain, but not for non-EOS
|
|
*/
|
|
switch (ctx->fct)
|
|
{
|
|
case FIB_FORW_CHAIN_TYPE_UNICAST_IP4:
|
|
case FIB_FORW_CHAIN_TYPE_UNICAST_IP6:
|
|
case FIB_FORW_CHAIN_TYPE_MCAST_IP4:
|
|
case FIB_FORW_CHAIN_TYPE_MCAST_IP6:
|
|
case FIB_FORW_CHAIN_TYPE_BIER:
|
|
/*
|
|
* EOS traffic with no label to stack, we need the IP Adj
|
|
*/
|
|
vec_add2(ctx->next_hops, nh, 1);
|
|
|
|
nh->path_index = path_index;
|
|
nh->path_weight = fib_path_get_weight(path_index);
|
|
fib_path_contribute_forwarding(path_index, ctx->fct,
|
|
ctx->payload_proto, &nh->path_dpo);
|
|
|
|
break;
|
|
case FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS:
|
|
if (fib_path_is_exclusive(path_index) ||
|
|
fib_path_is_deag(path_index))
|
|
{
|
|
vec_add2(ctx->next_hops, nh, 1);
|
|
|
|
nh->path_index = path_index;
|
|
nh->path_weight = fib_path_get_weight(path_index);
|
|
fib_path_contribute_forwarding(path_index,
|
|
FIB_FORW_CHAIN_TYPE_MPLS_NON_EOS,
|
|
ctx->payload_proto,
|
|
&nh->path_dpo);
|
|
}
|
|
break;
|
|
case FIB_FORW_CHAIN_TYPE_MPLS_EOS:
|
|
{
|
|
/*
|
|
* no label. we need a chain based on the payload. fixup.
|
|
*/
|
|
vec_add2(ctx->next_hops, nh, 1);
|
|
|
|
nh->path_index = path_index;
|
|
nh->path_weight = fib_path_get_weight(path_index);
|
|
fib_path_contribute_forwarding(path_index,
|
|
ctx->fct,
|
|
ctx->payload_proto,
|
|
&nh->path_dpo);
|
|
fib_path_stack_mpls_disp(path_index,
|
|
ctx->payload_proto,
|
|
FIB_MPLS_LSP_MODE_PIPE,
|
|
&nh->path_dpo);
|
|
|
|
break;
|
|
}
|
|
case FIB_FORW_CHAIN_TYPE_ETHERNET:
|
|
case FIB_FORW_CHAIN_TYPE_NSH:
|
|
ASSERT(0);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static fib_path_list_walk_rc_t
|
|
fib_entry_src_collect_forwarding (fib_node_index_t pl_index,
|
|
fib_node_index_t path_index,
|
|
void *arg)
|
|
{
|
|
fib_entry_src_collect_forwarding_ctx_t *ctx;
|
|
const fib_entry_src_t *esrc;
|
|
fib_path_ext_t *path_ext;
|
|
u32 n_nhs;
|
|
|
|
ctx = arg;
|
|
n_nhs = vec_len(ctx->next_hops);
|
|
|
|
/*
|
|
* walk the paths and extension of the best non-interpose source
|
|
*/
|
|
esrc = &ctx->fib_entry->fe_srcs[ctx->end_source_index];
|
|
|
|
/*
|
|
* if the path is not resolved, don't include it.
|
|
*/
|
|
if (!fib_path_is_resolved(path_index))
|
|
{
|
|
return (FIB_PATH_LIST_WALK_CONTINUE);
|
|
}
|
|
|
|
if (fib_path_is_recursive_constrained(path_index))
|
|
{
|
|
ctx->n_recursive_constrained += 1;
|
|
}
|
|
if (0xffff == ctx->preference)
|
|
{
|
|
/*
|
|
* not set a preference yet, so the first path we encounter
|
|
* sets the preference we are collecting.
|
|
*/
|
|
ctx->preference = fib_path_get_preference(path_index);
|
|
}
|
|
else if (ctx->preference != fib_path_get_preference(path_index))
|
|
{
|
|
/*
|
|
* this path does not belong to the same preference as the
|
|
* previous paths encountered. we are done now.
|
|
*/
|
|
return (FIB_PATH_LIST_WALK_STOP);
|
|
}
|
|
|
|
/*
|
|
* get the matching path-extension for the path being visited.
|
|
*/
|
|
path_ext = fib_path_ext_list_find_by_path_index(&esrc->fes_path_exts,
|
|
path_index);
|
|
|
|
if (NULL != path_ext)
|
|
{
|
|
switch (path_ext->fpe_type)
|
|
{
|
|
case FIB_PATH_EXT_MPLS:
|
|
if (fib_entry_src_valid_out_label(path_ext->fpe_label_stack[0].fml_value))
|
|
{
|
|
/*
|
|
* found a matching extension. stack it to obtain the forwarding
|
|
* info for this path.
|
|
*/
|
|
ctx->next_hops =
|
|
fib_path_ext_stack(path_ext,
|
|
ctx->payload_proto,
|
|
ctx->fct,
|
|
ctx->next_hops);
|
|
}
|
|
else
|
|
{
|
|
fib_entry_src_get_path_forwarding(path_index, ctx);
|
|
}
|
|
break;
|
|
case FIB_PATH_EXT_ADJ:
|
|
if (FIB_PATH_EXT_ADJ_FLAG_REFINES_COVER & path_ext->fpe_adj_flags)
|
|
{
|
|
fib_entry_src_get_path_forwarding(path_index, ctx);
|
|
}
|
|
/*
|
|
* else
|
|
* the path does not refine the cover, meaning that
|
|
* the adjacency does/does not match the sub-net on the link.
|
|
* So this path does not contribute forwarding.
|
|
*/
|
|
break;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
fib_entry_src_get_path_forwarding(path_index, ctx);
|
|
}
|
|
|
|
/*
|
|
* a this point 'ctx' has the DPO the path contributed, plus
|
|
* any labels from path extensions.
|
|
* check if there are any interpose sources that want to contribute
|
|
*/
|
|
if (n_nhs < vec_len(ctx->next_hops))
|
|
{
|
|
/*
|
|
* the path contributed a new choice.
|
|
*/
|
|
const fib_entry_src_vft_t *vft;
|
|
|
|
/*
|
|
* roll up the sources that are interposes
|
|
*/
|
|
i32 index;
|
|
|
|
for (index = ctx->end_source_index;
|
|
index >= ctx->start_source_index;
|
|
index--)
|
|
{
|
|
const dpo_id_t *interposer;
|
|
|
|
esrc = &ctx->fib_entry->fe_srcs[index];
|
|
vft = fib_entry_src_get_vft(esrc);
|
|
|
|
if (!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_CONTRIBUTING) ||
|
|
!(esrc->fes_entry_flags & FIB_ENTRY_FLAG_INTERPOSE))
|
|
continue;
|
|
|
|
ASSERT(vft->fesv_contribute_interpose);
|
|
interposer = vft->fesv_contribute_interpose(esrc, ctx->fib_entry);
|
|
|
|
if (NULL != interposer)
|
|
{
|
|
dpo_id_t clone = DPO_INVALID;
|
|
|
|
dpo_mk_interpose(interposer,
|
|
&ctx->next_hops[n_nhs].path_dpo,
|
|
&clone);
|
|
|
|
dpo_copy(&ctx->next_hops[n_nhs].path_dpo, &clone);
|
|
dpo_reset(&clone);
|
|
}
|
|
}
|
|
}
|
|
|
|
return (FIB_PATH_LIST_WALK_CONTINUE);
|
|
}
|
|
|
|
void
|
|
fib_entry_src_mk_lb (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
fib_forward_chain_type_t fct,
|
|
dpo_id_t *dpo_lb)
|
|
{
|
|
const fib_entry_src_t *esrc;
|
|
dpo_proto_t lb_proto;
|
|
u32 start, end;
|
|
|
|
/*
|
|
* The source passed here is the 'best', i.e. the one the client
|
|
* wants. however, if it's an interpose then it does not contribute
|
|
* the forwarding, the next best source that is not an interpose does.
|
|
* So roll down the sources, to find the best non-interpose
|
|
*/
|
|
vec_foreach_index (start, fib_entry->fe_srcs)
|
|
{
|
|
if (source == fib_entry->fe_srcs[start].fes_src)
|
|
break;
|
|
}
|
|
for (end = start; end < vec_len (fib_entry->fe_srcs); end++)
|
|
{
|
|
if (!(fib_entry->fe_srcs[end].fes_entry_flags &
|
|
FIB_ENTRY_FLAG_INTERPOSE) &&
|
|
(fib_entry->fe_srcs[end].fes_flags &
|
|
FIB_ENTRY_SRC_FLAG_CONTRIBUTING))
|
|
break;
|
|
}
|
|
if (end == vec_len(fib_entry->fe_srcs))
|
|
{
|
|
/* didn't find any contributing non-interpose sources */
|
|
end = start;
|
|
}
|
|
|
|
esrc = &fib_entry->fe_srcs[end];
|
|
|
|
/*
|
|
* If the entry has path extensions then we construct a load-balance
|
|
* by stacking the extensions on the forwarding chains of the paths.
|
|
* Otherwise we use the load-balance of the path-list
|
|
*/
|
|
fib_entry_src_collect_forwarding_ctx_t ctx = {
|
|
.fib_entry = fib_entry,
|
|
.next_hops = NULL,
|
|
.n_recursive_constrained = 0,
|
|
.fct = fct,
|
|
.preference = 0xffff,
|
|
.start_source_index = start,
|
|
.end_source_index = end,
|
|
.payload_proto = fib_prefix_get_payload_proto(&fib_entry->fe_prefix),
|
|
};
|
|
|
|
/*
|
|
* As an optimisation we allocate the vector of next-hops to be sized
|
|
* equal to the maximum number of paths we will need, which is also the
|
|
* most likely number we will need, since in most cases the paths are 'up'.
|
|
*/
|
|
vec_validate(ctx.next_hops, fib_path_list_get_n_paths(esrc->fes_pl));
|
|
vec_reset_length(ctx.next_hops);
|
|
|
|
lb_proto = fib_forw_chain_type_to_dpo_proto(fct);
|
|
|
|
fib_path_list_walk(esrc->fes_pl,
|
|
fib_entry_src_collect_forwarding,
|
|
&ctx);
|
|
|
|
if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_EXCLUSIVE)
|
|
{
|
|
/*
|
|
* the client provided the DPO that the entry should link to.
|
|
* all entries must link to a LB, so if it is an LB already
|
|
* then we can use it.
|
|
*/
|
|
if ((1 == vec_len(ctx.next_hops)) &&
|
|
(DPO_LOAD_BALANCE == ctx.next_hops[0].path_dpo.dpoi_type))
|
|
{
|
|
dpo_copy(dpo_lb, &ctx.next_hops[0].path_dpo);
|
|
dpo_reset(&ctx.next_hops[0].path_dpo);
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (!dpo_id_is_valid(dpo_lb))
|
|
{
|
|
/*
|
|
* first time create
|
|
*/
|
|
if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_MULTICAST)
|
|
{
|
|
dpo_set(dpo_lb,
|
|
DPO_REPLICATE,
|
|
lb_proto,
|
|
MPLS_IS_REPLICATE | replicate_create(0, lb_proto));
|
|
}
|
|
else
|
|
{
|
|
fib_protocol_t flow_hash_proto;
|
|
flow_hash_config_t fhc;
|
|
|
|
/*
|
|
* if the protocol for the LB we are building does not match that
|
|
* of the fib_entry (i.e. we are build the [n]EOS LB for an IPv[46]
|
|
* then the fib_index is not an index that relates to the table
|
|
* type we need. So get the default flow-hash config instead.
|
|
*/
|
|
flow_hash_proto = dpo_proto_to_fib(lb_proto);
|
|
if (fib_entry->fe_prefix.fp_proto != flow_hash_proto)
|
|
{
|
|
fhc = fib_table_get_default_flow_hash_config(flow_hash_proto);
|
|
}
|
|
else
|
|
{
|
|
fhc = fib_table_get_flow_hash_config(fib_entry->fe_fib_index,
|
|
flow_hash_proto);
|
|
}
|
|
|
|
dpo_set(dpo_lb,
|
|
DPO_LOAD_BALANCE,
|
|
lb_proto,
|
|
load_balance_create(0, lb_proto, fhc));
|
|
}
|
|
}
|
|
|
|
if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_MULTICAST)
|
|
{
|
|
/*
|
|
* MPLS multicast
|
|
*/
|
|
replicate_multipath_update(dpo_lb, ctx.next_hops);
|
|
}
|
|
else
|
|
{
|
|
load_balance_multipath_update(dpo_lb,
|
|
ctx.next_hops,
|
|
fib_entry_calc_lb_flags(&ctx, esrc));
|
|
vec_free(ctx.next_hops);
|
|
|
|
/*
|
|
* if this entry is sourced by the uRPF-exempt source then we
|
|
* append the always present local0 interface (index 0) to the
|
|
* uRPF list so it is not empty. that way packets pass the loose check.
|
|
*/
|
|
index_t ui = fib_path_list_get_urpf(esrc->fes_pl);
|
|
|
|
if ((fib_entry_is_sourced(fib_entry_get_index(fib_entry),
|
|
FIB_SOURCE_URPF_EXEMPT) ||
|
|
(esrc->fes_entry_flags & FIB_ENTRY_FLAG_LOOSE_URPF_EXEMPT))&&
|
|
(0 == fib_urpf_check_size(ui)))
|
|
{
|
|
/*
|
|
* The uRPF list we get from the path-list is shared by all
|
|
* other users of the list, but the uRPF exemption applies
|
|
* only to this prefix. So we need our own list.
|
|
*/
|
|
ui = fib_urpf_list_alloc_and_lock();
|
|
fib_urpf_list_append(ui, 0);
|
|
fib_urpf_list_bake(ui);
|
|
load_balance_set_urpf(dpo_lb->dpoi_index, ui);
|
|
fib_urpf_list_unlock(ui);
|
|
}
|
|
else
|
|
{
|
|
load_balance_set_urpf(dpo_lb->dpoi_index, ui);
|
|
}
|
|
load_balance_set_fib_entry_flags(dpo_lb->dpoi_index,
|
|
fib_entry_get_flags_i(fib_entry));
|
|
}
|
|
}
|
|
|
|
void
|
|
fib_entry_src_action_install (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
{
|
|
/*
|
|
* Install the forwarding chain for the given source into the forwarding
|
|
* tables
|
|
*/
|
|
fib_forward_chain_type_t fct;
|
|
int insert;
|
|
|
|
fct = fib_entry_get_default_chain_type(fib_entry);
|
|
|
|
/*
|
|
* Every entry has its own load-balance object. All changes to the entry's
|
|
* forwarding result in an inplace modify of the load-balance. This means
|
|
* the load-balance object only needs to be added to the forwarding
|
|
* DB once, when it is created.
|
|
*/
|
|
insert = !dpo_id_is_valid(&fib_entry->fe_lb);
|
|
|
|
fib_entry_src_mk_lb(fib_entry, source, fct, &fib_entry->fe_lb);
|
|
|
|
ASSERT(dpo_id_is_valid(&fib_entry->fe_lb));
|
|
FIB_ENTRY_DBG(fib_entry, "install: %d", fib_entry->fe_lb);
|
|
|
|
/*
|
|
* insert the adj into the data-plane forwarding trie
|
|
*/
|
|
if (insert)
|
|
{
|
|
fib_table_fwding_dpo_update(fib_entry->fe_fib_index,
|
|
&fib_entry->fe_prefix,
|
|
&fib_entry->fe_lb);
|
|
}
|
|
|
|
/*
|
|
* if any of the other chain types are already created they will need
|
|
* updating too
|
|
*/
|
|
fib_entry_delegate_type_t fdt;
|
|
fib_entry_delegate_t *fed;
|
|
|
|
FOR_EACH_DELEGATE_CHAIN(fib_entry, fdt, fed,
|
|
{
|
|
fib_entry_src_mk_lb(fib_entry, source,
|
|
fib_entry_delegate_type_to_chain_type(fdt),
|
|
&fed->fd_dpo);
|
|
});
|
|
}
|
|
|
|
void
|
|
fib_entry_src_action_uninstall (fib_entry_t *fib_entry)
|
|
{
|
|
/*
|
|
* uninstall the forwarding chain from the forwarding tables
|
|
*/
|
|
FIB_ENTRY_DBG(fib_entry, "uninstall");
|
|
|
|
if (dpo_id_is_valid(&fib_entry->fe_lb))
|
|
{
|
|
fib_table_fwding_dpo_remove(
|
|
fib_entry->fe_fib_index,
|
|
&fib_entry->fe_prefix,
|
|
&fib_entry->fe_lb);
|
|
|
|
vlib_worker_wait_one_loop();
|
|
dpo_reset(&fib_entry->fe_lb);
|
|
}
|
|
}
|
|
|
|
static void
|
|
fib_entry_recursive_loop_detect_i (fib_node_index_t path_list_index)
|
|
{
|
|
fib_node_index_t *entries = NULL;
|
|
|
|
fib_path_list_recursive_loop_detect(path_list_index, &entries);
|
|
|
|
vec_free(entries);
|
|
}
|
|
|
|
/*
|
|
* fib_entry_src_action_copy
|
|
*
|
|
* copy a source data from another entry to this one
|
|
*/
|
|
static fib_entry_t *
|
|
fib_entry_src_action_copy (fib_entry_t *fib_entry,
|
|
const fib_entry_src_t *orig_src)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find_or_create(fib_entry,
|
|
orig_src->fes_src,
|
|
orig_src->fes_entry_flags);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_copy,
|
|
(orig_src, fib_entry, esrc));
|
|
|
|
fib_path_list_unlock(esrc->fes_pl);
|
|
|
|
/*
|
|
* copy over all the data ...
|
|
*/
|
|
esrc->fes_flags = orig_src->fes_flags;
|
|
esrc->fes_pl = orig_src->fes_pl;
|
|
|
|
/*
|
|
* ... then update
|
|
*/
|
|
esrc->fes_ref_count = 1;
|
|
esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_INHERITED;
|
|
esrc->fes_flags &= ~(FIB_ENTRY_SRC_FLAG_ACTIVE |
|
|
FIB_ENTRY_SRC_FLAG_CONTRIBUTING);
|
|
esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_COVERED_INHERIT;
|
|
|
|
/*
|
|
* the source owns a lock on the entry
|
|
*/
|
|
fib_path_list_lock(esrc->fes_pl);
|
|
fib_entry_lock(fib_entry_get_index(fib_entry));
|
|
|
|
return (fib_entry);
|
|
}
|
|
|
|
/*
|
|
* fib_entry_src_action_update
|
|
*
|
|
* copy a source data from another entry to this one
|
|
*/
|
|
static fib_entry_src_t *
|
|
fib_entry_src_action_update_from_cover (fib_entry_t *fib_entry,
|
|
const fib_entry_src_t *orig_src)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find_or_create(fib_entry,
|
|
orig_src->fes_src,
|
|
orig_src->fes_entry_flags);
|
|
|
|
/*
|
|
* the source owns a lock on the entry
|
|
*/
|
|
fib_path_list_unlock(esrc->fes_pl);
|
|
esrc->fes_pl = orig_src->fes_pl;
|
|
fib_path_list_lock(esrc->fes_pl);
|
|
|
|
return (esrc);
|
|
}
|
|
|
|
static fib_table_walk_rc_t
|
|
fib_entry_src_covered_inherit_add_i (fib_entry_t *fib_entry,
|
|
const fib_entry_src_t *cover_src)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, cover_src->fes_src);
|
|
|
|
if (cover_src == esrc)
|
|
{
|
|
return (FIB_TABLE_WALK_CONTINUE);
|
|
}
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
/*
|
|
* the covered entry already has this source.
|
|
*/
|
|
if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT)
|
|
{
|
|
/*
|
|
* the covered source is itself a COVERED_INHERIT, i.e.
|
|
* it also pushes this source down the sub-tree.
|
|
* We consider this more specific covered to be the owner
|
|
* of the sub-tree from this point down.
|
|
*/
|
|
return (FIB_TABLE_WALK_SUB_TREE_STOP);
|
|
}
|
|
if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED)
|
|
{
|
|
/*
|
|
* The covered's source data has been inherited, presumably
|
|
* from this cover, i.e. this is a modify.
|
|
*/
|
|
fib_source_t best_source;
|
|
|
|
best_source = fib_entry_get_best_source(
|
|
fib_entry_get_index(fib_entry));
|
|
fib_entry_src_action_update_from_cover(fib_entry, cover_src);
|
|
fib_entry_source_change(fib_entry, best_source, cover_src->fes_src);
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* The covered's source was not inherited and it is also
|
|
* not inheriting. Nevertheless, it still owns the sub-tree from
|
|
* this point down.
|
|
*/
|
|
return (FIB_TABLE_WALK_SUB_TREE_STOP);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* The covered does not have this source - add it.
|
|
*/
|
|
fib_source_t best_source;
|
|
|
|
best_source = fib_entry_get_best_source(
|
|
fib_entry_get_index(fib_entry));
|
|
|
|
fib_entry_src_action_copy(fib_entry, cover_src);
|
|
fib_entry_source_change(fib_entry, best_source, cover_src->fes_src);
|
|
|
|
}
|
|
return (FIB_TABLE_WALK_CONTINUE);
|
|
}
|
|
|
|
static fib_table_walk_rc_t
|
|
fib_entry_src_covered_inherit_walk_add (fib_node_index_t fei,
|
|
void *ctx)
|
|
{
|
|
return (fib_entry_src_covered_inherit_add_i(fib_entry_get(fei), ctx));
|
|
}
|
|
|
|
static fib_table_walk_rc_t
|
|
fib_entry_src_covered_inherit_walk_remove (fib_node_index_t fei,
|
|
void *ctx)
|
|
{
|
|
fib_entry_src_t *cover_src, *esrc;
|
|
fib_entry_t *fib_entry;
|
|
|
|
fib_entry = fib_entry_get(fei);
|
|
|
|
cover_src = ctx;
|
|
esrc = fib_entry_src_find(fib_entry, cover_src->fes_src);
|
|
|
|
if (cover_src == esrc)
|
|
{
|
|
return (FIB_TABLE_WALK_CONTINUE);
|
|
}
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
/*
|
|
* the covered entry already has this source.
|
|
*/
|
|
if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT)
|
|
{
|
|
/*
|
|
* the covered source is itself a COVERED_INHERIT, i.e.
|
|
* it also pushes this source down the sub-tree.
|
|
* We consider this more specific covered to be the owner
|
|
* of the sub-tree from this point down.
|
|
*/
|
|
return (FIB_TABLE_WALK_SUB_TREE_STOP);
|
|
}
|
|
if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED)
|
|
{
|
|
/*
|
|
* The covered's source data has been inherited, presumably
|
|
* from this cover
|
|
*/
|
|
fib_entry_src_flag_t remaining;
|
|
|
|
remaining = fib_entry_special_remove(fei, cover_src->fes_src);
|
|
|
|
ASSERT(FIB_ENTRY_SRC_FLAG_ADDED == remaining);
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* The covered's source was not inherited and it is also
|
|
* not inheriting. Nevertheless, it still owns the sub-tree from
|
|
* this point down.
|
|
*/
|
|
return (FIB_TABLE_WALK_SUB_TREE_STOP);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* The covered does not have this source - that's an error,
|
|
* since it should have inherited, but there is nothing we can do
|
|
* about it now.
|
|
*/
|
|
}
|
|
return (FIB_TABLE_WALK_CONTINUE);
|
|
}
|
|
|
|
void
|
|
fib_entry_src_inherit (const fib_entry_t *cover,
|
|
fib_entry_t *covered)
|
|
{
|
|
CLIB_UNUSED(fib_source_t source);
|
|
const fib_entry_src_t *src;
|
|
|
|
FOR_EACH_SRC_ADDED(cover, src, source,
|
|
({
|
|
if ((src->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) ||
|
|
(src->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED))
|
|
{
|
|
fib_entry_src_covered_inherit_add_i(covered, src);
|
|
}
|
|
}))
|
|
}
|
|
|
|
static void
|
|
fib_entry_src_covered_inherit_add (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE);
|
|
|
|
if ((esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) ||
|
|
(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED))
|
|
{
|
|
fib_table_sub_tree_walk(fib_entry->fe_fib_index,
|
|
fib_entry->fe_prefix.fp_proto,
|
|
&fib_entry->fe_prefix,
|
|
fib_entry_src_covered_inherit_walk_add,
|
|
esrc);
|
|
}
|
|
}
|
|
|
|
static void
|
|
fib_entry_src_covered_inherit_remove (fib_entry_t *fib_entry,
|
|
fib_entry_src_t *esrc)
|
|
|
|
{
|
|
ASSERT(!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE));
|
|
|
|
if (esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT)
|
|
{
|
|
fib_table_sub_tree_walk(fib_entry->fe_fib_index,
|
|
fib_entry->fe_prefix.fp_proto,
|
|
&fib_entry->fe_prefix,
|
|
fib_entry_src_covered_inherit_walk_remove,
|
|
esrc);
|
|
}
|
|
}
|
|
|
|
void
|
|
fib_entry_src_action_activate (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
|
|
{
|
|
int houston_we_are_go_for_install;
|
|
const fib_entry_src_vft_t *vft;
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
ASSERT(!(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE));
|
|
ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ADDED);
|
|
|
|
esrc->fes_flags |= (FIB_ENTRY_SRC_FLAG_ACTIVE |
|
|
FIB_ENTRY_SRC_FLAG_CONTRIBUTING);
|
|
vft = fib_entry_src_get_vft(esrc);
|
|
|
|
if (NULL != vft->fesv_activate)
|
|
{
|
|
houston_we_are_go_for_install = vft->fesv_activate(esrc, fib_entry);
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* the source is not providing an activate function, we'll assume
|
|
* therefore it has no objection to installing the entry
|
|
*/
|
|
houston_we_are_go_for_install = !0;
|
|
}
|
|
|
|
/*
|
|
* link to the path-list provided by the source, and go check
|
|
* if that forms any loops in the graph.
|
|
*/
|
|
fib_entry->fe_parent = esrc->fes_pl;
|
|
fib_entry->fe_sibling =
|
|
fib_path_list_child_add(fib_entry->fe_parent,
|
|
FIB_NODE_TYPE_ENTRY,
|
|
fib_entry_get_index(fib_entry));
|
|
|
|
fib_entry_recursive_loop_detect_i(fib_entry->fe_parent);
|
|
|
|
FIB_ENTRY_DBG(fib_entry, "activate: %d",
|
|
fib_entry->fe_parent);
|
|
|
|
/*
|
|
* If this source should push its state to covered prefixs, do that now.
|
|
*/
|
|
fib_entry_src_covered_inherit_add(fib_entry, source);
|
|
|
|
if (0 != houston_we_are_go_for_install)
|
|
{
|
|
fib_entry_src_action_install(fib_entry, source);
|
|
}
|
|
else
|
|
{
|
|
fib_entry_src_action_uninstall(fib_entry);
|
|
}
|
|
}
|
|
|
|
void
|
|
fib_entry_src_action_deactivate (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
|
|
{
|
|
fib_node_index_t path_list_index;
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_deactivate,
|
|
(esrc, fib_entry));
|
|
|
|
esrc->fes_flags &= ~(FIB_ENTRY_SRC_FLAG_ACTIVE |
|
|
FIB_ENTRY_SRC_FLAG_CONTRIBUTING);
|
|
|
|
FIB_ENTRY_DBG(fib_entry, "deactivate: %d", fib_entry->fe_parent);
|
|
|
|
/*
|
|
* If this source should pull its state from covered prefixs, do that now.
|
|
* If this source also has the INHERITED flag set then it has a cover
|
|
* that wants to push down forwarding. We only want the covereds to see
|
|
* one update.
|
|
*/
|
|
fib_entry_src_covered_inherit_remove(fib_entry, esrc);
|
|
|
|
/*
|
|
* un-link from an old path-list. Check for any loops this will clear
|
|
*/
|
|
path_list_index = fib_entry->fe_parent;
|
|
fib_entry->fe_parent = FIB_NODE_INDEX_INVALID;
|
|
|
|
fib_entry_recursive_loop_detect_i(path_list_index);
|
|
|
|
/*
|
|
* this will unlock the path-list, so it may be invalid thereafter.
|
|
*/
|
|
fib_path_list_child_remove(path_list_index, fib_entry->fe_sibling);
|
|
fib_entry->fe_sibling = FIB_NODE_INDEX_INVALID;
|
|
}
|
|
|
|
static void
|
|
fib_entry_src_action_fwd_update (const fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
vec_foreach(esrc, fib_entry->fe_srcs)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_fwd_update,
|
|
(esrc, fib_entry, source));
|
|
}
|
|
}
|
|
|
|
void
|
|
fib_entry_src_action_reactivate (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
{
|
|
fib_node_index_t path_list_index;
|
|
const fib_entry_src_vft_t *vft;
|
|
fib_entry_src_t *esrc;
|
|
int remain_installed;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE);
|
|
|
|
FIB_ENTRY_DBG(fib_entry, "reactivate: %d to %d",
|
|
fib_entry->fe_parent,
|
|
esrc->fes_pl);
|
|
|
|
/*
|
|
* call the source to reactive and get the go/no-go to remain installed
|
|
*/
|
|
vft = fib_entry_src_get_vft(esrc);
|
|
|
|
if (NULL != vft->fesv_reactivate)
|
|
{
|
|
remain_installed = vft->fesv_reactivate(esrc, fib_entry);
|
|
}
|
|
else
|
|
{
|
|
remain_installed = 1;
|
|
}
|
|
|
|
if (fib_entry->fe_parent != esrc->fes_pl)
|
|
{
|
|
/*
|
|
* un-link from an old path-list. Check for any loops this will clear
|
|
*/
|
|
path_list_index = fib_entry->fe_parent;
|
|
fib_entry->fe_parent = FIB_NODE_INDEX_INVALID;
|
|
|
|
/*
|
|
* temporary lock so it doesn't get deleted when this entry is no
|
|
* longer a child.
|
|
*/
|
|
fib_path_list_lock(path_list_index);
|
|
|
|
/*
|
|
* this entry is no longer a child. after unlinking check if any loops
|
|
* were broken
|
|
*/
|
|
fib_path_list_child_remove(path_list_index,
|
|
fib_entry->fe_sibling);
|
|
|
|
fib_entry_recursive_loop_detect_i(path_list_index);
|
|
|
|
/*
|
|
* link to the path-list provided by the source, and go check
|
|
* if that forms any loops in the graph.
|
|
*/
|
|
fib_entry->fe_parent = esrc->fes_pl;
|
|
fib_entry->fe_sibling =
|
|
fib_path_list_child_add(fib_entry->fe_parent,
|
|
FIB_NODE_TYPE_ENTRY,
|
|
fib_entry_get_index(fib_entry));
|
|
|
|
fib_entry_recursive_loop_detect_i(fib_entry->fe_parent);
|
|
fib_path_list_unlock(path_list_index);
|
|
|
|
/*
|
|
* If this source should push its state to covered prefixs, do that now.
|
|
*/
|
|
fib_entry_src_covered_inherit_add(fib_entry, source);
|
|
}
|
|
|
|
if (!remain_installed)
|
|
{
|
|
fib_entry_src_action_uninstall(fib_entry);
|
|
}
|
|
else
|
|
{
|
|
fib_entry_src_action_install(fib_entry, source);
|
|
}
|
|
fib_entry_src_action_fwd_update(fib_entry, source);
|
|
}
|
|
|
|
fib_entry_t *
|
|
fib_entry_src_action_installed (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_installed,
|
|
(esrc, fib_entry));
|
|
|
|
fib_entry_src_action_fwd_update(fib_entry, source);
|
|
|
|
return (fib_entry);
|
|
}
|
|
|
|
/*
|
|
* fib_entry_src_action_add
|
|
*
|
|
* Adding a source can result in a new fib_entry being created, which
|
|
* can inturn mean the pool is realloc'd and thus the entry passed as
|
|
* an argument it also realloc'd
|
|
* @return the original entry
|
|
*/
|
|
fib_entry_t *
|
|
fib_entry_src_action_add (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
fib_entry_flag_t flags,
|
|
const dpo_id_t *dpo)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find_or_create(fib_entry, source, flags);
|
|
|
|
ASSERT(esrc->fes_ref_count < 255);
|
|
esrc->fes_ref_count++;
|
|
|
|
if (flags != esrc->fes_entry_flags)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_flags_change,
|
|
(esrc, fib_entry, flags));
|
|
}
|
|
esrc->fes_entry_flags = flags;
|
|
|
|
if (1 != esrc->fes_ref_count)
|
|
{
|
|
/*
|
|
* we only want to add the source on the 0->1 transition
|
|
*/
|
|
return (fib_entry);
|
|
}
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_add,
|
|
(esrc,
|
|
fib_entry,
|
|
flags,
|
|
fib_entry_get_dpo_proto(fib_entry),
|
|
dpo));
|
|
|
|
esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_ADDED;
|
|
|
|
fib_path_list_lock(esrc->fes_pl);
|
|
|
|
/*
|
|
* the source owns a lock on the entry
|
|
*/
|
|
fib_entry_lock(fib_entry_get_index(fib_entry));
|
|
|
|
return (fib_entry);
|
|
}
|
|
|
|
/*
|
|
* fib_entry_src_action_update
|
|
*
|
|
* Adding a source can result in a new fib_entry being created, which
|
|
* can inturn mean the pool is realloc'd and thus the entry passed as
|
|
* an argument it also realloc'd
|
|
* @return the original entry
|
|
*/
|
|
fib_entry_t *
|
|
fib_entry_src_action_update (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
fib_entry_flag_t flags,
|
|
const dpo_id_t *dpo)
|
|
{
|
|
fib_node_index_t old_path_list_index;
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find_or_create(fib_entry, source, flags);
|
|
|
|
if (NULL == esrc)
|
|
{
|
|
return (fib_entry_src_action_add(fib_entry, source, flags, dpo));
|
|
}
|
|
|
|
old_path_list_index = esrc->fes_pl;
|
|
esrc->fes_entry_flags = flags;
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_add,
|
|
(esrc,
|
|
fib_entry,
|
|
flags,
|
|
fib_entry_get_dpo_proto(fib_entry),
|
|
dpo));
|
|
|
|
esrc->fes_flags |= FIB_ENTRY_SRC_FLAG_ADDED;
|
|
|
|
fib_path_list_lock(esrc->fes_pl);
|
|
fib_path_list_unlock(old_path_list_index);
|
|
|
|
return (fib_entry);
|
|
}
|
|
|
|
fib_entry_src_flag_t
|
|
fib_entry_src_action_remove_or_update_inherit (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL == esrc)
|
|
return (FIB_ENTRY_SRC_FLAG_ACTIVE);
|
|
|
|
if ((esrc->fes_entry_flags & FIB_ENTRY_FLAG_COVERED_INHERIT) &&
|
|
(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_INHERITED))
|
|
{
|
|
fib_entry_src_t *cover_src;
|
|
fib_node_index_t coveri;
|
|
fib_entry_t *cover;
|
|
|
|
/*
|
|
* this source was pushing inherited state, but so is its
|
|
* cover. Now that this source is going away, we need to
|
|
* pull the covers forwarding and use it to update the covereds.
|
|
* Go grab the path-list from the cover, rather than start a walk from
|
|
* the cover, so we don't recursively update this entry.
|
|
*/
|
|
coveri = fib_table_get_less_specific(fib_entry->fe_fib_index,
|
|
&fib_entry->fe_prefix);
|
|
|
|
/*
|
|
* only the default route has itself as its own cover, but the
|
|
* default route cannot have inherited from something else.
|
|
*/
|
|
ASSERT(coveri != fib_entry_get_index(fib_entry));
|
|
|
|
cover = fib_entry_get(coveri);
|
|
cover_src = fib_entry_src_find(cover, source);
|
|
|
|
ASSERT(NULL != cover_src);
|
|
|
|
esrc = fib_entry_src_action_update_from_cover(fib_entry, cover_src);
|
|
esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_COVERED_INHERIT;
|
|
|
|
/*
|
|
* Now push the new state from the cover down to the covereds
|
|
*/
|
|
fib_entry_src_covered_inherit_add(fib_entry, source);
|
|
|
|
return (esrc->fes_flags);
|
|
}
|
|
else
|
|
{
|
|
return (fib_entry_src_action_remove(fib_entry, source));
|
|
}
|
|
}
|
|
|
|
fib_entry_src_flag_t
|
|
fib_entry_src_action_remove (fib_entry_t *fib_entry,
|
|
fib_source_t source)
|
|
|
|
{
|
|
fib_node_index_t old_path_list;
|
|
fib_entry_src_flag_t sflags;
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL == esrc)
|
|
return (FIB_ENTRY_SRC_FLAG_ACTIVE);
|
|
|
|
esrc->fes_ref_count--;
|
|
sflags = esrc->fes_flags;
|
|
|
|
if (0 != esrc->fes_ref_count)
|
|
{
|
|
/*
|
|
* only remove the source on the 1->0 transisition
|
|
*/
|
|
return (sflags);
|
|
}
|
|
|
|
if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ACTIVE)
|
|
{
|
|
fib_entry_src_action_deactivate(fib_entry, source);
|
|
}
|
|
else if (esrc->fes_flags & FIB_ENTRY_SRC_FLAG_CONTRIBUTING)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_deactivate,
|
|
(esrc, fib_entry));
|
|
esrc->fes_flags &= ~FIB_ENTRY_SRC_FLAG_CONTRIBUTING;
|
|
}
|
|
|
|
old_path_list = esrc->fes_pl;
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_remove, (esrc));
|
|
|
|
fib_path_list_unlock(old_path_list);
|
|
fib_entry_unlock(fib_entry_get_index(fib_entry));
|
|
|
|
sflags &= ~FIB_ENTRY_SRC_FLAG_ADDED;
|
|
fib_entry_src_action_deinit(fib_entry, source);
|
|
|
|
return (sflags);
|
|
}
|
|
|
|
/*
|
|
* fib_route_attached_cross_table
|
|
*
|
|
* Return true the the route is attached via an interface that
|
|
* is not in the same table as the route
|
|
*/
|
|
static int
|
|
fib_route_attached_cross_table (const fib_entry_t *fib_entry,
|
|
const fib_route_path_t *rpath)
|
|
{
|
|
const fib_prefix_t *pfx = &fib_entry->fe_prefix;
|
|
|
|
switch (pfx->fp_proto)
|
|
{
|
|
case FIB_PROTOCOL_MPLS:
|
|
/* MPLS routes are never imported/exported */
|
|
return (0);
|
|
case FIB_PROTOCOL_IP6:
|
|
/* Ignore link local addresses these also can't be imported/exported */
|
|
if (ip6_address_is_link_local_unicast (&pfx->fp_addr.ip6))
|
|
{
|
|
return (0);
|
|
}
|
|
break;
|
|
case FIB_PROTOCOL_IP4:
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* an attached path and entry's fib index not equal to interface's index
|
|
*/
|
|
if (fib_route_path_is_attached(rpath) &&
|
|
fib_entry->fe_fib_index !=
|
|
fib_table_get_index_for_sw_if_index(fib_entry_get_proto(fib_entry),
|
|
rpath->frp_sw_if_index))
|
|
{
|
|
return (!0);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
fib_path_list_flags_t
|
|
fib_entry_src_flags_2_path_list_flags (fib_entry_flag_t eflags)
|
|
{
|
|
fib_path_list_flags_t plf = FIB_PATH_LIST_FLAG_NONE;
|
|
|
|
if (eflags & FIB_ENTRY_FLAG_DROP)
|
|
{
|
|
plf |= FIB_PATH_LIST_FLAG_DROP;
|
|
}
|
|
if (eflags & FIB_ENTRY_FLAG_EXCLUSIVE)
|
|
{
|
|
plf |= FIB_PATH_LIST_FLAG_EXCLUSIVE;
|
|
}
|
|
if (eflags & FIB_ENTRY_FLAG_LOCAL)
|
|
{
|
|
plf |= FIB_PATH_LIST_FLAG_LOCAL;
|
|
}
|
|
|
|
return (plf);
|
|
}
|
|
|
|
static void
|
|
fib_entry_flags_update (const fib_entry_t *fib_entry,
|
|
const fib_route_path_t *rpaths,
|
|
fib_path_list_flags_t *pl_flags,
|
|
fib_entry_src_t *esrc)
|
|
{
|
|
const fib_route_path_t *rpath;
|
|
|
|
vec_foreach(rpath, rpaths)
|
|
{
|
|
if ((esrc->fes_src == FIB_SOURCE_API) ||
|
|
(esrc->fes_src == FIB_SOURCE_CLI))
|
|
{
|
|
if (fib_route_path_is_attached(rpath))
|
|
{
|
|
esrc->fes_entry_flags |= FIB_ENTRY_FLAG_ATTACHED;
|
|
}
|
|
else
|
|
{
|
|
esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_ATTACHED;
|
|
}
|
|
if (rpath->frp_flags & FIB_ROUTE_PATH_DEAG)
|
|
{
|
|
esrc->fes_entry_flags |= FIB_ENTRY_FLAG_LOOSE_URPF_EXEMPT;
|
|
}
|
|
if (rpath->frp_flags & FIB_ROUTE_PATH_DROP)
|
|
{
|
|
esrc->fes_entry_flags |= FIB_ENTRY_FLAG_NO_ATTACHED_EXPORT;
|
|
}
|
|
}
|
|
if (fib_route_attached_cross_table(fib_entry, rpath) &&
|
|
!(esrc->fes_entry_flags & FIB_ENTRY_FLAG_NO_ATTACHED_EXPORT))
|
|
{
|
|
esrc->fes_entry_flags |= FIB_ENTRY_FLAG_IMPORT;
|
|
}
|
|
else
|
|
{
|
|
esrc->fes_entry_flags &= ~FIB_ENTRY_FLAG_IMPORT;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* fib_entry_src_action_add
|
|
*
|
|
* Adding a source can result in a new fib_entry being created, which
|
|
* can inturn mean the pool is realloc'd and thus the entry passed as
|
|
* an argument it also realloc'd
|
|
* @return the entry
|
|
*/
|
|
fib_entry_t*
|
|
fib_entry_src_action_path_add (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
fib_entry_flag_t flags,
|
|
const fib_route_path_t *rpaths)
|
|
{
|
|
fib_node_index_t old_path_list;
|
|
fib_path_list_flags_t pl_flags;
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
if (NULL == esrc)
|
|
{
|
|
const dpo_id_t *dpo;
|
|
|
|
if (flags == FIB_ENTRY_FLAG_EXCLUSIVE) {
|
|
dpo = &rpaths->dpo;
|
|
} else {
|
|
dpo = drop_dpo_get(fib_entry_get_dpo_proto(fib_entry));
|
|
}
|
|
|
|
fib_entry =
|
|
fib_entry_src_action_add(fib_entry,
|
|
source,
|
|
flags,
|
|
dpo);
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
}
|
|
|
|
/*
|
|
* we are no doubt modifying a path-list. If the path-list
|
|
* is shared, and hence not modifiable, then the index returned
|
|
* will be for a different path-list. This FIB entry to needs
|
|
* to maintain its lock appropriately.
|
|
*/
|
|
old_path_list = esrc->fes_pl;
|
|
|
|
ASSERT(FIB_ENTRY_SRC_VFT_EXISTS(esrc, fesv_path_add));
|
|
|
|
pl_flags = fib_entry_src_flags_2_path_list_flags(fib_entry_get_flags_i(fib_entry));
|
|
fib_entry_flags_update(fib_entry, rpaths, &pl_flags, esrc);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_path_add,
|
|
(esrc, fib_entry, pl_flags, rpaths));
|
|
|
|
fib_path_list_lock(esrc->fes_pl);
|
|
fib_path_list_unlock(old_path_list);
|
|
|
|
return (fib_entry);
|
|
}
|
|
|
|
/*
|
|
* fib_entry_src_action_swap
|
|
*
|
|
* The source is providing new paths to replace the old ones.
|
|
* Adding a source can result in a new fib_entry being created, which
|
|
* can inturn mean the pool is realloc'd and thus the entry passed as
|
|
* an argument it also realloc'd
|
|
* @return the entry
|
|
*/
|
|
fib_entry_t*
|
|
fib_entry_src_action_path_swap (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
fib_entry_flag_t flags,
|
|
const fib_route_path_t *rpaths)
|
|
{
|
|
fib_node_index_t old_path_list;
|
|
fib_path_list_flags_t pl_flags;
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL == esrc)
|
|
{
|
|
const dpo_id_t *dpo;
|
|
|
|
if (flags == FIB_ENTRY_FLAG_EXCLUSIVE) {
|
|
dpo = &rpaths->dpo;
|
|
} else {
|
|
dpo = drop_dpo_get(fib_entry_get_dpo_proto(fib_entry));
|
|
}
|
|
|
|
fib_entry = fib_entry_src_action_add(fib_entry,
|
|
source,
|
|
flags,
|
|
dpo);
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
}
|
|
else
|
|
{
|
|
if (flags != esrc->fes_entry_flags)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_flags_change,
|
|
(esrc, fib_entry, flags));
|
|
}
|
|
esrc->fes_entry_flags = flags;
|
|
}
|
|
|
|
/*
|
|
* swapping paths may create a new path-list (or may use an existing shared)
|
|
* but we are certainly getting a different one. This FIB entry to needs
|
|
* to maintain its lock appropriately.
|
|
*/
|
|
old_path_list = esrc->fes_pl;
|
|
|
|
ASSERT(FIB_ENTRY_SRC_VFT_EXISTS(esrc, fesv_path_swap));
|
|
|
|
pl_flags = fib_entry_src_flags_2_path_list_flags(flags);
|
|
|
|
fib_entry_flags_update(fib_entry, rpaths, &pl_flags, esrc);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_path_swap,
|
|
(esrc, fib_entry,
|
|
pl_flags, rpaths));
|
|
|
|
fib_path_list_lock(esrc->fes_pl);
|
|
fib_path_list_unlock(old_path_list);
|
|
|
|
return (fib_entry);
|
|
}
|
|
|
|
fib_entry_src_flag_t
|
|
fib_entry_src_action_path_remove (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
const fib_route_path_t *rpaths)
|
|
{
|
|
fib_path_list_flags_t pl_flags;
|
|
fib_node_index_t old_path_list;
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
ASSERT(NULL != esrc);
|
|
ASSERT(esrc->fes_flags & FIB_ENTRY_SRC_FLAG_ADDED);
|
|
|
|
/*
|
|
* we no doubt modifying a path-list. If the path-list
|
|
* is shared, and hence not modifiable, then the index returned
|
|
* will be for a different path-list. This FIB entry to needs
|
|
* to maintain its lock appropriately.
|
|
*/
|
|
old_path_list = esrc->fes_pl;
|
|
|
|
ASSERT(FIB_ENTRY_SRC_VFT_EXISTS(esrc, fesv_path_remove));
|
|
|
|
pl_flags = fib_entry_src_flags_2_path_list_flags(fib_entry_get_flags_i(fib_entry));
|
|
fib_entry_flags_update(fib_entry, rpaths, &pl_flags, esrc);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_path_remove,
|
|
(esrc, pl_flags, rpaths));
|
|
|
|
/*
|
|
* lock the new path-list, unlock the old if it had one
|
|
*/
|
|
fib_path_list_unlock(old_path_list);
|
|
|
|
if (FIB_NODE_INDEX_INVALID != esrc->fes_pl) {
|
|
fib_path_list_lock(esrc->fes_pl);
|
|
return (FIB_ENTRY_SRC_FLAG_ADDED);
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* no more paths left from this source
|
|
*/
|
|
fib_entry_src_action_remove_or_update_inherit(fib_entry, source);
|
|
return (FIB_ENTRY_SRC_FLAG_NONE);
|
|
}
|
|
}
|
|
|
|
u8*
|
|
fib_entry_src_format (fib_entry_t *fib_entry,
|
|
fib_source_t source,
|
|
u8* s)
|
|
{
|
|
fib_entry_src_t *esrc;
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_format, (esrc, s));
|
|
|
|
return (s);
|
|
}
|
|
|
|
adj_index_t
|
|
fib_entry_get_adj_for_source (fib_node_index_t fib_entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
if (FIB_NODE_INDEX_INVALID == fib_entry_index)
|
|
return (ADJ_INDEX_INVALID);
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
if (FIB_NODE_INDEX_INVALID != esrc->fes_pl)
|
|
{
|
|
return (fib_path_list_get_adj(
|
|
esrc->fes_pl,
|
|
fib_entry_get_default_chain_type(fib_entry)));
|
|
}
|
|
}
|
|
return (ADJ_INDEX_INVALID);
|
|
}
|
|
|
|
const int
|
|
fib_entry_get_dpo_for_source (fib_node_index_t fib_entry_index,
|
|
fib_source_t source,
|
|
dpo_id_t *dpo)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
if (FIB_NODE_INDEX_INVALID == fib_entry_index)
|
|
return (0);
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
if (FIB_NODE_INDEX_INVALID != esrc->fes_pl)
|
|
{
|
|
fib_path_list_contribute_forwarding(
|
|
esrc->fes_pl,
|
|
fib_entry_get_default_chain_type(fib_entry),
|
|
FIB_PATH_LIST_FWD_FLAG_NONE,
|
|
dpo);
|
|
|
|
return (dpo_id_is_valid(dpo));
|
|
}
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
fib_node_index_t
|
|
fib_entry_get_path_list_for_source (fib_node_index_t fib_entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
if (FIB_NODE_INDEX_INVALID == fib_entry_index)
|
|
return FIB_NODE_INDEX_INVALID;
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (esrc)
|
|
return esrc->fes_pl;
|
|
|
|
return FIB_NODE_INDEX_INVALID;
|
|
}
|
|
|
|
u32
|
|
fib_entry_get_resolving_interface_for_source (fib_node_index_t entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
fib_entry = fib_entry_get(entry_index);
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
if (FIB_NODE_INDEX_INVALID != esrc->fes_pl)
|
|
{
|
|
return (fib_path_list_get_resolving_interface(esrc->fes_pl));
|
|
}
|
|
}
|
|
return (~0);
|
|
}
|
|
|
|
fib_entry_flag_t
|
|
fib_entry_get_flags_for_source (fib_node_index_t entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
fib_entry = fib_entry_get(entry_index);
|
|
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
return (esrc->fes_entry_flags);
|
|
}
|
|
|
|
return (FIB_ENTRY_FLAG_NONE);
|
|
}
|
|
|
|
fib_source_t
|
|
fib_entry_get_source_i (const fib_entry_t *fib_entry)
|
|
{
|
|
/* the vector of sources is deliberately arranged in priority order */
|
|
if (0 == vec_len(fib_entry->fe_srcs))
|
|
return (FIB_SOURCE_INVALID);
|
|
return (vec_elt(fib_entry->fe_srcs, 0).fes_src);
|
|
}
|
|
|
|
fib_entry_flag_t
|
|
fib_entry_get_flags_i (const fib_entry_t *fib_entry)
|
|
{
|
|
/* the vector of sources is deliberately arranged in priority order */
|
|
if (0 == vec_len(fib_entry->fe_srcs))
|
|
return (FIB_ENTRY_FLAG_NONE);
|
|
return (vec_elt(fib_entry->fe_srcs, 0).fes_entry_flags);
|
|
}
|
|
|
|
void
|
|
fib_entry_set_source_data (fib_node_index_t fib_entry_index,
|
|
fib_source_t source,
|
|
const void *data)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE(fib_entry, esrc, fesv_set_data,
|
|
(esrc, fib_entry, data));
|
|
}
|
|
}
|
|
|
|
const void*
|
|
fib_entry_get_source_data (fib_node_index_t fib_entry_index,
|
|
fib_source_t source)
|
|
{
|
|
fib_entry_t *fib_entry;
|
|
fib_entry_src_t *esrc;
|
|
|
|
fib_entry = fib_entry_get(fib_entry_index);
|
|
esrc = fib_entry_src_find(fib_entry, source);
|
|
|
|
if (NULL != esrc)
|
|
{
|
|
FIB_ENTRY_SRC_VFT_INVOKE_AND_RETURN(esrc, fesv_get_data,
|
|
(esrc, fib_entry));
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
void
|
|
fib_entry_src_module_init (void)
|
|
{
|
|
fib_entry_src_rr_register();
|
|
fib_entry_src_interface_register();
|
|
fib_entry_src_interpose_register();
|
|
fib_entry_src_drop_register();
|
|
fib_entry_src_simple_register();
|
|
fib_entry_src_api_register();
|
|
fib_entry_src_adj_register();
|
|
fib_entry_src_mpls_register();
|
|
fib_entry_src_lisp_register();
|
|
}
|