
Type: improvement Signed-off-by: Neale Ranns <nranns@cisco.com> Signed-off-by: Benoît Ganne <bganne@cisco.com> Change-Id: I2f30a4f04fd9a8635ce2d259b5fd5b0c85cee8c3
629 lines
15 KiB
C
629 lines
15 KiB
C
/*
|
|
* gbp.h : Group Based Policy
|
|
*
|
|
* Copyright (c) 2018 Cisco and/or its affiliates.
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <plugins/gbp/gbp.h>
|
|
#include <plugins/gbp/gbp_classify.h>
|
|
#include <plugins/gbp/gbp_policy_dpo.h>
|
|
#include <plugins/gbp/gbp_ext_itf.h>
|
|
#include <vnet/fib/ip4_fib.h>
|
|
#include <vnet/fib/ip6_fib.h>
|
|
#include <vnet/dpo/load_balance.h>
|
|
#include <vnet/l2/l2_input.h>
|
|
#include <vnet/l2/feat_bitmap.h>
|
|
#include <vnet/fib/fib_table.h>
|
|
#include <vnet/vxlan-gbp/vxlan_gbp_packet.h>
|
|
#include <vnet/ethernet/arp_packet.h>
|
|
|
|
/**
|
|
* per-packet trace data
|
|
*/
|
|
typedef struct gbp_classify_trace_t_
|
|
{
|
|
/* per-pkt trace data */
|
|
sclass_t sclass;
|
|
} gbp_classify_trace_t;
|
|
|
|
/*
|
|
* determine the SRC EPG form the input port
|
|
*/
|
|
always_inline uword
|
|
gbp_classify_inline (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame,
|
|
gbp_src_classify_type_t type, dpo_proto_t dproto)
|
|
{
|
|
gbp_src_classify_main_t *gscm = &gbp_src_classify_main;
|
|
u32 n_left_from, *from, *to_next;
|
|
u32 next_index;
|
|
|
|
next_index = 0;
|
|
n_left_from = frame->n_vectors;
|
|
from = vlib_frame_vector_args (frame);
|
|
|
|
while (n_left_from > 0)
|
|
{
|
|
u32 n_left_to_next;
|
|
|
|
vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
|
|
|
|
while (n_left_from > 0 && n_left_to_next > 0)
|
|
{
|
|
u32 next0, bi0, sw_if_index0;
|
|
const gbp_endpoint_t *ge0;
|
|
vlib_buffer_t *b0;
|
|
sclass_t sclass0;
|
|
|
|
bi0 = from[0];
|
|
to_next[0] = bi0;
|
|
from += 1;
|
|
to_next += 1;
|
|
n_left_from -= 1;
|
|
n_left_to_next -= 1;
|
|
|
|
b0 = vlib_get_buffer (vm, bi0);
|
|
|
|
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
|
|
vnet_buffer2 (b0)->gbp.flags = VXLAN_GBP_GPFLAGS_NONE;
|
|
|
|
if (GBP_SRC_CLASSIFY_NULL == type)
|
|
{
|
|
sclass0 = SCLASS_INVALID;
|
|
next0 =
|
|
vnet_l2_feature_next (b0, gscm->l2_input_feat_next[type],
|
|
L2INPUT_FEAT_GBP_NULL_CLASSIFY);
|
|
}
|
|
else
|
|
{
|
|
if (DPO_PROTO_ETHERNET == dproto)
|
|
{
|
|
const ethernet_header_t *h0;
|
|
|
|
h0 = vlib_buffer_get_current (b0);
|
|
next0 =
|
|
vnet_l2_feature_next (b0, gscm->l2_input_feat_next[type],
|
|
L2INPUT_FEAT_GBP_SRC_CLASSIFY);
|
|
ge0 = gbp_endpoint_find_mac (h0->src_address,
|
|
vnet_buffer (b0)->l2.bd_index);
|
|
}
|
|
else if (DPO_PROTO_IP4 == dproto)
|
|
{
|
|
const ip4_header_t *h0;
|
|
|
|
h0 = vlib_buffer_get_current (b0);
|
|
|
|
ge0 = gbp_endpoint_find_ip4
|
|
(&h0->src_address,
|
|
fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP4,
|
|
sw_if_index0));
|
|
|
|
|
|
/*
|
|
* Go straight to looukp, do not pass go, do not collect $200
|
|
*/
|
|
next0 = 0;
|
|
}
|
|
else if (DPO_PROTO_IP6 == dproto)
|
|
{
|
|
const ip6_header_t *h0;
|
|
|
|
h0 = vlib_buffer_get_current (b0);
|
|
|
|
ge0 = gbp_endpoint_find_ip6
|
|
(&h0->src_address,
|
|
fib_table_get_index_for_sw_if_index (FIB_PROTOCOL_IP6,
|
|
sw_if_index0));
|
|
|
|
|
|
/*
|
|
* Go straight to lookup, do not pass go, do not collect $200
|
|
*/
|
|
next0 = 0;
|
|
}
|
|
else
|
|
{
|
|
ge0 = NULL;
|
|
next0 = 0;
|
|
ASSERT (0);
|
|
}
|
|
|
|
if (PREDICT_TRUE (NULL != ge0))
|
|
sclass0 = ge0->ge_fwd.gef_sclass;
|
|
else
|
|
sclass0 = SCLASS_INVALID;
|
|
}
|
|
|
|
vnet_buffer2 (b0)->gbp.sclass = sclass0;
|
|
|
|
if (PREDICT_FALSE ((b0->flags & VLIB_BUFFER_IS_TRACED)))
|
|
{
|
|
gbp_classify_trace_t *t =
|
|
vlib_add_trace (vm, node, b0, sizeof (*t));
|
|
t->sclass = sclass0;
|
|
}
|
|
|
|
vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
|
|
to_next, n_left_to_next,
|
|
bi0, next0);
|
|
}
|
|
|
|
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
|
|
}
|
|
|
|
return frame->n_vectors;
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_src_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_classify_inline (vm, node, frame,
|
|
GBP_SRC_CLASSIFY_PORT, DPO_PROTO_ETHERNET));
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_null_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_classify_inline (vm, node, frame,
|
|
GBP_SRC_CLASSIFY_NULL, DPO_PROTO_ETHERNET));
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_ip4_src_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_classify_inline (vm, node, frame,
|
|
GBP_SRC_CLASSIFY_PORT, DPO_PROTO_IP4));
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_ip6_src_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_classify_inline (vm, node, frame,
|
|
GBP_SRC_CLASSIFY_PORT, DPO_PROTO_IP6));
|
|
}
|
|
|
|
|
|
/* packet trace format function */
|
|
static u8 *
|
|
format_gbp_classify_trace (u8 * s, va_list * args)
|
|
{
|
|
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
|
|
CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
|
|
gbp_classify_trace_t *t = va_arg (*args, gbp_classify_trace_t *);
|
|
|
|
s = format (s, "sclass:%d", t->sclass);
|
|
|
|
return s;
|
|
}
|
|
|
|
/* *INDENT-OFF* */
|
|
VLIB_REGISTER_NODE (gbp_null_classify_node) = {
|
|
.name = "gbp-null-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 0,
|
|
};
|
|
|
|
VLIB_REGISTER_NODE (gbp_src_classify_node) = {
|
|
.name = "gbp-src-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 0,
|
|
};
|
|
|
|
VLIB_REGISTER_NODE (gbp_ip4_src_classify_node) = {
|
|
.name = "ip4-gbp-src-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 1,
|
|
.next_nodes = {
|
|
[0] = "ip4-lookup"
|
|
},
|
|
};
|
|
|
|
VLIB_REGISTER_NODE (gbp_ip6_src_classify_node) = {
|
|
.name = "ip6-gbp-src-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 1,
|
|
.next_nodes = {
|
|
[0] = "ip6-lookup"
|
|
},
|
|
};
|
|
|
|
VNET_FEATURE_INIT (gbp_ip4_src_classify_feat_node, static) =
|
|
{
|
|
.arc_name = "ip4-unicast",
|
|
.node_name = "ip4-gbp-src-classify",
|
|
.runs_before = VNET_FEATURES ("nat44-out2in"),
|
|
};
|
|
VNET_FEATURE_INIT (gbp_ip6_src_classify_feat_node, static) =
|
|
{
|
|
.arc_name = "ip6-unicast",
|
|
.node_name = "ip6-gbp-src-classify",
|
|
.runs_before = VNET_FEATURES ("nat66-out2in"),
|
|
};
|
|
|
|
/* *INDENT-ON* */
|
|
|
|
typedef enum gbp_lpm_classify_next_t_
|
|
{
|
|
GPB_LPM_CLASSIFY_DROP,
|
|
} gbp_lpm_classify_next_t;
|
|
|
|
/**
|
|
* per-packet trace data
|
|
*/
|
|
typedef struct gbp_lpm_classify_trace_t_
|
|
{
|
|
sclass_t sclass;
|
|
index_t lbi;
|
|
ip46_address_t src;
|
|
} gbp_lpm_classify_trace_t;
|
|
|
|
/* packet trace format function */
|
|
static u8 *
|
|
format_gbp_lpm_classify_trace (u8 * s, va_list * args)
|
|
{
|
|
CLIB_UNUSED (vlib_main_t * vm) = va_arg (*args, vlib_main_t *);
|
|
CLIB_UNUSED (vlib_node_t * node) = va_arg (*args, vlib_node_t *);
|
|
gbp_lpm_classify_trace_t *t = va_arg (*args, gbp_lpm_classify_trace_t *);
|
|
|
|
s = format (s, "sclass:%d lb:%d src:%U",
|
|
t->sclass, t->lbi, format_ip46_address, &t->src, IP46_TYPE_ANY);
|
|
|
|
return s;
|
|
}
|
|
|
|
enum gbp_lpm_type
|
|
{
|
|
GBP_LPM_RECIRC,
|
|
GBP_LPM_EPG,
|
|
GBP_LPM_ANON
|
|
};
|
|
|
|
/*
|
|
* Determine the SRC EPG from a LPM
|
|
*/
|
|
always_inline uword
|
|
gbp_lpm_classify_inline (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame,
|
|
const dpo_proto_t dproto,
|
|
const enum gbp_lpm_type type)
|
|
{
|
|
gbp_src_classify_main_t *gscm = &gbp_src_classify_main;
|
|
u32 n_left_from, *from, *to_next;
|
|
u32 next_index;
|
|
|
|
next_index = 0;
|
|
n_left_from = frame->n_vectors;
|
|
from = vlib_frame_vector_args (frame);
|
|
|
|
while (n_left_from > 0)
|
|
{
|
|
u32 n_left_to_next;
|
|
|
|
vlib_get_next_frame (vm, node, next_index, to_next, n_left_to_next);
|
|
|
|
while (n_left_from > 0 && n_left_to_next > 0)
|
|
{
|
|
u32 bi0, sw_if_index0, fib_index0, lbi0;
|
|
const gbp_endpoint_t *ge0, *ge_lpm0;
|
|
gbp_lpm_classify_next_t next0;
|
|
const ethernet_header_t *eh0;
|
|
const gbp_policy_dpo_t *gpd0;
|
|
const ip4_address_t *ip4_0;
|
|
const ip6_address_t *ip6_0;
|
|
const gbp_recirc_t *gr0;
|
|
vlib_buffer_t *b0;
|
|
sclass_t sclass0;
|
|
|
|
bi0 = from[0];
|
|
to_next[0] = bi0;
|
|
from += 1;
|
|
to_next += 1;
|
|
n_left_from -= 1;
|
|
n_left_to_next -= 1;
|
|
ip4_0 = NULL;
|
|
ip6_0 = NULL;
|
|
next0 = GPB_LPM_CLASSIFY_DROP;
|
|
|
|
lbi0 = ~0;
|
|
eh0 = NULL;
|
|
b0 = vlib_get_buffer (vm, bi0);
|
|
|
|
sw_if_index0 = vnet_buffer (b0)->sw_if_index[VLIB_RX];
|
|
vnet_buffer2 (b0)->gbp.flags = VXLAN_GBP_GPFLAGS_NONE;
|
|
|
|
if (DPO_PROTO_IP4 == dproto)
|
|
ip4_0 =
|
|
&((ip4_header_t *) vlib_buffer_get_current (b0))->src_address;
|
|
else if (DPO_PROTO_IP6 == dproto)
|
|
ip6_0 =
|
|
&((ip6_header_t *) vlib_buffer_get_current (b0))->src_address;
|
|
else if (DPO_PROTO_ETHERNET == dproto)
|
|
{
|
|
eh0 = vlib_buffer_get_current (b0);
|
|
gbp_classify_get_ip_address (eh0, &ip4_0, &ip6_0,
|
|
GBP_CLASSIFY_GET_IP_SRC);
|
|
}
|
|
|
|
if (GBP_LPM_RECIRC == type)
|
|
{
|
|
gr0 = gbp_recirc_get (sw_if_index0);
|
|
fib_index0 = gr0->gr_fib_index[dproto];
|
|
ge0 = NULL;
|
|
|
|
vnet_feature_next (&next0, b0);
|
|
}
|
|
else
|
|
{
|
|
if (NULL == eh0)
|
|
{
|
|
/* packet should be l2 */
|
|
sclass0 = SCLASS_INVALID;
|
|
goto trace;
|
|
}
|
|
|
|
if (GBP_LPM_ANON == type)
|
|
{
|
|
/*
|
|
* anonymous LPM classification: only honour LPM as no EP
|
|
* were programmed
|
|
*/
|
|
gbp_ext_itf_t *gei = gbp_ext_itf_get (sw_if_index0);
|
|
if (ip4_0)
|
|
fib_index0 = gei->gx_fib_index[DPO_PROTO_IP4];
|
|
else if (ip6_0)
|
|
fib_index0 = gei->gx_fib_index[DPO_PROTO_IP6];
|
|
else
|
|
{
|
|
/* not IP so no LPM classify possible */
|
|
sclass0 = SCLASS_INVALID;
|
|
next0 = GPB_LPM_CLASSIFY_DROP;
|
|
goto trace;
|
|
}
|
|
next0 = vnet_l2_feature_next
|
|
(b0, gscm->l2_input_feat_next[GBP_SRC_CLASSIFY_LPM_ANON],
|
|
L2INPUT_FEAT_GBP_LPM_ANON_CLASSIFY);
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
* not an anonymous LPM classification: check it comes from
|
|
* an EP, and use EP RD info
|
|
*/
|
|
ge0 = gbp_endpoint_find_mac (eh0->src_address,
|
|
vnet_buffer (b0)->l2.bd_index);
|
|
|
|
if (NULL == ge0)
|
|
{
|
|
/* packet must have come from an EP's mac */
|
|
sclass0 = SCLASS_INVALID;
|
|
goto trace;
|
|
}
|
|
|
|
fib_index0 = ge0->ge_fwd.gef_fib_index;
|
|
|
|
if (~0 == fib_index0)
|
|
{
|
|
sclass0 = SCLASS_INVALID;
|
|
goto trace;
|
|
}
|
|
|
|
if (ip4_0)
|
|
{
|
|
ge_lpm0 = gbp_endpoint_find_ip4 (ip4_0, fib_index0);
|
|
}
|
|
else if (ip6_0)
|
|
{
|
|
ge_lpm0 = gbp_endpoint_find_ip6 (ip6_0, fib_index0);
|
|
}
|
|
else
|
|
{
|
|
ge_lpm0 = NULL;
|
|
}
|
|
|
|
next0 = vnet_l2_feature_next
|
|
(b0, gscm->l2_input_feat_next[GBP_SRC_CLASSIFY_LPM],
|
|
L2INPUT_FEAT_GBP_LPM_CLASSIFY);
|
|
|
|
/*
|
|
* if we found the EP by IP lookup, it must be from the EP
|
|
* not a network behind it
|
|
*/
|
|
if (NULL != ge_lpm0)
|
|
{
|
|
if (PREDICT_FALSE (ge0 != ge_lpm0))
|
|
{
|
|
/* an EP spoofing another EP */
|
|
sclass0 = SCLASS_INVALID;
|
|
next0 = GPB_LPM_CLASSIFY_DROP;
|
|
}
|
|
else
|
|
{
|
|
sclass0 = ge0->ge_fwd.gef_sclass;
|
|
}
|
|
goto trace;
|
|
}
|
|
}
|
|
}
|
|
|
|
gpd0 = gbp_classify_get_gpd (ip4_0, ip6_0, fib_index0);
|
|
if (0 == gpd0)
|
|
{
|
|
/* could not classify => drop */
|
|
sclass0 = SCLASS_INVALID;
|
|
next0 = GPB_LPM_CLASSIFY_DROP;
|
|
goto trace;
|
|
}
|
|
|
|
sclass0 = gpd0->gpd_sclass;
|
|
|
|
/* all packets from an external network should not be learned by the
|
|
* reciever. so set the Do-not-learn bit here */
|
|
vnet_buffer2 (b0)->gbp.flags = VXLAN_GBP_GPFLAGS_D;
|
|
|
|
trace:
|
|
vnet_buffer2 (b0)->gbp.sclass = sclass0;
|
|
|
|
if (PREDICT_FALSE ((b0->flags & VLIB_BUFFER_IS_TRACED)))
|
|
{
|
|
gbp_lpm_classify_trace_t *t =
|
|
vlib_add_trace (vm, node, b0, sizeof (*t));
|
|
t->sclass = sclass0;
|
|
t->lbi = lbi0;
|
|
if (ip4_0)
|
|
t->src.ip4 = *ip4_0;
|
|
if (ip6_0)
|
|
t->src.ip6 = *ip6_0;
|
|
}
|
|
|
|
vlib_validate_buffer_enqueue_x1 (vm, node, next_index,
|
|
to_next, n_left_to_next,
|
|
bi0, next0);
|
|
}
|
|
|
|
vlib_put_next_frame (vm, node, next_index, n_left_to_next);
|
|
}
|
|
|
|
return frame->n_vectors;
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_ip4_lpm_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_lpm_classify_inline
|
|
(vm, node, frame, DPO_PROTO_IP4, GBP_LPM_RECIRC));
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_ip6_lpm_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_lpm_classify_inline
|
|
(vm, node, frame, DPO_PROTO_IP6, GBP_LPM_RECIRC));
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_l2_lpm_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_lpm_classify_inline
|
|
(vm, node, frame, DPO_PROTO_ETHERNET, GBP_LPM_EPG));
|
|
}
|
|
|
|
VLIB_NODE_FN (gbp_l2_lpm_anon_classify_node) (vlib_main_t * vm,
|
|
vlib_node_runtime_t * node,
|
|
vlib_frame_t * frame)
|
|
{
|
|
return (gbp_lpm_classify_inline
|
|
(vm, node, frame, DPO_PROTO_ETHERNET, GBP_LPM_ANON));
|
|
}
|
|
|
|
/* *INDENT-OFF* */
|
|
VLIB_REGISTER_NODE (gbp_ip4_lpm_classify_node) = {
|
|
.name = "ip4-gbp-lpm-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_lpm_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 1,
|
|
.next_nodes = {
|
|
[GPB_LPM_CLASSIFY_DROP] = "ip4-drop"
|
|
},
|
|
};
|
|
|
|
VLIB_REGISTER_NODE (gbp_ip6_lpm_classify_node) = {
|
|
.name = "ip6-gbp-lpm-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_lpm_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 1,
|
|
.next_nodes = {
|
|
[GPB_LPM_CLASSIFY_DROP] = "ip6-drop"
|
|
},
|
|
};
|
|
|
|
VLIB_REGISTER_NODE (gbp_l2_lpm_classify_node) = {
|
|
.name = "l2-gbp-lpm-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_lpm_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 1,
|
|
.next_nodes = {
|
|
[GPB_LPM_CLASSIFY_DROP] = "error-drop"
|
|
},
|
|
};
|
|
|
|
VLIB_REGISTER_NODE (gbp_l2_lpm_anon_classify_node) = {
|
|
.name = "l2-gbp-lpm-anon-classify",
|
|
.vector_size = sizeof (u32),
|
|
.format_trace = format_gbp_lpm_classify_trace,
|
|
.type = VLIB_NODE_TYPE_INTERNAL,
|
|
|
|
.n_errors = 0,
|
|
.n_next_nodes = 1,
|
|
.next_nodes = {
|
|
[GPB_LPM_CLASSIFY_DROP] = "error-drop"
|
|
},
|
|
};
|
|
|
|
VNET_FEATURE_INIT (gbp_ip4_lpm_classify_feat_node, static) =
|
|
{
|
|
.arc_name = "ip4-unicast",
|
|
.node_name = "ip4-gbp-lpm-classify",
|
|
.runs_before = VNET_FEATURES ("nat44-out2in"),
|
|
};
|
|
VNET_FEATURE_INIT (gbp_ip6_lpm_classify_feat_node, static) =
|
|
{
|
|
.arc_name = "ip6-unicast",
|
|
.node_name = "ip6-gbp-lpm-classify",
|
|
.runs_before = VNET_FEATURES ("nat66-out2in"),
|
|
};
|
|
|
|
/* *INDENT-ON* */
|
|
|
|
/*
|
|
* fd.io coding-style-patch-verification: ON
|
|
*
|
|
* Local Variables:
|
|
* eval: (c-set-style "gnu")
|
|
* End:
|
|
*/
|