34fa0ce8f7
Check and skip VPP_EXCLUDED_PLUGINS tests for most of plugins. Type: improvement Signed-off-by: Dmitry Valter <d-valter@yandex-team.com> Change-Id: I23fd3666729251c639aa8da72a676058e3f5bb4e
2979 lines
91 KiB
Python
2979 lines
91 KiB
Python
#!/usr/bin/env python3
|
|
|
|
import unittest
|
|
|
|
from framework import VppTestCase
|
|
from asfframework import VppTestRunner, tag_fixme_vpp_workers
|
|
from vpp_ip import INVALID_INDEX
|
|
from vpp_ip_route import (
|
|
VppIpRoute,
|
|
VppRoutePath,
|
|
VppMplsRoute,
|
|
VppMplsIpBind,
|
|
VppIpMRoute,
|
|
VppMRoutePath,
|
|
VppIpTable,
|
|
VppMplsTable,
|
|
VppMplsLabel,
|
|
MplsLspMode,
|
|
find_mpls_route,
|
|
FibPathProto,
|
|
FibPathType,
|
|
FibPathFlags,
|
|
VppMplsLabel,
|
|
MplsLspMode,
|
|
)
|
|
from vpp_mpls_tunnel_interface import VppMPLSTunnelInterface
|
|
from vpp_papi import VppEnum
|
|
from config import config
|
|
|
|
import scapy.compat
|
|
from scapy.packet import Raw
|
|
from scapy.layers.l2 import Ether, ARP
|
|
from scapy.layers.inet import IP, UDP, ICMP, icmptypes, icmpcodes
|
|
from scapy.layers.inet6 import (
|
|
IPv6,
|
|
ICMPv6TimeExceeded,
|
|
ICMPv6EchoRequest,
|
|
ICMPv6PacketTooBig,
|
|
)
|
|
from scapy.contrib.mpls import MPLS
|
|
|
|
NUM_PKTS = 67
|
|
|
|
# scapy removed these attributes.
|
|
# we asked that they be restored: https://github.com/secdev/scapy/pull/1878
|
|
# semantic names have more meaning than numbers. so here they are.
|
|
ARP.who_has = 1
|
|
ARP.is_at = 2
|
|
|
|
|
|
def verify_filter(capture, sent):
|
|
if not len(capture) == len(sent):
|
|
# filter out any IPv6 RAs from the capture
|
|
for p in capture:
|
|
if p.haslayer(IPv6):
|
|
capture.remove(p)
|
|
return capture
|
|
|
|
|
|
def verify_mpls_stack(tst, rx, mpls_labels):
|
|
# the rx'd packet has the MPLS label popped
|
|
eth = rx[Ether]
|
|
tst.assertEqual(eth.type, 0x8847)
|
|
|
|
rx_mpls = rx[MPLS]
|
|
|
|
for ii in range(len(mpls_labels)):
|
|
tst.assertEqual(rx_mpls.label, mpls_labels[ii].value)
|
|
tst.assertEqual(rx_mpls.cos, mpls_labels[ii].exp)
|
|
tst.assertEqual(rx_mpls.ttl, mpls_labels[ii].ttl)
|
|
|
|
if ii == len(mpls_labels) - 1:
|
|
tst.assertEqual(rx_mpls.s, 1)
|
|
else:
|
|
# not end of stack
|
|
tst.assertEqual(rx_mpls.s, 0)
|
|
# pop the label to expose the next
|
|
rx_mpls = rx_mpls[MPLS].payload
|
|
|
|
|
|
@tag_fixme_vpp_workers
|
|
class TestMPLS(VppTestCase):
|
|
"""MPLS Test Case"""
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
super(TestMPLS, cls).setUpClass()
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestMPLS, cls).tearDownClass()
|
|
|
|
def setUp(self):
|
|
super(TestMPLS, self).setUp()
|
|
|
|
# create 2 pg interfaces
|
|
self.create_pg_interfaces(range(4))
|
|
|
|
# setup both interfaces
|
|
# assign them different tables.
|
|
table_id = 0
|
|
self.tables = []
|
|
|
|
tbl = VppMplsTable(self, 0)
|
|
tbl.add_vpp_config()
|
|
self.tables.append(tbl)
|
|
|
|
for i in self.pg_interfaces:
|
|
i.admin_up()
|
|
|
|
if table_id != 0:
|
|
tbl = VppIpTable(self, table_id)
|
|
tbl.add_vpp_config()
|
|
self.tables.append(tbl)
|
|
tbl = VppIpTable(self, table_id, is_ip6=1)
|
|
tbl.add_vpp_config()
|
|
self.tables.append(tbl)
|
|
|
|
i.set_table_ip4(table_id)
|
|
i.set_table_ip6(table_id)
|
|
i.config_ip4()
|
|
i.resolve_arp()
|
|
i.config_ip6()
|
|
i.resolve_ndp()
|
|
i.enable_mpls()
|
|
table_id += 1
|
|
|
|
def tearDown(self):
|
|
for i in self.pg_interfaces:
|
|
i.unconfig_ip4()
|
|
i.unconfig_ip6()
|
|
i.set_table_ip4(0)
|
|
i.set_table_ip6(0)
|
|
i.disable_mpls()
|
|
i.admin_down()
|
|
super(TestMPLS, self).tearDown()
|
|
|
|
# the default of 64 matches the IP packet TTL default
|
|
def create_stream_labelled_ip4(
|
|
self,
|
|
src_if,
|
|
mpls_labels,
|
|
ping=0,
|
|
ip_itf=None,
|
|
dst_ip=None,
|
|
chksum=None,
|
|
ip_ttl=64,
|
|
n=257,
|
|
):
|
|
self.reset_packet_infos()
|
|
pkts = []
|
|
for i in range(0, n):
|
|
info = self.create_packet_info(src_if, src_if)
|
|
payload = self.info_to_payload(info)
|
|
p = Ether(dst=src_if.local_mac, src=src_if.remote_mac)
|
|
|
|
for ii in range(len(mpls_labels)):
|
|
p = p / MPLS(
|
|
label=mpls_labels[ii].value,
|
|
ttl=mpls_labels[ii].ttl,
|
|
cos=mpls_labels[ii].exp,
|
|
)
|
|
if not ping:
|
|
if not dst_ip:
|
|
p = (
|
|
p
|
|
/ IP(src=src_if.local_ip4, dst=src_if.remote_ip4, ttl=ip_ttl)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(payload)
|
|
)
|
|
else:
|
|
p = (
|
|
p
|
|
/ IP(src=src_if.local_ip4, dst=dst_ip, ttl=ip_ttl)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(payload)
|
|
)
|
|
else:
|
|
p = (
|
|
p
|
|
/ IP(src=ip_itf.remote_ip4, dst=ip_itf.local_ip4, ttl=ip_ttl)
|
|
/ ICMP()
|
|
)
|
|
|
|
if chksum:
|
|
p[IP].chksum = chksum
|
|
info.data = p.copy()
|
|
pkts.append(p)
|
|
return pkts
|
|
|
|
def create_stream_ip4(
|
|
self, src_if, dst_ip, ip_ttl=64, ip_dscp=0, payload_size=None, n=257
|
|
):
|
|
self.reset_packet_infos()
|
|
pkts = []
|
|
for i in range(0, n):
|
|
dst = dst_ip[i % len(dst_ip)] if isinstance(dst_ip, list) else dst_ip
|
|
info = self.create_packet_info(src_if, src_if)
|
|
payload = self.info_to_payload(info)
|
|
p = (
|
|
Ether(dst=src_if.local_mac, src=src_if.remote_mac)
|
|
/ IP(src=src_if.remote_ip4, dst=dst, ttl=ip_ttl, tos=ip_dscp)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(payload)
|
|
)
|
|
info.data = p.copy()
|
|
if payload_size:
|
|
self.extend_packet(p, payload_size)
|
|
pkts.append(p)
|
|
return pkts
|
|
|
|
def create_stream_ip6(self, src_if, dst_ip, ip_ttl=64, ip_dscp=0, n=257):
|
|
self.reset_packet_infos()
|
|
pkts = []
|
|
for i in range(0, n):
|
|
dst = dst_ip[i % len(dst_ip)] if isinstance(dst_ip, list) else dst_ip
|
|
info = self.create_packet_info(src_if, src_if)
|
|
payload = self.info_to_payload(info)
|
|
p = (
|
|
Ether(dst=src_if.local_mac, src=src_if.remote_mac)
|
|
/ IPv6(src=src_if.remote_ip6, dst=dst, hlim=ip_ttl, tc=ip_dscp)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(payload)
|
|
)
|
|
info.data = p.copy()
|
|
pkts.append(p)
|
|
return pkts
|
|
|
|
def create_stream_labelled_ip6(
|
|
self, src_if, mpls_labels, hlim=64, dst_ip=None, ping=0, ip_itf=None
|
|
):
|
|
if dst_ip is None:
|
|
dst_ip = src_if.remote_ip6
|
|
self.reset_packet_infos()
|
|
pkts = []
|
|
for i in range(0, 257):
|
|
info = self.create_packet_info(src_if, src_if)
|
|
payload = self.info_to_payload(info)
|
|
p = Ether(dst=src_if.local_mac, src=src_if.remote_mac)
|
|
for l in mpls_labels:
|
|
p = p / MPLS(label=l.value, ttl=l.ttl, cos=l.exp)
|
|
|
|
if ping:
|
|
p = p / (
|
|
IPv6(src=ip_itf.remote_ip6, dst=ip_itf.local_ip6)
|
|
/ ICMPv6EchoRequest()
|
|
)
|
|
else:
|
|
p = p / (
|
|
IPv6(src=src_if.remote_ip6, dst=dst_ip, hlim=hlim)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(payload)
|
|
)
|
|
info.data = p.copy()
|
|
pkts.append(p)
|
|
return pkts
|
|
|
|
def verify_capture_ip4(
|
|
self, src_if, capture, sent, ping_resp=0, ip_ttl=None, ip_dscp=0
|
|
):
|
|
try:
|
|
capture = verify_filter(capture, sent)
|
|
|
|
self.assertEqual(len(capture), len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[i]
|
|
rx = capture[i]
|
|
|
|
# the rx'd packet has the MPLS label popped
|
|
eth = rx[Ether]
|
|
self.assertEqual(eth.type, 0x800)
|
|
|
|
tx_ip = tx[IP]
|
|
rx_ip = rx[IP]
|
|
|
|
if not ping_resp:
|
|
self.assertEqual(rx_ip.src, tx_ip.src)
|
|
self.assertEqual(rx_ip.dst, tx_ip.dst)
|
|
self.assertEqual(rx_ip.tos, ip_dscp)
|
|
if not ip_ttl:
|
|
# IP processing post pop has decremented the TTL
|
|
self.assertEqual(rx_ip.ttl + 1, tx_ip.ttl)
|
|
else:
|
|
self.assertEqual(rx_ip.ttl, ip_ttl)
|
|
else:
|
|
self.assertEqual(rx_ip.src, tx_ip.dst)
|
|
self.assertEqual(rx_ip.dst, tx_ip.src)
|
|
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_labelled_ip4(
|
|
self, src_if, capture, sent, mpls_labels, ip_ttl=None
|
|
):
|
|
try:
|
|
capture = verify_filter(capture, sent)
|
|
|
|
self.assertEqual(len(capture), len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[i]
|
|
rx = capture[i]
|
|
tx_ip = tx[IP]
|
|
rx_ip = rx[IP]
|
|
|
|
verify_mpls_stack(self, rx, mpls_labels)
|
|
|
|
self.assertEqual(rx_ip.src, tx_ip.src)
|
|
self.assertEqual(rx_ip.dst, tx_ip.dst)
|
|
if not ip_ttl:
|
|
# IP processing post pop has decremented the TTL
|
|
self.assertEqual(rx_ip.ttl + 1, tx_ip.ttl)
|
|
else:
|
|
self.assertEqual(rx_ip.ttl, ip_ttl)
|
|
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_labelled_ip6(
|
|
self, src_if, capture, sent, mpls_labels, ip_ttl=None
|
|
):
|
|
try:
|
|
capture = verify_filter(capture, sent)
|
|
|
|
self.assertEqual(len(capture), len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[i]
|
|
rx = capture[i]
|
|
tx_ip = tx[IPv6]
|
|
rx_ip = rx[IPv6]
|
|
|
|
verify_mpls_stack(self, rx, mpls_labels)
|
|
|
|
self.assertEqual(rx_ip.src, tx_ip.src)
|
|
self.assertEqual(rx_ip.dst, tx_ip.dst)
|
|
if not ip_ttl:
|
|
# IP processing post pop has decremented the TTL
|
|
self.assertEqual(rx_ip.hlim + 1, tx_ip.hlim)
|
|
else:
|
|
self.assertEqual(rx_ip.hlim, ip_ttl)
|
|
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_tunneled_ip4(self, src_if, capture, sent, mpls_labels):
|
|
try:
|
|
capture = verify_filter(capture, sent)
|
|
|
|
self.assertEqual(len(capture), len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[i]
|
|
rx = capture[i]
|
|
tx_ip = tx[IP]
|
|
rx_ip = rx[IP]
|
|
|
|
verify_mpls_stack(self, rx, mpls_labels)
|
|
|
|
self.assertEqual(rx_ip.src, tx_ip.src)
|
|
self.assertEqual(rx_ip.dst, tx_ip.dst)
|
|
# IP processing post pop has decremented the TTL
|
|
self.assertEqual(rx_ip.ttl + 1, tx_ip.ttl)
|
|
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_labelled(self, src_if, capture, sent, mpls_labels):
|
|
try:
|
|
capture = verify_filter(capture, sent)
|
|
|
|
self.assertEqual(len(capture), len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
rx = capture[i]
|
|
verify_mpls_stack(self, rx, mpls_labels)
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_ip6(
|
|
self, src_if, capture, sent, ip_hlim=None, ip_dscp=0, ping_resp=0
|
|
):
|
|
try:
|
|
self.assertEqual(len(capture), len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[i]
|
|
rx = capture[i]
|
|
|
|
# the rx'd packet has the MPLS label popped
|
|
eth = rx[Ether]
|
|
self.assertEqual(eth.type, 0x86DD)
|
|
|
|
tx_ip = tx[IPv6]
|
|
rx_ip = rx[IPv6]
|
|
|
|
if not ping_resp:
|
|
self.assertEqual(rx_ip.src, tx_ip.src)
|
|
self.assertEqual(rx_ip.dst, tx_ip.dst)
|
|
self.assertEqual(rx_ip.tc, ip_dscp)
|
|
# IP processing post pop has decremented the TTL
|
|
if not ip_hlim:
|
|
self.assertEqual(rx_ip.hlim + 1, tx_ip.hlim)
|
|
else:
|
|
self.assertEqual(rx_ip.hlim, ip_hlim)
|
|
else:
|
|
self.assertEqual(rx_ip.src, tx_ip.dst)
|
|
self.assertEqual(rx_ip.dst, tx_ip.src)
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_ip6_icmp(self, src_if, capture, sent):
|
|
try:
|
|
# rate limited ICMP
|
|
self.assertTrue(len(capture) <= len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[i]
|
|
rx = capture[i]
|
|
|
|
# the rx'd packet has the MPLS label popped
|
|
eth = rx[Ether]
|
|
self.assertEqual(eth.type, 0x86DD)
|
|
|
|
tx_ip = tx[IPv6]
|
|
rx_ip = rx[IPv6]
|
|
|
|
self.assertEqual(rx_ip.dst, tx_ip.src)
|
|
# ICMP sourced from the interface's address
|
|
self.assertEqual(rx_ip.src, src_if.local_ip6)
|
|
# hop-limit reset to 255 for IMCP packet
|
|
self.assertEqual(rx_ip.hlim, 255)
|
|
|
|
icmp = rx[ICMPv6TimeExceeded]
|
|
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_fragmented_labelled_ip4(
|
|
self, src_if, capture, sent, mpls_labels, ip_ttl=None
|
|
):
|
|
try:
|
|
capture = verify_filter(capture, sent)
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[0]
|
|
rx = capture[i]
|
|
tx_ip = tx[IP]
|
|
rx_ip = rx[IP]
|
|
|
|
verify_mpls_stack(self, rx, mpls_labels)
|
|
|
|
self.assertEqual(rx_ip.src, tx_ip.src)
|
|
self.assertEqual(rx_ip.dst, tx_ip.dst)
|
|
if not ip_ttl:
|
|
# IP processing post pop has decremented the TTL
|
|
self.assertEqual(rx_ip.ttl + 1, tx_ip.ttl)
|
|
else:
|
|
self.assertEqual(rx_ip.ttl, ip_ttl)
|
|
|
|
except:
|
|
raise
|
|
|
|
def verify_capture_fragmented_labelled_ip6(
|
|
self, src_if, capture, sent, mpls_labels, ip_ttl=None
|
|
):
|
|
try:
|
|
capture = verify_filter(capture, sent)
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[0]
|
|
rx = capture[i]
|
|
tx_ip = tx[IPv6]
|
|
rx.show()
|
|
rx_ip = IPv6(rx[MPLS].payload)
|
|
rx_ip.show()
|
|
|
|
verify_mpls_stack(self, rx, mpls_labels)
|
|
|
|
self.assertEqual(rx_ip.src, tx_ip.src)
|
|
self.assertEqual(rx_ip.dst, tx_ip.dst)
|
|
if not ip_ttl:
|
|
# IP processing post pop has decremented the hop-limit
|
|
self.assertEqual(rx_ip.hlim + 1, tx_ip.hlim)
|
|
else:
|
|
self.assertEqual(rx_ip.hlim, ip_ttl)
|
|
except:
|
|
raise
|
|
|
|
def test_swap(self):
|
|
"""MPLS label swap tests"""
|
|
|
|
#
|
|
# A simple MPLS xconnect - eos label in label out
|
|
#
|
|
route_32_eos = VppMplsRoute(
|
|
self,
|
|
32,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(33)]
|
|
)
|
|
],
|
|
)
|
|
route_32_eos.add_vpp_config()
|
|
|
|
self.assertTrue(
|
|
find_mpls_route(
|
|
self,
|
|
0,
|
|
32,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(33)],
|
|
)
|
|
],
|
|
)
|
|
)
|
|
|
|
#
|
|
# a stream that matches the route for 10.0.0.1
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(32, ttl=32, exp=1)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled(
|
|
self.pg0, rx, tx, [VppMplsLabel(33, ttl=31, exp=1)]
|
|
)
|
|
|
|
self.assertEqual(route_32_eos.get_stats_to()["packets"], 257)
|
|
|
|
#
|
|
# A simple MPLS xconnect - non-eos label in label out
|
|
#
|
|
route_32_neos = VppMplsRoute(
|
|
self,
|
|
32,
|
|
0,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(33)]
|
|
)
|
|
],
|
|
)
|
|
route_32_neos.add_vpp_config()
|
|
|
|
#
|
|
# a stream that matches the route for 10.0.0.1
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(32, ttl=21, exp=7), VppMplsLabel(99)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled(
|
|
self.pg0, rx, tx, [VppMplsLabel(33, ttl=20, exp=7), VppMplsLabel(99)]
|
|
)
|
|
self.assertEqual(route_32_neos.get_stats_to()["packets"], 257)
|
|
|
|
#
|
|
# A simple MPLS xconnect - non-eos label in label out, uniform mode
|
|
#
|
|
route_42_neos = VppMplsRoute(
|
|
self,
|
|
42,
|
|
0,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(43, MplsLspMode.UNIFORM)],
|
|
)
|
|
],
|
|
)
|
|
route_42_neos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(42, ttl=21, exp=7), VppMplsLabel(99)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled(
|
|
self.pg0, rx, tx, [VppMplsLabel(43, ttl=20, exp=7), VppMplsLabel(99)]
|
|
)
|
|
|
|
#
|
|
# An MPLS xconnect - EOS label in IP out
|
|
#
|
|
route_33_eos = VppMplsRoute(
|
|
self,
|
|
33,
|
|
1,
|
|
[VppRoutePath(self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[])],
|
|
)
|
|
route_33_eos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip4(self.pg0, [VppMplsLabel(33)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip4(self.pg0, rx, tx)
|
|
|
|
#
|
|
# disposed packets have an invalid IPv4 checksum
|
|
#
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(33)], dst_ip=self.pg0.remote_ip4, n=65, chksum=1
|
|
)
|
|
self.send_and_assert_no_replies(self.pg0, tx, "Invalid Checksum")
|
|
|
|
#
|
|
# An MPLS xconnect - EOS label in IP out, uniform mode
|
|
#
|
|
route_3333_eos = VppMplsRoute(
|
|
self,
|
|
3333,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(3, MplsLspMode.UNIFORM)],
|
|
)
|
|
],
|
|
)
|
|
route_3333_eos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(3333, ttl=55, exp=3)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip4(self.pg0, rx, tx, ip_ttl=54, ip_dscp=0x60)
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(3333, ttl=66, exp=4)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip4(self.pg0, rx, tx, ip_ttl=65, ip_dscp=0x80)
|
|
|
|
#
|
|
# An MPLS xconnect - EOS label in IPv6 out
|
|
#
|
|
route_333_eos = VppMplsRoute(
|
|
self,
|
|
333,
|
|
1,
|
|
[VppRoutePath(self.pg0.remote_ip6, self.pg0.sw_if_index, labels=[])],
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_IP6,
|
|
)
|
|
route_333_eos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip6(self.pg0, [VppMplsLabel(333)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip6(self.pg0, rx, tx)
|
|
|
|
#
|
|
# disposed packets have an TTL expired
|
|
#
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(333, ttl=64)], dst_ip=self.pg1.remote_ip6, hlim=1
|
|
)
|
|
rx = self.send_and_expect_some(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip6_icmp(self.pg0, rx, tx)
|
|
|
|
#
|
|
# An MPLS xconnect - EOS label in IPv6 out w imp-null
|
|
#
|
|
route_334_eos = VppMplsRoute(
|
|
self,
|
|
334,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip6, self.pg0.sw_if_index, labels=[VppMplsLabel(3)]
|
|
)
|
|
],
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_IP6,
|
|
)
|
|
route_334_eos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip6(self.pg0, [VppMplsLabel(334, ttl=64)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip6(self.pg0, rx, tx)
|
|
|
|
#
|
|
# An MPLS xconnect - EOS label in IPv6 out w imp-null in uniform mode
|
|
#
|
|
route_335_eos = VppMplsRoute(
|
|
self,
|
|
335,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip6,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(3, MplsLspMode.UNIFORM)],
|
|
)
|
|
],
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_IP6,
|
|
)
|
|
route_335_eos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(335, ttl=27, exp=4)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip6(self.pg0, rx, tx, ip_hlim=26, ip_dscp=0x80)
|
|
|
|
#
|
|
# disposed packets have an TTL expired
|
|
#
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(334)], dst_ip=self.pg1.remote_ip6, hlim=0
|
|
)
|
|
rx = self.send_and_expect_some(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip6_icmp(self.pg0, rx, tx)
|
|
|
|
#
|
|
# An MPLS xconnect - non-EOS label in IP out - an invalid configuration
|
|
# so this traffic should be dropped.
|
|
#
|
|
route_33_neos = VppMplsRoute(
|
|
self,
|
|
33,
|
|
0,
|
|
[VppRoutePath(self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[])],
|
|
)
|
|
route_33_neos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(33), VppMplsLabel(99)]
|
|
)
|
|
self.send_and_assert_no_replies(
|
|
self.pg0, tx, "MPLS non-EOS packets popped and forwarded"
|
|
)
|
|
|
|
#
|
|
# A recursive EOS x-connect, which resolves through another x-connect
|
|
# in pipe mode
|
|
#
|
|
route_34_eos = VppMplsRoute(
|
|
self,
|
|
34,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0",
|
|
0xFFFFFFFF,
|
|
nh_via_label=32,
|
|
labels=[VppMplsLabel(44), VppMplsLabel(45)],
|
|
)
|
|
],
|
|
)
|
|
route_34_eos.add_vpp_config()
|
|
self.logger.info(self.vapi.cli("sh mpls fib 34"))
|
|
|
|
tx = self.create_stream_labelled_ip4(self.pg0, [VppMplsLabel(34, ttl=3)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[VppMplsLabel(33), VppMplsLabel(44), VppMplsLabel(45, ttl=2)],
|
|
)
|
|
|
|
self.assertEqual(route_34_eos.get_stats_to()["packets"], 257)
|
|
self.assertEqual(route_32_neos.get_stats_via()["packets"], 257)
|
|
|
|
#
|
|
# A recursive EOS x-connect, which resolves through another x-connect
|
|
# in uniform mode
|
|
#
|
|
route_35_eos = VppMplsRoute(
|
|
self,
|
|
35,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0", 0xFFFFFFFF, nh_via_label=42, labels=[VppMplsLabel(44)]
|
|
)
|
|
],
|
|
)
|
|
route_35_eos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip4(self.pg0, [VppMplsLabel(35, ttl=3)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled(
|
|
self.pg0, rx, tx, [VppMplsLabel(43, ttl=2), VppMplsLabel(44, ttl=2)]
|
|
)
|
|
|
|
#
|
|
# A recursive non-EOS x-connect, which resolves through another
|
|
# x-connect
|
|
#
|
|
route_34_neos = VppMplsRoute(
|
|
self,
|
|
34,
|
|
0,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0",
|
|
0xFFFFFFFF,
|
|
nh_via_label=32,
|
|
labels=[VppMplsLabel(44), VppMplsLabel(46)],
|
|
)
|
|
],
|
|
)
|
|
route_34_neos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(34, ttl=45), VppMplsLabel(99)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
# it's the 2nd (counting from 0) label in the stack that is swapped
|
|
self.verify_capture_labelled(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[
|
|
VppMplsLabel(33),
|
|
VppMplsLabel(44),
|
|
VppMplsLabel(46, ttl=44),
|
|
VppMplsLabel(99),
|
|
],
|
|
)
|
|
|
|
#
|
|
# an recursive IP route that resolves through the recursive non-eos
|
|
# x-connect
|
|
#
|
|
ip_10_0_0_1 = VppIpRoute(
|
|
self,
|
|
"10.0.0.1",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0", 0xFFFFFFFF, nh_via_label=34, labels=[VppMplsLabel(55)]
|
|
)
|
|
],
|
|
)
|
|
ip_10_0_0_1.add_vpp_config()
|
|
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.1")
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[VppMplsLabel(33), VppMplsLabel(44), VppMplsLabel(46), VppMplsLabel(55)],
|
|
)
|
|
self.assertEqual(ip_10_0_0_1.get_stats_to()["packets"], 257)
|
|
|
|
ip_10_0_0_1.remove_vpp_config()
|
|
route_34_neos.remove_vpp_config()
|
|
route_34_eos.remove_vpp_config()
|
|
route_33_neos.remove_vpp_config()
|
|
route_33_eos.remove_vpp_config()
|
|
route_32_neos.remove_vpp_config()
|
|
route_32_eos.remove_vpp_config()
|
|
|
|
def test_bind(self):
|
|
"""MPLS Local Label Binding test"""
|
|
|
|
#
|
|
# Add a non-recursive route with a single out label
|
|
#
|
|
route_10_0_0_1 = VppIpRoute(
|
|
self,
|
|
"10.0.0.1",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(45)]
|
|
)
|
|
],
|
|
)
|
|
route_10_0_0_1.add_vpp_config()
|
|
|
|
# bind a local label to the route
|
|
binding = VppMplsIpBind(self, 44, "10.0.0.1", 32)
|
|
binding.add_vpp_config()
|
|
|
|
# non-EOS stream
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(44), VppMplsLabel(99)]
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled(
|
|
self.pg0, rx, tx, [VppMplsLabel(45, ttl=63), VppMplsLabel(99)]
|
|
)
|
|
|
|
# EOS stream
|
|
tx = self.create_stream_labelled_ip4(self.pg0, [VppMplsLabel(44)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled(self.pg0, rx, tx, [VppMplsLabel(45, ttl=63)])
|
|
|
|
# IP stream
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.1")
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(self.pg0, rx, tx, [VppMplsLabel(45)])
|
|
|
|
#
|
|
# cleanup
|
|
#
|
|
binding.remove_vpp_config()
|
|
route_10_0_0_1.remove_vpp_config()
|
|
|
|
def test_imposition(self):
|
|
"""MPLS label imposition test"""
|
|
|
|
#
|
|
# Add a non-recursive route with a single out label
|
|
#
|
|
route_10_0_0_1 = VppIpRoute(
|
|
self,
|
|
"10.0.0.1",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(32)]
|
|
)
|
|
],
|
|
)
|
|
route_10_0_0_1.add_vpp_config()
|
|
|
|
#
|
|
# a stream that matches the route for 10.0.0.1
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.1")
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(self.pg0, rx, tx, [VppMplsLabel(32)])
|
|
|
|
#
|
|
# Add a non-recursive route with a 3 out labels
|
|
#
|
|
route_10_0_0_2 = VppIpRoute(
|
|
self,
|
|
"10.0.0.2",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(32), VppMplsLabel(33), VppMplsLabel(34)],
|
|
)
|
|
],
|
|
)
|
|
route_10_0_0_2.add_vpp_config()
|
|
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.2", ip_ttl=44, ip_dscp=0xFF)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[VppMplsLabel(32), VppMplsLabel(33), VppMplsLabel(34)],
|
|
ip_ttl=43,
|
|
)
|
|
|
|
#
|
|
# Add a non-recursive route with a single out label in uniform mode
|
|
#
|
|
route_10_0_0_3 = VppIpRoute(
|
|
self,
|
|
"10.0.0.3",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(32, mode=MplsLspMode.UNIFORM)],
|
|
)
|
|
],
|
|
)
|
|
route_10_0_0_3.add_vpp_config()
|
|
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.3", ip_ttl=54, ip_dscp=0xBE)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(
|
|
self.pg0, rx, tx, [VppMplsLabel(32, ttl=53, exp=5)]
|
|
)
|
|
|
|
#
|
|
# Add a IPv6 non-recursive route with a single out label in
|
|
# uniform mode
|
|
#
|
|
route_2001_3 = VppIpRoute(
|
|
self,
|
|
"2001::3",
|
|
128,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip6,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(32, mode=MplsLspMode.UNIFORM)],
|
|
)
|
|
],
|
|
)
|
|
route_2001_3.add_vpp_config()
|
|
|
|
tx = self.create_stream_ip6(self.pg0, "2001::3", ip_ttl=54, ip_dscp=0xBE)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip6(
|
|
self.pg0, rx, tx, [VppMplsLabel(32, ttl=53, exp=5)]
|
|
)
|
|
|
|
#
|
|
# add a recursive path, with output label, via the 1 label route
|
|
#
|
|
route_11_0_0_1 = VppIpRoute(
|
|
self,
|
|
"11.0.0.1",
|
|
32,
|
|
[VppRoutePath("10.0.0.1", 0xFFFFFFFF, labels=[VppMplsLabel(44)])],
|
|
)
|
|
route_11_0_0_1.add_vpp_config()
|
|
|
|
#
|
|
# a stream that matches the route for 11.0.0.1, should pick up
|
|
# the label stack for 11.0.0.1 and 10.0.0.1
|
|
#
|
|
tx = self.create_stream_ip4(self.pg0, "11.0.0.1")
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(
|
|
self.pg0, rx, tx, [VppMplsLabel(32), VppMplsLabel(44)]
|
|
)
|
|
|
|
self.assertEqual(route_11_0_0_1.get_stats_to()["packets"], 257)
|
|
|
|
#
|
|
# add a recursive path, with 2 labels, via the 3 label route
|
|
#
|
|
route_11_0_0_2 = VppIpRoute(
|
|
self,
|
|
"11.0.0.2",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
"10.0.0.2", 0xFFFFFFFF, labels=[VppMplsLabel(44), VppMplsLabel(45)]
|
|
)
|
|
],
|
|
)
|
|
route_11_0_0_2.add_vpp_config()
|
|
|
|
#
|
|
# a stream that matches the route for 11.0.0.1, should pick up
|
|
# the label stack for 11.0.0.1 and 10.0.0.1
|
|
#
|
|
tx = self.create_stream_ip4(self.pg0, "11.0.0.2")
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[
|
|
VppMplsLabel(32),
|
|
VppMplsLabel(33),
|
|
VppMplsLabel(34),
|
|
VppMplsLabel(44),
|
|
VppMplsLabel(45),
|
|
],
|
|
)
|
|
|
|
self.assertEqual(route_11_0_0_2.get_stats_to()["packets"], 257)
|
|
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_labelled_ip4(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[
|
|
VppMplsLabel(32),
|
|
VppMplsLabel(33),
|
|
VppMplsLabel(34),
|
|
VppMplsLabel(44),
|
|
VppMplsLabel(45),
|
|
],
|
|
)
|
|
|
|
self.assertEqual(route_11_0_0_2.get_stats_to()["packets"], 514)
|
|
|
|
#
|
|
# cleanup
|
|
#
|
|
route_11_0_0_2.remove_vpp_config()
|
|
route_11_0_0_1.remove_vpp_config()
|
|
route_10_0_0_2.remove_vpp_config()
|
|
route_10_0_0_1.remove_vpp_config()
|
|
|
|
def test_imposition_fragmentation(self):
|
|
"""MPLS label imposition fragmentation test"""
|
|
|
|
#
|
|
# Add a ipv4 non-recursive route with a single out label
|
|
#
|
|
route_10_0_0_1 = VppIpRoute(
|
|
self,
|
|
"10.0.0.1",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(32)]
|
|
)
|
|
],
|
|
)
|
|
route_10_0_0_1.add_vpp_config()
|
|
route_1000_1 = VppIpRoute(
|
|
self,
|
|
"1000::1",
|
|
128,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip6, self.pg0.sw_if_index, labels=[VppMplsLabel(32)]
|
|
)
|
|
],
|
|
)
|
|
route_1000_1.add_vpp_config()
|
|
|
|
#
|
|
# a stream that matches the route for 10.0.0.1
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.1")
|
|
for i in range(0, 257):
|
|
self.extend_packet(tx[i], 10000)
|
|
|
|
#
|
|
# 5 fragments per packet (257*5=1285)
|
|
#
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0, 1285)
|
|
self.verify_capture_fragmented_labelled_ip4(
|
|
self.pg0, rx, tx, [VppMplsLabel(32)]
|
|
)
|
|
|
|
# packets with DF bit set generate ICMP
|
|
for t in tx:
|
|
t[IP].flags = "DF"
|
|
rxs = self.send_and_expect_some(self.pg0, tx, self.pg0)
|
|
|
|
for rx in rxs:
|
|
self.assertEqual(icmptypes[rx[ICMP].type], "dest-unreach")
|
|
self.assertEqual(
|
|
icmpcodes[rx[ICMP].type][rx[ICMP].code], "fragmentation-needed"
|
|
)
|
|
# the link MTU is 9000, the MPLS over head is 4 bytes
|
|
self.assertEqual(rx[ICMP].nexthopmtu, 9000 - 4)
|
|
|
|
self.assertEqual(
|
|
self.statistics.get_err_counter("/err/mpls-frag/dont_fragment_set"),
|
|
len(tx),
|
|
)
|
|
#
|
|
# a stream that matches the route for 1000::1/128
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_ip6(self.pg0, "1000::1")
|
|
for i in range(0, 257):
|
|
self.extend_packet(tx[i], 10000)
|
|
|
|
rxs = self.send_and_expect_some(self.pg0, tx, self.pg0)
|
|
for rx in rxs:
|
|
self.assertEqual(rx[ICMPv6PacketTooBig].mtu, 9000 - 4)
|
|
|
|
#
|
|
# cleanup
|
|
#
|
|
route_10_0_0_1.remove_vpp_config()
|
|
|
|
def test_tunnel_pipe(self):
|
|
"""MPLS Tunnel Tests - Pipe"""
|
|
|
|
#
|
|
# Create a tunnel with two out labels
|
|
#
|
|
mpls_tun = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(44), VppMplsLabel(46)],
|
|
)
|
|
],
|
|
)
|
|
mpls_tun.add_vpp_config()
|
|
mpls_tun.admin_up()
|
|
|
|
#
|
|
# add an unlabelled route through the new tunnel
|
|
#
|
|
route_10_0_0_3 = VppIpRoute(
|
|
self, "10.0.0.3", 32, [VppRoutePath("0.0.0.0", mpls_tun._sw_if_index)]
|
|
)
|
|
route_10_0_0_3.add_vpp_config()
|
|
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.3")
|
|
self.pg0.add_stream(tx)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg0.get_capture()
|
|
self.verify_capture_tunneled_ip4(
|
|
self.pg0, rx, tx, [VppMplsLabel(44), VppMplsLabel(46)]
|
|
)
|
|
|
|
#
|
|
# add a labelled route through the new tunnel
|
|
#
|
|
route_10_0_0_4 = VppIpRoute(
|
|
self,
|
|
"10.0.0.4",
|
|
32,
|
|
[VppRoutePath("0.0.0.0", mpls_tun._sw_if_index, labels=[33])],
|
|
)
|
|
route_10_0_0_4.add_vpp_config()
|
|
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.4")
|
|
self.pg0.add_stream(tx)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg0.get_capture()
|
|
self.verify_capture_tunneled_ip4(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[VppMplsLabel(44), VppMplsLabel(46), VppMplsLabel(33, ttl=255)],
|
|
)
|
|
|
|
#
|
|
# change tunnel's MTU to a low value
|
|
#
|
|
mpls_tun.set_l3_mtu(1200)
|
|
|
|
# send IP into the tunnel to be fragmented
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.3", payload_size=1500)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0, len(tx) * 2)
|
|
|
|
fake_tx = []
|
|
for p in tx:
|
|
fake_tx.append(p)
|
|
fake_tx.append(p)
|
|
self.verify_capture_tunneled_ip4(
|
|
self.pg0, rx, fake_tx, [VppMplsLabel(44), VppMplsLabel(46)]
|
|
)
|
|
|
|
# send MPLS into the tunnel to be fragmented
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.4", payload_size=1500)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0, len(tx) * 2)
|
|
|
|
fake_tx = []
|
|
for p in tx:
|
|
fake_tx.append(p)
|
|
fake_tx.append(p)
|
|
self.verify_capture_tunneled_ip4(
|
|
self.pg0,
|
|
rx,
|
|
fake_tx,
|
|
[VppMplsLabel(44), VppMplsLabel(46), VppMplsLabel(33, ttl=255)],
|
|
)
|
|
|
|
def test_tunnel_uniform(self):
|
|
"""MPLS Tunnel Tests - Uniform"""
|
|
|
|
#
|
|
# Create a tunnel with a single out label
|
|
# The label stack is specified here from outer to inner
|
|
#
|
|
mpls_tun = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[
|
|
VppMplsLabel(44, ttl=32),
|
|
VppMplsLabel(46, MplsLspMode.UNIFORM),
|
|
],
|
|
)
|
|
],
|
|
)
|
|
mpls_tun.add_vpp_config()
|
|
mpls_tun.admin_up()
|
|
|
|
#
|
|
# add an unlabelled route through the new tunnel
|
|
#
|
|
route_10_0_0_3 = VppIpRoute(
|
|
self, "10.0.0.3", 32, [VppRoutePath("0.0.0.0", mpls_tun._sw_if_index)]
|
|
)
|
|
route_10_0_0_3.add_vpp_config()
|
|
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.3", ip_ttl=24)
|
|
self.pg0.add_stream(tx)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg0.get_capture()
|
|
self.verify_capture_tunneled_ip4(
|
|
self.pg0, rx, tx, [VppMplsLabel(44, ttl=32), VppMplsLabel(46, ttl=23)]
|
|
)
|
|
|
|
#
|
|
# add a labelled route through the new tunnel
|
|
#
|
|
route_10_0_0_4 = VppIpRoute(
|
|
self,
|
|
"10.0.0.4",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0", mpls_tun._sw_if_index, labels=[VppMplsLabel(33, ttl=47)]
|
|
)
|
|
],
|
|
)
|
|
route_10_0_0_4.add_vpp_config()
|
|
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.4")
|
|
self.pg0.add_stream(tx)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg0.get_capture()
|
|
self.verify_capture_tunneled_ip4(
|
|
self.pg0,
|
|
rx,
|
|
tx,
|
|
[
|
|
VppMplsLabel(44, ttl=32),
|
|
VppMplsLabel(46, ttl=47),
|
|
VppMplsLabel(33, ttl=47),
|
|
],
|
|
)
|
|
|
|
def test_tunnel_ecmp(self):
|
|
"""MPLS Tunnel Tests - ECMP"""
|
|
|
|
#
|
|
# Create a tunnel with multiple paths and labels
|
|
#
|
|
self.pg0.generate_remote_hosts(2)
|
|
self.pg0.configure_ipv4_neighbors()
|
|
mpls_tun = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_hosts[0].ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(3)],
|
|
),
|
|
VppRoutePath(
|
|
self.pg0.remote_hosts[1].ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(44)],
|
|
),
|
|
],
|
|
)
|
|
mpls_tun.add_vpp_config()
|
|
mpls_tun.admin_up()
|
|
|
|
self.vapi.cli("clear trace")
|
|
pkts = self.create_stream_ip4(
|
|
self.pg0, ["10.0.0.%d" % i for i in range(NUM_PKTS)], n=NUM_PKTS
|
|
)
|
|
|
|
def send_and_expect_mpls_lb(pkts, path_labels, min_ratio):
|
|
self.pg0.add_stream(pkts)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg0.get_capture()
|
|
|
|
paths = {}
|
|
for packet in rx:
|
|
eth = packet[Ether]
|
|
self.assertEqual(eth.type, 0x8847)
|
|
|
|
mpls = packet[MPLS]
|
|
labels = []
|
|
while True:
|
|
labels.append(mpls.label)
|
|
if mpls.s == 1:
|
|
break
|
|
mpls = mpls[MPLS].payload
|
|
self.assertIn(labels, path_labels)
|
|
|
|
key = "{}-{}".format(eth.dst, "-".join(str(i) for i in labels))
|
|
paths[key] = paths.get(key, 0) + 1
|
|
|
|
#
|
|
# Check distribution over multiple mpls paths
|
|
#
|
|
self.assertEqual(len(paths), len(path_labels))
|
|
for n in paths.values():
|
|
self.assertGreaterEqual(n, NUM_PKTS / len(paths) * min_ratio)
|
|
|
|
#
|
|
# Add labelled route through the new tunnel,
|
|
# traffic should be balanced over all tunnel paths only.
|
|
#
|
|
route_10_0_0_0 = VppIpRoute(
|
|
self,
|
|
"10.0.0.0",
|
|
16,
|
|
[VppRoutePath("0.0.0.0", mpls_tun._sw_if_index, labels=[33])],
|
|
)
|
|
route_10_0_0_0.add_vpp_config()
|
|
send_and_expect_mpls_lb(pkts, [[33], [44, 33]], 0.85)
|
|
|
|
#
|
|
# Add labelled multipath route through the new tunnel,
|
|
# traffic should be balanced over both paths first and
|
|
# then over all tunnel paths.
|
|
#
|
|
route_10_0_0_0 = VppIpRoute(
|
|
self,
|
|
"10.0.0.0",
|
|
16,
|
|
[
|
|
VppRoutePath("0.0.0.1", mpls_tun._sw_if_index, labels=[33]),
|
|
VppRoutePath("0.0.0.2", mpls_tun._sw_if_index, labels=[34]),
|
|
],
|
|
)
|
|
route_10_0_0_0.add_vpp_config()
|
|
send_and_expect_mpls_lb(pkts, [[33], [44, 33], [34], [44, 34]], 0.70)
|
|
|
|
def test_mpls_tunnel_many(self):
|
|
"""MPLS Multiple Tunnels"""
|
|
|
|
for ii in range(100):
|
|
mpls_tun = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[
|
|
VppMplsLabel(44, ttl=32),
|
|
VppMplsLabel(46, MplsLspMode.UNIFORM),
|
|
],
|
|
)
|
|
],
|
|
)
|
|
mpls_tun.add_vpp_config()
|
|
mpls_tun.admin_up()
|
|
for ii in range(100):
|
|
mpls_tun = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[
|
|
VppMplsLabel(44, ttl=32),
|
|
VppMplsLabel(46, MplsLspMode.UNIFORM),
|
|
],
|
|
)
|
|
],
|
|
is_l2=1,
|
|
)
|
|
mpls_tun.add_vpp_config()
|
|
mpls_tun.admin_up()
|
|
|
|
def test_v4_exp_null(self):
|
|
"""MPLS V4 Explicit NULL test"""
|
|
|
|
#
|
|
# The first test case has an MPLS TTL of 0
|
|
# all packet should be dropped
|
|
#
|
|
tx = self.create_stream_labelled_ip4(self.pg0, [VppMplsLabel(0, ttl=0)])
|
|
self.send_and_assert_no_replies(self.pg0, tx, "MPLS TTL=0 packets forwarded")
|
|
|
|
#
|
|
# a stream with a non-zero MPLS TTL
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_labelled_ip4(self.pg0, [VppMplsLabel(0)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip4(self.pg0, rx, tx)
|
|
|
|
#
|
|
# a stream with a non-zero MPLS TTL
|
|
# PG1 is in table 1
|
|
# we are ensuring the post-pop lookup occurs in the VRF table
|
|
#
|
|
tx = self.create_stream_labelled_ip4(self.pg1, [VppMplsLabel(0)])
|
|
rx = self.send_and_expect(self.pg1, tx, self.pg1)
|
|
self.verify_capture_ip4(self.pg1, rx, tx)
|
|
|
|
def test_v6_exp_null(self):
|
|
"""MPLS V6 Explicit NULL test"""
|
|
|
|
#
|
|
# a stream with a non-zero MPLS TTL
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_labelled_ip6(self.pg0, [VppMplsLabel(2)])
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip6(self.pg0, rx, tx)
|
|
|
|
#
|
|
# a stream with a non-zero MPLS TTL
|
|
# PG1 is in table 1
|
|
# we are ensuring the post-pop lookup occurs in the VRF table
|
|
#
|
|
tx = self.create_stream_labelled_ip6(self.pg1, [VppMplsLabel(2)])
|
|
rx = self.send_and_expect(self.pg1, tx, self.pg1)
|
|
self.verify_capture_ip6(self.pg0, rx, tx)
|
|
|
|
@unittest.skipIf(
|
|
"ping" in config.excluded_plugins, "Exclude tests requiring Ping plugin"
|
|
)
|
|
def test_deag(self):
|
|
"""MPLS Deagg"""
|
|
|
|
#
|
|
# A de-agg route - next-hop lookup in default table
|
|
#
|
|
route_34_eos = VppMplsRoute(
|
|
self, 34, 1, [VppRoutePath("0.0.0.0", 0xFFFFFFFF, nh_table_id=0)]
|
|
)
|
|
route_34_eos.add_vpp_config()
|
|
|
|
#
|
|
# ping an interface in the default table
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(34)], ping=1, ip_itf=self.pg0
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip4(self.pg0, rx, tx, ping_resp=1)
|
|
|
|
#
|
|
# A de-agg route - next-hop lookup in non-default table
|
|
#
|
|
route_35_eos = VppMplsRoute(
|
|
self, 35, 1, [VppRoutePath("0.0.0.0", 0xFFFFFFFF, nh_table_id=1)]
|
|
)
|
|
route_35_eos.add_vpp_config()
|
|
route_356_eos = VppMplsRoute(
|
|
self,
|
|
356,
|
|
1,
|
|
[VppRoutePath("0::0", 0xFFFFFFFF, nh_table_id=1)],
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_IP6,
|
|
)
|
|
route_356_eos.add_vpp_config()
|
|
|
|
#
|
|
# ping an interface in the non-default table
|
|
# PG0 is in the default table. packet arrive labelled in the
|
|
# default table and egress unlabelled in the non-default
|
|
#
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(35)], ping=1, ip_itf=self.pg1
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg1)
|
|
self.verify_capture_ip4(self.pg1, rx, tx, ping_resp=1)
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(356)], ping=1, ip_itf=self.pg1
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg1)
|
|
self.verify_capture_ip6(self.pg1, rx, tx, ping_resp=1)
|
|
|
|
#
|
|
# Double pop
|
|
#
|
|
route_36_neos = VppMplsRoute(self, 36, 0, [VppRoutePath("0.0.0.0", 0xFFFFFFFF)])
|
|
route_36_neos.add_vpp_config()
|
|
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(36), VppMplsLabel(35)], ping=1, ip_itf=self.pg1
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg1)
|
|
self.verify_capture_ip4(self.pg1, rx, tx, ping_resp=1)
|
|
|
|
route_36_neos.remove_vpp_config()
|
|
route_35_eos.remove_vpp_config()
|
|
route_34_eos.remove_vpp_config()
|
|
|
|
def test_interface_rx(self):
|
|
"""MPLS Interface Receive"""
|
|
|
|
#
|
|
# Add a non-recursive route that will forward the traffic
|
|
# post-interface-rx
|
|
#
|
|
route_10_0_0_1 = VppIpRoute(
|
|
self,
|
|
"10.0.0.1",
|
|
32,
|
|
table_id=1,
|
|
paths=[VppRoutePath(self.pg1.remote_ip4, self.pg1.sw_if_index)],
|
|
)
|
|
route_10_0_0_1.add_vpp_config()
|
|
|
|
#
|
|
# An interface receive label that maps traffic to RX on interface
|
|
# pg1
|
|
# by injecting the packet in on pg0, which is in table 0
|
|
# doing an interface-rx on pg1 and matching a route in table 1
|
|
# if the packet egresses, then we must have swapped to pg1
|
|
# so as to have matched the route in table 1
|
|
#
|
|
route_34_eos = VppMplsRoute(
|
|
self,
|
|
34,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0",
|
|
self.pg1.sw_if_index,
|
|
type=FibPathType.FIB_PATH_TYPE_INTERFACE_RX,
|
|
)
|
|
],
|
|
)
|
|
route_34_eos.add_vpp_config()
|
|
|
|
#
|
|
# ping an interface in the default table
|
|
# PG0 is in the default table
|
|
#
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="10.0.0.1"
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg1)
|
|
self.verify_capture_ip4(self.pg1, rx, tx)
|
|
|
|
def test_mcast_mid_point(self):
|
|
"""MPLS Multicast Mid Point"""
|
|
|
|
#
|
|
# Add a non-recursive route that will forward the traffic
|
|
# post-interface-rx
|
|
#
|
|
route_10_0_0_1 = VppIpRoute(
|
|
self,
|
|
"10.0.0.1",
|
|
32,
|
|
table_id=1,
|
|
paths=[VppRoutePath(self.pg1.remote_ip4, self.pg1.sw_if_index)],
|
|
)
|
|
route_10_0_0_1.add_vpp_config()
|
|
|
|
#
|
|
# Add a mcast entry that replicate to pg2 and pg3
|
|
# and replicate to a interface-rx (like a bud node would)
|
|
#
|
|
route_3400_eos = VppMplsRoute(
|
|
self,
|
|
3400,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
self.pg2.remote_ip4,
|
|
self.pg2.sw_if_index,
|
|
labels=[VppMplsLabel(3401)],
|
|
),
|
|
VppRoutePath(
|
|
self.pg3.remote_ip4,
|
|
self.pg3.sw_if_index,
|
|
labels=[VppMplsLabel(3402)],
|
|
),
|
|
VppRoutePath(
|
|
"0.0.0.0",
|
|
self.pg1.sw_if_index,
|
|
type=FibPathType.FIB_PATH_TYPE_INTERFACE_RX,
|
|
),
|
|
],
|
|
is_multicast=1,
|
|
)
|
|
route_3400_eos.add_vpp_config()
|
|
|
|
#
|
|
# ping an interface in the default table
|
|
# PG0 is in the default table
|
|
#
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(3400, ttl=64)], n=257, dst_ip="10.0.0.1"
|
|
)
|
|
self.pg0.add_stream(tx)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg1.get_capture(257)
|
|
self.verify_capture_ip4(self.pg1, rx, tx)
|
|
|
|
rx = self.pg2.get_capture(257)
|
|
self.verify_capture_labelled(self.pg2, rx, tx, [VppMplsLabel(3401, ttl=63)])
|
|
rx = self.pg3.get_capture(257)
|
|
self.verify_capture_labelled(self.pg3, rx, tx, [VppMplsLabel(3402, ttl=63)])
|
|
|
|
def test_mcast_head(self):
|
|
"""MPLS Multicast Head-end"""
|
|
|
|
MRouteItfFlags = VppEnum.vl_api_mfib_itf_flags_t
|
|
MRouteEntryFlags = VppEnum.vl_api_mfib_entry_flags_t
|
|
|
|
#
|
|
# Create a multicast tunnel with two replications
|
|
#
|
|
mpls_tun = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg2.remote_ip4, self.pg2.sw_if_index, labels=[VppMplsLabel(42)]
|
|
),
|
|
VppRoutePath(
|
|
self.pg3.remote_ip4, self.pg3.sw_if_index, labels=[VppMplsLabel(43)]
|
|
),
|
|
],
|
|
is_multicast=1,
|
|
)
|
|
mpls_tun.add_vpp_config()
|
|
mpls_tun.admin_up()
|
|
|
|
#
|
|
# add an unlabelled route through the new tunnel
|
|
#
|
|
route_10_0_0_3 = VppIpRoute(
|
|
self, "10.0.0.3", 32, [VppRoutePath("0.0.0.0", mpls_tun._sw_if_index)]
|
|
)
|
|
route_10_0_0_3.add_vpp_config()
|
|
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_ip4(self.pg0, "10.0.0.3")
|
|
self.pg0.add_stream(tx)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg2.get_capture(257)
|
|
self.verify_capture_tunneled_ip4(self.pg0, rx, tx, [VppMplsLabel(42)])
|
|
rx = self.pg3.get_capture(257)
|
|
self.verify_capture_tunneled_ip4(self.pg0, rx, tx, [VppMplsLabel(43)])
|
|
|
|
#
|
|
# An an IP multicast route via the tunnel
|
|
# A (*,G).
|
|
# one accepting interface, pg0, 1 forwarding interface via the tunnel
|
|
#
|
|
route_232_1_1_1 = VppIpMRoute(
|
|
self,
|
|
"0.0.0.0",
|
|
"232.1.1.1",
|
|
32,
|
|
MRouteEntryFlags.MFIB_API_ENTRY_FLAG_NONE,
|
|
[
|
|
VppMRoutePath(
|
|
self.pg0.sw_if_index, MRouteItfFlags.MFIB_API_ITF_FLAG_ACCEPT
|
|
),
|
|
VppMRoutePath(
|
|
mpls_tun._sw_if_index, MRouteItfFlags.MFIB_API_ITF_FLAG_FORWARD
|
|
),
|
|
],
|
|
)
|
|
route_232_1_1_1.add_vpp_config()
|
|
self.logger.info(self.vapi.cli("sh ip mfib index 0"))
|
|
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_ip4(self.pg0, "232.1.1.1")
|
|
self.pg0.add_stream(tx)
|
|
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg2.get_capture(257)
|
|
self.verify_capture_tunneled_ip4(self.pg0, rx, tx, [VppMplsLabel(42)])
|
|
rx = self.pg3.get_capture(257)
|
|
self.verify_capture_tunneled_ip4(self.pg0, rx, tx, [VppMplsLabel(43)])
|
|
|
|
def test_mcast_ip4_tail(self):
|
|
"""MPLS IPv4 Multicast Tail"""
|
|
|
|
MRouteItfFlags = VppEnum.vl_api_mfib_itf_flags_t
|
|
MRouteEntryFlags = VppEnum.vl_api_mfib_entry_flags_t
|
|
|
|
#
|
|
# Add a multicast route that will forward the traffic
|
|
# post-disposition
|
|
#
|
|
route_232_1_1_1 = VppIpMRoute(
|
|
self,
|
|
"0.0.0.0",
|
|
"232.1.1.1",
|
|
32,
|
|
MRouteEntryFlags.MFIB_API_ENTRY_FLAG_NONE,
|
|
table_id=1,
|
|
paths=[
|
|
VppMRoutePath(
|
|
self.pg1.sw_if_index, MRouteItfFlags.MFIB_API_ITF_FLAG_FORWARD
|
|
)
|
|
],
|
|
)
|
|
route_232_1_1_1.add_vpp_config()
|
|
|
|
#
|
|
# An interface receive label that maps traffic to RX on interface
|
|
# pg1
|
|
# by injecting the packet in on pg0, which is in table 0
|
|
# doing an rpf-id and matching a route in table 1
|
|
# if the packet egresses, then we must have matched the route in
|
|
# table 1
|
|
#
|
|
route_34_eos = VppMplsRoute(
|
|
self,
|
|
34,
|
|
1,
|
|
[VppRoutePath("0.0.0.0", 0xFFFFFFFF, nh_table_id=1, rpf_id=55)],
|
|
is_multicast=1,
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_IP4,
|
|
)
|
|
|
|
route_34_eos.add_vpp_config()
|
|
|
|
#
|
|
# Drop due to interface lookup miss
|
|
#
|
|
self.vapi.cli("clear trace")
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="232.1.1.1", n=1
|
|
)
|
|
self.send_and_assert_no_replies(self.pg0, tx, "RPF-ID drop none")
|
|
|
|
#
|
|
# set the RPF-ID of the entry to match the input packet's
|
|
#
|
|
route_232_1_1_1.update_rpf_id(55)
|
|
self.logger.info(self.vapi.cli("sh ip mfib index 1 232.1.1.1"))
|
|
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="232.1.1.1"
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg1)
|
|
self.verify_capture_ip4(self.pg1, rx, tx)
|
|
|
|
#
|
|
# disposed packets have an invalid IPv4 checksum
|
|
#
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="232.1.1.1", n=65, chksum=1
|
|
)
|
|
self.send_and_assert_no_replies(self.pg0, tx, "Invalid Checksum")
|
|
|
|
#
|
|
# set the RPF-ID of the entry to not match the input packet's
|
|
#
|
|
route_232_1_1_1.update_rpf_id(56)
|
|
tx = self.create_stream_labelled_ip4(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="232.1.1.1"
|
|
)
|
|
self.send_and_assert_no_replies(self.pg0, tx, "RPF-ID drop 56")
|
|
|
|
def test_mcast_ip6_tail(self):
|
|
"""MPLS IPv6 Multicast Tail"""
|
|
|
|
MRouteItfFlags = VppEnum.vl_api_mfib_itf_flags_t
|
|
MRouteEntryFlags = VppEnum.vl_api_mfib_entry_flags_t
|
|
|
|
#
|
|
# Add a multicast route that will forward the traffic
|
|
# post-disposition
|
|
#
|
|
route_ff = VppIpMRoute(
|
|
self,
|
|
"::",
|
|
"ff01::1",
|
|
32,
|
|
MRouteEntryFlags.MFIB_API_ENTRY_FLAG_NONE,
|
|
table_id=1,
|
|
paths=[
|
|
VppMRoutePath(
|
|
self.pg1.sw_if_index,
|
|
MRouteItfFlags.MFIB_API_ITF_FLAG_FORWARD,
|
|
proto=FibPathProto.FIB_PATH_NH_PROTO_IP6,
|
|
)
|
|
],
|
|
)
|
|
route_ff.add_vpp_config()
|
|
|
|
#
|
|
# An interface receive label that maps traffic to RX on interface
|
|
# pg1
|
|
# by injecting the packet in on pg0, which is in table 0
|
|
# doing an rpf-id and matching a route in table 1
|
|
# if the packet egresses, then we must have matched the route in
|
|
# table 1
|
|
#
|
|
route_34_eos = VppMplsRoute(
|
|
self,
|
|
34,
|
|
1,
|
|
[VppRoutePath("::", 0xFFFFFFFF, nh_table_id=1, rpf_id=55)],
|
|
is_multicast=1,
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_IP6,
|
|
)
|
|
|
|
route_34_eos.add_vpp_config()
|
|
|
|
#
|
|
# Drop due to interface lookup miss
|
|
#
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="ff01::1"
|
|
)
|
|
self.send_and_assert_no_replies(self.pg0, tx, "RPF Miss")
|
|
|
|
#
|
|
# set the RPF-ID of the entry to match the input packet's
|
|
#
|
|
route_ff.update_rpf_id(55)
|
|
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="ff01::1"
|
|
)
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg1)
|
|
self.verify_capture_ip6(self.pg1, rx, tx)
|
|
|
|
#
|
|
# disposed packets have hop-limit = 1
|
|
#
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="ff01::1", hlim=1
|
|
)
|
|
rx = self.send_and_expect_some(self.pg0, tx, self.pg0)
|
|
self.verify_capture_ip6_icmp(self.pg0, rx, tx)
|
|
|
|
#
|
|
# set the RPF-ID of the entry to not match the input packet's
|
|
#
|
|
route_ff.update_rpf_id(56)
|
|
tx = self.create_stream_labelled_ip6(
|
|
self.pg0, [VppMplsLabel(34)], dst_ip="ff01::1"
|
|
)
|
|
self.send_and_assert_no_replies(self.pg0, tx, "RPF-ID drop 56")
|
|
|
|
def test_6pe(self):
|
|
"""MPLS 6PE"""
|
|
|
|
#
|
|
# Add a non-recursive route with a single out label
|
|
#
|
|
route_10_0_0_1 = VppIpRoute(
|
|
self,
|
|
"10.0.0.1",
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(45)]
|
|
)
|
|
],
|
|
)
|
|
route_10_0_0_1.add_vpp_config()
|
|
|
|
# bind a local label to the route
|
|
binding = VppMplsIpBind(self, 44, "10.0.0.1", 32)
|
|
binding.add_vpp_config()
|
|
|
|
#
|
|
# a labelled v6 route that resolves through the v4
|
|
#
|
|
route_2001_3 = VppIpRoute(
|
|
self,
|
|
"2001::3",
|
|
128,
|
|
[VppRoutePath("10.0.0.1", INVALID_INDEX, labels=[VppMplsLabel(32)])],
|
|
)
|
|
route_2001_3.add_vpp_config()
|
|
|
|
tx = self.create_stream_ip6(self.pg0, "2001::3")
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
|
|
self.verify_capture_labelled_ip6(
|
|
self.pg0, rx, tx, [VppMplsLabel(45), VppMplsLabel(32)]
|
|
)
|
|
|
|
#
|
|
# and a v4 recursive via the v6
|
|
#
|
|
route_20_3 = VppIpRoute(
|
|
self,
|
|
"20.0.0.3",
|
|
32,
|
|
[VppRoutePath("2001::3", INVALID_INDEX, labels=[VppMplsLabel(99)])],
|
|
)
|
|
route_20_3.add_vpp_config()
|
|
|
|
tx = self.create_stream_ip4(self.pg0, "20.0.0.3")
|
|
rx = self.send_and_expect(self.pg0, tx, self.pg0)
|
|
|
|
self.verify_capture_labelled_ip4(
|
|
self.pg0, rx, tx, [VppMplsLabel(45), VppMplsLabel(32), VppMplsLabel(99)]
|
|
)
|
|
|
|
def test_attached(self):
|
|
"""Attach Routes with Local Label"""
|
|
|
|
#
|
|
# test that if a local label is associated with an attached/connected
|
|
# prefix, that we can reach hosts in the prefix.
|
|
#
|
|
binding = VppMplsIpBind(
|
|
self, 44, self.pg0._local_ip4_subnet, self.pg0.local_ip4_prefix_len
|
|
)
|
|
binding.add_vpp_config()
|
|
|
|
tx = (
|
|
Ether(src=self.pg0.remote_mac, dst=self.pg0.local_mac)
|
|
/ MPLS(label=44, ttl=64)
|
|
/ IP(src=self.pg0.remote_ip4, dst=self.pg0.remote_ip4)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
rxs = self.send_and_expect(self.pg0, [tx], self.pg0)
|
|
for rx in rxs:
|
|
# if there's an ARP then the label is linked to the glean
|
|
# which is wrong.
|
|
self.assertFalse(rx.haslayer(ARP))
|
|
# it should be unicasted to the host
|
|
self.assertEqual(rx[Ether].dst, self.pg0.remote_mac)
|
|
self.assertEqual(rx[IP].dst, self.pg0.remote_ip4)
|
|
|
|
|
|
class TestMPLSDisabled(VppTestCase):
|
|
"""MPLS disabled"""
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
super(TestMPLSDisabled, cls).setUpClass()
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestMPLSDisabled, cls).tearDownClass()
|
|
|
|
def setUp(self):
|
|
super(TestMPLSDisabled, self).setUp()
|
|
|
|
# create 2 pg interfaces
|
|
self.create_pg_interfaces(range(2))
|
|
|
|
self.tbl = VppMplsTable(self, 0)
|
|
self.tbl.add_vpp_config()
|
|
|
|
# PG0 is MPLS enabled
|
|
self.pg0.admin_up()
|
|
self.pg0.config_ip4()
|
|
self.pg0.resolve_arp()
|
|
self.pg0.enable_mpls()
|
|
|
|
# PG 1 is not MPLS enabled
|
|
self.pg1.admin_up()
|
|
|
|
def tearDown(self):
|
|
for i in self.pg_interfaces:
|
|
i.unconfig_ip4()
|
|
i.admin_down()
|
|
|
|
self.pg0.disable_mpls()
|
|
super(TestMPLSDisabled, self).tearDown()
|
|
|
|
def test_mpls_disabled(self):
|
|
"""MPLS Disabled"""
|
|
|
|
self.logger.info(self.vapi.cli("show mpls interface"))
|
|
self.logger.info(self.vapi.cli("show mpls interface pg1"))
|
|
self.logger.info(self.vapi.cli("show mpls interface pg0"))
|
|
|
|
tx = (
|
|
Ether(src=self.pg1.remote_mac, dst=self.pg1.local_mac)
|
|
/ MPLS(label=32, ttl=64)
|
|
/ IPv6(src="2001::1", dst=self.pg0.remote_ip6)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
|
|
#
|
|
# A simple MPLS xconnect - eos label in label out
|
|
#
|
|
route_32_eos = VppMplsRoute(
|
|
self,
|
|
32,
|
|
1,
|
|
[VppRoutePath(self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[33])],
|
|
)
|
|
route_32_eos.add_vpp_config()
|
|
|
|
#
|
|
# PG1 does not forward IP traffic
|
|
#
|
|
self.send_and_assert_no_replies(self.pg1, tx, "MPLS disabled")
|
|
|
|
#
|
|
# MPLS enable PG1
|
|
#
|
|
self.pg1.enable_mpls()
|
|
|
|
self.logger.info(self.vapi.cli("show mpls interface"))
|
|
self.logger.info(self.vapi.cli("show mpls interface pg1"))
|
|
|
|
#
|
|
# Now we get packets through
|
|
#
|
|
self.pg1.add_stream(tx)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx = self.pg0.get_capture(1)
|
|
|
|
#
|
|
# Disable PG1
|
|
#
|
|
self.pg1.disable_mpls()
|
|
|
|
#
|
|
# PG1 does not forward IP traffic
|
|
#
|
|
self.send_and_assert_no_replies(self.pg1, tx, "IPv6 disabled")
|
|
self.send_and_assert_no_replies(self.pg1, tx, "IPv6 disabled")
|
|
|
|
|
|
class TestMPLSPIC(VppTestCase):
|
|
"""MPLS Prefix-Independent Convergence (PIC) edge convergence"""
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
super(TestMPLSPIC, cls).setUpClass()
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestMPLSPIC, cls).tearDownClass()
|
|
|
|
def setUp(self):
|
|
super(TestMPLSPIC, self).setUp()
|
|
|
|
# create 2 pg interfaces
|
|
self.create_pg_interfaces(range(4))
|
|
|
|
mpls_tbl = VppMplsTable(self, 0)
|
|
mpls_tbl.add_vpp_config()
|
|
tbl4 = VppIpTable(self, 1)
|
|
tbl4.add_vpp_config()
|
|
tbl6 = VppIpTable(self, 1, is_ip6=1)
|
|
tbl6.add_vpp_config()
|
|
|
|
# core links
|
|
self.pg0.admin_up()
|
|
self.pg0.config_ip4()
|
|
self.pg0.resolve_arp()
|
|
self.pg0.enable_mpls()
|
|
|
|
self.pg1.admin_up()
|
|
self.pg1.config_ip4()
|
|
self.pg1.resolve_arp()
|
|
self.pg1.enable_mpls()
|
|
|
|
# VRF (customer facing) link
|
|
self.pg2.admin_up()
|
|
self.pg2.set_table_ip4(1)
|
|
self.pg2.config_ip4()
|
|
self.pg2.resolve_arp()
|
|
self.pg2.set_table_ip6(1)
|
|
self.pg2.config_ip6()
|
|
self.pg2.resolve_ndp()
|
|
|
|
self.pg3.admin_up()
|
|
self.pg3.set_table_ip4(1)
|
|
self.pg3.config_ip4()
|
|
self.pg3.resolve_arp()
|
|
self.pg3.set_table_ip6(1)
|
|
self.pg3.config_ip6()
|
|
self.pg3.resolve_ndp()
|
|
|
|
def tearDown(self):
|
|
self.pg0.disable_mpls()
|
|
self.pg1.disable_mpls()
|
|
for i in self.pg_interfaces:
|
|
i.unconfig_ip4()
|
|
i.unconfig_ip6()
|
|
i.set_table_ip4(0)
|
|
i.set_table_ip6(0)
|
|
i.admin_down()
|
|
super(TestMPLSPIC, self).tearDown()
|
|
|
|
def test_mpls_ibgp_pic(self):
|
|
"""MPLS iBGP Prefix-Independent Convergence (PIC) edge convergence
|
|
|
|
1) setup many iBGP VPN routes via a pair of iBGP peers.
|
|
2) Check EMCP forwarding to these peers
|
|
3) withdraw the IGP route to one of these peers.
|
|
4) check forwarding continues to the remaining peer
|
|
"""
|
|
|
|
#
|
|
# IGP+LDP core routes
|
|
#
|
|
core_10_0_0_45 = VppIpRoute(
|
|
self,
|
|
"10.0.0.45",
|
|
32,
|
|
[VppRoutePath(self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[45])],
|
|
)
|
|
core_10_0_0_45.add_vpp_config()
|
|
|
|
core_10_0_0_46 = VppIpRoute(
|
|
self,
|
|
"10.0.0.46",
|
|
32,
|
|
[VppRoutePath(self.pg1.remote_ip4, self.pg1.sw_if_index, labels=[46])],
|
|
)
|
|
core_10_0_0_46.add_vpp_config()
|
|
|
|
#
|
|
# Lot's of VPN routes. We need more the 64 so VPP will build
|
|
# the fast convergence indirection
|
|
#
|
|
vpn_routes = []
|
|
pkts = []
|
|
for ii in range(NUM_PKTS):
|
|
dst = "192.168.1.%d" % ii
|
|
vpn_routes.append(
|
|
VppIpRoute(
|
|
self,
|
|
dst,
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
"10.0.0.45",
|
|
0xFFFFFFFF,
|
|
labels=[145],
|
|
flags=FibPathFlags.FIB_PATH_FLAG_RESOLVE_VIA_HOST,
|
|
),
|
|
VppRoutePath(
|
|
"10.0.0.46",
|
|
0xFFFFFFFF,
|
|
labels=[146],
|
|
flags=FibPathFlags.FIB_PATH_FLAG_RESOLVE_VIA_HOST,
|
|
),
|
|
],
|
|
table_id=1,
|
|
)
|
|
)
|
|
vpn_routes[ii].add_vpp_config()
|
|
|
|
pkts.append(
|
|
Ether(dst=self.pg2.local_mac, src=self.pg2.remote_mac)
|
|
/ IP(src=self.pg2.remote_ip4, dst=dst)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
|
|
#
|
|
# Send the packet stream (one pkt to each VPN route)
|
|
# - expect a 50-50 split of the traffic
|
|
#
|
|
self.pg2.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg0._get_capture(NUM_PKTS)
|
|
rx1 = self.pg1._get_capture(NUM_PKTS)
|
|
|
|
# not testing the LB hashing algorithm so we're not concerned
|
|
# with the split ratio, just as long as neither is 0
|
|
self.assertNotEqual(0, len(rx0))
|
|
self.assertNotEqual(0, len(rx1))
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0) + len(rx1),
|
|
"Expected all (%s) packets across both ECMP paths. "
|
|
"rx0: %s rx1: %s." % (len(pkts), len(rx0), len(rx1)),
|
|
)
|
|
|
|
#
|
|
# use a test CLI command to stop the FIB walk process, this
|
|
# will prevent the FIB converging the VPN routes and thus allow
|
|
# us to probe the interim (post-fail, pre-converge) state
|
|
#
|
|
self.vapi.ppcli("test fib-walk-process disable")
|
|
|
|
#
|
|
# Withdraw one of the IGP routes
|
|
#
|
|
core_10_0_0_46.remove_vpp_config()
|
|
|
|
#
|
|
# now all packets should be forwarded through the remaining peer
|
|
#
|
|
self.vapi.ppcli("clear trace")
|
|
self.pg2.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg0.get_capture(NUM_PKTS)
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0),
|
|
"Expected all (%s) packets across single path. "
|
|
"rx0: %s." % (len(pkts), len(rx0)),
|
|
)
|
|
|
|
#
|
|
# enable the FIB walk process to converge the FIB
|
|
#
|
|
self.vapi.ppcli("test fib-walk-process enable")
|
|
|
|
#
|
|
# packets should still be forwarded through the remaining peer
|
|
#
|
|
self.pg2.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg0.get_capture(NUM_PKTS)
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0),
|
|
"Expected all (%s) packets across single path. "
|
|
"rx0: %s." % (len(pkts), len(rx0)),
|
|
)
|
|
|
|
#
|
|
# Add the IGP route back and we return to load-balancing
|
|
#
|
|
core_10_0_0_46.add_vpp_config()
|
|
|
|
self.pg2.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg0._get_capture(NUM_PKTS)
|
|
rx1 = self.pg1._get_capture(NUM_PKTS)
|
|
self.assertNotEqual(0, len(rx0))
|
|
self.assertNotEqual(0, len(rx1))
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0) + len(rx1),
|
|
"Expected all (%s) packets across both ECMP paths. "
|
|
"rx0: %s rx1: %s." % (len(pkts), len(rx0), len(rx1)),
|
|
)
|
|
|
|
def test_mpls_ebgp_pic(self):
|
|
"""MPLS eBGP Prefix-Independent Convergence (PIC) edge convergence
|
|
|
|
1) setup many eBGP VPN routes via a pair of eBGP peers.
|
|
2) Check EMCP forwarding to these peers
|
|
3) withdraw one eBGP path - expect LB across remaining eBGP
|
|
"""
|
|
|
|
#
|
|
# Lot's of VPN routes. We need more the 64 so VPP will build
|
|
# the fast convergence indirection
|
|
#
|
|
vpn_routes = []
|
|
vpn_bindings = []
|
|
pkts = []
|
|
for ii in range(NUM_PKTS):
|
|
dst = "192.168.1.%d" % ii
|
|
local_label = 1600 + ii
|
|
vpn_routes.append(
|
|
VppIpRoute(
|
|
self,
|
|
dst,
|
|
32,
|
|
[
|
|
VppRoutePath(
|
|
self.pg2.remote_ip4,
|
|
0xFFFFFFFF,
|
|
nh_table_id=1,
|
|
flags=FibPathFlags.FIB_PATH_FLAG_RESOLVE_VIA_ATTACHED,
|
|
),
|
|
VppRoutePath(
|
|
self.pg3.remote_ip4,
|
|
0xFFFFFFFF,
|
|
nh_table_id=1,
|
|
flags=FibPathFlags.FIB_PATH_FLAG_RESOLVE_VIA_ATTACHED,
|
|
),
|
|
],
|
|
table_id=1,
|
|
)
|
|
)
|
|
vpn_routes[ii].add_vpp_config()
|
|
|
|
vpn_bindings.append(
|
|
VppMplsIpBind(self, local_label, dst, 32, ip_table_id=1)
|
|
)
|
|
vpn_bindings[ii].add_vpp_config()
|
|
|
|
pkts.append(
|
|
Ether(dst=self.pg0.local_mac, src=self.pg0.remote_mac)
|
|
/ MPLS(label=local_label, ttl=64)
|
|
/ IP(src=self.pg0.remote_ip4, dst=dst)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
|
|
#
|
|
# Send the packet stream (one pkt to each VPN route)
|
|
# - expect a 50-50 split of the traffic
|
|
#
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg2._get_capture(NUM_PKTS)
|
|
rx1 = self.pg3._get_capture(NUM_PKTS)
|
|
|
|
# not testing the LB hashing algorithm so we're not concerned
|
|
# with the split ratio, just as long as neither is 0
|
|
self.assertNotEqual(0, len(rx0))
|
|
self.assertNotEqual(0, len(rx1))
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0) + len(rx1),
|
|
"Expected all (%s) packets across both ECMP paths. "
|
|
"rx0: %s rx1: %s." % (len(pkts), len(rx0), len(rx1)),
|
|
)
|
|
|
|
#
|
|
# use a test CLI command to stop the FIB walk process, this
|
|
# will prevent the FIB converging the VPN routes and thus allow
|
|
# us to probe the interim (post-fail, pre-converge) state
|
|
#
|
|
self.vapi.ppcli("test fib-walk-process disable")
|
|
|
|
#
|
|
# withdraw the connected prefix on the interface.
|
|
#
|
|
self.pg2.unconfig_ip4()
|
|
|
|
#
|
|
# now all packets should be forwarded through the remaining peer
|
|
#
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg3.get_capture(NUM_PKTS)
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0),
|
|
"Expected all (%s) packets across single path. "
|
|
"rx0: %s." % (len(pkts), len(rx0)),
|
|
)
|
|
|
|
#
|
|
# enable the FIB walk process to converge the FIB
|
|
#
|
|
self.vapi.ppcli("test fib-walk-process enable")
|
|
|
|
#
|
|
# packets should still be forwarded through the remaining peer
|
|
#
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg3.get_capture(NUM_PKTS)
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0),
|
|
"Expected all (%s) packets across single path. "
|
|
"rx0: %s." % (len(pkts), len(rx0)),
|
|
)
|
|
|
|
#
|
|
# put the connected routes back
|
|
#
|
|
self.pg2.config_ip4()
|
|
self.pg2.resolve_arp()
|
|
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg2._get_capture(NUM_PKTS)
|
|
rx1 = self.pg3._get_capture(NUM_PKTS)
|
|
self.assertNotEqual(0, len(rx0))
|
|
self.assertNotEqual(0, len(rx1))
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0) + len(rx1),
|
|
"Expected all (%s) packets across both ECMP paths. "
|
|
"rx0: %s rx1: %s." % (len(pkts), len(rx0), len(rx1)),
|
|
)
|
|
|
|
def test_mpls_v6_ebgp_pic(self):
|
|
"""MPLSv6 eBGP Prefix-Independent Convergence (PIC) edge convergence
|
|
|
|
1) setup many eBGP VPNv6 routes via a pair of eBGP peers
|
|
2) Check EMCP forwarding to these peers
|
|
3) withdraw one eBGP path - expect LB across remaining eBGP
|
|
"""
|
|
|
|
#
|
|
# Lot's of VPN routes. We need more the 64 so VPP will build
|
|
# the fast convergence indirection
|
|
#
|
|
vpn_routes = []
|
|
vpn_bindings = []
|
|
pkts = []
|
|
for ii in range(NUM_PKTS):
|
|
dst = "3000::%d" % ii
|
|
local_label = 1600 + ii
|
|
vpn_routes.append(
|
|
VppIpRoute(
|
|
self,
|
|
dst,
|
|
128,
|
|
[
|
|
VppRoutePath(
|
|
self.pg2.remote_ip6,
|
|
0xFFFFFFFF,
|
|
nh_table_id=1,
|
|
flags=FibPathFlags.FIB_PATH_FLAG_RESOLVE_VIA_ATTACHED,
|
|
),
|
|
VppRoutePath(
|
|
self.pg3.remote_ip6,
|
|
0xFFFFFFFF,
|
|
nh_table_id=1,
|
|
flags=FibPathFlags.FIB_PATH_FLAG_RESOLVE_VIA_ATTACHED,
|
|
),
|
|
],
|
|
table_id=1,
|
|
)
|
|
)
|
|
vpn_routes[ii].add_vpp_config()
|
|
|
|
vpn_bindings.append(
|
|
VppMplsIpBind(self, local_label, dst, 128, ip_table_id=1)
|
|
)
|
|
vpn_bindings[ii].add_vpp_config()
|
|
|
|
pkts.append(
|
|
Ether(dst=self.pg0.local_mac, src=self.pg0.remote_mac)
|
|
/ MPLS(label=local_label, ttl=64)
|
|
/ IPv6(src=self.pg0.remote_ip6, dst=dst)
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
self.logger.info(self.vapi.cli("sh ip6 fib %s" % dst))
|
|
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg2._get_capture(NUM_PKTS)
|
|
rx1 = self.pg3._get_capture(NUM_PKTS)
|
|
self.assertNotEqual(0, len(rx0))
|
|
self.assertNotEqual(0, len(rx1))
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0) + len(rx1),
|
|
"Expected all (%s) packets across both ECMP paths. "
|
|
"rx0: %s rx1: %s." % (len(pkts), len(rx0), len(rx1)),
|
|
)
|
|
|
|
#
|
|
# use a test CLI command to stop the FIB walk process, this
|
|
# will prevent the FIB converging the VPN routes and thus allow
|
|
# us to probe the interim (post-fail, pre-converge) state
|
|
#
|
|
self.vapi.ppcli("test fib-walk-process disable")
|
|
|
|
#
|
|
# withdraw the connected prefix on the interface.
|
|
# and shutdown the interface so the ND cache is flushed.
|
|
#
|
|
self.pg2.unconfig_ip6()
|
|
self.pg2.admin_down()
|
|
|
|
#
|
|
# now all packets should be forwarded through the remaining peer
|
|
#
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg3.get_capture(NUM_PKTS)
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0),
|
|
"Expected all (%s) packets across single path. "
|
|
"rx0: %s." % (len(pkts), len(rx0)),
|
|
)
|
|
|
|
#
|
|
# enable the FIB walk process to converge the FIB
|
|
#
|
|
self.vapi.ppcli("test fib-walk-process enable")
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg3.get_capture(NUM_PKTS)
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0),
|
|
"Expected all (%s) packets across single path. "
|
|
"rx0: %s." % (len(pkts), len(rx0)),
|
|
)
|
|
|
|
#
|
|
# put the connected routes back
|
|
#
|
|
self.logger.info(self.vapi.cli("sh log"))
|
|
self.pg2.admin_up()
|
|
self.pg2.config_ip6()
|
|
self.pg2.resolve_ndp()
|
|
|
|
self.pg0.add_stream(pkts)
|
|
self.pg_enable_capture(self.pg_interfaces)
|
|
self.pg_start()
|
|
|
|
rx0 = self.pg2._get_capture(NUM_PKTS)
|
|
rx1 = self.pg3._get_capture(NUM_PKTS)
|
|
self.assertNotEqual(0, len(rx0))
|
|
self.assertNotEqual(0, len(rx1))
|
|
self.assertEqual(
|
|
len(pkts),
|
|
len(rx0) + len(rx1),
|
|
"Expected all (%s) packets across both ECMP paths. "
|
|
"rx0: %s rx1: %s." % (len(pkts), len(rx0), len(rx1)),
|
|
)
|
|
|
|
|
|
class TestMPLSL2(VppTestCase):
|
|
"""MPLS-L2"""
|
|
|
|
@classmethod
|
|
def setUpClass(cls):
|
|
super(TestMPLSL2, cls).setUpClass()
|
|
|
|
@classmethod
|
|
def tearDownClass(cls):
|
|
super(TestMPLSL2, cls).tearDownClass()
|
|
|
|
def setUp(self):
|
|
super(TestMPLSL2, self).setUp()
|
|
|
|
# create 2 pg interfaces
|
|
self.create_pg_interfaces(range(2))
|
|
|
|
# create the default MPLS table
|
|
self.tables = []
|
|
tbl = VppMplsTable(self, 0)
|
|
tbl.add_vpp_config()
|
|
self.tables.append(tbl)
|
|
|
|
# use pg0 as the core facing interface, don't resolve ARP
|
|
self.pg0.admin_up()
|
|
self.pg0.config_ip4()
|
|
self.pg0.enable_mpls()
|
|
|
|
# use the other 2 for customer facing L2 links
|
|
for i in self.pg_interfaces[1:]:
|
|
i.admin_up()
|
|
|
|
def tearDown(self):
|
|
for i in self.pg_interfaces[1:]:
|
|
i.admin_down()
|
|
|
|
self.pg0.disable_mpls()
|
|
self.pg0.unconfig_ip4()
|
|
self.pg0.admin_down()
|
|
super(TestMPLSL2, self).tearDown()
|
|
|
|
def verify_capture_tunneled_ethernet(self, capture, sent, mpls_labels):
|
|
capture = verify_filter(capture, sent)
|
|
|
|
self.assertEqual(len(capture), len(sent))
|
|
|
|
for i in range(len(capture)):
|
|
tx = sent[i]
|
|
rx = capture[i]
|
|
|
|
# the MPLS TTL is 255 since it enters a new tunnel
|
|
verify_mpls_stack(self, rx, mpls_labels)
|
|
|
|
tx_eth = tx[Ether]
|
|
rx_eth = Ether(scapy.compat.raw(rx[MPLS].payload))
|
|
|
|
self.assertEqual(rx_eth.src, tx_eth.src)
|
|
self.assertEqual(rx_eth.dst, tx_eth.dst)
|
|
|
|
def verify_arp_req(self, rx, smac, sip, dip):
|
|
ether = rx[Ether]
|
|
self.assertEqual(ether.dst, "ff:ff:ff:ff:ff:ff")
|
|
self.assertEqual(ether.src, smac)
|
|
|
|
arp = rx[ARP]
|
|
self.assertEqual(arp.hwtype, 1)
|
|
self.assertEqual(arp.ptype, 0x800)
|
|
self.assertEqual(arp.hwlen, 6)
|
|
self.assertEqual(arp.plen, 4)
|
|
self.assertEqual(arp.op, ARP.who_has)
|
|
self.assertEqual(arp.hwsrc, smac)
|
|
self.assertEqual(arp.hwdst, "00:00:00:00:00:00")
|
|
self.assertEqual(arp.psrc, sip)
|
|
self.assertEqual(arp.pdst, dip)
|
|
|
|
def test_vpws(self):
|
|
"""Virtual Private Wire Service"""
|
|
|
|
#
|
|
# Create an MPLS tunnel that pushes 1 label
|
|
# For Ethernet over MPLS the uniform mode is irrelevant since ttl/cos
|
|
# information is not in the packet, but we test it works anyway
|
|
#
|
|
mpls_tun_1 = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4,
|
|
self.pg0.sw_if_index,
|
|
labels=[VppMplsLabel(42, MplsLspMode.UNIFORM)],
|
|
)
|
|
],
|
|
is_l2=1,
|
|
)
|
|
mpls_tun_1.add_vpp_config()
|
|
mpls_tun_1.admin_up()
|
|
|
|
#
|
|
# Create a label entry to for 55 that does L2 input to the tunnel
|
|
#
|
|
route_55_eos = VppMplsRoute(
|
|
self,
|
|
55,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0",
|
|
mpls_tun_1.sw_if_index,
|
|
type=FibPathType.FIB_PATH_TYPE_INTERFACE_RX,
|
|
proto=FibPathProto.FIB_PATH_NH_PROTO_ETHERNET,
|
|
)
|
|
],
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_ETHERNET,
|
|
)
|
|
route_55_eos.add_vpp_config()
|
|
|
|
#
|
|
# Cross-connect the tunnel with one of the customers L2 interfaces
|
|
#
|
|
self.vapi.sw_interface_set_l2_xconnect(
|
|
self.pg1.sw_if_index, mpls_tun_1.sw_if_index, enable=1
|
|
)
|
|
self.vapi.sw_interface_set_l2_xconnect(
|
|
mpls_tun_1.sw_if_index, self.pg1.sw_if_index, enable=1
|
|
)
|
|
|
|
#
|
|
# inject a packet from the core
|
|
#
|
|
pcore = (
|
|
Ether(dst=self.pg0.local_mac, src=self.pg0.remote_mac)
|
|
/ MPLS(label=55, ttl=64)
|
|
/ Ether(dst="00:00:de:ad:ba:be", src="00:00:de:ad:be:ef")
|
|
/ IP(src="10.10.10.10", dst="11.11.11.11")
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
|
|
tx0 = pcore * NUM_PKTS
|
|
rx0 = self.send_and_expect(self.pg0, tx0, self.pg1)
|
|
payload = pcore[MPLS].payload
|
|
|
|
self.assertEqual(rx0[0][Ether].dst, payload[Ether].dst)
|
|
self.assertEqual(rx0[0][Ether].src, payload[Ether].src)
|
|
|
|
#
|
|
# Inject a packet from the customer/L2 side
|
|
# there's no resolved ARP entry so the first packet we see should be
|
|
# an ARP request
|
|
#
|
|
tx1 = pcore[MPLS].payload
|
|
rx1 = self.send_and_expect(self.pg1, [tx1], self.pg0)
|
|
|
|
self.verify_arp_req(
|
|
rx1[0], self.pg0.local_mac, self.pg0.local_ip4, self.pg0.remote_ip4
|
|
)
|
|
|
|
#
|
|
# resolve the ARP entries and send again
|
|
#
|
|
self.pg0.resolve_arp()
|
|
tx1 = pcore[MPLS].payload * NUM_PKTS
|
|
rx1 = self.send_and_expect(self.pg1, tx1, self.pg0)
|
|
|
|
self.verify_capture_tunneled_ethernet(rx1, tx1, [VppMplsLabel(42)])
|
|
|
|
def test_vpls(self):
|
|
"""Virtual Private LAN Service"""
|
|
|
|
# we skipped this in the setup
|
|
self.pg0.resolve_arp()
|
|
|
|
#
|
|
# Create a L2 MPLS tunnels
|
|
#
|
|
mpls_tun1 = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(42)]
|
|
)
|
|
],
|
|
is_l2=1,
|
|
)
|
|
mpls_tun1.add_vpp_config()
|
|
mpls_tun1.admin_up()
|
|
|
|
mpls_tun2 = VppMPLSTunnelInterface(
|
|
self,
|
|
[
|
|
VppRoutePath(
|
|
self.pg0.remote_ip4, self.pg0.sw_if_index, labels=[VppMplsLabel(43)]
|
|
)
|
|
],
|
|
is_l2=1,
|
|
)
|
|
mpls_tun2.add_vpp_config()
|
|
mpls_tun2.admin_up()
|
|
|
|
#
|
|
# Create a label entries, 55 and 56, that do L2 input to the tunnel
|
|
# the latter includes a Psuedo Wire Control Word
|
|
#
|
|
route_55_eos = VppMplsRoute(
|
|
self,
|
|
55,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0",
|
|
mpls_tun1.sw_if_index,
|
|
type=FibPathType.FIB_PATH_TYPE_INTERFACE_RX,
|
|
proto=FibPathProto.FIB_PATH_NH_PROTO_ETHERNET,
|
|
)
|
|
],
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_ETHERNET,
|
|
)
|
|
|
|
route_56_eos = VppMplsRoute(
|
|
self,
|
|
56,
|
|
1,
|
|
[
|
|
VppRoutePath(
|
|
"0.0.0.0",
|
|
mpls_tun2.sw_if_index,
|
|
type=FibPathType.FIB_PATH_TYPE_INTERFACE_RX,
|
|
flags=FibPathFlags.FIB_PATH_FLAG_POP_PW_CW,
|
|
proto=FibPathProto.FIB_PATH_NH_PROTO_ETHERNET,
|
|
)
|
|
],
|
|
eos_proto=FibPathProto.FIB_PATH_NH_PROTO_ETHERNET,
|
|
)
|
|
|
|
# move me
|
|
route_56_eos.add_vpp_config()
|
|
route_55_eos.add_vpp_config()
|
|
|
|
self.logger.info(self.vapi.cli("sh mpls fib 56"))
|
|
|
|
#
|
|
# add to tunnel to the customers bridge-domain
|
|
#
|
|
self.vapi.sw_interface_set_l2_bridge(
|
|
rx_sw_if_index=mpls_tun1.sw_if_index, bd_id=1
|
|
)
|
|
self.vapi.sw_interface_set_l2_bridge(
|
|
rx_sw_if_index=mpls_tun2.sw_if_index, bd_id=1
|
|
)
|
|
self.vapi.sw_interface_set_l2_bridge(
|
|
rx_sw_if_index=self.pg1.sw_if_index, bd_id=1
|
|
)
|
|
|
|
#
|
|
# Packet from host on the customer interface to each host
|
|
# reachable over the core, and vice-versa
|
|
#
|
|
p_cust1 = (
|
|
Ether(dst="00:00:de:ad:ba:b1", src="00:00:de:ad:be:ef")
|
|
/ IP(src="10.10.10.10", dst="11.11.11.11")
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
p_cust2 = (
|
|
Ether(dst="00:00:de:ad:ba:b2", src="00:00:de:ad:be:ef")
|
|
/ IP(src="10.10.10.10", dst="11.11.11.12")
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
p_core1 = (
|
|
Ether(dst=self.pg0.local_mac, src=self.pg0.remote_mac)
|
|
/ MPLS(label=55, ttl=64)
|
|
/ Ether(src="00:00:de:ad:ba:b1", dst="00:00:de:ad:be:ef")
|
|
/ IP(dst="10.10.10.10", src="11.11.11.11")
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
p_core2 = (
|
|
Ether(dst=self.pg0.local_mac, src=self.pg0.remote_mac)
|
|
/ MPLS(label=56, ttl=64)
|
|
/ Raw(b"\x01" * 4)
|
|
/ Ether(src="00:00:de:ad:ba:b2", dst="00:00:de:ad:be:ef") # PW CW
|
|
/ IP(dst="10.10.10.10", src="11.11.11.12")
|
|
/ UDP(sport=1234, dport=1234)
|
|
/ Raw(b"\xa5" * 100)
|
|
)
|
|
|
|
#
|
|
# The BD is learning, so send in one of each packet to learn
|
|
#
|
|
|
|
# 2 packets due to BD flooding
|
|
rx = self.send_and_expect(self.pg1, p_cust1, self.pg0, n_rx=2)
|
|
rx = self.send_and_expect(self.pg1, p_cust2, self.pg0, n_rx=2)
|
|
|
|
# we've learnt this so expect it be be forwarded not flooded
|
|
rx = self.send_and_expect(self.pg0, [p_core1], self.pg1)
|
|
self.assertEqual(rx[0][Ether].dst, p_cust1[Ether].src)
|
|
self.assertEqual(rx[0][Ether].src, p_cust1[Ether].dst)
|
|
|
|
rx = self.send_and_expect(self.pg0, [p_core2], self.pg1)
|
|
self.assertEqual(rx[0][Ether].dst, p_cust2[Ether].src)
|
|
self.assertEqual(rx[0][Ether].src, p_cust2[Ether].dst)
|
|
|
|
#
|
|
# now a stream in each direction from each host
|
|
#
|
|
rx = self.send_and_expect(self.pg1, p_cust1 * NUM_PKTS, self.pg0)
|
|
self.verify_capture_tunneled_ethernet(
|
|
rx, p_cust1 * NUM_PKTS, [VppMplsLabel(42)]
|
|
)
|
|
|
|
rx = self.send_and_expect(self.pg1, p_cust2 * NUM_PKTS, self.pg0)
|
|
self.verify_capture_tunneled_ethernet(
|
|
rx, p_cust2 * NUM_PKTS, [VppMplsLabel(43)]
|
|
)
|
|
|
|
rx = self.send_and_expect(self.pg0, p_core1 * NUM_PKTS, self.pg1)
|
|
rx = self.send_and_expect(self.pg0, p_core2 * NUM_PKTS, self.pg1)
|
|
|
|
#
|
|
# remove interfaces from customers bridge-domain
|
|
#
|
|
self.vapi.sw_interface_set_l2_bridge(
|
|
rx_sw_if_index=mpls_tun1.sw_if_index, bd_id=1, enable=0
|
|
)
|
|
self.vapi.sw_interface_set_l2_bridge(
|
|
rx_sw_if_index=mpls_tun2.sw_if_index, bd_id=1, enable=0
|
|
)
|
|
self.vapi.sw_interface_set_l2_bridge(
|
|
rx_sw_if_index=self.pg1.sw_if_index, bd_id=1, enable=0
|
|
)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
unittest.main(testRunner=VppTestRunner)
|