vpp/docs/usecases/contiv/VPP_PACKET_TRACING_K8S.rst
Nathan Skrzypczak 9ad39c026c docs: better docs, mv doxygen to sphinx
This patch refactors the VPP sphinx docs
in order to make it easier to consume
for external readers as well as VPP developers.

It also makes sphinx the single source
of documentation, which simplifies maintenance
and operation.

Most important updates are:

- reformat the existing documentation as rst
- split RELEASE.md and move it into separate rst files
- remove section 'events'
- remove section 'archive'
- remove section 'related projects'
- remove section 'feature by release'
- remove section 'Various links'
- make (Configuration reference, CLI docs,
  developer docs) top level items in the list
- move 'Use Cases' as part of 'About VPP'
- move 'Troubleshooting' as part of 'Getting Started'
- move test framework docs into 'Developer Documentation'
- add a 'Contributing' section for gerrit,
  docs and other contributer related infos
- deprecate doxygen and test-docs targets
- redirect the "make doxygen" target to "make docs"

Type: refactor

Change-Id: I552a5645d5b7964d547f99b1336e2ac24e7c209f
Signed-off-by: Nathan Skrzypczak <nathan.skrzypczak@gmail.com>
Signed-off-by: Andrew Yourtchenko <ayourtch@gmail.com>
2021-10-13 23:22:32 +00:00

536 lines
28 KiB
ReStructuredText
Raw Permalink Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

How to do VPP Packet Tracing in Kubernetes
==========================================
This document describes the steps to do *manual* packet tracing
(capture) using VPP in Kubernetes. Contiv/VPP also ships with a simple
bash script
`vpptrace.sh <https://github.com/contiv/vpp/blob/master/scripts/vpptrace.sh>`__,
which allows to *continuously* trace and *filter* packets incoming
through a given set of interface types. Documentation for vpptrace.sh is
available
`here <https://github.com/contiv/vpp/blob/master/docs/VPPTRACE.md>`__.
More information about VPP packet tracing is in:
- https://wiki.fd.io/view/VPP/Command-line_Interface_(CLI)_Guide#packet_tracer
- https://wiki.fd.io/view/VPP/How_To_Use_The_Packet_Generator_and_Packet_Tracer
- https://wiki.fd.io/view/VPP/Tutorial_Routing_and_Switching
SSH into the Node
-----------------
Perform the following commands to SSH into the node:
::
cd vpp/vagrant/vagrant-scripts/
vagrant ssh k8s-worker1
Check the VPP Graph Nodes (Input and Output Queues)
---------------------------------------------------
The following content shows what is running on VPP, via the ``show run``
command
::
vagrant@k8s-worker1:~$ sudo vppctl
_______ _ _ _____ ___
__/ __/ _ \ (_)__ | | / / _ \/ _ \
_/ _// // / / / _ \ | |/ / ___/ ___/
/_/ /____(_)_/\___/ |___/_/ /_/
vpp# show run
Time 1026791.9, average vectors/node 1.12, last 128 main loops 0.00 per node 0.00
vector rates in 1.6459e-4, out 1.5485e-4, drop 1.3635e-5, punt 0.0000e0
Name State Calls Vectors Suspends Clocks Vectors/Call
GigabitEthernet0/8/0-output active 56 69 0 1.34e3 1.23
GigabitEthernet0/8/0-tx active 54 67 0 8.09e5 1.24
acl-plugin-fa-cleaner-process event wait 0 0 1 2.84e4 0.00
admin-up-down-process event wait 0 0 1 4.59e3 0.00
api-rx-from-ring any wait 0 0 3316292 1.24e5 0.00
arp-input active 3 3 0 2.53e5 1.00
bfd-process event wait 0 0 1 5.94e3 0.00
cdp-process any wait 0 0 145916 1.36e4 0.00
dhcp-client-process any wait 0 0 10268 3.65e4 0.00
dns-resolver-process any wait 0 0 1027 5.86e4 0.00
dpdk-input polling 8211032318951 93 0 1.48e13 0.00
dpdk-ipsec-process done 1 0 0 2.10e5 0.00
dpdk-process any wait 0 0 342233 9.86e6 0.00
error-drop active 12 14 0 6.67e3 1.17
ethernet-input active 60 74 0 5.81e3 1.23
fib-walk any wait 0 0 513322 1.59e4 0.00
flow-report-process any wait 0 0 1 1.45e3 0.00
flowprobe-timer-process any wait 0 0 1 6.34e3 0.00
ikev2-manager-process any wait 0 0 1026484 1.18e4 0.00
interface-output active 2 2 0 3.23e3 1.00
ioam-export-process any wait 0 0 1 1.98e3 0.00
ip-route-resolver-process any wait 0 0 10268 3.02e4 0.00
ip4-arp active 1 1 0 1.49e4 1.00
ip4-input active 223 248 0 3.39e3 1.11
ip4-load-balance active 106 132 0 5.34e3 1.25
ip4-local active 86 92 0 2.46e3 1.07
ip4-local-end-of-arc active 86 92 0 1.00e3 1.07
ip4-lookup active 223 248 0 3.31e3 1.11
ip4-rewrite active 190 222 0 1.92e3 1.17
ip4-udp-lookup active 86 92 0 3.76e3 1.07
ip6-drop active 6 7 0 2.29e3 1.17
ip6-icmp-neighbor-discovery-ev any wait 0 0 1026484 1.13e4 0.00
ip6-input active 6 7 0 3.33e3 1.17
l2-flood active 2 2 0 4.42e3 1.00
l2-fwd active 138 157 0 2.13e3 1.14
l2-input active 140 159 0 2.41e3 1.14
l2-learn active 86 92 0 3.64e4 1.07
l2-output active 54 67 0 3.05e3 1.24
l2fib-mac-age-scanner-process event wait 0 0 85 5.01e4 0.00
lisp-retry-service any wait 0 0 513322 1.62e4 0.00
lldp-process event wait 0 0 1 5.02e4 0.00
loop0-output active 54 67 0 1.66e3 1.24
loop0-tx active 54 0 0 2.49e3 0.00
memif-process event wait 0 0 1 1.70e4 0.00
nat-det-expire-walk done 1 0 0 3.79e3 0.00
nat44-classify active 171 183 0 2.49e3 1.07
nat44-hairpinning active 86 92 0 1.80e3 1.07
nat44-in2out active 171 183 0 4.45e3 1.07
nat44-in2out-slowpath active 171 183 0 3.98e3 1.07
nat44-out2in active 52 65 0 1.28e4 1.25
nat64-expire-walk any wait 0 0 102677 5.95e4 0.00
nat64-expire-worker-walk interrupt wa 102676 0 0 7.39e3 0.00
send-garp-na-process event wait 0 0 1 1.28e3 0.00
startup-config-process done 1 0 1 4.19e3 0.00
tapcli-0-output active 1 1 0 6.97e3 1.00
tapcli-0-tx active 1 1 0 7.32e4 1.00
tapcli-1-output active 57 63 0 1.66e3 1.11
tapcli-1-tx active 57 63 0 1.35e5 1.11
tapcli-2-output active 28 28 0 3.26e3 1.00
tapcli-2-tx active 28 28 0 4.06e5 1.00
tapcli-rx interrupt wa 62 76 0 6.58e4 1.23
udp-ping-process any wait 0 0 1 1.79e4 0.00
unix-cli-127.0.0.1:43282 active 2 0 455 1.26e15 0.00
unix-epoll-input polling 8010763239 0 0 8.17e2 0.00
vhost-user-process any wait 0 0 1 1.96e3 0.00
vhost-user-send-interrupt-proc any wait 0 0 1 3.85e3 0.00
vpe-link-state-process event wait 0 0 8 9.79e4 0.00
vpe-oam-process any wait 0 0 503263 1.21e4 0.00
vxlan-gpe-ioam-export-process any wait 0 0 1 2.91e3 0.00
vxlan4-encap active 54 67 0 3.55e3 1.24
vxlan4-input active 86 92 0 3.79e3 1.07
wildcard-ip4-arp-publisher-pro event wait 0 0 1 6.44e3 0.00
``tapcli-rx`` above is the node-level input queue for incoming packets
into all the pods on the node. There is one ``tapcli-rx`` input queue
for every node.
The following are the input and output queues for each pod and the node:
::
tapcli-0-output
tapcli-0-tx
tapcli-1-output
tapcli-1-tx
tapcli-2-output
tapcli-2-tx
Each pod and node has two queues, one for rx (``tapcli-X-output``), and
one for tx (``tapcli-X-tx``). The above output is with two ``nginx``
pods in kubernetes.
Clear Existing VPP Packet Trace
-------------------------------
Enter the following command:
::
vpp# clear trace
How to Turn on VPP Packet Tracing
---------------------------------
Enter the following commands:
::
vpp# trace add <input or output queue name> <number of packets to capture>
vpp# trace add dpdk-input 1000
vpp# trace add tapcli-rx 1000
Send Traffic to the Pods
------------------------
Open another terminal, SSH into the master node, refer the documentation
in ``vpp/vagrant/README.md`` and send traffic to the two ``nginx`` pods
using ``wget``.
::
cd vpp/vagrant/vagrant-scripts/
vagrant ssh k8s-master
vagrant@k8s-master:~$ kubectl get pods -o wide
NAME READY STATUS RESTARTS AGE IP NODE
nginx-8586cf59-768qw 1/1 Running 0 11d 10.1.2.3 k8s-worker1
nginx-8586cf59-d27h2 1/1 Running 0 11d 10.1.2.2 k8s-worker1
vagrant@k8s-master:~$ wget 10.1.2.2
--2018-02-08 16:46:01-- http://10.1.2.2/
Connecting to 10.1.2.2:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 612 [text/html]
Saving to: index.html
index.html 100%[=========================================================>] 612 --.-KB/s in 0.004s
2018-02-08 16:46:01 (162 KB/s) - index.html saved [612/612]
vagrant@k8s-master:~$ wget 10.1.2.3
--2018-02-08 16:46:02-- http://10.1.2.3/
Connecting to 10.1.2.3:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 612 [text/html]
Saving to: index.html.1
index.html.1 100%[=========================================================>] 612 --.-KB/s in 0.004s
2018-02-08 16:46:02 (143 KB/s) - index.html.1 saved [612/612]
Check the Packets Captured by VPP
---------------------------------
Back in the first terminal, check the packets captured by VPP.
::
vpp# show trace
...
...
Packet 33
21:34:51:476110: tapcli-rx
tapcli-2
21:34:51:476115: ethernet-input
IP4: 00:00:00:00:00:02 -> 02:fe:72:95:66:c7
21:34:51:476117: ip4-input
TCP: 10.1.2.3 -> 172.30.1.2
tos 0x00, ttl 64, length 52, checksum 0x6fb4
fragment id 0x11ec, flags DONT_FRAGMENT
TCP: 80 -> 58430
seq. 0x5db741c8 ack 0x709defa7
flags 0x11 FIN ACK, tcp header: 32 bytes
window 235, checksum 0x55c3
21:34:51:476118: nat44-out2in
NAT44_OUT2IN: sw_if_index 6, next index 1, session index -1
21:34:51:476120: ip4-lookup
fib 0 dpo-idx 23 flow hash: 0x00000000
TCP: 10.1.2.3 -> 172.30.1.2
tos 0x00, ttl 64, length 52, checksum 0x6fb4
fragment id 0x11ec, flags DONT_FRAGMENT
TCP: 80 -> 58430
seq. 0x5db741c8 ack 0x709defa7
flags 0x11 FIN ACK, tcp header: 32 bytes
window 235, checksum 0x55c3
21:34:51:476121: ip4-load-balance
fib 0 dpo-idx 23 flow hash: 0x00000000
TCP: 10.1.2.3 -> 172.30.1.2
tos 0x00, ttl 64, length 52, checksum 0x6fb4
fragment id 0x11ec, flags DONT_FRAGMENT
TCP: 80 -> 58430
seq. 0x5db741c8 ack 0x709defa7
flags 0x11 FIN ACK, tcp header: 32 bytes
window 235, checksum 0x55c3
21:34:51:476122: ip4-rewrite
tx_sw_if_index 3 dpo-idx 5 : ipv4 via 192.168.30.1 loop0: 1a2b3c4d5e011a2b3c4d5e020800 flow hash: 0x00000000
00000000: 1a2b3c4d5e011a2b3c4d5e0208004500003411ec40003f0670b40a010203ac1e
00000020: 01020050e43e5db741c8709defa7801100eb55c300000101080a0f4b
21:34:51:476123: loop0-output
loop0
IP4: 1a:2b:3c:4d:5e:02 -> 1a:2b:3c:4d:5e:01
TCP: 10.1.2.3 -> 172.30.1.2
tos 0x00, ttl 63, length 52, checksum 0x70b4
fragment id 0x11ec, flags DONT_FRAGMENT
TCP: 80 -> 58430
seq. 0x5db741c8 ack 0x709defa7
flags 0x11 FIN ACK, tcp header: 32 bytes
window 235, checksum 0x55c3
21:34:51:476124: l2-input
l2-input: sw_if_index 3 dst 1a:2b:3c:4d:5e:01 src 1a:2b:3c:4d:5e:02
21:34:51:476125: l2-fwd
l2-fwd: sw_if_index 3 dst 1a:2b:3c:4d:5e:01 src 1a:2b:3c:4d:5e:02 bd_index 1
21:34:51:476125: l2-output
l2-output: sw_if_index 4 dst 1a:2b:3c:4d:5e:01 src 1a:2b:3c:4d:5e:02 data 08 00 45 00 00 34 11 ec 40 00 3f 06
21:34:51:476126: vxlan4-encap
VXLAN encap to vxlan_tunnel0 vni 10
21:34:51:476126: ip4-load-balance
fib 4 dpo-idx 22 flow hash: 0x00000103
UDP: 192.168.16.2 -> 192.168.16.1
tos 0x00, ttl 254, length 102, checksum 0x1b33
fragment id 0x0000
UDP: 24320 -> 4789
length 82, checksum 0x0000
21:34:51:476127: ip4-rewrite
tx_sw_if_index 1 dpo-idx 4 : ipv4 via 192.168.16.1 GigabitEthernet0/8/0: 080027b2610908002733fb6f0800 flow hash: 0x00000103
00000000: 080027b2610908002733fb6f08004500006600000000fd111c33c0a81002c0a8
00000020: 10015f0012b5005200000800000000000a001a2b3c4d5e011a2b3c4d
21:34:51:476127: GigabitEthernet0/8/0-output
GigabitEthernet0/8/0
IP4: 08:00:27:33:fb:6f -> 08:00:27:b2:61:09
UDP: 192.168.16.2 -> 192.168.16.1
tos 0x00, ttl 253, length 102, checksum 0x1c33
fragment id 0x0000
UDP: 24320 -> 4789
length 82, checksum 0x0000
21:34:51:476128: GigabitEthernet0/8/0-tx
GigabitEthernet0/8/0 tx queue 0
buffer 0xfa7f: current data -50, length 116, free-list 0, clone-count 0, totlen-nifb 0, trace 0x20
l2-hdr-offset 0 l3-hdr-offset 14
PKT MBUF: port 255, nb_segs 1, pkt_len 116
buf_len 2176, data_len 116, ol_flags 0x0, data_off 78, phys_addr 0x569ea040
packet_type 0x0 l2_len 0 l3_len 0 outer_l2_len 0 outer_l3_len 0
IP4: 08:00:27:33:fb:6f -> 08:00:27:b2:61:09
UDP: 192.168.16.2 -> 192.168.16.1
tos 0x00, ttl 253, length 102, checksum 0x1c33
fragment id 0x0000
UDP: 24320 -> 4789
length 82, checksum 0x0000
In the above captured packet, we can see:
- Input queue name ``tapcli-rx``
- Pods IP address ``10.1.2.3``
- IP address of the master node ``172.30.1.2``, which sent the ``wget``
traffic to the two pods
- HTTP port ``80``, destination port and TCP protocol
(``TCP: 80 -> 58430``)
- NAT queue name ``nat44-out2in``
- VXLAN VNI ID ``VXLAN encap to vxlan_tunnel0 vni 10``
- VXLAN UDP port ``4789``
- IP address of ``GigabitEthernet0/8/0`` interface (``192.168.16.2``)
- Packet on the outgoing queue ``GigabitEthernet0/8/0-tx``
Find IP Addresses of GigabitEthernet and the Tap Interfaces
-----------------------------------------------------------
Enter the following commands to find the IP addresses and Tap
interfaces:
::
vpp# show int address
GigabitEthernet0/8/0 (up):
L3 192.168.16.2/24
local0 (dn):
loop0 (up):
L2 bridge bd-id 1 idx 1 shg 0 bvi
L3 192.168.30.2/24
tapcli-0 (up):
L3 172.30.2.1/24
tapcli-1 (up):
L3 10.2.1.2/32
tapcli-2 (up):
L3 10.2.1.3/32
vxlan_tunnel0 (up):
L2 bridge bd-id 1 idx 1 shg 0
Other Useful VPP CLIs
---------------------
Enter the following commands to see additional information about VPP:
::
vpp# show int
Name Idx State Counter Count
GigabitEthernet0/8/0 1 up rx packets 138
rx bytes 18681
tx packets 100
tx bytes 29658
drops 1
ip4 137
tx-error 2
local0 0 down drops 1
loop0 3 up rx packets 137
rx bytes 9853
tx packets 200
tx bytes 49380
drops 1
ip4 136
tapcli-0 2 up rx packets 8
rx bytes 600
tx packets 1
tx bytes 42
drops 9
ip6 7
tapcli-1 5 up rx packets 56
rx bytes 13746
tx packets 78
tx bytes 6733
drops 1
ip4 56
tapcli-2 6 up rx packets 42
rx bytes 10860
tx packets 58
tx bytes 4996
drops 1
ip4 42
vxlan_tunnel0 4 up rx packets 137
rx bytes 11771
tx packets 100
tx bytes 28290
vpp# show hardware
Name Idx Link Hardware
GigabitEthernet0/8/0 1 up GigabitEthernet0/8/0
Ethernet address 08:00:27:33:fb:6f
Intel 82540EM (e1000)
carrier up full duplex speed 1000 mtu 9216
rx queues 1, rx desc 1024, tx queues 1, tx desc 1024
cpu socket 0
tx frames ok 100
tx bytes ok 29658
rx frames ok 138
rx bytes ok 19233
extended stats:
rx good packets 138
tx good packets 100
rx good bytes 19233
tx good bytes 29658
local0 0 down local0
local
loop0 3 up loop0
Ethernet address 1a:2b:3c:4d:5e:02
tapcli-0 2 up tapcli-0
Ethernet address 02:fe:95:07:df:9c
tapcli-1 5 up tapcli-1
Ethernet address 02:fe:3f:5f:0f:9a
tapcli-2 6 up tapcli-2
Ethernet address 02:fe:72:95:66:c7
vxlan_tunnel0 4 up vxlan_tunnel0
VXLAN
vpp# show bridge-domain
BD-ID Index BSN Age(min) Learning U-Forwrd UU-Flood Flooding ARP-Term BVI-Intf
1 1 1 off on on on on off loop0
vpp# show bridge-domain 1 detail
BD-ID Index BSN Age(min) Learning U-Forwrd UU-Flood Flooding ARP-Term BVI-Intf
1 1 1 off on on on on off loop0
Interface If-idx ISN SHG BVI TxFlood VLAN-Tag-Rewrite
loop0 3 3 0 * * none
vxlan_tunnel0 4 1 0 - * none
vpp# show l2fib verbose
Mac-Address BD-Idx If-Idx BSN-ISN Age(min) static filter bvi Interface-Name
1a:2b:3c:4d:5e:02 1 3 0/0 - * - * loop0
1a:2b:3c:4d:5e:01 1 4 1/1 - - - - vxlan_tunnel0
L2FIB total/learned entries: 2/1 Last scan time: 0.0000e0sec Learn limit: 4194304
vpp# show ip fib
ipv4-VRF:0, fib_index:0, flow hash:[src dst sport dport proto ] locks:[src:(nil):2, src:adjacency:3, src:default-route:1, ]
0.0.0.0/0
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:1 buckets:1 uRPF:21 to:[0:0]]
[0] [@5]: ipv4 via 172.30.2.2 tapcli-0: def35b93961902fe9507df9c0800
0.0.0.0/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:2 buckets:1 uRPF:1 to:[0:0]]
[0] [@0]: dpo-drop ip4
10.1.1.0/24
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:24 buckets:1 uRPF:29 to:[0:0]]
[0] [@10]: dpo-load-balance: [proto:ip4 index:23 buckets:1 uRPF:28 to:[0:0] via:[98:23234]]
[0] [@5]: ipv4 via 192.168.30.1 loop0: 1a2b3c4d5e011a2b3c4d5e020800
10.1.2.2/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:27 buckets:1 uRPF:12 to:[78:5641]]
[0] [@5]: ipv4 via 10.1.2.2 tapcli-1: 00000000000202fe3f5f0f9a0800
10.1.2.3/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:29 buckets:1 uRPF:32 to:[58:4184]]
[0] [@5]: ipv4 via 10.1.2.3 tapcli-2: 00000000000202fe729566c70800
10.2.1.2/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:26 buckets:1 uRPF:31 to:[0:0]]
[0] [@2]: dpo-receive: 10.2.1.2 on tapcli-1
10.2.1.3/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:28 buckets:1 uRPF:33 to:[0:0]]
[0] [@2]: dpo-receive: 10.2.1.3 on tapcli-2
172.30.1.0/24
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:25 buckets:1 uRPF:29 to:[98:23234]]
[0] [@10]: dpo-load-balance: [proto:ip4 index:23 buckets:1 uRPF:28 to:[0:0] via:[98:23234]]
[0] [@5]: ipv4 via 192.168.30.1 loop0: 1a2b3c4d5e011a2b3c4d5e020800
172.30.2.0/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:14 buckets:1 uRPF:15 to:[0:0]]
[0] [@0]: dpo-drop ip4
172.30.2.0/24
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:13 buckets:1 uRPF:14 to:[0:0]]
[0] [@4]: ipv4-glean: tapcli-0
172.30.2.1/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:16 buckets:1 uRPF:19 to:[0:0]]
[0] [@2]: dpo-receive: 172.30.2.1 on tapcli-0
172.30.2.2/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:17 buckets:1 uRPF:18 to:[0:0]]
[0] [@5]: ipv4 via 172.30.2.2 tapcli-0: def35b93961902fe9507df9c0800
172.30.2.255/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:15 buckets:1 uRPF:17 to:[0:0]]
[0] [@0]: dpo-drop ip4
192.168.16.0/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:10 buckets:1 uRPF:9 to:[0:0]]
[0] [@0]: dpo-drop ip4
192.168.16.1/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:22 buckets:1 uRPF:34 to:[0:0] via:[100:28290]]
[0] [@5]: ipv4 via 192.168.16.1 GigabitEthernet0/8/0: 080027b2610908002733fb6f0800
192.168.16.0/24
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:9 buckets:1 uRPF:30 to:[0:0]]
[0] [@4]: ipv4-glean: GigabitEthernet0/8/0
192.168.16.2/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:12 buckets:1 uRPF:13 to:[137:16703]]
[0] [@2]: dpo-receive: 192.168.16.2 on GigabitEthernet0/8/0
192.168.16.255/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:11 buckets:1 uRPF:11 to:[0:0]]
[0] [@0]: dpo-drop ip4
192.168.30.0/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:19 buckets:1 uRPF:23 to:[0:0]]
[0] [@0]: dpo-drop ip4
192.168.30.1/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:23 buckets:1 uRPF:28 to:[0:0] via:[98:23234]]
[0] [@5]: ipv4 via 192.168.30.1 loop0: 1a2b3c4d5e011a2b3c4d5e020800
192.168.30.0/24
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:18 buckets:1 uRPF:22 to:[0:0]]
[0] [@4]: ipv4-glean: loop0
192.168.30.2/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:21 buckets:1 uRPF:27 to:[0:0]]
[0] [@2]: dpo-receive: 192.168.30.2 on loop0
192.168.30.255/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:20 buckets:1 uRPF:25 to:[0:0]]
[0] [@0]: dpo-drop ip4
224.0.0.0/4
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:4 buckets:1 uRPF:3 to:[0:0]]
[0] [@0]: dpo-drop ip4
240.0.0.0/4
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:3 buckets:1 uRPF:2 to:[0:0]]
[0] [@0]: dpo-drop ip4
255.255.255.255/32
unicast-ip4-chain
[@0]: dpo-load-balance: [proto:ip4 index:5 buckets:1 uRPF:4 to:[0:0]]
[0] [@0]: dpo-drop ip4