:lastproofread: 2026-02-18 .. _vpp_troubleshooting: .. include:: /_include/need_improvement.txt ############################# VPP Dataplane Troubleshooting ############################# This page shows you how to collect diagnostic information to troubleshoot VPP dataplane issues. These techniques help you resolve problems yourself and provide support teams with the information they need. Collecting the right diagnostic data is crucial for effective troubleshooting. Packet Capture (PCAP) ===================== Packet capture is a valuable debugging tool for analyzing network traffic and identifying issues with packet processing, routing, and filtering. ``pcap trace`` in VPP captures packets at different states: received (rx), transmitted (tx), and dropped (drop). Starting Packet Capture ----------------------- **Command syntax:** .. stop_vyoslinter .. opcmd:: sudo vppctl pcap trace [rx] [tx] [drop] [max ] [intfc ] [file ] [max-bytes-per-pkt ] .. start_vyoslinter **Parameters:** - ``rx`` - Capture received packets - ``tx`` - Capture transmitted packets - ``drop`` - Capture dropped packets - ``max `` - Depth of the local buffer. After ``n`` packets arrive, the buffer flushes to file. When the next ``n`` packets arrive, the file overwrites with new data. (default: 100) - ``intfc `` - Specify an interface or use ``any`` for all interfaces (default: any) - ``file `` - Output filename. The PCAP file is stored in the ``/tmp/`` directory. - ``max-bytes-per-pkt `` - Maximum bytes to capture per packet (must be >= 32, <= 9000) **Examples:** .. code-block:: none # Start capturing tx packets with specific parameters sudo vppctl pcap trace tx max 35 intfc eth1 file vpp_eth1.pcap # Capture all packet types from any interface sudo vppctl pcap trace rx tx drop max 1000 intfc any file vpp_capture.pcap max-bytes-per-pkt 128 Monitoring Capture Status ------------------------- To check the capture status: .. opcmd:: sudo vppctl pcap trace status This command displays: - Whether capture is active - Capture parameters - Number of packets captured - Output file location Stopping Packet Capture ----------------------- .. warning:: VPP does not automatically stop packet captures. If left running, captures consume resources indefinitely. Always stop captures when you're done with them. To stop the active packet capture: .. opcmd:: sudo vppctl pcap trace off Example output when stopping: .. code-block:: none Write 35 packets to /tmp/vpp_eth1.pcap, and stop capture... **Notes:** - PCAP files are stored in the ``/tmp/`` directory. - Existing files are overwritten. - If you don't specify a filename, default names are used: ``/tmp/rx.pcap``, ``/tmp/tx.pcap``, and ``/tmp/rxandtx.pcap``. - Large captures consume significant disk space—monitor available space. - Stop captures promptly to avoid filling storage. Packet Tracing ============== VPP packet tracing shows how packets flow through the VPP processing graph, including which nodes process each packet and what transformations occur. .. warning:: Tracing generates large amounts of data, especially on high-traffic systems. Limit the number of traced packets to avoid overwhelming the system. Basic Packet Tracing Commands ----------------------------- Start tracing ^^^^^^^^^^^^^ To start tracing packets at a specific graph node: .. opcmd:: sudo vppctl trace add [verbose] - ```` - Graph node name where tracing starts (for example, ``dpdk-input``, ``ethernet-input``, or ``ip4-input``). - ```` - Number of packets to trace (for example, 100). - ``[verbose]`` - Optional flag to include detailed buffer information in the trace output. **Common node names for tracing:** - ``dpdk-input``: Packets received from DPDK interfaces - ``ethernet-input``: Ethernet frame processing - ``ip4-input``: IPv4 packet processing - ``ip6-input``: IPv6 packet processing - ``ip4-lookup``: IPv4 routing table lookup - ``ip6-lookup``: IPv6 routing table lookup View traces ^^^^^^^^^^^ After packets are traced, view the results: .. opcmd:: sudo vppctl show trace [max COUNT] - ``[max COUNT]`` - Optional limit on number of packets to display (default: all) Clear traces ^^^^^^^^^^^^ After reviewing traces, clear them to free up resources: .. opcmd:: sudo vppctl clear trace Example Workflow ^^^^^^^^^^^^^^^^ .. code-block:: none # Add traces for 100 packets on dpdk-input node sudo vppctl trace add dpdk-input 100 # Send some traffic, then view results sudo vppctl show trace # Clear traces for next test sudo vppctl clear trace Understanding Trace Output -------------------------- Trace output shows how packets flow through VPP processing nodes: .. code-block:: none Packet 1 01:00:09:508438: dpdk-input eth2 rx queue 0 buffer 0x8533: current data 0, length 98, buffer-pool 0, ref-count 1, trace handle 0x1000000 ext-hdr-valid PKT MBUF: port 1, nb_segs 1, pkt_len 98 buf_len 1828, data_len 98, ol_flags 0x0, data_off 128, phys_addr 0x78814d40 packet_type 0x0 l2_len 0 l3_len 0 outer_l2_len 0 outer_l3_len 0 rss 0x0 fdir.hi 0x0 fdir.lo 0x0 IP4: 0c:87:6c:4e:00:01 -> 0c:de:0d:e2:00:02 ICMP: 192.168.102.2 -> 192.168.99.3 tos 0x00, ttl 64, length 84, checksum 0xb88d dscp CS0 ecn NON_ECN fragment id 0x37c5, flags DONT_FRAGMENT ICMP echo_request checksum 0x64e id 3024 01:00:09:508449: ethernet-input frame: flags 0x1, hw-if-index 2, sw-if-index 2 IP4: 0c:87:6c:4e:00:01 -> 0c:de:0d:e2:00:02 01:00:09:508455: ip4-input ICMP: 192.168.102.2 -> 192.168.99.3 tos 0x00, ttl 64, length 84, checksum 0xb88d dscp CS0 ecn NON_ECN fragment id 0x37c5, flags DONT_FRAGMENT ICMP echo_request checksum 0x64e id 3024 01:00:09:508458: ip4-sv-reassembly-feature [not-fragmented] 01:00:09:508460: nat-pre-in2out in2out next_index 2 arc_next_index 10 01:00:09:508462: nat44-ed-in2out NAT44_IN2OUT_ED_FAST_PATH: sw_if_index 2, next index 10, session 0, translation result 'success' via i2of i2of match: saddr 192.168.102.2 sport 3024 daddr 192.168.99.3 dport 3024 proto ICMP fib_idx 0 rewrite: saddr 192.168.99.1 daddr 192.168.99.3 icmp-id 3024 txfib 0 o2if match: saddr 192.168.99.3 sport 3024 daddr 192.168.99.1 dport 3024 proto ICMP fib_idx 0 rewrite: saddr 192.168.99.3 daddr 192.168.102.2 icmp-id 3024 txfib 0 search key local 192.168.102.2:3024 remote 192.168.99.3:3024 proto ICMP fib 0 thread-index 0 session-index 0 01:00:09:508469: ip4-lookup fib 0 dpo-idx 10 flow hash: 0x00000000 ICMP: 192.168.99.1 -> 192.168.99.3 tos 0x00, ttl 64, length 84, checksum 0xbb8e dscp CS0 ecn NON_ECN fragment id 0x37c5, flags DONT_FRAGMENT ICMP echo_request checksum 0x64e id 3024 01:00:09:508472: ip4-rewrite tx_sw_if_index 1 dpo-idx 10 : ipv4 via 192.168.99.3 eth1: mtu:1500 next:5 flags:[] 0ccea70400010cde0de200010800 flow hash: 0x00000000 00000000: 0ccea70400010cde0de2000108004500005437c540003f01bc8ec0a86301c0a8 00000020: 63030800064e0bd00d9a52c2d26800000000f4490000000000001011 01:00:09:508474: eth1-output eth1 flags 0x0038000d IP4: 0c:de:0d:e2:00:01 -> 0c:ce:a7:04:00:01 ICMP: 192.168.99.1 -> 192.168.99.3 tos 0x00, ttl 63, length 84, checksum 0xbc8e dscp CS0 ecn NON_ECN fragment id 0x37c5, flags DONT_FRAGMENT ICMP echo_request checksum 0x64e id 3024 01:00:09:508477: eth1-tx eth1 tx queue 0 buffer 0x8533: current data 0, length 98, buffer-pool 0, ref-count 1, trace handle 0x1000000 ext-hdr-valid natted l2-hdr-offset 0 l3-hdr-offset 14 PKT MBUF: port 1, nb_segs 1, pkt_len 98 buf_len 1828, data_len 98, ol_flags 0x0, data_off 128, phys_addr 0x78814d40 packet_type 0x0 l2_len 0 l3_len 0 outer_l2_len 0 outer_l3_len 0 rss 0x0 fdir.hi 0x0 fdir.lo 0x0 IP4: 0c:de:0d:e2:00:01 -> 0c:ce:a7:04:00:01 ICMP: 192.168.99.1 -> 192.168.99.3 tos 0x00, ttl 63, length 84, checksum 0xbc8e dscp CS0 ecn NON_ECN fragment id 0x37c5, flags DONT_FRAGMENT ICMP echo_request checksum 0x64e id 3024 In this example, the trace shows: - The packet is received on ``eth2`` interface at the ``dpdk-input`` node. - It flows through ``ethernet-input`` and ``ip4-input`` nodes. - NAT translation occurs at the ``nat44-ed-in2out`` node, changing the source IP. - The packet is routed through ``ip4-lookup`` and ``ip4-rewrite`` nodes. - It transmits out of ``eth1`` interface at the ``eth1-tx`` node. Additional Diagnostic Information ================================= When reporting issues to support teams or performing advanced troubleshooting, collect additional diagnostic information. Before/After Traffic Analysis ----------------------------- Before you send traffic: .. code-block:: none sudo vppctl clear hardware-interfaces sudo vppctl clear interfaces sudo vppctl clear error sudo vppctl clear runtime After you send traffic: .. code-block:: none sudo vppctl show version verbose sudo vppctl show hardware-interfaces sudo vppctl show interface address sudo vppctl show interface sudo vppctl show runtime sudo vppctl show error Core System Information ----------------------- **Memory and buffer information:** .. code-block:: none sudo vppctl show memory api-segment stats-segment numa-heaps main-heap map verbose sudo vppctl show buffers sudo vppctl show physmem detail sudo vppctl show physmem map **Runtime and performance data:** .. code-block:: none sudo vppctl show cpu sudo vppctl show threads sudo vppctl show runtime sudo vppctl show node counters Protocol-Specific Information ----------------------------- **Layer 2 data (if configured):** .. code-block:: none sudo vppctl show l2fib sudo vppctl show bridge-domain **IPv4 data (if configured):** .. code-block:: none sudo vppctl show ip fib sudo vppctl show ip neighbors **IPv6 data (if configured):** .. code-block:: none sudo vppctl show ip6 fib sudo vppctl show ip6 neighbors **MPLS data (if configured):** .. code-block:: none sudo vppctl show mpls fib sudo vppctl show mpls tunnel Creating Support Packages ========================= Use the automated diagnostic collection script to gather comprehensive VPP troubleshooting information when contacting support or reporting issues. VPP Diagnostic Collection Script -------------------------------- Create the diagnostic collection script: .. code-block:: python #!/usr/bin/env python3 """VyOS VPP Diagnostic Collection Script""" import datetime import shutil import subprocess import tarfile from pathlib import Path def run_cmd(cmd, output_file, diag_dir): """Run command and save output to file.""" try: result = subprocess.run( cmd, shell=True, capture_output=True, text=True, timeout=30 ) content = f"Command: {cmd}\nExit code: {result.returncode}\nTimestamp: {datetime.datetime.now()}\n{'-' * 50}\n" if result.stdout: content += f"\nSTDOUT:\n{result.stdout}" if result.stderr: content += f"\nSTDERR:\n{result.stderr}" (diag_dir / output_file).write_text(content) except Exception as e: (diag_dir / output_file).write_text(f"Command: {cmd}\nERROR: {e}") def collect_diagnostics(): """Collect all VPP diagnostics and create archive.""" timestamp = datetime.datetime.now().strftime("%Y%m%d-%H%M%S") diag_dir = Path.home() / f"vpp-diagnostics-{timestamp}" # VPP commands to collect commands = [ ("sudo vppctl show version verbose cmdline", "vpp-version.txt"), ("sudo vppctl show hardware-interfaces", "hardware-interfaces.txt"), ("sudo vppctl show interface address", "interface-addresses.txt"), ("sudo vppctl show interface", "interfaces.txt"), ("sudo vppctl show errors", "errors.txt"), ("sudo vppctl show runtime", "runtime.txt"), ( "sudo vppctl show memory api-segment stats-segment numa-heaps main-heap map verbose", "memory.txt", ), ("sudo vppctl show buffers", "buffers.txt"), ("sudo vppctl show physmem detail", "physmem.txt"), ("sudo vppctl show physmem map", "physmem-map.txt"), ("sudo vppctl show cpu", "cpu.txt"), ("sudo vppctl show threads", "threads.txt"), ("sudo vppctl show node counters", "node-counters.txt"), ("sudo vppctl show l2fib", "l2fib.txt"), ("sudo vppctl show bridge-domain", "bridge-domains.txt"), ("sudo vppctl show ip fib", "ip4-fib.txt"), ("sudo vppctl show ip neighbors", "ip4-neighbors.txt"), ("sudo vppctl show ip6 fib", "ip6-fib.txt"), ("sudo vppctl show ip6 neighbors", "ip6-neighbors.txt"), ("sudo vppctl show mpls fib", "mpls-fib.txt"), ("sudo vppctl show mpls tunnel", "mpls-tunnels.txt"), ("sudo vppctl show trace", "packet-traces.txt"), ] try: # Create diagnostics directory diag_dir.mkdir(parents=True, exist_ok=True) # Collect VPP data for cmd, output_file in commands: run_cmd(cmd, output_file, diag_dir) # Collect PCAP files pcap_files = list(Path("/tmp").glob("*.pcap")) if pcap_files: pcap_dir = diag_dir / "pcap-files" pcap_dir.mkdir(exist_ok=True) for pcap_file in pcap_files: try: shutil.copy2(pcap_file, pcap_dir) except (PermissionError, OSError): pass # Create archive archive_name = f"vpp-diagnostics-{timestamp}.tar.gz" archive_path = Path.home() / archive_name with tarfile.open(archive_path, "w:gz") as tar: tar.add(diag_dir, arcname=diag_dir.name) # Cleanup shutil.rmtree(diag_dir) print(f"VPP diagnostics collected: {archive_path}") return archive_path except Exception as e: if diag_dir.exists(): shutil.rmtree(diag_dir) print(f"Collection failed: {e}") return None def main(): """Main function.""" collect_diagnostics() if __name__ == "__main__": main() Save this script as ``/config/scripts/vpp-collect-diagnostics`` Installation and Usage ---------------------- **1. Make the script executable** .. opcmd:: sudo chmod +x /config/scripts/vpp-collect-diagnostics **2. Run VPP diagnostic collection** The script automatically collects all diagnostics and stores them in your home directory. .. opcmd:: /config/scripts/vpp-collect-diagnostics **3. Generate VyOS tech-support archive separately** You can also generate a tech-support archive with system-wide diagnostics: .. opcmd:: generate tech-support archive What the Script Collects ------------------------ - **System information**: Version details, build information, command-line parameters. - **Interface data**: Hardware interfaces, interface addresses, statistics, and configurations. - **Performance metrics**: Runtime statistics, error counters, node counters, CPU, and thread information. - **Memory analysis**: Memory usage (API segment, stats segment, NUMA heaps, main heap), buffers, and physical memory. - **Layer 2 data**: L2 forwarding table (L2FIB) and bridge domain configurations. - **IPv4 data**: IPv4 forwarding table (FIB) and IPv4 neighbor table. - **IPv6 data**: IPv6 forwarding table (FIB) and IPv6 neighbor table. - **MPLS data**: MPLS forwarding table (FIB) and MPLS tunnel information. - **Packet traces**: Captured packet traces (if available). - **Packet captures**: PCAP files from the ``/tmp`` directory (if available).