BackgroundThe client company is a retail company with about 200 employees. The company's network is a classic full-gigabit three-layer network architecture, with an export iKuai soft router + an intranet switch. The general topology is shown in the figure below: Typical topology The network segment planning is as follows:
The current problem is: the company's IT often finds that the CPU of the iKuai export router often soars to more than 90%, accessing pages is particularly slow, and at the same time, the terminals below have problems such as lag and delay when accessing the Internet, and the entire network has collapsed. Existing analysis
The above tests show that there should be no major problems with routers, switches and other devices. When encountering this kind of high throughput statistics problem, it is speculated that there is a loop in the network, causing a broadcast storm to consume all the bandwidth and destroy the address table? Troubleshooting AnalysisStep 1: Confirm the problem We use the VLAN30 office area PC to ping Baidu IP: You can see that the packet loss is quite serious. Step 2: Confirm the loop problem in the network The most common loops we encounter are Layer 2 loops, which have two main manifestations:
As shown in the topology above, because only the WAN and LAN ports of the router are in use, from the previous traffic statistics analysis, "there is no large traffic on the WAN port, but the traffic on the LAN port is abnormally large". We only need to check the data statistics on the uplink interface of the core switch to confirm whether this large traffic is broadcast & multicast! The command is as follows: From the above figure, we can see that when the problem reappears, the multicast and broadcast packets on the core switch uplink GE0/0/1 interface increase very slowly within 5 seconds, which basically eliminates the possible existence of Layer 2 loops and packet flooding. But please note: if you look closely at the statistics of the core switch uplink interface, 35,000 unicast packets are sent and received per second within 5 seconds, which is equivalent to 7,000 bidirectional packets sent and received per second. This amount is amazing! How can unicast packets have such a large throughput when the entire network is paralyzed? The next step is to capture the data flow analysis. Step 3: Intranet trunk data flow analysis We mirror the flow of the upstream egress routing interface on the core switch: The captured packets are as follows: The analysis from this flow is as follows: This flow is a UDP flow from the source VLAN20 monitoring network to the destination 172.16.40.0/24. According to statistics, the throughput of these flows is as high as 1Gbps, which is the main reason for the full Gigabit link bandwidth and the main reason for the router to be fully loaded with sending and receiving packets, resulting in a continuous high CPU. So why do these "abnormal flows" appear between the core switch and the router and flood? Let's take a look at the MAC address of the data packet and compare it: Let’s look at the first one in chronological order: Look at the second one in chronological order: The analysis shows that this UDP packet circulates between the switch and the router, that is, SW—>R—>SW—>R... It keeps circulating until TTL=0 and then the UDP packet is discarded! Conclusion: There is a routing loop between the router and the switch. The destination network segment accessed by the monitoring network is 172.16.40.0/24. But strangely, this network segment does not exist in the intranet. Even if the monitoring device tries to access this network segment, it should be sent out from the WAN port of the router. Why does it bounce back? It must be related to the configuration! Step 4: Check the routing tables of core switches and routers The core switch routing table is as follows: The core switch configuration is standard and there is no problem. Let’s take a look at the iKuai routing table: For the sake of convenience, the IT staff configured the return route on iKuai as 172.16.0.0/16 (including all network segments of the intranet) with the next hop as the core. This is a disaster! Once the intranet accesses 172.16.1.X, 172.16.2.X, and 172.16.200.X, the trunk route will be looped and the link bandwidth will be exploded! Summary and solutionsThe summary is as follows
Solution The return route must include details, do not use large network segments! As follows: |
>>: One minute to show you: the complete process from entering a URL to displaying the page
In 2019, the global market for SD-WAN grew by 70%...
Power over Ethernet (PoE) is a revolutionary tech...
[[442879]] Hello everyone, I am Xuanyuan. That da...
How far has 5G construction progressed? [[424068]...
The most eye-catching thing in the communications...
The term "remote" itself has a connotat...
[51CTO.com original article] With the development...
At present, the network security environment is d...
2020 is a critical year for 5G construction, and ...
[Original article from 51CTO.com] Hello, my frien...
[51CTO.com original article] In what direction wi...
I guess everyone here is already familiar with th...
With the rise of IoT applications and the increas...
For most enterprises, IT teams will deploy fifth-...
5G technology is the most anticipated network upd...