Let’s talk about HTTP pipelining

Let’s talk about HTTP pipelining

HTTP pipelining is a technique for sending multiple HTTP requests in batches without waiting for the server's response.

Request result pipelining makes the loading time of HTML web pages increase dramatically, especially in specific high-latency connection environments, such as satellite Internet access. In broadband connections, the acceleration is not so significant, because the server side needs to apply the HTTP/1.1 protocol, and the server side must reply to requests in the order of the client's requests, so that the entire connection is still first-in-first-out, and head-of-line blocking (HOL blocking) may occur, causing delays. Future asynchronous operations in HTTP/2.0 or SPDY will solve this problem. Because it may fill multiple HTTP requests in one TCP packet, HTTP pipelining requires fewer TCP packets to be transmitted on the network, reducing the network load.

The pipelining mechanism must be completed through a persistent connection, and only requests such as GET and HEAD can be pipelined. Non-idempotent methods such as POST will not be pipelined. Consecutive GET and HEAD requests can always be pipelined. Whether a series of idempotent requests, such as GET, HEAD, PUT, DELETE, can be pipelined depends on whether the series of requests depends on others. In addition, the pipelining mechanism should not be activated when the connection is first established, because the other party (server) may not support the HTTP/1.1 version of the protocol.

HTTP pipelining relies on support from both the client and the server. Servers that comply with HTTP/1.1 support pipelining. This does not mean that the server is required to provide pipelined responses, but only that it should not fail when receiving pipelined requests.

What is http pipelining

Usually, http requests are always sent sequentially, and the next request will only be sent when the response to the current request is fully received. Due to network latency and bandwidth limitations, this will cause a large delay in the server sending the next response.

HTTP/1.1 allows multiple http requests to be output simultaneously through a socket without waiting for the corresponding response. The requester will then wait for their respective responses, which arrive in the order of the previous requests. (Me: All requests maintain a FIFO queue. After a request is sent, there is no need to wait for the response of this request to be received. The next request can be issued again; at the same time, the server returns the responses to these requests in FIFO order.) Pipelining can greatly improve the speed of page loading, especially in high-latency connections.

Pipelining can also reduce the size of TCP/IP packets. Usually the MSS size is between 536-1460 bytes, so it is possible to put many http requests in one TCP/IP packet. Reducing the number of packets required to load a web page can benefit the network as a whole, because fewer packets means less burden on routers and the network.

HTTP/1.1 requires the server to support pipelining as well. This does not mean that the server needs to pipeline responses, but that the server will not fail to respond when a client issues a pipelined request. This obviously has the potential to introduce a new category of evangelism bugs, since only modern browsers support pipelining.

When should we pipeline requests?

Only idempotent requests (see Note 1) can be pipelined, such as GET and HEAD. POST and PUT should not be pipelined. We should also not issue pipelined requests when establishing a new connection, because we cannot be sure whether the origin service or proxy supports HTTP/1.1. Therefore, pipeline can only use existing keep-alive connections.

How many requests should be pipelined

If the connection is closed too early, it is not worthwhile to pipeline many requests, because we will spend a lot of time writing the request to the network, and then have to rewrite it on the new connection. Moreover, if the earlier request takes a long time to complete, a long pipeline will actually cause the user to perceive longer delays. The HTTP/1.1 standard does not provide any guidance on the ideal number of pipelined requests. In fact, we recommend no more than 2 keep-alive connections per server. Obviously, this depends on the application. For the reasons mentioned above, browsers may not need a particularly long pipeline. 2 may be a good value, but it remains to be tested.

What happens if a request is cancelled?

If a request is canceled, does that mean the entire pipeline is canceled? Or does that mean the response to the canceled request should simply be discarded so that other requests in the pipeline are not forced to be resent? The answer depends on many factors, including how much of the response to the canceled request has not yet been received. The most naive approach might be to simply cancel the pipeline and resend all requests. This is only possible if the requests are idempotent. This naive approach can also have a positive impact, because the requests being sent in the pipeline may belong to the same page load group that is being canceled.

What happens if the connection fails?

If the connection fails or the server is interrupted while downloading a piped response, the browser must be able to restart the request that was lost. This situation is equivalent to the cancellation example discussed above.

Note

  1. Idempotence of HTTP/methods: This means that requesting a resource once or multiple times should have the same side effects.
    Idempotent requests are actually requests that do not change the result after multiple operations. For example, with GET, I can get resources from the same place multiple times, but the resource itself will not change. If I GET 10 times or GET 100 times, the resource will not change. However, with POST, if I submit the form 10 times or 100 times, the result will be different, at least the newly added data in the database will be different.

explain

  1. In fact, HTTP pipelining is to move the client's FIFO queue to the server. The client can send all the requests to be sent in sequence (of course, these requests are under the same domain). After a request is sent, there is no need to wait for the response of this request to be received before the next request can be sent again. The FIFO queue maintained on the server is arranged according to the importance of the resources. For example, HTML is returned before CSS, and JS and CSS are returned before images.
  2. There is a buffer on the server side to store responses that have been processed but have not yet been sent. For example, the server receives two requests, A and B. A resource has a higher priority than B resource. It takes 10ms to process A and 1ms to process B. Assuming that the server can process requests in parallel, then B's response must be processed first, but B's response cannot be sent out first. It must stay in the buffer and wait until A's response is processed. A's response can be sent out first, and then B's response can be sent out. Because the server must follow the FIFO principle.
  3. HTTP pipelining is not part of HTTP2, but an improvement to the HTTP1.1 protocol where the server cannot handle parallel requests well.
  4. The order of pipeline and the order of TCP are fundamentally different. The order of pipeline is the order between messages. The order in TCP is the order between multiple segments that make up a message. To make an inappropriate analogy, it is like whether student A or student B finishes lunch first and can go to play computer. Suppose student A goes to the cafeteria first and student B goes to the cafeteria later. They eat at the same speed. Then according to the FIFO principle, no matter whether student A eats apples, pears, rice, vegetables, or student B only eats apples and rice. Although student B finishes eating first and eats less, in pipeline, student A must go to play computer first, and student B follows. In TCP, it is like describing the internal order of this meal, whether student A eats apples, pears, rice, vegetables first or vegetables, rice, pears first.
  5. What the pipeline does, as far as I understand it, is to create a place where the next request can be sent without waiting for the response of the previous request. As for what to pay attention to, apart from knowing that some devices do not support it, I have no practical experience with other things.

The article comes from: Front-end Restaurant. If you wish to reprint this article, please contact the Front-end Restaurant ReTech Toutiao account.

github: https://github.com/zuopf769​​

<<:  Smart Manufacturing: Ensuring a Smart Future for Manufacturing

>>:  36.2%! H3C leads the Chinese campus switch market

Recommend

China's 4G is a global sensation! Two figures tell the whole story

The press conference on the major project "N...

Summary information: CUBECLOUD/zorocloud/Eurasia Cloud/Bluemi Cloud

Recently, we have received product promotion info...

Z-Wave not concerned about potential threats from Project CHIP

This year marks the 20th anniversary of Z-Wave be...

A brief analysis of Web real-time communication technology!

Web-based instant messaging The server can immedi...

Li-Fi, which failed to beat WiFi, may be the savior of 5G

Regarding the technical solutions for future comm...

It’s time to consider leaf-spine network architecture

With the changes in traffic flows used in modern ...

Four Best Practices for Network Cable Management

If a cabling project is to be successful, you fir...

iWebFusion: $9.38/month KVM-4GB/30GB/2TB/5 data centers including Los Angeles

iWebFusion is a site under the old foreign host c...