HTTP1.0 HTTP version 1.0 is a stateless, connectionless application layer protocol. HTTP 1.0 stipulates that the browser and server maintain a short-lived connection. Each time the browser makes a request, it needs to establish a TCP connection with the server. The server disconnects the TCP connection immediately after processing (no connection). The server does not track each client order or record past requests (stateless). This statelessness can be used with the help of cookie/session mechanism for identity authentication and status recording. Problems with HTTP 1.0 Unable to reuse connection Each time a request is sent, a TCP connection is required, and the TCP connection release process is relatively time-consuming. This connectionless feature will reduce the utilization rate of the network. Head of line blocking Since HTTP1.0 stipulates that the next request must be sent before the response to the previous request arrives, assuming that the response to the previous request has not arrived, the next request will not be sent and the subsequent requests will be blocked. HTTP1.1 HTTP1.1 inherits the simplicity of HTTP1.0 and overcomes the performance problems of HTTP1.0. Long Connection HTTP1.1 adds the Connection field, which keeps the HTTP connection connected by setting Keep-Alive. This avoids the need to repeatedly establish and release a TCP connection between the client and the server for each request, thus improving network utilization. If the client wants to close the HTTP connection, it can carry Connection:false in the request header to tell the server to close the request. Pipelining — Awkward pseudo-parallelism HTTP1.1 supports request pipelining. Based on the long connection of HTTP1.1, request pipelining is possible, which enables requests to be transmitted in parallel. For example: If the body of the response is an HTML page that contains many imgs, keep-alive is very useful. It can send multiple requests "in parallel". (Note that "in parallel" here does not mean parallel transmission in the true sense) It should be noted that the server must return the corresponding results in the order of the client's requests to ensure that the client can distinguish the response content of each request. In other words, HTTP pipelining allows us to migrate the first-in-first-out queue from the client (request queue) to the server (response queue). If the client sends two requests at the same time to obtain HTML and CSS respectively, if the server's CSS resource is ready first, the server will send HTML first and then CSS. In other words, the CSS response resource will not be transmitted until the HTML response resource is completely transmitted. Two parallel responses are not allowed at the same time. It can be seen that HTTP1.1 still cannot solve the problem of head of line blocking. At the same time, there are various problems with the "pipelining" technology, so many browsers either do not support it at all, or directly turn it off by default, and the conditions for turning it on are very harsh... and it seems that it is actually useless. True parallel transmission — browser optimization strategy HTTP1.1 supports pipelining, but the server must also send back responses one by one, which is a big flaw. In fact, browser manufacturers at this stage have adopted another approach, which allows us to open multiple TCP sessions. In other words, the parallelism we see in the above figure is actually HTTP requests and responses on different TCP connections. This is true parallelism! Many people think that the number of connections is: Actual situation (China): Cache processing — strong cache, negotiation cache, heuristic cache (new) In addition, HTTP1.1 also adds cache processing (strong cache and negotiated cache), new fields such as cache-control, support for breakpoint transmission, and an increase in the Host field (enabling a server to create multiple Web sites) HTTP2.0 Binary framing HTTP2.0 breaks through the performance limitations of HTTP1.1 and improves transmission performance by adding a binary layered frame between the application layer and the transport layer. Multiplexing (link sharing) — true parallel transmission
All HTTP 2.0 communications are done over a TCP link, which can carry bidirectional data flows of any traffic. Each data stream is sent in the form of a message, which consists of one or more frames. These frames can be sent out of order and then repackaged according to the stream identifier (Stream_id) in the header of each frame. Multiplexing (connection sharing) may cause keywords to be blocked. In HTTP2.0, each data stream can set priority and dependency. High-priority data streams will be processed by the server first and returned to the client. Data streams can also depend on other sub-data streams. It can be seen that HTTP2.0 realizes true parallel transmission, which can make any number of HTTP requests on one TCP. This powerful function is based on the feature of "two-level framing". Header Compression In HTTP 1.X, header metadata is sent as plain text, typically adding 500-8000 bytes of payload to each request. For example, cookies. By default, the browser will attach the cookie to the header and send it to the server every time a request is made. HTTP2.0 uses encoder to reduce the size of headers that need to be transmitted. Both communicating parties cache a header_files table, which not only avoids the transmission of duplicate headers, but also reduces the size that needs to be transmitted. Efficient compression algorithms can greatly compress the header, reduce the number of packets sent, and thus reduce latency. Server Push In addition to the response to the initial request, the server can also push additional resources to the client without the client's explicit request. HTTP3.0 Google developed a QUIC protocol based on the UDP protocol and used it on HTTP/3. The previous name of HTTP/3 was HTTP-over-QUIC. The early QUIC protocol had two versions, IETF and Google, until it was confirmed and named HTTP 3.0 The QUIC working group of IETF created the QUIC transport protocol. QUIC is a protocol that uses UDP instead of TCP. Initially, Google started to support QUIC, and later QUIC was more commonly known as "HTTP/2-encrypted-over-UDP". People in the community have been using informal names like iQUIC and gQUIC to refer to these different versions of the protocol to separate the QUIC protocol from the IETF and Google (because they differ so much in details). The protocol for sending HTTP over "iQUIC" was called "HQ" (HTTP-over-QUIC) for a long time. On November 7, 2018, Dmitri from Litespeed announced that they and Facebook had successfully completed the first interoperation between two HTTP/3 implementations. Mike Bihop's subsequent presentation in the HTTPBIS session on the topic can be seen here. The session ended with a consensus that the new name is HTTP/3! 0-RTT — The biggest advantage of QUIC protocol over HTTP2.0 Cache the context of the current session. The next time you resume the session, you only need to pass the previous cache to the server. Once it is verified, it can be transmitted. 0-RTT connection establishment can be said to be the biggest performance advantage of QUIC compared to HTTP2. What is 0-RTT connection?
Multiplexing QUIC is based on UDP. There is no dependency between multiple streams on a connection. Even if packets are lost, only the lost packets need to be resent without retransmitting the entire connection. Better mobile performance QUIC performs better than TCP on mobile devices because TCP identifies connections based on IP, while QUIC identifies connections by ID. No matter how the network environment changes, as long as the ID is inconvenient, you can quickly reconnect. The root of encryption authentication — armed to the teeth The TCP protocol header is not encrypted or authenticated, and can be easily tampered with, injected, and eavesdropped by intermediate network devices during transmission. QUIC packets are armed to the teeth. Except for individual messages, such as PUBLIC_RESET and CHLO, all message headers are authenticated and message bodies are encrypted. Therefore, as long as any changes are made to QUIC, the receiving end can detect them in time, effectively reducing security risks. Forward Error Correction The QUIC protocol has a very unique feature called Forward Error Correction (FEC). Each data packet includes the data of other data packets in addition to its own content. Therefore, a small amount of packet loss can be directly assembled through the redundant data of other packets without retransmission. Forward error correction sacrifices the upper limit of the data that can be sent per data packet, but the improvement it brings is greater than the data retransmission caused by packet loss, because data retransmission will consume more time (including the time consumed in steps such as confirming packet loss, requesting retransmission, and waiting for new data packets). For example:
Problem Summary Are HTTP1.1 merge requests (such as CSSsprites) applicable to HTTP2.0? no need. In header compression technology, both the client and the server maintain two identical static dictionaries and dynamic dictionaries. The static dictionary contains common header name and value combinations. The static dictionary can be used during the first request. Now the header fields can be abbreviated as the index of the corresponding field in the static dictionary. The dynamic dictionary is related to the context of the connection. The dynamic dictionary maintained by each HTTP/2 connection is different. The dynamic dictionary can be continuously updated during the connection. In other words, the key values or fields of the original complete HTTP message header can now be converted into an index due to the existence of the dictionary, and then searched and restored at the corresponding end, which plays a compression role. Therefore, the more requests and responses generated on the same link, the more complete the dynamic dictionary is accumulated, and the better the header compression effect is. Therefore, for HTTP/2 websites, the best practice is not to merge resources. In addition, HTTP2.0 multiplexing allows requests to be transmitted in parallel, and one of the reasons why HTTP1.1 merges requests is to prevent blocking problems caused by too many HTTP requests. Now that HTTP2.0 can be transmitted in parallel, there is no need to merge requests. Why HTTP 3.0: Problems caused by the limitations of HTTP/2's underlying TCP Since HTTP/2 uses multiplexing, generally speaking, only one TCP link is needed for the same domain name. However, when packet loss occurs in this connection, the performance of HTTP/2 will be worse than that of HTTP/2. The reason is: when packet loss occurs, the entire TCP starts waiting for retransmission, causing all subsequent data to be blocked. However, for HTTP/1.1, multiple TCP connections can be opened. This situation will only affect one of the connections, and the remaining TCP links can still transmit data normally. Because modifying the TCP protocol is an impossible task. How to enable QUIC protocol in Chrome MTF has enabled HTTP3.0 protocol on both resource servers and content distribution nodes. Based on the backward compatibility of user browsers, it is strongly recommended that you enable experimental QUICK protocol support in Chrome browser to experience the acceleration effect: In the browser address bar: enter chrome://flags Find Experimental QUIC protocol and change Default to Enabled Summarize HTTP 1.0
HTTP 1.1
HTTP 2.0
HTTP 3.0
UDP-based implementation 0RTT connection UDP-based multiplexing Encrypted and authenticated messages Forward Error Correction |
<<: Getting started with SD-WAN, just read this article
On June 8, a research organization recently relea...
This series of articles will start with an introd...
As the construction of 5G networks continues to a...
Compared with LTE, 5G networks that introduce fea...
At the end of last month, we shared some of RAKsm...
HostKvm is a foreign VPS service provider founded...
Let me first state the following conclusion. 1 Sp...
Today, the Internet world is slowly transitioning...
RAKsmart's cloud server and VPS host flash sa...
Many see 5G wireless technology as the next wave ...
【51CTO.com Quick Translation】 Living in the prese...
HostXen is a domestic hosting merchant founded in...
According to Gartner's forecast, global IT sp...
The construction industry is an important pillar ...
[51CTO.com original article] If you download a mo...