Interviewer: Can you tell me how Nginx handles requests?When a client initiates a request, the Nginx worker process listens to the network port and receives the client's connection request. The following is the specific process of Nginx processing the request: (1) Receiving a connection request: After Nginx receives a connection request from a client, it allocates a connection object (ngx_connection_t) for the connection. The connection object contains connection status information, read and write event handlers, etc. (2) Reading request header information: Nginx reads request header information from the client. The request header contains the HTTP method (such as GET, POST), URL, HTTP version, and various request header fields (such as Host, User-Agent, Content-Length, etc.). (3) Parsing request header information: Nginx parses the request header information and extracts necessary parameters, such as request method, URI, Host, etc. The parsed request header information is stored in the ngx_http_request_t structure. (4) Find matching virtual hosts and location blocks:
(5) Execution processing phase: Nginx's request processing is divided into multiple phases, each of which can be processed by multiple modules. These phases include:
(6) Generate and send a response:
(7) Close the connection: Once the response is sent, Nginx closes the connection. If keep-alive connections are enabled, the connection can remain open for subsequent requests. Interviewer: What is the process architecture of Nginx? Why doesn't Nginx use a multi-threaded model?1. Process ModelNginx uses a Master-Worker multi-process architecture, which is designed to ensure separation of responsibilities in order to better manage system resources, concurrent request processing, and fault recovery. (1) Master-Worker architecture: ① Master Process:
② Worker Process
(2) Inter-process collaboration:
(3) Load balancing: Nginx implements load balancing through a multi-process model. When new client connection requests arrive, these connections are evenly distributed to each Worker process to achieve load balancing. This design ensures that Nginx can efficiently handle a large number of concurrent connections and avoids a single process becoming a bottleneck. (4) High availability: Nginx's multi-process model also provides high availability. When a Worker process fails, the Master process will automatically restart a new Worker process to replace the original process, thereby ensuring high availability of the server. This design enables Nginx to run stably under high load and complex environments. 2. Why doesn’t Nginx use a multi-threaded model?Nginx chooses not to use a multi-threaded model, but an event-driven model with multiple processes and asynchronous non-blocking I/O, mainly based on the following reasons: (1) Resource isolation and stability
(2) Utilizing multi-core CPUs
(3) Avoid thread competition and deadlock
(4) Simplify the programming model
(5) Design intention
Interviewer: What are forward proxy and reverse proxy? How does Nginx implement forward proxy and reverse proxy functions?The reverse proxy function means that the proxy server accepts connection requests from the Internet, then forwards these requests to the server on the internal network, and returns the response obtained from the internal server to the client requesting the connection on the Internet. In this process, the proxy server appears as a server to the outside world. Generally speaking, the proxy server and the backend service in a reverse proxy are in the same group and are bound together. The client does not know who it is actually requesting. Real-world examples of reverse proxies include load balancing servers, network security protection (anti-DDoS attacks), and content distribution networks (CDNs). Nginx implements reverse proxy function mainly by configuring the Nginx server to make it an intermediary between the client and the target server. The following are the specific steps and key points for Nginx to implement reverse proxy function: (1) Configure Nginx: Nginx's reverse proxy configuration is mainly performed in the nginx.conf file, or in the included sub-configuration files.
(2) Configuration example: A basic reverse proxy configuration example is as follows: A forward proxy is used to forward the client's request to the target server and return the server's response to the client. In this way, the client sends the request to the proxy server, which sends the request to the target server on behalf of the client and returns the target server's response to the client. Through a forward proxy, the client can directly access the external network without establishing a direct connection with the target server. Real-world examples of forward proxies include: VPN. As a high-performance Web server and reverse proxy server, Nginx can also implement forward proxy function. The following are the specific steps and configuration methods for Nginx to implement forward proxy: How to configure Nginx forward proxySince the default Nginx release does not support the forward proxy function, you need to use the third-party plug-in ngx_http_proxy_connect_module to complete it. Add the forward proxy configuration in the configuration file, for example: In the above configuration, the listen directive specifies the port that Nginx listens on, the resolver directive specifies the address of the DNS resolver, the proxy_connect directive and other directives are used to configure the processing of CONNECT requests, and the location directive and proxy_pass directive are used to specify which target server to forward the request to. Interviewer: What is the separation of dynamic resources and static resources? Why do we need to separate dynamic and static resources? How does Nginx separate dynamic and static resources?Separation of dynamic resources from static resources (referred to as dynamic and static separation) is a common Web application architecture pattern. 1. Dynamic resources and static resources(1) Static resources When a user accesses a resource multiple times, if the source code of the resource does not change, then the resource is called a static resource. Common static resources include images (img), style sheets (css), script files (js), videos (mp4), etc. These resources can usually be cached by browsers and CDN (content delivery network) to reduce repeated requests to the server. (2) Dynamic resources If the source code of a resource may change when a user accesses it multiple times, then the resource is called a dynamic resource. Common dynamic resources include server-side scripts or template files such as JSP and FTL. These resources usually need to dynamically generate response content based on user requests. 2. Reasons for separation of movement and stillness(1) Performance optimization There are differences in the processing and distribution of dynamic content and static content. Separating them allows them to be optimized separately, thereby improving overall performance. For example, static resources can be directly processed and served by a dedicated static resource server (such as Nginx), while dynamic resources are processed by the application server. This can significantly reduce the number of requests to the dynamic resource server and reduce its load. (2) Cache Management Static resources are easy to cache, while dynamic resources are usually not suitable for caching or require more sophisticated cache control. By separating dynamic and static resources, you can better manage cache strategies, improve cache hit rates, and reduce server response time and bandwidth consumption. (3) Load Balancing After separation, resources can be optimally allocated according to content type to achieve more effective load balancing. For example, the number and scale of dynamic resource servers can be increased in a targeted manner according to the access volume and characteristics of dynamic resources to cope with high concurrent access requirements. (4) Security enhancement Static content servers usually do not need to execute complex program code, so the attack surface is smaller and security risks can be reduced. Separating them from servers that execute dynamic code can reduce potential security threats. The following is the specific implementation of Nginx dynamic and static separation: 3. Configuration methodIn Nginx, you can use the location directive to achieve dynamic and static separation. The location directive is used to match the request URI and distribute the request to different processing modules according to different paths. (1) Static resource processing In the above configuration, location /static/ and location /images/ match requests starting with /static/ and /images/ respectively, and Nginx will search for the corresponding files in the specified root directory. If the file exists, Nginx will directly return the file to the client. location ~* \.(jpg|jpeg|png|gif|ico|css|js)$ uses regular expressions to match all requests ending with these extensions and sets the cache expiration time and Cache-Control header. Static resources are usually processed directly by Nginx, so the directory of static resources can be specified in the location block. Use regular expressions to match file extensions of static resources, such as .jpg, .jpeg, .png, .gif, .ico, .css, .js, etc. Configuration example: (2) Dynamic request processing For dynamic requests, the requests can be forwarded to the backend application server (such as PHP-FPM, Django, Node.js, etc.) for processing. Configuration example: In the above configuration, location ~ \.php$ uses a regular expression to match all requests ending with .php and treats them as dynamic requests. fastcgi_pass specifies that these requests are forwarded to the PHP-FPM server for processing, and 127.0.0.1:9000 is the listening address of PHP-FPM. Other parameters are used to set FastCGI parameters and the file path of the request. 4. Notes(1) Nginx location matching priority Nginx's location matching follows specific priority rules, including exact matching (using the = symbol), regular expression matching (using ~ or ~*), and prefix matching (ordinary location). When configuring multiple location blocks, you need to carefully consider the priority relationship between them to ensure correct request routing. (2) Cache strategy By properly configuring the cache, you can significantly improve website performance. Nginx provides a powerful and flexible cache function, which can be implemented through the proxy_cache_path and proxy_cache directives. For static resources, you can set a longer cache expiration time to reduce repeated requests to the server. (3) Security By separating static and dynamic content, you can separate static content servers from servers that execute dynamic code, thereby reducing security risks. Static content servers usually do not need to execute complex program codes, so the attack surface is smaller. Interviewer: What are the algorithm strategies for Nginx load balancing?1. Round RobinPrinciple: The polling algorithm distributes requests in the order of the server list. When a new request arrives, Nginx will assign it to the next server in the list. If it reaches the end of the list, the cycle will start again. Features:
2. Weighted Round RobinPrinciple: Based on round-robin, a weight value is assigned to each server. The higher the weight value of the server, the more requests it receives. Nginx will calculate the proportion of requests received by each server based on the weight value. Features:
3. IP HashPrinciple: Requests are assigned based on the hash value of the client IP address. Nginx calculates the hash value of each client IP address and uses the hash value to select a backend server. Since the hash value of the same IP address is the same, requests from the same IP address are always assigned to the same backend server. Features:
4. Least ConnectionsPrinciple: Distribute requests to the server with the least number of current connections. Nginx monitors the current number of connections of each backend server and distributes new requests to the server with the least number of connections. Features:
5. Fair (third party)Principle: Allocate requests based on the response time of the backend server. Nginx monitors the response time of each backend server and allocates new requests to the server with the shortest response time. Features:
6. URL Hash (third party)Principle: Requests are assigned based on the hash value of the request URL. Nginx calculates the hash value of each request URL and uses the hash value to select a backend server. Since the hash value of the same URL is the same, requests for the same URL are always assigned to the same backend server. Features:
|
<<: A "right remedy" for Ethernet Mac and Phy layer problems
CloudCone has once again released this year's...
As businesses expand globally to gain access to n...
If you have used the earliest generations of smar...
[51CTO.com original article] The boss said that d...
DediPath has launched a Cyber Monday promotion....
The Hong Kong International series VPS hosts prov...
[[348075]] We still have a long way to go before ...
RAKsmart's year-end discount is coming. In ad...
Entering 2021, with the continuous acceleration o...
As digital transformation is in full swing, the n...
I haven't shared information about Digital-vm...
After years of sustained rapid growth, my country...
[51CTO.com original article] On a weekend in earl...
According to phoronix, Li-Fi technology supplier ...
Preface The gateway is the entrance for traffic r...