白丝美女被狂躁免费视频网站,500av导航大全精品,yw.193.cnc爆乳尤物未满,97se亚洲综合色区,аⅴ天堂中文在线网官网

Load balancing web service by rejecting connections

專利號
US11159406B2
公開日期
2021-10-26
申請人
Facebook, Inc.(US CA Menlo Park)
發(fā)明人
Robert M. English; Jeffrey Rothschild
IPC分類
H04L12/26; G06F9/50; H04L12/911
技術(shù)領(lǐng)域
server,balancer,queue,130a,tcp,requests,rejection,load,in,request
地域: CA CA Menlo Park

摘要

A load balancer allocates requests to a pool of web servers configured to have low queue capacities. If the queue capacity of a web server is reached, the web server responds to an additional request with a rejection notification to the load balancer, which enables the load balancer to quickly send the rejected request to another web server. Each web server self-monitors its rejection rate. If the rejection rate exceeds a threshold, the number of processes concurrently running on the web server is increased. If the rejection rate falls below a threshold, the number of processes concurrently running on the web server is decreased.

說明書

Keeping the queue capacity low functions to limit how disproportionate the workload between servers 130A-D in the server pool 139 can be. In other words, if the queue 134A of one server 130A of the server pool 139 is large, it is easier for a large queue of TCP requests to accumulate at that server while another server 103B-D may have no work to do. A low queue capacity caps the number of requests that are ageing in one queue rather than being processed by another server 130B-D that may be less busy. Another reason for a low queue capacity is to limit the time requests spend in the queue before processing starts. As a result, configuring servers 130A-D with low queue capacities can result in an overall improvement in response time performance. For example, if the request rates is low (for example one request per second average service rate), a queue length of 1 allows requests to sit in the queue unprocessed for up to a second (half of a second on average). However, at higher request rates, the time spent in the queue is less, so the difference in performance between a queue length of zero and one can become insignificant.

In one variation, although a queue capacity of one is preferred in some implementations, an exception would be cases where the system as a whole needs to serve a given number of requests, but the number of requests an individual server can handle is limited by resource constraints. In this case, instead of increasing the number of processes, the feedback loop, which will be described in more detail below, keeps the queue depth balanced across all nodes by increasing the queue depths, so that individual requests are served as quickly as possible.

權(quán)利要求

1
微信群二維碼
意見反饋