Fact-checked by Grok 2 weeks ago

Network load balancing

Network Load Balancing (NLB) is a clustering technology that allows multiple servers to be managed as a single virtual cluster, distributing incoming TCP/IP traffic across the nodes to improve availability and scalability for applications such as web servers, FTP, and VPNs. Primarily implemented as a software feature in Windows Server, NLB operates by having cluster hosts respond to client requests using a shared , functioning at the network and transport layers of the . In NLB, traffic distribution occurs distributively among nodes through heartbeats for status monitoring, with equal load balancing across available hosts based on configured rules. It supports session using client for consistent routing and enables dynamic scaling by allowing hosts to be added or removed without downtime. NLB uses virtual addresses to present the as a unified entity and operates in or modes to handle traffic efficiently in enterprise and environments. NLB enhances reliability through automatic , redistributing traffic from failed hosts within about 10 seconds, and supports for handling variable loads in networked applications.

Fundamentals

Definition and Purpose

Network load balancing (NLB) is a technique used to distribute incoming network traffic across multiple servers or resources in a , ensuring that no single server becomes overwhelmed and acts as a . This method treats the as a single virtual entity, allowing client requests to be evenly spread to optimize performance and prevent failures due to overload. At its core, NLB operates within the client-server architecture, where clients—such as web browsers or applications—send requests for services or data to servers that process and respond to those requests. In this model, network traffic flows from clients to a central point (the load balancer), which then directs the requests to available backend servers based on predefined criteria, maintaining smooth communication and resource access without assuming advanced prior knowledge of protocols. The primary purposes of NLB include enhancing to accommodate growing traffic volumes, providing through redundancy to minimize , and improving resource utilization in environments like data centers and applications. By distributing workloads, NLB ensures that applications remain responsive under high demand, reducing and supporting if one fails. NLB typically focuses on Layer 4 () balancing, where decisions are made based on IP addresses and ports without inspecting application data, distinguishing it from Layer 7 () proxies that analyze content for more granular routing.

Historical Development

Network load balancing emerged in the mid- amid the rapid growth of the , servers, and early platforms, which generated unprecedented traffic spikes that overwhelmed single-server architectures. Early approaches relied on simple techniques like DNS , where multiple IP addresses were assigned to a single and rotated sequentially to distribute requests across servers, providing a basic precursor to more sophisticated balancing methods. This was driven by the need to scale websites during the dot-com boom of the late , when surging online demand necessitated affordable ways to handle high volumes of concurrent users without hardware failures. A pivotal came with the of Microsoft's Network Load Balancing (NLB) in , offering a software-based clustering solution that enabled TCP/IP traffic distribution across multiple hosts for without dedicated hardware. In the early 2000s, hardware appliances from vendors like F5 and gained prominence, providing robust Layer 4 traffic management with health checks and to route requests away from overloaded or failed servers, improving by up to 25% over DNS methods. These developments were influenced by , which exponentially reduced server hardware costs, making clustered deployments economically viable for enterprises scaling beyond individual machines. The mid-2000s saw trends, led by VMware's advancements since 1999, integrate load balancing into virtual environments, allowing dynamic resource allocation across virtual machines and paving the way for software-defined solutions. Post-2010, the rise of shifted focus to elastic, software-based balancing; launched Elastic Load Balancing in 2009 to automatically distribute traffic across EC2 instances in scalable clusters. This transition from on-premises hardware to cloud-native services enabled seamless handling of variable loads, reflecting broader adoption in distributed architectures.

Core Mechanisms

Traffic Distribution Techniques

Network load balancing employs various techniques to distribute incoming traffic across multiple , ensuring efficient resource utilization and . One foundational method is IP-based distribution, where traffic is routed by hashing attributes such as the client's source (and often the destination port) to deterministically select a backend from the pool. This approach, known as IP hashing, generates a from the IP addresses of both client and , mapping the request to a specific server to maintain without requiring session state tracking at the load balancer. Session persistence, also referred to as sticky sessions, complements hashing by ensuring that subsequent requests from the same client are directed to the same server, preserving application state for stateful protocols like HTTP sessions. This is achieved through rules based on client , source port, or higher-layer identifiers such as cookies, preventing disruptions in user sessions while allowing load distribution across the . In implementations, inactivity timeouts are applied to release after a period, balancing persistence with even load spreading. Health checks are integral to traffic distribution, enabling the load balancer to continuously probe server availability and remove unhealthy nodes from the rotation. Probes typically operate at different layers: Layer 3 using ICMP pings for basic connectivity verification, Layer 4 via or connections to check port responsiveness, and Layer 7 through HTTP requests to validate application-level functionality. Failed probes trigger immediate traffic rerouting to available servers, maintaining cluster reliability. At Layer 4, traffic distribution focuses on transport-layer protocols like and , enabling port-based where connections are balanced based on the 4-tuple (source IP, , destination IP, destination port). This allows for connection multiplexing, in which multiple client connections are aggregated and shared over fewer server links, optimizing usage in high-throughput environments. Such techniques ensure stateless operation while supporting protocols requiring low-latency forwarding. Cluster synchronization facilitates dynamic load redistribution through mechanisms like heartbeat protocols, where nodes periodically exchange status messages to detect failures and share load information. Upon detecting a node failure via missed heartbeats, the cluster updates its membership view, prompting surviving nodes to absorb the redistributed traffic according to predefined rules. This accrual-based detection estimates failure probability from heartbeat arrival times, enabling proactive adjustments without centralized coordination. A representative workflow for traffic distribution begins with the load balancer inspecting an incoming packet's header for source details. An rule or then selects a target ; if the 's passes, the packet is forwarded, potentially multiplexed with others. In case of failure—detected via or —the traffic is rerouted to an alternative , ensuring seamless (visualize this as a : client packet → inspection/ → forward/reroute → response). Load balancing algorithms, such as those optimizing for least , inform these decisions but are detailed separately.

Load Balancing Algorithms

Load balancing algorithms determine how incoming is distributed across multiple to optimize utilization, minimize response times, and prevent overload on any single . These algorithms can be broadly classified into static methods, which make decisions based on predefined configurations without considering server states, and dynamic methods, which adapt to current load conditions for more efficient distribution. A comprehensive survey of load balancing techniques in environments highlights that static algorithms like are suitable for homogeneous server clusters, while dynamic ones such as least connections excel in heterogeneous setups with varying workloads. Among the most common algorithms is , which sequentially assigns incoming requests to servers in a , ensuring an even distribution over time. This method is particularly effective for environments where servers have identical processing capabilities and request handling times are uniform, as it promotes fairness without requiring ongoing monitoring. However, round-robin does not account for current server loads, potentially leading to inefficiencies if some servers become temporarily overloaded. The least connections algorithm, a dynamic approach, routes new requests to the with the fewest active at the of arrival, aiming to balance the workload more precisely in scenarios with persistent or long-duration sessions. This method assumes that connections indicate processing load and is ideal for applications like web servers where connection counts correlate with resource usage. Its primary advantage is improved fairness under uneven loads, though it incurs overhead from continuous tracking of connection states across the . Weighted extends the basic by assigning proportional weights to s based on their , such as CPU power or , allowing higher- s to receive more . For instance, a with twice the of another might be assigned a weight of 2, receiving roughly double the requests in the rotation. This static variant enhances distribution in heterogeneous environments but lacks adaptability to runtime changes in server performance. Advanced methods include hash, which generates a value from the client and IP addresses (and optionally ports) to deterministically map requests from the same client to the same , preserving session affinity without storing state. This ensures consistent routing for sticky sessions in applications requiring it, such as carts, but can result in uneven loads if client IP distributions are skewed, such as in environments. Least response time builds on dynamic balancing by selecting the with the lowest measured response time for recent requests, often combined with connection counts to avoid overburdening slow servers. It directly targets end-user by prioritizing speed, making it suitable for latency-sensitive applications like video streaming, though it requires active health checks and can introduce slight delays in decision-making due to measurements. For highly variable traffic patterns, dynamic algorithms incorporating and forecast future loads using historical data and real-time metrics to proactively allocate resources. These approaches, such as those employing temporal graph neural networks for state prediction and for task scheduling, enable anticipation of spikes, reducing reactive adjustments. They offer superior handling of bursty workloads but demand significant computational resources for model training and inference. The mathematical foundation of the least connections algorithm can be expressed as selecting the server i that minimizes the current number of active connections: i = \arg\min_{j \in \text{servers}} \text{connections}_j Pseudocode for its implementation upon a new request arrival is as follows:
function selectServer(request):
    min_conn = infinity
    selected_server = None
    for server in cluster_servers:
        if connections[server] < min_conn:
            min_conn = connections[server]
            selected_server = server
    route(request, selected_server)
    connections[selected_server] += 1
    return selected_server
This logic ensures balanced distribution by favoring underutilized servers, promoting fairness in connection-heavy scenarios at the cost of monitoring overhead. In handling uneven loads, such as during traffic spikes, dynamic algorithms like least connections and -based predictors outperform static ones like by adapting to conditions, achieving throughput improvements of 20-24% and response time reductions of up to 40% in simulated environments with heterogeneous workloads. For example, in a study of server clusters, least connections yielded up to 24% higher throughput compared to non-adaptive methods under imbalanced conditions. variants further enhance this by loads, demonstrating 20% throughput gains and 35% reductions over traditional heuristics in dynamic graph-based models.

Operational Modes

Microsoft Network Load Balancing (NLB) operational modes, including and , are deprecated as of and no longer actively developed; alternatives like software load balancers are recommended.

Unicast Mode

In mode, all cluster nodes share a single and respond to requests for that by advertising the same virtual cluster , a process akin to that causes the network switch to associate the MAC with multiple ports. Incoming traffic directed to the virtual IP is then flooded by the switch to all connected cluster nodes, where an internal load balancing mechanism selects one node to process the packets while the others discard them. This emulation makes the cluster appear as a single network entity to upstream devices. Configuration in environments like Windows involves selecting unicast mode during cluster creation via the Network Load Balancing Manager, which binds the NLB driver to the designated network adapters and overrides their original hardware addresses with the . Switches connected to the must support this setup by allowing the same on multiple ports, often requiring the disabling of features that enforce unique learning per port to prevent blocking; unlike , is irrelevant and should not be enabled for operations. Nodes typically connect to a dedicated switch or to isolate flooding. A primary advantage of unicast mode is its straightforward integration with standard infrastructure, as the presents itself as one logical device without requiring multicast-enabled hardware or protocols, making it ideal for legacy or non-multicast-supporting environments. However, unicast mode can lead to inefficiencies, including duplication where inbound packets are broadcast to every , roughly doubling the load on the local as non-selected nodes receive and drop unnecessary copies. Without proper switch , such as isolating the cluster on a dedicated segment, this flooding risks broadcast storms, excessive consumption, or even loops if redundant paths exist.

Multicast Mode

In multicast mode, Network Load Balancing (NLB) assigns a shared multicast MAC address (in the format 03-BF-XX-XX-XX-XX, derived from the virtual IP address octets in hexadecimal) to the cluster's virtual IP address, while each node retains its original unicast MAC address. Incoming traffic destined for the virtual IP is resolved via ARP to this multicast MAC, causing network switches to flood the packets to all ports in the VLAN unless IGMP snooping is enabled. Each node in the cluster joins the corresponding multicast group and receives the flooded traffic, after which the NLB driver filters it according to predefined port rules to determine which node processes the request. To implement multicast mode, administrators enable it through the NLB Manager console during , which modifies the network adapter settings to support operations. Network interfaces must have enabled, and for optimal , IGMP mode is recommended, where nodes send IGMP membership reports to join the group (typically mapped to a IP like 239.255.x.y, with x.y derived from the virtual IP's last two octets). Switches capable of are required to dynamically build tables based on these reports; in environments without an IGMP querier (often provided by a router or designated switch), manual or enabling a querier may be necessary to maintain group membership and prevent traffic flooding. This mode offers benefits such as efficient bandwidth utilization by avoiding the traffic duplication common in mode, where all nodes share a single leading to switch port blocking or replication overhead. It supports high-throughput scenarios by leveraging native delivery, reducing performance impacts on interconnected switches, and permits direct node-to-node communication within the since individual addresses are preserved. However, multicast mode introduces drawbacks including incompatibility with switches that block or poorly handle multicast traffic, potentially causing packet drops or excessive flooding. It also adds complexity to routing tables, as the multicast MAC requires static ARP entries on routers and switches without IGMP support, and some network devices may not forward multicast packets correctly without additional configuration.

Implementations

Microsoft NLB

Microsoft Network Load Balancing (NLB) is a clustering technology introduced as the Windows Load Balancing Service (WLBS) with Server 4.0 Enterprise Edition in 1997, functioning as a kernel-mode driver that enables up to 32 nodes to operate as a single virtual cluster for distributing / traffic. It primarily supports stateless /UDP-based services such as HTTP for servers and FTP, allowing seamless load distribution across cluster hosts without requiring shared storage. Key features of NLB include automatic failover, where the cluster detects a failed host and redistributes traffic to remaining nodes within 10 seconds, ensuring minimal disruption for high-availability scenarios. It supports port-specific rules to define load balancing behavior for individual / ports or port ranges, such as directing all HTTP traffic () to multiple hosts while restricting other ports to a single host for affinity-based handling. NLB is compatible with , enabling virtualized clusters where multiple virtual machines on hosts can form an NLB without needing multihomed physical servers, thus supporting scalable deployments in virtual environments. Configuration of an NLB cluster begins with installing the feature through Server Manager via the Add Roles and Features Wizard or using the PowerShell cmdlet Install-WindowsFeature NLB -IncludeManagementTools, followed by creating the with tools like the NLB Manager (nlbmgr.exe) or the New-NLBCluster cmdlet specifying parameters such as the cluster and virtual name. Port rules and host priorities are then defined in the NLB Manager interface, with settings configurable as none (for stateless distribution), single (routing all requests from a client to one ), or class C (network address-based for broader client grouping). Once configured, the can operate in or mode to handle traffic routing. NLB integrates natively with editions from 2000 through 2022, providing built-in support for on-premises clustering in enterprise environments. However, as of 2025, NLB is deprecated and no longer under active development, with Microsoft recommending migration to cloud-native alternatives like Load Balancer for modern, scalable deployments.

Alternative Solutions

Software solutions for network load balancing include open-source options like , which has supported both layer 4 () and layer 7 (HTTP) balancing since its initial release in 2001. operates as a high-performance , distributing traffic based on configurable algorithms such as or least connections, and is widely used for its reliability in handling high-traffic environments. Another open-source alternative is the (LVS), a kernel-based module using IP Virtual Server (IPVS) for Layer 4 load balancing, enabling efficient distribution of / traffic across multiple nodes in Linux environments. , originally released in 2004, functions as a with built-in load balancing modules like ngx_http_upstream_module for HTTP, , and traffic. supports methods including , least connections, and IP hash, making it suitable for web applications requiring session persistence. For commercial software, F5 BIG-IP provides enterprise-grade load balancing through its , offering advanced features like traffic management, security, and global server load balancing across on-premises and cloud deployments. Hardware appliances deliver dedicated network load balancing with optimized processing. Citrix (now ADC) hardware platforms, such as the MPX series, provide high-speed balancing with up to 200 Gbps of layer 7 throughput in a single appliance, leveraging for low-latency traffic distribution in enterprise data centers. These devices support layer 4 and 7 protocols, including SSL offloading and content switching. Although ACE was a prominent load balancer offering up to 16 Gbps throughput with ASIC-based processing, it reached end-of-sale in 2014 and end-of-support in 2019, prompting migrations to modern alternatives like or F5 hardware. Cloud-native solutions emphasize seamless integration and scalability in distributed environments. AWS Network Load Balancer (NLB), launched in , operates at layer 4 to handle , , and TLS traffic, supporting millions of requests per second with automatic based on demand. It preserves client addresses and integrates with services like Amazon EC2 and containers. Microsoft Load Balancer, available since 2010 and updated for modern features, provides Layer 4 load balancing for and traffic, with automatic and integration across Virtual Machines, containers, and virtual networks. Cloud's Proxy Load Balancer provides global distribution for traffic, using backend services that auto-scale with compute instances or clusters in Google Engine (GKE). This setup enables low-latency routing across regions without manual intervention for traffic spikes. In comparisons, software solutions like , LVS, and offer cost-effectiveness and high scalability by running on commodity hardware or virtual machines, allowing easy horizontal scaling in hybrid cloud setups, though they may introduce slightly higher due to general-purpose processing. Hardware appliances such as excel in low- scenarios with dedicated throughput exceeding 100 Gbps, but incur higher upfront costs and less flexibility for rapid scaling in dynamic hybrid clouds. For instance, organizations deploying across AWS and on-premises often combine software balancers for cost savings with cloud-native options like AWS NLB or Load Balancer for auto-scaling, contrasting the Windows-specific focus of NLB.

Applications and Considerations

Use Cases

Network load balancing plays a pivotal role in web and environments by distributing incoming traffic across multiple servers to handle massive surges, such as those during seasonal sales events. For instance, utilizes Elastic Load Balancing (ELB) from AWS to automatically scale and route traffic across availability zones during high-demand periods like Prime Day, which in 2016 processed over 85 billion clickstream log entries in 40 hours, representing 74% of U.S. volume and doubling mobile orders from the previous year. This approach ensures seamless performance for high-traffic sites, mitigating overloads during events akin to , where temporary server additions and intelligent routing prevent bottlenecks and maintain user access to shopping platforms. In database and API services within microservices architectures, network load balancing distributes queries and requests evenly to prevent any single instance from becoming overwhelmed, promoting and reliability. AWS ELB, for example, supports hybrid environments by balancing traffic across AWS resources and on-premises databases, allowing applications to scale dynamically without manual intervention. Research on microservices highlights how client-side load balancing, such as , enables efficient HTTP-based communication among service instances, reducing and ensuring in distributed systems. This is particularly vital for gateways, where load balancers route requests to available containers or virtual machines, optimizing resource utilization in cloud-native setups. For gaming and streaming applications, network load balancing facilitates to support low-latency interactions and content delivery. In multiplayer servers, techniques like Network Load Balancers handle connections for player traffic, enabling auto-scaling during peak times and distributing loads across to maintain for thousands of concurrent users. integrates load balancing within its Open Connect CDN, which employs horizontal scaling and traffic distribution across appliances to deliver video streams to over 300 million paid subscribers globally, as of 2025, selecting optimal servers based on proximity and load to minimize buffering and ensure high-quality playback. Enterprise deployments in and healthcare leverage network load balancing for high-availability configurations that minimize disruptions in mission-critical systems. In financial trading platforms, Layer 4 load balancers like the Netberg Aurora 610 distribute multi-terabit traffic for electronic (eFX) and multi-asset trading, supporting millions of persistent connections with sub-second and embedded health checks to sustain operations during volatile market hours. Similarly, in healthcare (EHR) systems, load balancing enhances performability in medical information systems (MIS) by employing strategies such as shortest-queue distribution across fog nodes and virtual machines, which improves throughput and reduces response times while ensuring service continuity through fail-over mechanisms. For EHR implementations, load balancers automatically redirect traffic upon server failure, achieving robust availability for patient data access and clinical workflows. These setups have demonstrated uptime exceeding 99.9% in optimized trading environments, underscoring downtime reductions in regulated sectors.

Benefits and Limitations

Network load balancing provides enhanced by distributing across multiple , enabling zero-downtime when a server fails, as is automatically rerouted to healthy nodes without interrupting . This redundancy minimizes outages and ensures continuous availability, particularly in high- environments where single-server failures could otherwise cause significant disruptions. It also improves overall performance by optimizing resource utilization and maximizing throughput in server clusters, allowing systems to handle increased loads more efficiently through even traffic distribution. For instance, in clustered setups, load balancing can scale throughput proportionally to the number of added nodes, potentially achieving substantial gains in for demanding applications. Additionally, it enables cost savings by leveraging commodity hardware for , reducing the need for expensive systems while maintaining high performance. Despite these advantages, network load balancing introduces limitations, such as the potential for the load balancer itself to become a if not properly configured, though this can be mitigated through high-availability (HA) pairs that provide redundancy and automatic between balancers. Configuration complexity poses another challenge, as misconfigurations can lead to uneven traffic distribution or outages; studies indicate that up to 75% of issues stem from such errors. Health checks, essential for monitoring status, add operational overhead, including moderate CPU utilization on both the balancer and backend servers to perform periodic probes. Security considerations are critical, as load balancers often serve as the public-facing , exposing them to distributed denial-of-service (DDoS) attacks that can overwhelm resources unless protected by firewalls or integrated mitigation tools. Furthermore, SSL termination at the load balancer—where is decrypted before forwarding traffic—centralizes security management but requires robust protocols to prevent vulnerabilities in the unencrypted internal traffic between the balancer and backend servers. Looking ahead, future trends in network load balancing include integration with for predictive , which analyzes patterns to proactively distribute loads and prevent bottlenecks, enhancing efficiency in dynamic environments. This approach also addresses limitations in , such as bandwidth constraints and latency, by enabling more adaptive balancing closer to data sources.

References

  1. [1]
    What is load balancing? | How load balancers work - Cloudflare
    Load balancing is the process of distributing traffic among multiple servers to improve a service or application's performance and reliability.
  2. [2]
    What Is Load Balancing? | IBM
    Load balancing is the process of distributing network traffic efficiently among multiple servers to optimize application availability.What is load balancing? · How it works
  3. [3]
    Network Load Balancing | Microsoft Learn
    Jul 29, 2021 · The Network Load Balancing (NLB) feature distributes traffic across several servers by using the TCP/IP networking protocol. By combining two or ...Practical applications · Important functionality
  4. [4]
    What is network load balancing (NLB)? | Definition from TechTarget
    Nov 15, 2023 · NLB is used to distribute or balance network traffic across multiple servers or VMs. In doing so, it prevents the overloading of any single host.
  5. [5]
    What Is the Client/Server Model? - Akamai
    The client/server model refers to a basic concept in networking where the client is a device or software that requests information or services.
  6. [6]
    The client/server model - IBM
    Client/server is a model of interaction in which a program sends a request to another program and awaits a response. The requesting program is called a client; ...Missing: definition | Show results with:definition
  7. [7]
    What is a Network Load Balancer? - AWS Documentation
    A load balancer serves as the single point of contact for clients. The load balancer distributes incoming traffic across multiple targets, such as Amazon EC2 ...Network Load Balancer... · Network Load Balancer overview
  8. [8]
    Layer 4 vs. Layer 7 Load Balancing: Learn the Difference - Kemp
    Feb 23, 2023 · Layer 4 load balancing gets used in situations where the contents of the data packets aren't needed when making routing decisions.
  9. [9]
    What Is Load Balancing & How Do Load Balancers Work - Kemp
    A load balancer acts as a 'reverse-proxy' to represent the application servers to the client through a virtual IP address (VIP).<|control11|><|separator|>
  10. [10]
    Load Balancing Evolution - The Basics - Austral Tech
    Jan 17, 2017 · Server load balancing grew out of the need to scale websites in the 1990s and is the foundation of today's modern application delivery controller.
  11. [11]
    A History of Load Balancing - Packet Pushers
    Aug 19, 2014 · A visual representation of the company and, to a lesser extent, product history of the load balancing/application delivery field.
  12. [12]
    Classic.Ars: Understanding Moore's Law - Ars Technica
    Sep 27, 2008 · "The number of transistors per chip that yields the minimum cost per transistor has increased at a rate of roughly a factor of two per year.".
  13. [13]
    The 3 Generations of Load Balancing, According to a Lifer
    Apr 16, 2020 · The first load balancers offered connection-based traffic management between clients and servers. These hardware appliances were used primarily ...Missing: 2000s | Show results with:2000s
  14. [14]
    Timeline of Amazon Web Services - Wikipedia
    ... Amazon S3. 2009, May 18, Product (compute), Amazon introduces Elastic Load Balancing (ELB) (which makes it easy for users to distribute web traffic across ...
  15. [15]
    What Is Load Balancing & How Does It Work? - CDNetworks
    Mar 22, 2021 · A Brief History of Load Balancing. Load balancing came to prominence in the 1990s as hardware appliances distributing traffic across a network.
  16. [16]
  17. [17]
    Layer 4-7 Server Load Balancing - IEEE Web Hosting
    • Load balancing at the “transaction” boundaries. IP. Network. IP. Network ... • UDP session persistence using Layer 3/4. ➢ Source IP & port, Destination ...
  18. [18]
    Ananta: cloud scale load balancing: ACM SIGCOMM Computer Communication Review: Vol 43, No 4
    ### Summary of Health Checks in Ananta Paper for Load Balancing
  19. [19]
    [PDF] FiDe: Reliable and Fast Crash Failure Detection to Boost Datacenter ...
    Jul 9, 2025 · Timeouts are at the center of the prominent heartbeat crash FD design which has dis- tributed processes periodically send “I am alive” messages ...
  20. [20]
    Load balancing algorithms in cloud computing: A survey of modern ...
    This paper surveys the state of the art load balancing tools and techniques over the period of 2004-2015. We group existing approaches aimed at providing load ...
  21. [21]
    Types of load balancing algorithms - Cloudflare
    Round robin: Round robin load balancing distributes traffic to a list of servers in rotation using the Domain Name System (DNS).
  22. [22]
    What is Load Balancing? - Load Balancing Algorithm Explained - AWS
    Round-robin method. Servers have IP addresses that tell the client where to send requests. · Weighted round-robin method · IP hash method · Least connection method.
  23. [23]
    Load Balancing Algorithms and Techniques - Kemp Technologies
    Weighted round robin is similar to the round-robin load balancing algorithm, adding the ability to spread the incoming client requests across the server farm ...
  24. [24]
    HTTP Load Balancing | NGINX Documentation
    NGINX Open Source supports four load balancing methods: Round Robin, Least Connections, IP Hash, and Generic Hash. ... least average time to receive the response ...
  25. [25]
    What are load-balancing algorithms? - HAProxy Technologies
    Least response time – Requests are routed to servers with the quickest response times, provided they also have the least number of connections. This takes ...
  26. [26]
    Dynamic load balancing in cloud computing using predictive graph ...
    Jul 1, 2025 · More recently, machine learning (ML) and deep learning (DL) methods have been explored to manage load balancing issues in cloud environments.
  27. [27]
  28. [28]
    Configure network infrastructure to support the NLB operation mode
    Jan 15, 2025 · In unicast mode, NLB uses the NLB MAC address to replace the original hardware MAC address of each adapter in each node of the cluster.
  29. [29]
    Catalyst Switches for Microsoft Network Load Balancing ... - Cisco
    Aug 6, 2014 · This document describes how to configure the Cisco Catalyst switches in order to interact with Microsoft Network Load Balancing (NLB).
  30. [30]
    Network Load Balancing concept and notes - Windows Server
    Jan 15, 2025 · NLB enables each host to detect and receive incoming TCP/IP traffic. This traffic is received by all the hosts in cluster and NLB driver filter ...
  31. [31]
    Configure Catalyst Switches for Microsoft NLB - Cisco
    Oct 16, 2020 · The IGMP snooping programs the virtual MAC address for you once the switch receives a membership report from a member in the cluster. An Mrouter ...
  32. [32]
    Microsoft Network Load Balancing Multicast and Unicast operation ...
    Apr 21, 2025 · Unicast mode induces switch flooding, where all switch ports are flooded with NLB traffic, even ports to which non-NLB servers are attached.
  33. [33]
    Cisco Nexus 9000 Series NX-OS Multicast Routing Configuration ...
    Sep 11, 2023 · This section describes how to configure a Cisco Nexus 9000 series switches for multicast and IGMP multicast mode NLB.
  34. [34]
    Microsoft Expands Clustering Capabilities Of Windows NT Server ...
    The Windows NT Load Balancing Service allows customers to cluster their TCP/IP-based network services across up to 32 systems, which then appear as a single ...
  35. [35]
    nlbmgr | Microsoft Learn
    Nov 1, 2024 · The nlbmgr command manages Network Load Balancing clusters and hosts. It is started with `nlbmgr.exe` and can replicate cluster configurations.
  36. [36]
    Features Removed or No Longer Developed in Windows Server
    Sep 24, 2025 · NLB is no longer in active feature development and is deprecated. ... 2025 update, Windows Server 2025 no longer includes Windows PowerShell 2.0.
  37. [37]
    The History of HAProxy
    Nov 8, 2019 · Willy released the first version of HAProxy in 2001, welcomed the first contribution in 2004, and became a Linux kernel maintainer in 2006. His ...Missing: official | Show results with:official
  38. [38]
    Layer 4 vs Layer 7 Load Balancing (Differences Explained)
    Jun 9, 2011 · With HAProxy you can switch between proxying traffic at layer 4 (TCP) or layer 7 (HTTP). This blog post describes the features available to you ...Missing: 2000 | Show results with:2000
  39. [39]
    Module ngx_http_upstream_module - nginx
    The ngx_http_upstream_module module is used to define groups of servers that can be referenced by the proxy_pass, fastcgi_pass, uwsgi_pass, scgi_pass, ...
  40. [40]
    Load Balancing - F5
    F5 offers load balancing via hardware, software (like NGINX), and cloud solutions, including BIG-IP, BIG-IP LTM, and Distributed Cloud DNS Load Balancer.F5 Load Balancing Deployment... · Load Balancing Services · Resources
  41. [41]
    High-Performance Load Balancing Use Cases - NetScaler
    A single NetScaler hardware ADC achieves up to 200 Gbps L7 throughput, and a single software ADC achieves up to 100 Gbps L7 throughput. ... NetScaler doesn't ...
  42. [42]
    Hardware platforms | NetScaler MPX™ - Product Documentation
    The NetScaler hardware platforms range from the single 10-core processor MPX 9100 to the high-capacity MPX two 16-core processor MPX 16000.Missing: throughput ASIC
  43. [43]
    Cisco ACE Application Control Engine Module for Cisco Catalyst ...
    The Cisco ACE Module provides best-in-industry scalability and throughput for managing application traffic, up to 16 Gbps in a single module.Missing: ASIC | Show results with:ASIC
  44. [44]
    Cisco ACE Application Control Engine Module
    The Cisco ACE Application Control Engine Module has been retired and is no longer supported. End-of-Sale Date: 2014-01-24. End-of-Support Date: 2019-01-31.<|separator|>
  45. [45]
    New Network Load Balancer – Effortless Scaling to Millions of ...
    Sep 7, 2017 · It is designed to handle tens of millions of requests per second while maintaining high throughput at ultra low latency, with no effort on your part.
  46. [46]
    Proxy Network Load Balancer overview
    Proxy Network Load Balancers are layer 4 reverse proxy load balancers that distribute TCP traffic to backends in your Google Cloud Virtual Private Cloud (VPC) ...
  47. [47]
    Hardware vs Software Load Balancers: Key Differences - Serverion
    Jun 7, 2025 · Explore the key differences between hardware and software load balancers, including performance, scalability, and cost to find the right fit ...
  48. [48]
    Software Load Balancers vs Appliances (Differences Explained)
    Feb 15, 2024 · We'll take a look at some core differences between software load balancers and load balancing appliances, and explain how HAProxy's unique approach to building ...Missing: hybrid | Show results with:hybrid
  49. [49]
    What is Hardware Load Balancer (HLD) | Box vs Cloud - Imperva
    The main difference is the higher cost of purchasing hardware compared to a LBaaS subscription fee. In addition, lack of HLD scalability may hinder performance, ...Missing: software | Show results with:software
  50. [50]
    How AWS Powered Amazon's Biggest Day Ever
    Jul 25, 2016 · Running the Amazon website and mobile app on AWS makes short-term, large scale global events like Prime Day technically feasible and economically viable.
  51. [51]
    Load Balancer - Elastic Load Balancing (ELB) - AWS
    ### Summary of Elastic Load Balancing Use Cases
  52. [52]
    [PDF] Load Balancing across Microservices - Fangming Liu
    Using. HTTP, a microservice instance can initiate a connection to anyone without relying on an intermediate broker. However, since HTTP communication requires ...
  53. [53]
    Running your game servers at scale for up to 90% lower compute cost
    Mar 5, 2019 · For game servers that use TCP as the transport network layer, AWS offers Network Load Balancers as an option for distributing player traffic ...
  54. [54]
    Netflix | Open Connect
    The Netflix Open Connect program provides opportunities for ISP partners to improve their customers' Netflix user experience by localizing Netflix traffic.Peering With Open Connect · English (GB) · Appliances · Partner Portal
  55. [55]
    How Netflix Delivers High-Quality Streaming to 269 Million Viewers
    Jun 10, 2024 · CDNs use load balancing to distribute traffic across multiple servers to prevent any single server from becoming overloaded. This technique is ...
  56. [56]
    Layer 4 Load Balancer in eFX and Multi-asset Class Trading Case ...
    Considerable resources are invested in deploying of hardware and software load balancing for scale-out. Increased infrastructure costs and the burden of life ...
  57. [57]
    Performability Evaluation of Load Balancing and Fail-over Strategies ...
    Sep 17, 2021 · This study can help improve the design of MIS systems integrated with different load-balancing techniques and fail-over mechanisms to maintain ...
  58. [58]
    Load balancing Epic EHR: A step-by-step guide - Loadbalancer.org
    Mar 3, 2023 · High Availability (HA): By distributing traffic across multiple servers, if one server fails, the load balancer automatically redirects users ...
  59. [59]
    System Optimization & Real-Time Analytics for a Financial Trading ...
    Mar 30, 2025 · Cloud migration, load balancing, and system redundancy together ensured the platform achieved 99.9% uptime even during peak trading hours. List ...
  60. [60]
    What is server failover? | Failover meaning - Cloudflare
    Cloudflare Load Balancing achieves fast failover by actively monitoring servers and instantly rerouting traffic when an issue is detected, resulting in zero ...
  61. [61]
    What Is High Availability? - Cisco
    Fault tolerance aims for zero downtime, while high availability is focused on delivering minimal downtime. A high-availability system designed to provide 99.999 ...
  62. [62]
    Load Balancing Options - Azure Architecture Center | Microsoft Learn
    The term load balancing refers to the distribution of processing across multiple computing resources. You load balance to optimize resource usage, maximize ...Azure load balancing services · Service categorizations
  63. [63]
    Throughput and load balancing optimization in heterogeneous ...
    Oct 9, 2025 · The outcomes are examined to see how various approaches, which seek to increase throughput while guaranteeing equitable resource distribution ...
  64. [64]
    5 Scaling Strategies for Load Balancers - Serverion
    May 24, 2025 · "Clustering is a cost-effective way to improve a website or application's performance, reliability, and scalability using commodity hardware.
  65. [65]
    How to avoid a single point of failure - Loadbalancer.org
    Feb 8, 2022 · Here I'll walk you through how you can maximize your uptime by avoiding what we call "a single point of failure."
  66. [66]
    How to test for and help prevent bad network connectivity
    Mar 3, 2020 · By some estimates, 75% of network outages and performance issues are the result of a misconfiguration, and more often than not, ...
  67. [67]
    Troubleshoot a Classic Load Balancer: Health checks
    Check the monitoring graph for over-utilization of CPU. · Check the utilization of other application resources, such as memory or limits, by connecting to your ...Missing: overhead | Show results with:overhead
  68. [68]
    How to prevent DDoS attacks | Methods and tools - Cloudflare
    Several methods for reducing this exposure include restricting traffic to specific locations, implementing a load balancer, and blocking communication from ...
  69. [69]
    Should SSL be terminated at a load balancer?
    Feb 6, 2013 · A second reason SSL should terminate at the load balancer is because it offers a centralized place to correct SSL attacks such as CRIME or BEAST ...
  70. [70]
    AI Load Balancing for Data Centers: How It Works - Serverion
    Oct 20, 2025 · AI load balancing keeps an eye on network activity and adjusts resource allocation on the fly to ensure everything runs smoothly.
  71. [71]
    Edge Computing: the Future of Load Balancers - skudonet
    Rating 5.0 (2) Mar 14, 2024 · Issues such as network latency, bandwidth limitations, and network congestion can impact the overall performance of load balancers. SKUDONET ...