Conceptly
← All Concepts
⚖️

Azure Load Balancer

NetworkingLayer 4 traffic distribution for backend server pools

Azure Load Balancer spreads TCP and UDP traffic across healthy backend instances. It uses health probes to remove unresponsive targets and keeps a single entry point in front of multiple backends.

Architecture Diagram

🔗 Relationship

Dashed line animations indicate the flow direction of data or requests

Why do you need it?

A single backend host becomes a bottleneck and a single point of failure. Adding more backends solves only half the problem, because users still need one stable address and the system still needs a way to stop sending traffic to dead targets.

Why did this approach emerge?

Horizontal scaling created the need for a traffic distribution layer that could keep pace with rapidly changing backend counts. In on-prem environments that often meant dedicated hardware appliances. In cloud environments, the balancing layer also had to become elastic and software-defined so it could adjust as fast as compute did.

How does it work inside?

Azure Load Balancer combines a frontend IP, a backend pool, balancing rules, and health probes. The frontend accepts traffic, the backend pool holds the actual targets, the rules map ports between them, and the probes keep checking whether each backend is still healthy enough to receive requests.

What is it often confused with?

Azure Load Balancer and Application Gateway both spread traffic, but they operate at different layers. Load Balancer works at the transport layer and does not inspect URLs or headers. Application Gateway and other layer-7 gateways add path-based routing, TLS termination, and web application filtering at the cost of more protocol awareness. When routing depends on URL paths, host headers, or TLS offloading, a layer-7 gateway handles it. When you need raw TCP/UDP distribution across backends without inspecting application content, Load Balancer is the right layer.

When should you use it?

A load balancer is the basic entry layer when several backend servers need to behave like one service endpoint. It is common in high-availability server pools and private service tiers where traffic should be spread evenly across healthy targets. It is not enough on its own when routing must depend on URLs, TLS termination rules, or cross-region delivery policy.

Distribute traffic across multiple backend serversKeep services available during instance failuresBalance private traffic between internal service tiersForward specific inbound ports to controlled backend targets