DYNAMIC Burden Adjusting IN WEBSERVERS and PARALLEL PCs By Vidhya Balasubramanian - PowerPoint PPT Presentation

slide1 n.
Skip this Video
Loading SlideShow in 5 Seconds..
DYNAMIC Burden Adjusting IN WEBSERVERS and PARALLEL PCs By Vidhya Balasubramanian PowerPoint Presentation
DYNAMIC Burden Adjusting IN WEBSERVERS and PARALLEL PCs By Vidhya Balasubramanian

play fullscreen
1 / 37
Download
Download Presentation

DYNAMIC Burden Adjusting IN WEBSERVERS and PARALLEL PCs By Vidhya Balasubramanian

Presentation Transcript

  1. DYNAMIC LOAD BALANCING IN WEBSERVERS & PARALLEL COMPUTERS By Vidhya Balasubramanian

  2. Dynamic Load Balancing on Highly Parallel Computers • - dynamic balancing schemes which seek to minimize total execution time of a single application running in parallel on a multiprocessor system • 1. Sender Initiated Diffusion (SID) • 2. Receiver Initiated Diffusion(RID) • 3. Hierarchical Balancing Method (HBM) • 4. Gradient Model (GM) • 5. Dynamic Exchange method (DEM) • Dynamic Load Balancing on Web Servers • dynamic load balancing techniques in distributed web-server architectures , by scheduling client requests among multiple nodes in a transparent way • 1. Client-based approach • 2. DNS-Based approach • 3. Dispatcher-based approach • 4. Server-based approach

  3. Load balancing on Highly Parallel computers • load balancing is needed to solve non-uniform problems on multiprocessor systems • load balancing to minimize total execution time of a single application running in parallel on a multicomputer system • General Model for dynamic load balancing includes four phases • * process load evaluation • * load balancing profitability determination • * task migration strategy • * task selection strategy • 1st and 4th phase application dependent and hence can be done independently • load balancing overhead includes :- • - communication costs of acquiring load information • - informing processors of load migration decisions • - processing costs of evaluating load information to determine task transfers

  4. Issues in DLB Strategies • 1. Sender or Receiver initiation of balancing • 2. Size and type of balancing domains • 3. Degree of knowledge used in the decision process • 4. Overhead , distribution and complexity • General DLB Model • Assumption – each task is estimated to require equal computation time • process load evaluation – count of number of tasks pending execution • task selection simple – no distinction between tasks • inaccuracy of task requirements estimates leads to unbalanced load distributions • imbalance detected in phase 2, and appropriate migration strategy devised in phase 3. • centralized vs. distributed approach – • centralized –more accurate, high degree of knowledge, but requires synchronization which incurs an overhead and delay • distributed – less accurate, lesser overhead

  5. Load Balancing Terminology Load Imbalance Factor ( f(t) ) : It is a measure of potential speedup obtainable through load balancing at time t It is defined as the maximum processor loads before and after load balancing , Lmax, and Lbal respectively f(t) = Lmax - Lbal Profitability: Load Balancing is profitable if the savings is greater than load balancing overhead Loverhead i.e., f(t) > Loverhead Simplifying assumption : One the processor’s load drops below a preset threshold , Koverhead any balancing will improve the system performance Balancing Domains: system partitioned into individual groups of processors Larger domains – more accurate migration strategies : smaller domains – reduced complexity

  6. Gradient Model • Under loaded processors inform other processors in the system of their state and overloaded processors respond by sending a portion of the load to the nearest lightly loaded processor • threshold parameters – Low-Water-Mark(LWM) , High-Water-Mark(HWM) • processors state light if less than LWM, and high if greater than HWM • Proximity of a process : defined as the shortest distance from itself to the nearest lightly loaded node in the system • wmax - initial proximity, the diameter of the system • proximity of system is 0 if state becomes light • Proximity of p with ni neighbors computed as : • proximity(p) = mini ( proximity(ni )) + 1 • Load balancing profitable if : • Lp – Lq> HWM – LWM • Complexity: • 1. May perform inefficiently when too mulch or too little work is sent to an under loaded processor • 2. In the worst case an update would require NlogN messages (dependent on network topology) • 3. Since ultimate destination of migrating tasks is not explicitly known , intermediate processors must be interrupted to do the migration • 4. Proximity map might change during a task’s migration altering its destination

  7. 3 3 2 2 3 Overloaded d d 0 1 Moderately Overloaded 1 Underloaded 1 2 2 3

  8. Sender Initiated Diffusion • Local, near- neighbor diffusion approach which employs overlapping balancing domains to achieve global balancing • balancing performed when a processor receives a load update message from a neighbor indicating that the neighbors load li < L low where L low is preset threshold • Average load in domain Lp _ k • Lp = 1 / (k+1) ( lp + S lk ) • k=1 • Profitability: Profitable if _ Lp – Lp > Lthreshold • Each neighbor assigned a weight hk depending on its load • the weights hk are summed to find the local deficiency Hp • The portion of processor p’s excess load that is apportioned to neighbor k is given by dk = ( lp – Lp) hk / Hp • Complexity 1. Number of messages for update = KN 2. Overhead incurred by each processor = K messages 3. Communication overhead for migration = N/2 k transfers

  9. 0 8 4 6 Average load L =10 Domain deficiency H = 20 Surplus load S = 21

  10. Receiver Initiated Diffusion • under loaded processors request load from overloaded processors • initiated by any processor whose load drops below a prespecified threshold Llow • processor will fulfill request only upto half of its current load. • underloaded processors take on majority of load balancing overhead • dk = ( lp – Lp) hk / Hp same as SID, except it is amount of load requested. • balancing activated when load drops below threshold and there are no outstanding requests. • Complexity • Num of messages for update = KN • Communication overhead for task migration = Nk messages + N/2 K transfers • (due to extra messages for requests) • As in SID, number of iterations to achieve global balancing is dependent on topology and application

  11. Hierarchical Balancing Method • processors in charge of balancing process at level li , receive load information from both lower level li-1 domains • size of balancing domains double from one level to the next • subtree load information is computed at intermediate nodes and propagated to the root • The absolute value of difference between the left domain LL and right domain LR is compared to Lthreshold • | LL – LR | > Lthreshold • Processors within the overloaded subtree , send a designated amount of load to matching neighbor in corresponding subtree • Complexity: • 1. Load transfer request messages = N/2 • 2. Total messages required = N(log N+1) • 3. Avg cost per processor = log N+1 sends and receives • 4. Cost at leaves = 1 send + log N receives • 5 . Cost at root = log N receives + N-1 sends + log N receives

  12. Dimension Exchange Method • small domains balanced first, then entire system is balanced • synchronized approach • in N processor hypercube, balancing performed iteratively in each logN dimensions • balancing initiated by processor with load that drops below threshold • Complexity • 1. Total communication overhead = 3N log N messages

  13. Summary of Comparison Analysis U = load update factor: if u = ½ then processor must send update messages whenever load has doubled or halved from last update

  14. Performance Analysis Graphs

  15. Speedup Vs Number of Processors

  16. Dynamic Load Balancing on Web Servers • load balancing is required to route requests among distributed web server nodes in a transparent way • this helps in improving throughput and provides high scalability and availability • user: one who accesses the information • client: a program, typically a web browser • client obtains IP address of a web server node through an address mapping request to the DNS server • there are intermediate name server, local gateways and browsers , that can cache the address mapping for sometime

  17. Requirements of the web server: • transparency • scalability • load balancing • availability • applicability to existing Web standards (backward compatibility) • geographic scalability (i.e., solutions applicable to both LAN and WAN distributed systems)

  18. Client –Based Approach • In this approach it is the client side itself that routes the request to one of the servers in the cluster. This can be done by the Web-browser or by the client-side proxy-server. • 1 . Web Clients • assume web clients know the existence of replicated servers of the web server system • based on protocol centered description • web client selects the node of a cluster , resolves the address and submits requests to selected node • Example: • 1. Netscape • * Picks random server i • * not scalable • 2. Smart Clients • * Java applet monitors node states and network delays • * scalable, but large network traffic

  19. Client –Based Approach-contd • Client Side Proxies • combined caching and server replication • Web Location and Information service can keep track of replicated URL addresses and route client requests appropriately • Advantages and Disadvantages: • -Scalable and high availability • -Limited applicability • -Lack of portability on the client side

  20. DNS –Based Approach • cluster DNS – routes requests to the corresponding server • transparency at URL level • through the translation process from the symbolic name to IP address , it can select any node of the cluster • DNS it also specifies, a validity period known as Time-to-Live, TTL • After expiration of TTL, address mapping request forwarded to cluster DNS • limited factors affecting DNS • * TTL does not work on browser caching • * no cooperative intermediate name servers • * can become potential bottleneck • Two DNS based System of algorithms • * Constant TTL Algorithms • * Adaptive TTL algorithms

  21. A DNS-based Web server cluster

  22. DNS-Based Approach

  23. Constant TTL Algorithms • classified based on system state information and constant TTL value • System Stateless Algorithms: - Round Robin DNS by NCSA • - load distribution not very balanced, overloaded server nodes • - ignores sever capacity and availability • Server State Based Algorithms: - simple feedback alarm mechanism - selects server with lightest load - limited applicability • Client State Based Algorithms - typical load that can come from each connected domain - Hidden Load , measure of average number of data requests sent from each domain to a Web site during the TTL caching period - geographical location of the client - Cisco DistributedDirector – takes into account relative client-to-server topological proximity, and client-to-server link latency - Internet2 Distributed Storage Infrastructureuses round trip delays • Server and Client State Based Algorithm -Distributed Director DNS - both server availability and client proximity

  24. Adaptive TTL Algorithm • By base of dynamic information from servers and/or clients to assign different TTL • Two step process • * DNS selects server node similar to hidden load weight algorithms • * DNS chooses appropriate value for the TTL period • TTL values inversely proportional to the domain request rate • popular domains have shorter TTL intervals • scalable from LAN to WAN distributed Web Server systems

  25. Dispatcher Based Approach • provides full control on client requests and masks the request routing among multiple servers • cluster has only one virtual IP address the IP address of the dispatcher • dispatcher identifies the servers through unique private IP addresses • Classes of routing • 1. Packet single-rewriting by the dispatcher • 2. Packet double-rewriting by the dispatcher • 3. Packet forwarding by the dispatcher • 4. HTTP redirection

  26. Packet Single Rewriting • dispatcher reroutes client-to-server packets by rewriting their IP address • requires modification of the kernel code of the servers, since IP address substitution occurs at TCP/IP level • Provides high system availability

  27. Packet Double Rewriting • -modification of all IP addresses, including that in the response packets carried out by dispatcher • two architectures based on this: • * Magicrouter (fast packet interposing where user level process,acting as a switchboard, intercepts client-to-server and server-to-client packets and modifies them) • * LocalDirector ( modifies IP address of client-server packets according to a dynamic mapping table)

  28. Packet Forwarding * forwards client packets to servers instead of rewriting IP address * Network Dispatcher - use MAC address - dispatcher and servers share same IP-SVA address - for WAN, two level dispatcher (first level packet rewriting) - transparent to both the client and server * ONE-IP address - publicizes the same secondary IP addresses of all Web-server nodes as IP-SVA of the Web-server cluster - routing based dispatching : destination server selected based on hash function - broadcast based dispatching: router broadcasts the packets to every server in the cluster - using hash function restricts dynamic load balancing - does not account for server heterogeneity

  29. HTTP Redirection • Distribute requests among web-servers through HTTP redirection mechanism • redirection transparent to user • Server State based dispatching • - each server periodically reports both the number of processes in its run queue and number of received requests per second • Location based dispatching • can be finely applied to LAN and WAN distributed Web Server Systems • duplicates the number of necessary TCP connections

  30. Server Based Approach • - uses two level dispatching mechanism • - cluster DNS assigns requests to a server • - server may redirect request to another server in the cluster • allows all servers to participate in load balancing (distributed) • Redirection is done in two ways • - HTTP redirection • - Packet redirection by packet rewriting

  31. HTTP Redirection by the Server

  32. Packet Redirection • transparent to client • Two balancing algorithms • use RR-DNS to schedule request (static routing) • periodic communication among servers about their current load

  33. Main Pros and Cons

  34. Performance of various distributed architectures • Exponential distribution model

  35. 2. Heavy-tailed distribution model

  36. Conclusions • consider performance constraints due to network bandwidth than server node capacity • account for network load as well as client proximity