Learn More
A linear time-delay system is used to model load balancing in a cluster of computer nodes used for parallel computations. The linear model is analyzed for stability in terms of the delays in the transfer of information between nodes and the gains in the load balancing algorithm. This model is compared with an experimental implementation of the algorithm on(More)
A linear time-delay system is proposed to model load balancing in a cluster of computer nodes used for parallel computations. The linear model is analyzed for stability in terms of the delays in the transfer of information between nodes and the gains in the load balancing algorithm. This model is compared with an experimental implementation of the algorithm(More)
  • 1