This paper studies two problems that arise in distributed computing. We deal with these problems from a game theoretical approach. We are interested in the convergence to the Nash equilibrium of algorithms based on the best reply strategy in a special case of linear costs. We present three specific types of algorithm that converge to the equilibrium. In our first model, composed of two processors, the convergence is established through monotonicity of the sequence of updates generated by each of the three algorithms. In the second model, made up of N processors, the convergence is due to the contraction of the algorithms.