Our research interest lies in studing how local strategies about partner agent selection using reinforcement learning with variable exploitation-versus- exploration parameters influence the overall efficiency of multi-agent systems (MAS). An agent often has to select appropriate agents to assign tasks that are not locally executable. Unfortunately no agent in an open environment can understand the all states of all agents, so this selection must be done according to local information. In this paper we investigate how the overall performance of MAS is affected by their individual learning parameters for adaptive partner selections for collaboration. We show experimental results using simulation and discuss why the overall performance of MAS varies.