This paper investigates incremental multiagent learning in static or evolving structured networks. Learning examples are incrementally distributed among the agents, and the objective is to build a common hypothesis that is consistent with all the examples present in the system, despite communication constraints. Recently, a first mechanism was proposed to deal with static networks, but its accuracy was reduced in some topologies. We propose here several possible improvements of this mechanism, whose different behaviors with respect to some efficiency requirements (redundancy, computational cost and communicational cost) are experimentally investigated. Then, we provide an experimental analysis of some variants for evolving networks.