This research presents a navigation and obstacle avoidance policy network for multi-robot systems using deep reinforcement learning. The network is first designed and trained for a dual-robot setup. By incorporating nonholonomic constraints and priority rules, reinforcement learning is used to train the network with respect to the kinematics of mobile robots, enabling effective navigation and collision avoidance. An innovative expansion architecture is introduced, leveraging the social-force model to extend the dual-robot policy to multi-robot scenarios with moderate computational cost. Although the network is trained in an open environment, it can be applied to general map environments by using virtual robots to simulate walls and compartments. Simulations and indoor experiments validate the feasibility and performance of the proposed multi-robot navigation and obstacle avoidance policy.
