The invention discloses a collision avoidance 
planning method for mobile robots based on deep 
reinforcement learning in a dynamic environment, and belongs to the technical field of 
mobile robot navigation. The method of the invention includes the following steps of: collecting 
raw data through a 
laser rangefinder, 
processing the 
raw data as input of a neural network, and building an LSTM neural network; through an A3C 
algorithm, outputting corresponding parameters by the neural network, and 
processing the corresponding parameters to obtain the action of each step of the 
robot. The scheme of the invention does not need to model the environment, is more suitable for an unknown obstacle environment, adopts an actor-critic framework and a temporal difference 
algorithm, is more suitable for a continuous 
motion space while realizing low variance, and realizes the effect of learning while training. The scheme of the invention designs the continuous 
motion space with a heading angle limitationand uses 4 threads for 
parallel learning and training, so that compared with general deep 
reinforcement learning methods, the learning and 
training time is greatly improved, the sample correlation isreduced, the high utilization of exploration spaces and the diversity of exploration strategies are guaranteed, and thus the 
algorithm convergence, stability and the success rate of 
obstacle avoidance can be improved.