[Objective] With the rapid popularization of new energy vehicles and the vigorous development of autonomous driving technology, 3D object detection algorithms play a pivotal role in real road scenes. The LiDAR point clouds contain precise position and geometric structure information of the object, which can accurately describe the 3D space position of the target. Moreover, LiDAR makes environmental perception and route planning of unmanned vehicles a reality. However, cars in real scenes often fall into complex and difficult situations, such as occlusion and truncation of objects, which contribute to highly sparse clouds and incomplete contours. Therefore, the effective use of disordered and unevenly distributed point clouds for accurate 3D object detection has important research significance and practical value for the safety of autonomous driving. [Methods] This paper uses LiDAR point clouds in an autonomous driving scene to conduct in-depth research on a high-performance 3D object detection algorithm based on deep learning. The 3D object detection algorithm based on multibranch feature fusion (PSANet) is designed to improve the capacity and viability of autonomous driving technology. After the disordered point clouds are divided into regular voxels, the voxel feature coding module and convolutional neural network are used to learn the voxel features, and the sparse 3D data are compressed into a dense 2D bird's-eye view. Furthermore, the multiscale bird's-eye view features are deeply fused through the coarse and fine branches of the 2D backbone network. The splitting and aggregation feature pyramid module in the fine branch splits and aggregates the bird's-eye view features at different levels and realizes the deep fusion of semantic information, texture information, and context information of multiscale features to obtain more expressive features. The multiscale features in the coarse branch are fused after transposed convolution, and the precise original spatial location information is preserved. After the feature extraction of coarse and fine branches, the element-wise addition method can obtain more accurate and complete features for object classification, position regression, and orientation prediction. [Results] The experimental results on the KITTI dataset show that the average precision of PSANet in 3D object detection and bird's-eye view object detection tasks reach 81.72% and 88.25%, respectively. The inference speed on a single GTX 1080Ti GPU can reach 24 frames per second, and it shows strong robustness in complex scenes. Compared with the two-stage target detection algorithms MV3D, AVOD-FPN, F-PointNet, and IPOD, the average accuracy of 3D target detection of this algorithm increased by 18.21%, 5.89%, 8.94%, and 3.12%, respectively. Compared with the one-stage target detection algorithms VoxelNet, SECOND, PointPillars, and VoTr-SSD, the average accuracy of 3D target detection of this algorithm increased by 11.63%, 4.05%, 3.19%, 0.98%, 1.16%, and 0.7%, respectively. The detection speed of this algorithm improved by 14 frames per second compared with the two-stage algorithm PointRCNN with similar accuracy. [Conclusions] In comparison with other advanced algorithms, this algorithm exhibits strong performance, in which it can better balance the accuracy and speed of target detection in autonomous driving scenarios. Higher accuracy is required in the algorithm, whether one- or two-stage, for novel applications. The method with the highest efficiency must be used in an autonomous car. [ABSTRACT FROM AUTHOR]