• Laser & Optoelectronics Progress
  • Vol. 62, Issue 8, 0815012 (2025)
Fanna Meng1,*, ZouYongjia1, Yang Cao1, Jin Lü2, and Hongfei Yu1
Author Affiliations
  • 1School of Artificial Intelligence and Software, Liaoning Petrochemical University, Fushun 113000, Liaoning , China
  • 2Neusoft Reach Automotive Technology (Shenyang) Co., Ltd., Shenyang 110179, Liaoning , China
  • show less
    DOI: 10.3788/LOP241894 Cite this Article Set citation alerts
    Fanna Meng, ZouYongjia, Yang Cao, Jin Lü, Hongfei Yu. Stereo Matching Algorithm Based on Adaptive Spatial Convolution[J]. Laser & Optoelectronics Progress, 2025, 62(8): 0815012 Copy Citation Text show less

    Abstract

    Stereo matching, a significant research focus in computer vision, has wide-ranging applications in fields such as autonomous driving, medical imaging, and robotic navigation. To address the issue of poor matching performance in ill-posed regions within image sequences, this paper presents a stereo matching algorithm based on adaptive spatial convolution. Initially, an adaptive spatial convolution block is incorporated into the context network. Through weighted aggregation of multiple convolution kernel responses, the context network's capability to capture pathological regions in complex scenes is enhanced, achieving accurate feature representation. Attention maps of the input features are then obtained along the channel dimension. Subsequently, the algorithm uses a multiscale gated recurrent unit (GRU) network structure to optimize the initial disparity results, and the attention maps generated by adaptive spatial convolution are employed to weight the GRU's output, effectively suppressing noise and further improving the accuracy of disparity estimation. Experimental results show that the proposed algorithm achieves an average endpoint error of 0.46 pixel on the Scene Flow dataset, reducing the error by 14.81% compared to benchmark methods. On the KITTI dataset, it achieves a 3-pixel error rate of 1.40% across all regions, a 15.66% reduction compared to benchmark methods, and delivers superior disparity estimation performance. Notably, in ill-posed scenarios such as occluded or reflective regions, the algorithm effectively retains detailed image features.