The site is for the paper appear in CVPR2023 Photogrammetric Computer Vision Workshop.
The 8 minute presentation can be found here.
I also make a poster, but I did not participate in person because of visa issue, the poster can be found here.
In the paper, we use two dataset with high dense LiDAR.
dataset | Image GSD(cm) | LiDAR density( |
---|---|---|
DublinCity | 3.4 | 250-348 |
Toulouse2020 | 5 | 50 |
DublinCity is an open dataset, the original aerial and LiDAR point cloud can be downloaded, the origin dataset is very large.
Origin DublinCity coverage |
Because the origin dataset use Terrasolid for the orientation, the origin orientation is not accurate, so the registration step is mondatory, the experiment area is shown in
DublinCity experiment coverage |
Toulouse2020 is a dataset collected by IGN (French Mapping Agency) for AI4GEO project, the origin dataset is very large.
Origin Toulouse2020 coverage |
Because the whole area is too large, in order to registration the image and LiDAR, we select the center city of Toulouse, the experiment area is shown in
Toulouse2020 experiment coverage |
The data is generated using our previous work, the detail introduction can also be found on Github. The training and testing splitting is :
DublinCity | Toulouse2020 |
We will also publish the dataset for public use, because the original dataset is too large, at present, we will only publish the used training and testing dataset in the paper.
All the dataset are host by Zenodo, the download site is here.
In the paper, we propose a method based on PSMNet1, based on the stereo work, like the work2 in computer vision, we use the TIN expansion for remote sensing data. The newwork is shown :
PSMNet-FusionX3 |
In the experiment, we compare our method with GCNet3, PSMNet1, GuidedStereo4, and GCNet-CCVNorm5. There is no official code for GCNet, so the GCNet result is from the code of GCNet-CCVNorm.
PSMNet-FusionX3 and the methods compared in the paper are all provided here.
The pre-trained models will be also available.
For the other methods, because our dataset is different from the computer vision dataset, we will also put the revised code in this repository.
We revise the official code to adopt to the remote sensing dataset, the detail can be found in folder.
We revise the official code to adopt to the remote sensing dataset, the detail can be found in folder.
We also release the code of our method, the detail can be found in folder.
Because the input guidance is sampled from the origin dense disparity, and for the TIN based interpolation, this is processed before training, the detail can be found in folder.
@inproceedings{wu2023psmnet, title={PSMNet-FusionX3: LiDAR-Guided Deep Learning Stereo Dense Matching on Aerial Images}, author={Wu, Teng and Vallet, Bruno and Pierrot-Deseilligny, Marc}, booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition}, pages={6526--6535}, year={2023} }
If you think you have any problem, contact Teng Wu [email protected]
Footnotes
-
Chang, Jia-Ren, and Yong-Sheng Chen. "Pyramid stereo matching network." Proceedings of the IEEE conference on computer vision and pattern recognition. 2018. ↩ ↩2
-
Huang, Yu-Kai, et al. "S3: Learnable sparse signal superdensity for guided depth estimation." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021. ↩
-
Kendall, Alex, et al. "End-to-end learning of geometry and context for deep stereo regression." Proceedings of the IEEE international conference on computer vision. 2017. ↩
-
Poggi, Matteo, et al. "Guided stereo matching." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2019. ↩
-
Wang, Tsun-Hsuan, et al. "3d lidar and stereo fusion using stereo matching network with conditional cost volume normalization." 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2019. ↩