GLane3D: Detecting Lanes with Graph of 3D Keypoints

Togg/Trutek AI Team
CVPR 2025
Model architecture of GLane3D

Figure: Overview of the GLane3D model architecture.

Abstract

Accurate and efficient lane detection in 3D space is essential for autonomous driving systems, where robust generalization is the foremost requirement for 3D lane detection algorithms. Considering the extensive variation in lane structures worldwide, achieving high generalization capacity is particularly challenging, as algorithms must accurately identify a wide variety of lane patterns worldwide. Traditional top-down approaches rely heavily on learning lane characteristics from training datasets, often struggling with lanes exhibiting previously unseen attributes. To address this generalization limitation, we propose a method that detects keypoints of lanes and subsequently predicts sequential connections between them to construct complete 3D lanes. Each key point is essential for maintaining lane continuity, and we predict multiple proposals per keypoint by allowing adjacent grids to predict the same keypoint using an offset mechanism. PointNMS is employed to eliminate overlapping proposal keypoints, reducing redundancy in the estimated BEV graph and minimizing computational overhead from connection estimations. Our model surpasses previous state-of-the-art methods on both the Apollo and OpenLane datasets, demonstrating superior F1 scores and a strong generalization capacity when models trained on OpenLane are evaluated on the Apollo dataset, compared to prior approaches.

Method Overview

Stage A
(a)
Stage B
(b)
Stage C
(c)
Stage D
(d)
  • a) Multiple Proposals: Multiple keypoint proposals are generated for each lane by selecting the top N anchor points with the highest scores from the predicted segmentation map 𝑀seg. These proposals are sampled within a lateral distance dₓ from the target lanes, increasing robustness to ensure that no keypoints are missed due to sparse or noisy predictions.
  • b) Keypoint Refinement: Each proposed keypoint is refined toward its corresponding location on the lane by applying a predicted lateral offset. This adjustment aligns proposals more precisely to the actual lane geometry, improving spatial accuracy and continuity along the lane.
  • c) Redundancy Suppression (PointNMS): To reduce redundancy and computational load, PointNMS is applied to suppress overlapping proposals. Only the top S keypoints with the highest confidence scores are retained, ensuring that the remaining keypoints are both distinct and reliable.
  • d) Connection Estimation: Directed connections between the retained keypoints are estimated to construct a lane graph. A relationship head predicts an adjacency matrix 𝑨, encoding which keypoints are connected in a sequential manner to represent continuous lane segments.
  • Lane Extraction: Final lanes are extracted by applying Dijkstra’s shortest path algorithm over the graph of keypoints. This graph-based decoding traces the most plausible paths between start and end keypoints, enabling robust reconstruction of complete lane topologies.

Quantitative Results @ 1.5m (OpenLane Dataset)

Method Backbone F1-Score ↑ X-error near (m) ↓ X-error far (m) ↓ Z-error near (m) ↓ Z-error far (m) ↓
PersFormerEffNet-B750.50.4850.5530.3640.431
BEV-LaneDetResNet-3458.40.3090.6590.2440.631
Anchor3DLaneEffNet-B353.70.2760.3110.1070.138
LATRResNet-5061.90.2190.2590.0750.104
LaneCPPEffNet-B760.30.2640.3100.0770.117
PVALaneResNet-5062.70.2320.2590.0920.118
PVALaneSwin-B63.40.2260.2570.0930.119
GLane3D-Lite (Ours)ResNet-1861.50.2210.2520.0730.101
GLane3D-Base (Ours)ResNet-5063.90.1930.2340.0650.090
GLane3D-Large (Ours)Swin-B66.00.1700.2030.0630.087

Quantitative Results @ 0.5m (OpenLane Dataset)

Method Backbone F1-Score ↑ X-error near (m) ↓ X-error far (m) ↓ Z-error near (m) ↓ Z-error far (m) ↓
PersFormerEffNet-B736.50.3430.2630.1610.115
Anchor3DLaneEffNet-B334.90.3440.2640.1810.134
PersFormerResNet-5043.20.2290.2450.0780.106
LATRResNet-5054.00.1710.2010.0720.099
DV-3DLane(Camera)ResNet-3452.90.1730.2120.0690.098
GLane3D-Lite (Ours)ResNet-1853.80.1820.2060.0700.095
GLane3D-Base (Ours)ResNet-5057.90.1570.1790.0670.087
GLane3D-Large (Ours)Swin-B61.10.1420.1670.0610.084

Cross-Dataset Evaluation on Apollo Balanced Scenes

Table: Cross-Dataset Evaluation on Apollo Balanced Scenes
Dₜ Methods F1 X-error (m) ↓ Z-error (m) ↓
Near Far Near Far
1.5mPersFormer53.20.4070.8130.1220.453
LATR34.30.3270.7370.1420.500
GLane3D58.90.2890.7010.0860.479
0.5mPersFormer17.40.2460.3810.0980.214
LATR19.00.2010.3130.1160.220
GLane3D42.60.1620.2960.0630.198
Apollo Cross-Dataset Comparison
Figure: Visual comparison on Apollo dataset.

Apollo Dataset Results

Table: Quantitative Results on the Apollo 3D Synthetic Dataset
Subset Methods Backbone F1-Score (%) ↑ AP (%) ↑ X-error near (m) ↓ X-error far (m) ↓ Z-error near (m) ↓ Z-error far (m) ↓
Balanced ScenePersFormerEffNet-B792.9-0.0540.3560.0100.234
BEVLaneDetResNet-3496.9-0.0160.2420.0200.216
LaneCPPEffNet-B797.499.50.0300.2770.0110.206
LATRResNet-5096.897.90.0220.2530.0070.202
DV-3DLaneResNet-5096.497.60.0460.2990.0160.213
GLane3D (Ours)ResNet-5098.198.80.0210.2500.0070.213
Rare ScenePersFormerEffNet-B787.5-0.1070.7820.0240.602
BEVLaneDetResNet-3497.6-0.0310.5940.0400.556
LaneCPPEffNet-B796.298.60.0730.6510.0230.543
LATRResNet-5096.197.30.0500.6000.0150.532
DV-3DLaneResNet-5095.597.20.0710.6640.0250.568
GLane3D (Ours)ResNet-5098.499.10.0440.6210.0230.566
Visual VariationsPersFormerEffNet-B789.6-0.0740.4300.0150.266
BEVLaneDetResNet-3495.0-0.0270.3200.0310.256
LaneCPPEffNet-B790.493.70.0540.3270.0200.222
LATRResNet-5095.196.60.0450.3150.0160.228
DV-3DLaneResNet-5091.393.40.0950.4170.0400.320
GLane3D (Ours)ResNet-5092.794.80.0460.3640.0200.317

OpenLane Qualitative Results

OpenLane Qualitative Results

More Cross-Dataset Qualitative Results

More Cross-Dataset Qualitative Results

BibTeX

@inproceedings{ozturk2025glane3d,
          title={GLane3D: Detecting Lanes with Graph of 3D Keypoints},
          author={{\"O}zt{\"u}rk, Halil {\.I}brahim and Kalfao{\u{g}}lu, Muhammet Esat and Kilinc, Ozsel},
          booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
          pages={27508--27518},
          year={2025}
        }