TY - JOUR
T1 - HWLane
T2 - HW-Transformer for Lane Detection
AU - Zhao, Jing
AU - Qiu, Zengyu
AU - Hu, Huiqin
AU - Sun, Shiliang
N1 - Publisher Copyright:
© 2000-2011 IEEE.
PY - 2024
Y1 - 2024
N2 - Lane detection is one of the most fundamental tasks in autonomous driving perception, but it still faces many challenges in some special driving scenarios. For example, in dazzling light, crowded roads, etc., lane detection is very dependent on surrounding visual cues. Previous segmentation-based lane detection methods have not paid enough attention to the surrounding visual range, resulting in poor performance. In this paper, we design a novel lane detection network namely HW-Transformer, which is based on row and column multi-head self-Attention. It restricts the attention only to their respective rows and columns, and transfers information across rows and columns by intersection features. In this way, the attention to the visual range around the lane is greatly expanded, and the communication of global information can be achieved through intersecting features. In addition, we further propose a self-Attention knowledge distillation (SAKD) method for the Transformer model, where higher-level attention guides lower-level attention to learn. SAKD not only helps to improve the performance of lane detection, but also has universality in better learning semantic features from general images. Extensive experiments on BDD100K, TuSimple, CULane, and VIL100 datasets demonstrate that our method outperforms the state-of-The-Art segmentation-based lane detection methods. We also apply the proposed SAKD to DeiT-Tiny, and it achieves 1.51 Top-1 accuracy improvement on ImageNet-1K dataset. Our code will be available at https://github.com/Cuibaby/HWLane.
AB - Lane detection is one of the most fundamental tasks in autonomous driving perception, but it still faces many challenges in some special driving scenarios. For example, in dazzling light, crowded roads, etc., lane detection is very dependent on surrounding visual cues. Previous segmentation-based lane detection methods have not paid enough attention to the surrounding visual range, resulting in poor performance. In this paper, we design a novel lane detection network namely HW-Transformer, which is based on row and column multi-head self-Attention. It restricts the attention only to their respective rows and columns, and transfers information across rows and columns by intersection features. In this way, the attention to the visual range around the lane is greatly expanded, and the communication of global information can be achieved through intersecting features. In addition, we further propose a self-Attention knowledge distillation (SAKD) method for the Transformer model, where higher-level attention guides lower-level attention to learn. SAKD not only helps to improve the performance of lane detection, but also has universality in better learning semantic features from general images. Extensive experiments on BDD100K, TuSimple, CULane, and VIL100 datasets demonstrate that our method outperforms the state-of-The-Art segmentation-based lane detection methods. We also apply the proposed SAKD to DeiT-Tiny, and it achieves 1.51 Top-1 accuracy improvement on ImageNet-1K dataset. Our code will be available at https://github.com/Cuibaby/HWLane.
KW - Deep learning
KW - lane detection
KW - self-knowledge distillation
KW - transformer
UR - https://www.scopus.com/pages/publications/85191311649
U2 - 10.1109/TITS.2024.3386531
DO - 10.1109/TITS.2024.3386531
M3 - 文章
AN - SCOPUS:85191311649
SN - 1524-9050
VL - 25
SP - 9321
EP - 9331
JO - IEEE Transactions on Intelligent Transportation Systems
JF - IEEE Transactions on Intelligent Transportation Systems
IS - 8
ER -