应用科学学报 ›› 2024, Vol. 42 ›› Issue (1): 103-118.doi: 10.3969/j.issn.0255-8297.2024.01.009

• 计算机应用专辑 • 上一篇    下一篇

基于车辆外观特征和帧间光流的目标跟踪算法

李绍骞1, 程鑫1, 周经美2, 赵祥模1   

  1. 1. 长安大学信息工程学院, 陕西 西安 710064;
    2. 长安大学电子与控制工程学院, 陕西 西安 710064
  • 收稿日期:2023-06-29 出版日期:2024-01-30 发布日期:2024-02-02
  • 通信作者: 程鑫,副教授,研究方向为深度学习与计算机视觉、人工智能与车联网技术。E-mail:xincheng@chd.edu.cn E-mail:xincheng@chd.edu.cn
  • 基金资助:
    国家重点研发计划(No. 2021YFB2501200);国家自然科学基金(No. 52102452);陕西省重点研发计划(No. 2023-YBGY-119);陕西省自然科学基础研究计划面上项目(No. 2023-JC-YB-523);陕西省创新能力支撑计划(No. 2022KJXX-02);陕西省交通运输厅交通科研项目(No. 21-05X);陕西省高校科协青年人才托举计划(No. 20210122);中央高校基本科研业务费专项资金项目(No. 300102242203)资助

Object Tracking Algorithm Based on Vehicle Appearance Features and Inter-frame Optical Flow

LI Shaoqian1, CHENG Xin1, ZHOU Jingmei2, ZHAO Xiangmo1   

  1. 1. College of Information Engineering, Chang'an University, Xi'an 710064, Shaanxi, China;
    2. College of Electronic and Control Engineering, Chang'an University, Xi'an 710064, Shaanxi, China
  • Received:2023-06-29 Online:2024-01-30 Published:2024-02-02

摘要: 在复杂道路场景下,车辆目标之间频繁遮挡、车辆目标之间相似的外观、目标整个运动过程中采用静态预设参数都会引起跟踪准确率下降等问题。该文提出了一种基于车辆外观特征和帧间光流的目标跟踪算法。首先,通过YOLOv5算法中的YOLOv5x网络模型获得车辆目标框的位置信息;其次,利用RAFT(recurrent all-pairs field transforms for opticalflow)算法计算当前帧和前一帧之间的光流,并根据得到的位置信息对光流图进行裁剪;最后,在卡尔曼滤波过程中利用帧间光流进行补偿得到更精确的运动状态信息,并利用车辆外观特征和交并比特征完成轨迹匹配。实验结果表明,基于车辆外观特征和帧间光流的目标跟踪算法在MOT16数据集上表现良好,相较于跟踪算法DeepSORT,成功跟踪帧数占比提高了1.6%,跟踪准确度提升了1.3%,跟踪精度提升了0.6%,改进的车辆外观特征提取模型准确率在训练集和验证集上分别提高了1.7%、6.3%。因此,基于高精度的车辆外观特征模型结合关联帧间光流的运动状态信息能够有效实现交通场景下的车辆目标跟踪。

关键词: 目标跟踪, 车辆外观特征, 帧间光流, 卡尔曼滤波

Abstract: In complex road scenes, frequent occlusions and similar appearances between vehicle targets, coupled with the use of static preset parameters used throughout the entire movement of the targets collectively contribute to a decline in tracking accuracy. This paper proposes an object tracking algorithm based on vehicle appearance features and inter-frame optical flow. Firstly, the position information of the vehicle target frame is obtained through the YOLOv5x network model. Secondly, the optical flow between the current frame and the previous frame is calculated using the RAFT (recurrent all-pairs field transforms for optical flow) algorithm, and the optical flow map is clipped according to the obtained position information. Finally, in the process of Kalman filtering, inter-frame optical flow is used to compensate for more accurate motion state information, while vehicle appearance features and intersection over union (IOU) features are used to complete trajectory matching. Experimental results show that the tracking algorithm correlating inter-frame optical flow performs well on the MOT16 data set. Compared with simple online and realtime tracking with a deep association metric (DeepSORT), mostly tracked trajectories (MT) has increased by 1.6%, multiple object tracking accuracy (MOTA) has increased by 1.3%, and multiple object tracking precision (MOTP) has increased by 0.6%. The accuracy of the improved vehicle appearance feature extraction model has been improved by 1.7% and 6.3% on the training and verification sets, respectively. Consequently, leveraging the high-precision vehicle appearance feature model and motion state information from the associated inter-frame optical flow enables effective vehicle target tracking in traffic scenes.

Key words: object tracking, vehicle exterior features, inter-frame optical flow, Kalman filter

中图分类号: