The original paper is in English. Non-English content has been machine-translated and may contain typographical errors or mistranslations. ex. Some numerals are expressed as "XNUMX".
Copyrights notice
The original paper is in English. Non-English content has been machine-translated and may contain typographical errors or mistranslations. Copyrights notice
Ces dernières années, l'attention visuelle du conducteur a été activement étudiée pour la technologie d'automatisation de la conduite. Cependant, le nombre de modèles est faible pour percevoir une compréhension approfondie de l'attention du conducteur à différents moments. Tous les modèles d'attention traitent des représentations d'images à plusieurs niveaux par un réseau à deux flux/multi-flux, augmentant le coût de calcul en raison d'un incrément de paramètres du modèle. Cependant, la représentation d'images à plusieurs niveaux, telle que le flux optique, joue un rôle essentiel dans les tâches impliquant des vidéos. Par conséquent, afin de réduire le coût de calcul d'un réseau à deux flux et d'utiliser une représentation d'image à plusieurs niveaux, ce travail propose un modèle d'attention visuelle du conducteur à flux unique pour une situation critique. L’expérience a été menée à l’aide d’un ensemble de données de conduite critique accessible au public nommé BDD-A. Les résultats qualitatifs confirment l'efficacité du modèle proposé. De plus, les résultats quantitatifs soulignent que le modèle proposé surpasse les modèles d'attention visuelle de pointe selon CC et SIM. Des études d'ablation approfondies vérifient la présence du flux optique dans le modèle, la position du flux optique dans le réseau spatial, les couches de convolution pour traiter le flux optique et le coût de calcul par rapport à un modèle à deux flux.
Rebeka SULTANA
Shizuoka University
Gosuke OHASHI
Shizuoka University
The copyright of the original papers published on this site belongs to IEICE. Unauthorized use of the original or translated papers is prohibited. See IEICE Provisions on Copyright for details.
Copier
Rebeka SULTANA, Gosuke OHASHI, "Prediction of Driver's Visual Attention in Critical Moment Using Optical Flow" in IEICE TRANSACTIONS on Information,
vol. E106-D, no. 5, pp. 1018-1026, May 2023, doi: 10.1587/transinf.2022EDP7146.
Abstract: In recent years, driver's visual attention has been actively studied for driving automation technology. However, the number of models is few to perceive an insight understanding of driver's attention in various moments. All attention models process multi-level image representations by a two-stream/multi-stream network, increasing the computational cost due to an increment of model parameters. However, multi-level image representation such as optical flow plays a vital role in tasks involving videos. Therefore, to reduce the computational cost of a two-stream network and use multi-level image representation, this work proposes a single stream driver's visual attention model for a critical situation. The experiment was conducted using a publicly available critical driving dataset named BDD-A. Qualitative results confirm the effectiveness of the proposed model. Moreover, quantitative results highlight that the proposed model outperforms state-of-the-art visual attention models according to CC and SIM. Extensive ablation studies verify the presence of optical flow in the model, the position of optical flow in the spatial network, the convolution layers to process optical flow, and the computational cost compared to a two-stream model.
URL: https://global.ieice.org/en_transactions/information/10.1587/transinf.2022EDP7146/_p
Copier
@ARTICLE{e106-d_5_1018,
author={Rebeka SULTANA, Gosuke OHASHI, },
journal={IEICE TRANSACTIONS on Information},
title={Prediction of Driver's Visual Attention in Critical Moment Using Optical Flow},
year={2023},
volume={E106-D},
number={5},
pages={1018-1026},
abstract={In recent years, driver's visual attention has been actively studied for driving automation technology. However, the number of models is few to perceive an insight understanding of driver's attention in various moments. All attention models process multi-level image representations by a two-stream/multi-stream network, increasing the computational cost due to an increment of model parameters. However, multi-level image representation such as optical flow plays a vital role in tasks involving videos. Therefore, to reduce the computational cost of a two-stream network and use multi-level image representation, this work proposes a single stream driver's visual attention model for a critical situation. The experiment was conducted using a publicly available critical driving dataset named BDD-A. Qualitative results confirm the effectiveness of the proposed model. Moreover, quantitative results highlight that the proposed model outperforms state-of-the-art visual attention models according to CC and SIM. Extensive ablation studies verify the presence of optical flow in the model, the position of optical flow in the spatial network, the convolution layers to process optical flow, and the computational cost compared to a two-stream model.},
keywords={},
doi={10.1587/transinf.2022EDP7146},
ISSN={1745-1361},
month={May},}
Copier
TY - JOUR
TI - Prediction of Driver's Visual Attention in Critical Moment Using Optical Flow
T2 - IEICE TRANSACTIONS on Information
SP - 1018
EP - 1026
AU - Rebeka SULTANA
AU - Gosuke OHASHI
PY - 2023
DO - 10.1587/transinf.2022EDP7146
JO - IEICE TRANSACTIONS on Information
SN - 1745-1361
VL - E106-D
IS - 5
JA - IEICE TRANSACTIONS on Information
Y1 - May 2023
AB - In recent years, driver's visual attention has been actively studied for driving automation technology. However, the number of models is few to perceive an insight understanding of driver's attention in various moments. All attention models process multi-level image representations by a two-stream/multi-stream network, increasing the computational cost due to an increment of model parameters. However, multi-level image representation such as optical flow plays a vital role in tasks involving videos. Therefore, to reduce the computational cost of a two-stream network and use multi-level image representation, this work proposes a single stream driver's visual attention model for a critical situation. The experiment was conducted using a publicly available critical driving dataset named BDD-A. Qualitative results confirm the effectiveness of the proposed model. Moreover, quantitative results highlight that the proposed model outperforms state-of-the-art visual attention models according to CC and SIM. Extensive ablation studies verify the presence of optical flow in the model, the position of optical flow in the spatial network, the convolution layers to process optical flow, and the computational cost compared to a two-stream model.
ER -