首页 | 本学科首页   官方微博 | 高级检索  
     检索      

基于膨胀卷积的多模态融合视线估计
引用本文:罗元,陈顺,张毅.基于膨胀卷积的多模态融合视线估计[J].重庆邮电大学学报(自然科学版),2021,33(4):637-644.
作者姓名:罗元  陈顺  张毅
作者单位:重庆邮电大学 光电工程学院,重庆400065;重庆邮电大学 先进制造学院,重庆400065
基金项目:国家自然科学基金(61801061)
摘    要:基于表观的视线估计方法主要是在二维的三原色(red green blue,RGB)图像上进行,当头部在自由运动时视线估计精度较低,且目前基于卷积神经网络的表观视线估计都普遍使用池化来增大特征图中像素点的感受野,导致了特征图的信息损失,提出一种基于膨胀卷积神经网络的多模态融合视线估计模型.在该模型中,利用膨胀卷积设计了一种叫GENet(gaze estimation network)的网络提取眼睛的RGB和深度图像的特征图,并利用卷积神经网络的全连接层自动融合头部姿态和2种图像的特征图,从而进行视线估计.实验部分在公开数据集Eyediap上验证了设计的模型,并将设计的模型同其他视线估计模型进行比较.实验结果表明,提出的视线估计模型可以在自由的头部运动下准确地估计视线方向.

关 键 词:视线估计  膨胀卷积  三原色(RGB)图像  深度图像
收稿时间:2019/10/21 0:00:00
修稿时间:2021/3/11 0:00:00

Dilated convolution-based gaze estimation via multimodel fusion
LUO Yuan,CHEN Shun,ZHANG Yi.Dilated convolution-based gaze estimation via multimodel fusion[J].Journal of Chongqing University of Posts and Telecommunications,2021,33(4):637-644.
Authors:LUO Yuan  CHEN Shun  ZHANG Yi
Institution:College of Optoelectronic Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, P. R. China; College of Advanced Manufacturing, Chongqing University of Posts and Telecommunications, Chongqing, 400065, P. R. China
Abstract:Appearance-based gaze estimation is mainly performed on two-dimensional red green blue(RGB)images. When the head is free to move, the gaze estimation accuracy is low, and the current appearance-based gaze estimation via the convolutional neural network generally uses pooling to increase the receptive field of pixel points in the feature map, resulting in information loss of the feature map. In this paper, a dilated convolution-based gaze estimation via multimodel fusion is proposed. In this model, a feature map of RGB and depth images of the eye are extracted by a network called gaze estimation network (GENet) using dilated convolution, and the head pose and the feature map of the two images are automatically fused using the fully connected layer of the convolutional neural network to make gaze estimation. The experimental validated the designed model on the public dataset Eyediap and compared the designed model with other gaze estimation models. The experimental results show that the proposed gaze estimation model can accurately estimate the direction of gaze under free head motion.
Keywords:gaze estimation  dilated convolution  red green blue(RGB)image  depth image
本文献已被 万方数据 等数据库收录!
点击此处可从《重庆邮电大学学报(自然科学版)》浏览原始摘要信息
点击此处可从《重庆邮电大学学报(自然科学版)》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号