首页 | 本学科首页   官方微博 | 高级检索  
     检索      

深度学习与一致性表示空间学习的跨媒体检索
引用本文:邹辉,杜吉祥,翟传敏,王靖.深度学习与一致性表示空间学习的跨媒体检索[J].华侨大学学报(自然科学版),2018,0(1):127-132.
作者姓名:邹辉  杜吉祥  翟传敏  王靖
作者单位:华侨大学 计算机科学与技术学院, 福建 厦门 361021
摘    要:提出一种基于深度学习与一致性表示空间学习的方法,针对图像与文本2种模态,分别采用卷积神经网络模型和潜在狄利克雷分布算法学习图像的深度特征和文档的主题概率分布;通过一个概率模型将两个高度异构的向量空间非线性映射到一个一致性表示空间;采用中心相关性算法计算不同模态信息在此空间的距离.在Wikipedia Dataset上的实验结果表明:在单模态输入检索中,文中方法的平均准确率为38.43%,相比于其他方法有明显提高.

关 键 词:跨模态  跨媒体  深度学习  卷积神经网络  一致性表示空间  中心相关性

Cross-Modal Multimedia Retrieval Based Deep Learning and Shared Representation Space Learning
ZOU Hui,DU Jixiang,ZHAI Chuanmin,WANG Jing.Cross-Modal Multimedia Retrieval Based Deep Learning and Shared Representation Space Learning[J].Journal of Huaqiao University(Natural Science),2018,0(1):127-132.
Authors:ZOU Hui  DU Jixiang  ZHAI Chuanmin  WANG Jing
Institution:College of Computer Science and Technology, Huaqiao University, Xiamen 361021, China
Abstract:A new learning method based deep learning and shared representation space learning is proposed in this paper. Using image and text as an example, we learn the deep learning features of images by convolution neural networks, and learn the text topic probability distribution by a latent Dirichlet allocation model respectively. Then nonlinear mapping the two features spaces into a shared presentation space by a probability model. At last, we adopt centered correlation to measure the distance between them. The experimental results in the Wikipedia Dataset show that our approach is better than that of the similar experiments for single mode input retrieval in recent years and its mean average precision reaches 38.43%.
Keywords:cross-modal  cross-media  deep learning  convolution neural networks  shared presentation space  centered correlation
本文献已被 CNKI 等数据库收录!
点击此处可从《华侨大学学报(自然科学版)》浏览原始摘要信息
点击此处可从《华侨大学学报(自然科学版)》下载免费的PDF全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号