Multimedia Questions and Answering Using Web Data Mining
Multimedia Questions and Answering Using Web Data Mining
I. INTRODUCTION
The amount of information on the web has increase year to
year with content covering almost any topic. As a result, when
looking for information, user becomes overloaded to find the
correct information from the current search engine. Users
usually have to painstakingly browse through a long list of
results to look for a precise answer [1]. Therefore questions
answering system solve these problems. It avoids the
painstaking browsing the vast quantity of information returned
by the search engines for the correct answers.
QA only focus on the textual data. Therefore its time to
extend the concept of text to multimedia data. Multimedia
data are helpful to user to quickly understand the content of
information. Multimedia questions answering (MMQA)
provides the textual answers along with the media
format(image and video) according to the questions.
Multimedia answers are more helpful for some questions
likeWhat are the steps to download a Firefox browser. In
this type of questions if multimedia videos are available users
quickly understand the answers. Textual answers cannot give
more information and user gets painstaking to understand the
RELATED WORK
ASPECTS OF MMQA
[5] Y.-S. Lee, Y.-C. Wu, and J.-C. Yang, Bvideoqa : Online English/Chinese
bilingual video question answering, Amer. Soc. Inf. Sci. Technol., vol.
60, no. 3, pp. 509-525, 2009.
C. Presenting Answers
Traditional system presents results using a sorted list of
descending relevancy. In the traditional system, the list of
related document is search from the web according to the
questions, after retrieving of the documents, the top related
documents are selected and analysis the questions then present
a answer according to the questions. In multimedia question
answering (MMQA) can use semantic summarization to
present an answer by summarizing the retrieved potential
answers from various sources (text, image, audio, video, or a
hybrid) at the semantic level.
IV.
CONCLUSION
[4] Y.-C. Wu, C.-H. Chang, and Y.-S. Lee, Cross-Language Video
Questions/Answering System, in Proc. IEEE Int. Symp. Multimedia
Software Engineering, 2004, pp. 294-301.
[6] Y.-C. Wu and J.-C. Yang, A robust passage retrieval algorithm for video
question answering , IEEE Trans. Circuits Syst. Video Technol., vol. 18,
no. 10, pp. 1411-1421, 2008.
[7] T. Yeh, J.J. Lee, and T. Darrell, Photo-Based Question Answering,
Proc. 16th ACM IntI Conf. Multimedia, ACM Press, 2008, pp. 389-398.
[8] Trec: The Text Retrieval Conf. [Online]. Available: http://trec.nist.gov/.
[9] S. A. Quarteroni and S. Manandhar, Designing an interactive open
domain question answering system, J. Natural Lang. Eng., vol. 15, no. 1,
pp. 73-95, 2008.
[10] D. Molla and J.L. Vicedo, Question answering in restricted domains: An
overview, Computat. Linguist, vol. 13, no. 1, pp. 41-61, 2007.
[11] H. Cui, M.-Y. Kan, and T.-S. Chua, Soft pattern matching models for
definitional question answering, ACM Trans. Inf. Syst., vol. 25, no. 2,
pp. 30-30, 2007.
[12] R.C. Wang, N. Schlaefer, W. W. Cohen, and E. Nyberg, Automatic set
expansion for list question answering, in Proc. Int. Conf. Empirical
Methods in Natural Language Processing, 2008.
[13] R.Hong, M. Wang, G. Li, L. Nie, Z.-J. Zha, and T.-S Chua, Multimedia
Question Answering IEEE 2012.
[14] L. Nie et al., Multimedia Answering: Enriching Text QA with Media
Information, Proc. 34th IntI ACM SIGIR Conf. Research and
Development in Information Retrieval, ACM Press, 2011, pp. 695-704.
[15]G. Kacmarcik, Multi-Modal Question-Answering: Questions Without
Keyboards, Asia Federation of Natural Language Processing, 2005.
[16] M. Wang, X.S. Hua, R. Hong, J. Tang, G.J. Qi, Y. Song, and L.R. Dai,
Unified video annotation via multi-graph learning, IEEE Trans. Circuits
Syst. Video Technol., vol. 19, no. 5, pp. 733-749, 2009.
[17] M. Wang, X. S. Hua, T. Mei, R. Hong, G. J. Qi, Y. Song, and L. R. Dai,
Semi-supervised kernel density estimation for video annotation,
Comput. Vision Image Understand., vol. 113, no. 3, pp. 384-396, 2009.
[18] J. Tang, R. Hong, S. Yan, T. S. Chua, G. J. Qi, and R. Jain, Image
annotation by KNN- sparse graph-based label propagation over noisytagged web images, ACM Trans. Intell. Syst. Technol., vol. 2, no. 2, pp.
1-15, 2011.
[19] J. Tang, X. S. Hua, M. Wang, Z. Gu, G. J. Qi, and X. Wu, Correlative
linear neighborhood propagation for video annotation, IEEE Trans.
Syst., Man, Cybern. B, vol. 39, no. 2, pp. 409-416, 2009.
[20] Z.-J. Zha, X.-S. Hua, T. Mei, J. Wang, G. J. Qi, and Z. Wang, Joint
multi-label multi-instance learning for image classification, in Proc.
IEEE Conf. Computer Vision and Pattern Recognition, 2008, pp. 1-8.
REFERENCES
[1]G.Li, H. Li, Z. Ming, R. Hong, S. Tang, and T.-S. Chua, Question
answering over community contributed web videos , IEEE Multimedia,
vol. 17, no. 4, pp. 46-57, 2010.
[2] H. Yang et al., Structured Use of External Knowledge for Event-based
Open-Domain Question-Answering, Proc 26th Ann. IntI ACM SIGIR
Conf. Research and Development in Information Retrieval, ACM Press,
2003, pp 33-40.
[21] R. Datta et al., Image Retrieval: Ideas, Influence, and Trends of the
New Age, ACM Computing Surveys, vol. 40, no. 2, 2008, article no. 5.
[22] A. Kotav and C. Zhai, Towards Natural Question Guided Search,
Proc. 19th IntI Conf. World Wide Web(WWW), ACM Press, 2010, pp.
541-550.
[23]S. Huston and W. B. Croft, Evaluating Verbose Query Processing
Techniques, Proc. 33rd Int I ACM SIGIR Conf. Research and
Development in Information Retrieval, ACM Press. 2010, pp. 291-298.