Journal articles on the topic 'Visual question answering (VQA)'
Create a spot-on reference in APA, MLA, Chicago, Harvard, and other styles
Consult the top 50 journal articles for your research on the topic 'Visual question answering (VQA).'
Next to every source in the list of references, there is an 'Add to bibliography' button. Press on it, and we will generate automatically the bibliographic reference to the chosen work in the citation style you need: APA, MLA, Harvard, Chicago, Vancouver, etc.
You can also download the full text of the academic publication as pdf and read online its abstract whenever available in the metadata.
Browse journal articles on a wide variety of disciplines and organise your bibliography correctly.
Agrawal, Aishwarya, Jiasen Lu, Stanislaw Antol, Margaret Mitchell, C. Lawrence Zitnick, Devi Parikh, and Dhruv Batra. "VQA: Visual Question Answering." International Journal of Computer Vision 123, no. 1 (November 8, 2016): 4–31. http://dx.doi.org/10.1007/s11263-016-0966-6.
Full textLei, Chenyi, Lei Wu, Dong Liu, Zhao Li, Guoxin Wang, Haihong Tang, and Houqiang Li. "Multi-Question Learning for Visual Question Answering." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 07 (April 3, 2020): 11328–35. http://dx.doi.org/10.1609/aaai.v34i07.6794.
Full textShah, Sanket, Anand Mishra, Naganand Yadati, and Partha Pratim Talukdar. "KVQA: Knowledge-Aware Visual Question Answering." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 8876–84. http://dx.doi.org/10.1609/aaai.v33i01.33018876.
Full textGuo, Zihan, Dezhi Han, and Kuan-Ching Li. "Double-layer affective visual question answering network." Computer Science and Information Systems, no. 00 (2020): 38. http://dx.doi.org/10.2298/csis200515038g.
Full textWu, Chenfei, Jinlai Liu, Xiaojie Wang, and Ruifan Li. "Differential Networks for Visual Question Answering." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 8997–9004. http://dx.doi.org/10.1609/aaai.v33i01.33018997.
Full textZhou, Yiyi, Rongrong Ji, Jinsong Su, Xiaoshuai Sun, and Weiqiu Chen. "Dynamic Capsule Attention for Visual Question Answering." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 9324–31. http://dx.doi.org/10.1609/aaai.v33i01.33019324.
Full textEt. al., K. P. Moholkar,. "Visual Question Answering using Convolutional Neural Networks." Turkish Journal of Computer and Mathematics Education (TURCOMAT) 12, no. 1S (April 11, 2021): 170–75. http://dx.doi.org/10.17762/turcomat.v12i1s.1602.
Full textGuo, Wenya, Ying Zhang, Xiaoping Wu, Jufeng Yang, Xiangrui Cai, and Xiaojie Yuan. "Re-Attention for Visual Question Answering." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 01 (April 3, 2020): 91–98. http://dx.doi.org/10.1609/aaai.v34i01.5338.
Full textBoukhers, Zeyd, Timo Hartmann, and Jan Jürjens. "COIN: Counterfactual Image Generation for Visual Question Answering Interpretation." Sensors 22, no. 6 (March 14, 2022): 2245. http://dx.doi.org/10.3390/s22062245.
Full textLi, Qun, Fu Xiao, Bir Bhanu, Biyun Sheng, and Richang Hong. "Inner Knowledge-based Img2Doc Scheme for Visual Question Answering." ACM Transactions on Multimedia Computing, Communications, and Applications 18, no. 3 (August 31, 2022): 1–21. http://dx.doi.org/10.1145/3489142.
Full textQiu, Yue, Yutaka Satoh, Ryota Suzuki, Kenji Iwata, and Hirokatsu Kataoka. "Multi-View Visual Question Answering with Active Viewpoint Selection." Sensors 20, no. 8 (April 17, 2020): 2281. http://dx.doi.org/10.3390/s20082281.
Full textR, Lokesh, Madhusudan C, Darshan T, and Sunil kumar N. "VISUAL QUESTIONING AND ANSWERING." International Journal of Innovative Research in Advanced Engineering 9, no. 8 (August 12, 2022): 312–15. http://dx.doi.org/10.26562/ijirae.2022.v0908.29.
Full textLee, Doyup, Yeongjae Cheon, and Wook-Shin Han. "Regularizing Attention Networks for Anomaly Detection in Visual Question Answering." Proceedings of the AAAI Conference on Artificial Intelligence 35, no. 3 (May 18, 2021): 1845–53. http://dx.doi.org/10.1609/aaai.v35i3.16279.
Full textKim, Incheol. "Visual Experience-Based Question Answering with Complex Multimodal Environments." Mathematical Problems in Engineering 2020 (November 19, 2020): 1–18. http://dx.doi.org/10.1155/2020/8567271.
Full textLi, Haiyan, and Dezhi Han. "Multimodal encoders and decoders with gate attention for visual question answering." Computer Science and Information Systems 18, no. 3 (2021): 1023–40. http://dx.doi.org/10.2298/csis201120032l.
Full textHe, Shirong, and Dezhi Han. "An Effective Dense Co-Attention Networks for Visual Question Answering." Sensors 20, no. 17 (August 30, 2020): 4897. http://dx.doi.org/10.3390/s20174897.
Full textAlizadeh, Mehrdad, and Barbara Di Eugenio. "Incorporating Verb Semantic Information in Visual Question Answering Through Multitask Learning Paradigm." International Journal of Semantic Computing 14, no. 02 (June 2020): 223–48. http://dx.doi.org/10.1142/s1793351x20400085.
Full textGarcia, Noa, Mayu Otani, Chenhui Chu, and Yuta Nakashima. "KnowIT VQA: Answering Knowledge-Based Questions about Videos." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 07 (April 3, 2020): 10826–34. http://dx.doi.org/10.1609/aaai.v34i07.6713.
Full textYan, Feng, Wushouer Silamu, and Yanbing Li. "Deep Modular Bilinear Attention Network for Visual Question Answering." Sensors 22, no. 3 (January 28, 2022): 1045. http://dx.doi.org/10.3390/s22031045.
Full textYang, Cheng, Weijia Wu, Yuxing Wang, and Hong Zhou. "Multi-Modality Global Fusion Attention Network for Visual Question Answering." Electronics 9, no. 11 (November 9, 2020): 1882. http://dx.doi.org/10.3390/electronics9111882.
Full textLi, Qifeng, Xinyi Tang, and Yi Jian. "Adversarial Learning with Bidirectional Attention for Visual Question Answering." Sensors 21, no. 21 (October 28, 2021): 7164. http://dx.doi.org/10.3390/s21217164.
Full textZhang, Pufen, Hong Lan, and Muhammad Asim Khan. "Multiple Context Learning Networks for Visual Question Answering." Scientific Programming 2022 (February 9, 2022): 1–11. http://dx.doi.org/10.1155/2022/4378553.
Full textGuo, Zihan, and Dezhi Han. "Multi-Modal Explicit Sparse Attention Networks for Visual Question Answering." Sensors 20, no. 23 (November 26, 2020): 6758. http://dx.doi.org/10.3390/s20236758.
Full textShen, Xiang, Dezhi Han, Chongqing Chen, Gaofeng Luo, and Zhongdai Wu. "An effective spatial relational reasoning networks for visual question answering." PLOS ONE 17, no. 11 (November 28, 2022): e0277693. http://dx.doi.org/10.1371/journal.pone.0277693.
Full textJing, Chenchen, Yuwei Wu, Xiaoxun Zhang, Yunde Jia, and Qi Wu. "Overcoming Language Priors in VQA via Decomposed Linguistic Representations." Proceedings of the AAAI Conference on Artificial Intelligence 34, no. 07 (April 3, 2020): 11181–88. http://dx.doi.org/10.1609/aaai.v34i07.6776.
Full textYanqing Cui, Yanqing Cui, Guangjie Han Yanqing Cui, and Hongbo Zhu Guangjie Han. "A Novel Online Teaching Effect Evaluation Model Based on Visual Question Answering." 網際網路技術學刊 23, no. 1 (January 2022): 093–100. http://dx.doi.org/10.53106/160792642022012301009.
Full textHuang, Jia-Hong, Cuong Duc Dao, Modar Alfadly, and Bernard Ghanem. "A Novel Framework for Robustness Analysis of Visual QA Models." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 8449–56. http://dx.doi.org/10.1609/aaai.v33i01.33018449.
Full textXiang, Yingxin, Chengyuan Zhang, Zhichao Han, Hao Yu, Jiaye Li, and Lei Zhu. "Path-Wise Attention Memory Network for Visual Question Answering." Mathematics 10, no. 18 (September 7, 2022): 3244. http://dx.doi.org/10.3390/math10183244.
Full textBen-younes, Hedi, Remi Cadene, Nicolas Thome, and Matthieu Cord. "BLOCK: Bilinear Superdiagonal Fusion for Visual Question Answering and Visual Relationship Detection." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 8102–9. http://dx.doi.org/10.1609/aaai.v33i01.33018102.
Full textWang, Ruiping, Shihong Wu, and Xiaoping Wang. "The Core of Smart Cities: Knowledge Representation and Descriptive Framework Construction in Knowledge-Based Visual Question Answering." Sustainability 14, no. 20 (October 14, 2022): 13236. http://dx.doi.org/10.3390/su142013236.
Full textZhu, Han, Xiaohai He, Meiling Wang, Mozhi Zhang, and Linbo Qing. "Medical visual question answering via corresponding feature fusion combined with semantic attention." Mathematical Biosciences and Engineering 19, no. 10 (2022): 10192–212. http://dx.doi.org/10.3934/mbe.2022478.
Full textLiu, Yibing, Yangyang Guo, Jianhua Yin, Xuemeng Song, Weifeng Liu, Liqiang Nie, and Min Zhang. "Answer Questions with Right Image Regions: A Visual Attention Regularization Approach." ACM Transactions on Multimedia Computing, Communications, and Applications 18, no. 4 (November 30, 2022): 1–18. http://dx.doi.org/10.1145/3498340.
Full textZhang, Xu, DeZhi Han, and Chin-Chen Chang. "RDMMFET: Representation of Dense Multimodality Fusion Encoder Based on Transformer." Mobile Information Systems 2021 (October 18, 2021): 1–9. http://dx.doi.org/10.1155/2021/2662064.
Full textAcharya, Manoj, Kushal Kafle, and Christopher Kanan. "TallyQA: Answering Complex Counting Questions." Proceedings of the AAAI Conference on Artificial Intelligence 33 (July 17, 2019): 8076–84. http://dx.doi.org/10.1609/aaai.v33i01.33018076.
Full textLei, Zhi, Guixian Zhang, Lijuan Wu, Kui Zhang, and Rongjiao Liang. "A Multi-level Mesh Mutual Attention Model for Visual Question Answering." Data Science and Engineering 7, no. 4 (October 30, 2022): 339–53. http://dx.doi.org/10.1007/s41019-022-00200-9.
Full textPark, Sungho, Sunhee Hwang, Jongkwang Hong, and Hyeran Byun. "Fair-VQA: Fairness-Aware Visual Question Answering Through Sensitive Attribute Prediction." IEEE Access 8 (2020): 215091–99. http://dx.doi.org/10.1109/access.2020.3041503.
Full textLiu, Feng, Tao Xiang, Timothy M. Hospedales, Wankou Yang, and Changyin Sun. "Inverse Visual Question Answering: A New Benchmark and VQA Diagnosis Tool." IEEE Transactions on Pattern Analysis and Machine Intelligence 42, no. 2 (February 1, 2020): 460–74. http://dx.doi.org/10.1109/tpami.2018.2880185.
Full textYang, Zhengyuan, Zhe Gan, Jianfeng Wang, Xiaowei Hu, Yumao Lu, Zicheng Liu, and Lijuan Wang. "An Empirical Study of GPT-3 for Few-Shot Knowledge-Based VQA." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 3 (June 28, 2022): 3081–89. http://dx.doi.org/10.1609/aaai.v36i3.20215.
Full textCao, Qingqing, Prerna Khanna, Nicholas D. Lane, and Aruna Balasubramanian. "MobiVQA." Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies 6, no. 2 (July 4, 2022): 1–23. http://dx.doi.org/10.1145/3534619.
Full textLi, Mingxiao, and Marie-Francine Moens. "Dynamic Key-Value Memory Enhanced Multi-Step Graph Reasoning for Knowledge-Based Visual Question Answering." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 10 (June 28, 2022): 10983–92. http://dx.doi.org/10.1609/aaai.v36i10.21346.
Full textNarayanan, Abhishek, Abijna Rao, Abhishek Prasad, and Natarajan S. "VQA as a factoid question answering problem: A novel approach for knowledge-aware and explainable visual question answering." Image and Vision Computing 116 (December 2021): 104328. http://dx.doi.org/10.1016/j.imavis.2021.104328.
Full textLiang, Haotian, and Zhanqing Wang. "Hierarchical Attention Networks for Multimodal Machine Learning." Journal of Physics: Conference Series 2218, no. 1 (March 1, 2022): 012020. http://dx.doi.org/10.1088/1742-6596/2218/1/012020.
Full textYuan, Desen, Lei Wang, Qingbo Wu, Fanman Meng, King Ngi Ngan, and Linfeng Xu. "Language Bias-Driven Self-Knowledge Distillation with Generalization Uncertainty for Reducing Language Bias in Visual Question Answering." Applied Sciences 12, no. 15 (July 28, 2022): 7588. http://dx.doi.org/10.3390/app12157588.
Full textGaidamavičius, Dainius, and Tomas Iešmantas. "Deep learning method for visual question answering in the digital radiology domain." Mathematical Models in Engineering 8, no. 2 (June 26, 2022): 58–71. http://dx.doi.org/10.21595/mme.2022.22737.
Full textLi, Xuewei, Dezhi Han, and Chin-Chen Chang. "Pre-training Model Based on Parallel Cross-Modality Fusion Layer." PLOS ONE 17, no. 2 (February 3, 2022): e0260784. http://dx.doi.org/10.1371/journal.pone.0260784.
Full textLobry, S., D. Marcos, B. Kellenberger, and D. Tuia. "BETTER GENERIC OBJECTS COUNTING WHEN ASKING QUESTIONS TO IMAGES: A MULTITASK APPROACH FOR REMOTE SENSING VISUAL QUESTION ANSWERING." ISPRS Annals of Photogrammetry, Remote Sensing and Spatial Information Sciences V-2-2020 (August 3, 2020): 1021–27. http://dx.doi.org/10.5194/isprs-annals-v-2-2020-1021-2020.
Full textWu, Jialin, Jiasen Lu, Ashish Sabharwal, and Roozbeh Mottaghi. "Multi-Modal Answer Validation for Knowledge-Based VQA." Proceedings of the AAAI Conference on Artificial Intelligence 36, no. 3 (June 28, 2022): 2712–21. http://dx.doi.org/10.1609/aaai.v36i3.20174.
Full textLi, Qifeng, Xinyi Tang, and Yi Jian. "Learning to Reason on Tree Structures for Knowledge-Based Visual Question Answering." Sensors 22, no. 4 (February 17, 2022): 1575. http://dx.doi.org/10.3390/s22041575.
Full textMa, Zhiyang, Wenfeng Zheng, Xiaobing Chen, and Lirong Yin. "Joint embedding VQA model based on dynamic word vector." PeerJ Computer Science 7 (March 3, 2021): e353. http://dx.doi.org/10.7717/peerj-cs.353.
Full textZhu, He, Ren Togo, Takahiro Ogawa, and Miki Haseyama. "Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation." Sensors 23, no. 3 (January 17, 2023): 1057. http://dx.doi.org/10.3390/s23031057.
Full text