In recent years, social media platforms have evolved into primary sources of travel inspiration, with travelers increasingly relying on short-form video content to discover unique, ”off-the-beaten-path” locations. Despite their popularity, these videos often lack the structured and reliable information necessary for practical travel planning. Critical details such as precise locations, accessibility, and logistical guidance are frequently fragmented across captions and comments or omitted entirely, creating a significant gap between visual discovery and actionable decision-making. This paper reviews existing research in social media mining, multimodal information extraction, and Natural Language Processing (NLP) to determine how Artificial Intelligence can bridge this gap. Building on insights from the reviewed literature, the study presents the conceptual design of ”ReelScout,” an AI-driven platform that integrates Computer Vision and NLP to analyze social media reels for identifying hidden Points of Interest (POIs). By synthesizing multimodal cues from visual content, audio narration, and textual metadata, the proposed framework aims to organize unstructured social media data into meaningful travel knowledge. Finally, this review highlights current methodological trends, limitations, and future research directions for the development of intelligent, social-media-driven travel discovery systems.
Introduction
The text reviews how the rapid growth of short-form social media videos (reels and shorts) has transformed travel inspiration and location discovery, while also highlighting the limitations of such content for practical travel planning. Although visually engaging, social media posts often lack structured and reliable information such as exact locations, accessibility, safety, and logistics, making it difficult for users to convert inspiration into actionable decisions.
The large volume of unstructured, multimodal data (text, images, audio, and video) generated on social media poses significant challenges for traditional data analysis methods. Recent advances in Artificial Intelligence, including Natural Language Processing (NLP), Computer Vision (CV), multimodal learning, and large language models, have enabled more effective extraction of insights from social media content. In particular, multimodal approaches that combine text, visuals, and audio outperform unimodal methods, while Retrieval-Augmented Generation (RAG) improves factual accuracy by grounding AI outputs in external knowledge sources—an important requirement for travel recommendation and decision-support systems.
The literature review traces the evolution of research from early text- and metadata-based POI discovery to vision-based, multimodal, and RAG-based systems. Text-based approaches are scalable but limited by noisy and informal language, while vision-based methods excel at recognizing well-known landmarks but struggle with ambiguous or lesser-known locations. Multimodal systems provide richer contextual understanding but introduce higher computational complexity. RAG-based systems further enhance reliability and adaptability by incorporating up-to-date external information.
The paper proposes a taxonomy of existing approaches—text-based, vision-based, multimodal, and RAG-based—and presents a comparative analysis highlighting trade-offs among accuracy, scalability, and system complexity. Despite methodological advances, key challenges remain, including noisy and incomplete data, multimodal fusion complexity, high computational cost, lack of standardized evaluation benchmarks, and ethical and privacy concerns related to implicit geolocation inference.
Finally, the review identifies future research directions, emphasizing the need for advanced multimodal fusion strategies, explainable and interpretable AI systems, privacy-preserving methods, and robust evaluation frameworks. Overall, the text concludes that while AI-driven multimodal and retrieval-grounded approaches show strong potential for social media-based travel discovery and recommendation, further research is required to ensure scalability, reliability, and ethical deployment in real-world applications.
Conclusion
This review presented a comprehensive and structured analysis of artificial intelligence-based approaches for social media- driven travel discovery and recommendation systems. By systematically categorizing existing literature into text-based, vision- based, multimodal, and retrieval-augmented approaches, the paper highlighted the evolution of methodologies and the growing reliance on content-aware and data-driven techniques. The comparative analysis further revealed the strengths and limitations of representative studies, illustrating a clear shift from metadata-driven methods toward multimodal and retrieval-grounded frameworks capable of richer contextual understanding. The discussion of challenges emphasized critical issues related to data noise, multimodal fusion complexity, scalability, evaluation, and ethical considerations, underscoring the limitations of current solutions in real-world deployment scenarios. By identifying these open research problems and outlining future research directions, this review provides a structured foundation for advancing intelligent, reliable, and responsible travel discovery systems.
Overall, the insights presented in this paper aim to support researchers and practitioners in designing next-generation AI solutions that effectively bridge the gap between social media-based visual inspiration and actionable travel planning. As social media platforms continue to evolve, the synthesis offered by this review is intended to guide future work at the intersection of social media analytics, artificial intelligence, and intelligent tourism systems.
References
[1] J. Chen, L. Wang, and C. Hsieh, “Mining user-generated content on social media for discovering hidden points of interest,” IEEE Trans. Knowledge and Data Engineering, 2021.
[2] B. Bekhouche, “The impact of user-generated content on tourism destinations: A case study on Instagram Reels,” International Tourism Journal, vol. 52, no. 3, pp. 999-1011, 2025.
[3] D. J. Crandall et al., “Mapping the world’s photos,” in Proc. Int. World Wide Web Conf. (WWW), 2009.
[4] A. Gupta, S. Kumar, and A. Zisserman, “Multi-modal information extraction from short-form video content for geo-localization and activity recognition,” in Proc. IEEE CVPR, 2022.
[5] Y. Wang, “Multimodal analysis: Researching short-form videos on TikTok,” ScholarSpace, University of Hawaii, 2021.
[6] Z. Yang et al., “Evaluation of geolocation capabilities of multimodal large language models,” arXiv preprint arXiv: 2406.12348, 2024.
[7] P. Lewis, E. Perez, A. Piktus et al., “Retrieval-augmented generation for knowledge-intensive NLP tasks,” in Advances in Neural Information Processing Systems (NeurIPS), 2020.
[8] J. Li et al., “TP-RAG: Benchmarking retrieval-augmented LLM agents for travel planning,” in Proc. EMNLP, 2025.
[9] Y. Gao et al., “Retrieval-augmented generation for large language models: A survey,” arXiv preprint arXiv:2312.10997, 2023.
[10] S. Smith and J. Doe, “Transformer-based named entity recognition for place name extraction,” Int. J. Geographical Information Science, 2022.
[11] X. Zhou, X. Liu, and Y. Zhang, “Survey of deep learning-based recommender systems,” ACM Computing Surveys, vol. 54, no. 7, 2021.
[12] H. Wang, F. Wang, J. Liu, and S. Chen, “Social media analytics for tourism: A survey,” Information Processing & Management, vol. 57, no. 6, 2020.
[13] Y. Li, T. Yao, and T. Mei, “Deep learning for multimedia content analysis: A review,” ACM Multimedia, 2019.
[14] A. Graves, G. Mohamed, and G. Hinton, “Speech recognition with deep recurrent neural networks,” in Proc. IEEE ICASSP, 2013.
[15] A. Radford et al., “Learning transferable visual models from natural language supervision,” in Proc. ICML, 2021.
[16] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proc. NAACL-HLT, 2019.
[17] A. Dosovitskiy et al., “An image is worth 16x16 words: Transformers for image recognition,” in Proc. ICLR, 2021.
[18] S. Ruder, “Neural transfer learning for natural language processing,” Ph.D. dissertation, NUI Galway, 2019.
[19] Z. Wu et al., “A comprehensive survey on graph neural networks,” IEEE Trans. Neural Networks and Learning Systems, 2021.
[20] Y. Bengio, A. Courville, and P. Vincent, “Representation learning: A review and new perspectives,” IEEE TPAMI, vol. 35, no. 8, 2013.
[21] C. H. Chen, J. Zhan, and M. Lee, “Location recognition from social media images,” Pattern Recognition, vol. 96, 2019.
[22] M. Zhang and Y. Liu, “Multimodal deep learning: A survey,” IEEE Access, vol. 7, 2019.
[23] H. Liu, X. Hu, and M. Zhang, “Mining social media for tourism recommendation: A survey,” Expert Systems with Applications, vol. 150, 2020.
[24] S. Balakrishnan and S. Chopra, “Automatic location tagging from short video content,” in Proc. ACM Multimedia, 2021.
[25] J. Huang et al., “Survey on sentiment analysis for social media,” IEEE Access, vol. 8, 2020.
[26] R. K. Gupta and P. Kumar, “Geolocation inference from multimedia content,” Multimedia Tools and Applications, vol. 78, 2019.
[27] T. Mikolov et al., “Distributed representations of words and phrases,” in Advances in Neural Information Processing Systems, 2013.
[28] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, no. 8, 1997.
[29] A. Vaswani et al., “Attention is all you need,” in Advances in Neural Information Processing Systems, 2017.
[30] K. Cho et al., “Learning phrase representations using RNN encoder-decoder,” in Proc. EMNLP, 2014.
[31] A. Karpathy and L. Fei-Fei, “Deep visual-semantic alignments for generating image descriptions,” in Proc. CVPR, 2015.
[32] Y. Sun, B. Mobasher, and R. Burke, “Social recommendation: A review,” ACM Computing Surveys, vol. 53, no. 4, 2020.
[33] J. Leskovec, A. Rajaraman, and J. Ullman, Mining of Massive Datasets, Cambridge Univ. Press, 2014.
[34] C. C. Aggarwal, Machine Learning for Text, Springer, 2018.
[35] M. Allahyari et al., “A brief survey of text mining,” ACM SIGKDD Explorations, vol. 19, no. 2, 2017.
[36] A. Madani et al., “Multimodal deep learning for video understanding: A survey,” IEEE Access, vol. 8, 2020.
[37] L. Chen et al., “Recommender systems: A survey,” ACM Computing Surveys, vol. 54, no. 3, 2021.
[38] M. S. Hossain et al., “Toward AI-driven tourism systems,” Future Generation Computer Systems, vol. 122, 2021.
[39] S. Zhang, L. Yao, A. Sun, and Y. Tay, “Deep learning-based recommender systems: A survey,” ACM Computing Surveys, 2019.
[40] D. Jurafsky and J. H. Martin, Speech and Language Processing, 3rd ed., Pearson, 2023.