Digging Deeper With Machine Learning for Unbalanced Multimedia Data Categorization
DOI:
https://doi.org/10.35671/jmtt.v3i1.47Keywords:
Classification, Imbalance Data, Video, TrackingAbstract
Since many real-world data sets have skewed class distributions—in which the majority of data instances (examples) belong to one class and considerably fewer instances belong to others—classifying unbalanced data is an important area of research. While minority instances (fraud in banking operations, abnormal cells in medical data, etc.) in many applications actually represent the concept of interest, a classifier induced from an imbalanced data set is more likely to be biassed towards the majority class and show very poor classification accuracy for the minority class. Unbalanced data classification, particularly for multimedia data, continues to be one of the most difficult issues in data mining and machine learning, despite substantial research efforts. In this research, we present an extended deep learning strategy to address this difficulty and get encouraging results in the classification of skewed multimedia data sets. In particular, we examine the combination of advanced empirical research on convolutional neural networks (CNNs), a cutting-edge deep learning technique, and bootstrapping techniques. Given that deep learning techniques, like CNNs, are typically computationally costly, we suggest feeding low-level features to CNNs and demonstrate that this may be done in a way that saves a significant amount of training time while still producing promising results. The experimental findings demonstrate how well our methodology performs in the TRECVID data set when it comes to categorising highly unbalanced data.
Downloads
References
L. Wang, T. Liu, G. Wang, K. L. Chan, and Q. Yang, “Video tracking using learned hierarchical features,” IEEE Trans. Image Process., vol. 24, no. 4, pp. 1424–1435, 2015.
H. Xue, Y. Liu, D. Cai, and X. He, “Tracking people in RGBD videos using deep learning and motion clues,” Neurocomputing, vol. 204, pp. 70–76, 2016.
D. Zhang, H. Maei, X. Wang, and Y.-F. Wang, “Deep reinforcement learning for visual object tracking in videos,” arXiv Prepr. arXiv1701.08936, 2017.
G. Ciaparrone, F. L. Sánchez, S. Tabik, L. Troiano, R. Tagliaferri, and F. Herrera, “Deep learning in video multi-object tracking: A survey,” Neurocomputing, vol. 381, pp. 61–88, 2020.
G. Zheng and Y. Xu, “Efficient face detection and tracking in video sequences based on deep learning,” Inf. Sci. (Ny)., vol. 568, pp. 265–285, 2021.
S. M. Marvasti-Zadeh, L. Cheng, H. Ghanei-Yakhdan, and S. Kasaei, “Deep learning for visual tracking: A comprehensive survey,” IEEE Trans. Intell. Transp. Syst., vol. 23, no. 5, pp. 3943–3968, 2021.
S. Pang, J. J. del Coz, Z. Yu, O. Luaces, and J. D’iez, “Deep learning to frame objects for visual target tracking,” Eng. Appl. Artif. Intell., vol. 65, pp. 406–420, 2017.
G. Chandan, A. Jain, H. Jain, and others, “Real time object detection and tracking using Deep Learning and OpenCV,” in 2018 International Conference on inventive research in computing applications (ICIRCA), 2018, pp. 1305–1308.
Y. Yoon et al., “Analyzing basketball movements and pass relationships using realtime object tracking techniques based on deep learning,” IEEE Access, vol. 7, pp. 56564–56576, 2019.
S. Pang, J. J. Del Coz, Z. Yu, O. Luaces, and J. D’iez, “Deep learning and preference learning for object tracking: a combined approach,” Neural Process. Lett., vol. 47, pp. 859–876, 2018.
Y. Li et al., “Deep learning-based object tracking in satellite videos: A comprehensive survey with a new dataset,” IEEE Geosci. Remote Sens. Mag., vol. 10, no. 4, pp. 181–212, 2022.
H. V. R. Aradhya and others, “Object detection and tracking using deep learning and artificial intelligence for video surveillance applications,” Int. J. Adv. Comput. Sci. Appl., vol. 10, no. 12, 2019.
P. R. Kamble, A. G. Keskar, and K. M. Bhurchandi, “A deep learning ball tracking system in soccer videos,” Opto-Electronics Rev., vol. 27, no. 1, pp. 58–69, 2019.
X.-Q. Zhang, R.-H. Jiang, C.-X. Fan, T.-Y. Tong, T. Wang, and P.-C. Huang, “Advances in deep learning methods for visual tracking: Literature review and fundamentals,” Int. J. Autom. Comput., vol. 18, no. 3, pp. 311–333, 2021.
L. Jiao, D. Wang, Y. Bai, P. Chen, and F. Liu, “Deep learning in visual tracking: A review,” IEEE Trans. neural networks Learn. Syst., vol. 34, no. 9, pp. 5497–5516, 2021.
S. K. Pal, A. Pramanik, J. Maiti, and P. Mitra, “Deep learning in multi-object detection and tracking: state of the art,” Appl. Intell., vol. 51, pp. 6400–6429, 2021.
D. Meimetis, I. Daramouskas, I. Perikos, and I. Hatzilygeroudis, “Real-time multiple object tracking using deep learning methods,” Neural Comput. Appl., vol. 35, no. 1, pp. 89–118, 2023.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2024 Nihayatusyifa, Dita Febrianti
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.