Journal of Applied Science and Engineering

Published by Tamkang University Press

1.30

Impact Factor

2.10

CiteScore

Xinjie ZhuThis email address is being protected from spambots. You need JavaScript enabled to view it.

School of Foreign Languages, Zhengzhou University of Science and Technology, Zhengzhou, 450064, China


 

 

Received: March 21, 2024
Accepted: September 1, 2024
Publication Date: October 7, 2024

 Copyright The Author(s). This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are cited.


Download Citation: ||https://doi.org/10.6180/jase.202507_28(7).0007  


In multimedia pattern recognition, particularly in English education, the sharing of local features between classes and their varying classification reliability are often overlooked in existing methods, which diminishes feature discrimination and complicates the handling of small inter-class variations. In this paper, a multi-level feature learning method based on enhanced local descriptors is proposed for mining multimedia patterns in English educations (MFL). Specifically, multi-scale global information is extracted using the pyramid aggregation network and fused with local features to enhance inter-class uniqueness. During classification, local descriptors that better distinguish between classes are emphasized, resulting in improved inter-class discrimination. MFL achieves a significant accuracy improvement over baseline methods across three datasets, offering a new perspective for analyzing multimedia content in English education.


Keywords: English education; multimedia image recognition; multi-level feature learning


  1. [1] J. Y. Lim, K. M. Lim, C. P. Lee, and Y. X. Tan, (2023) “SCL: Self-supervised Contrastive Learning for Few-shot Image Classification" Neural Networks: 19–30. DOI: 10.1016/j.neunet.2023.05.037.
  2. [2] J. Gao, P. Li, A. A. Laghari, G. Srivastava, T. R. Gadekallu, S. Abbas, and J. Zhang, (2024) “Incomplete multiview clustering via semidiscrete optimal transport for multimedia data mining in IoT" ACM Transactions on Multimedia Computing, Communications and Applications 20(6): 1–20. DOI: 10.1145/3625548.
  3. [3] Y. Dong, H. Zhang, C. Wang, and Y. Wang. “Finegrained Ship Classification Based on Deep Residual Learning for High-resolution SAR Images”. In: International Conference on Learning Representations, Remote Sens. 2019, 1095–1104. DOI: 10.1080/2150704X.2019.1650982.
  4. [4] X. Wei et al., (2022) “Fine-grained Image Analysis with Deep Learning: A Survey" IEEE Trans. Pattern Anal. Mach. Intell. 8927–8948. DOI: 10.1109/TPAMI.2021.3126648.
  5. [5] P. Li, A. A. Laghari, M. Rashid, J. Gao, T. R. Gadekallu, A. R. Javed, and S. Yin, (2022) “A deep multimodal adversarial cycle-consistent network for smart enterprise system" IEEE Transactions on Industrial Informatics 19(1): 693–702. DOI: 10.1109/TII.2022.3197201.
  6. [6] P. Chikontwe, S. Kim, and S. H. Park. “CAD: Coadapting Discriminative Features for Improved Fewshot Classification”. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022, 14554–14563.
  7. [7] H. J. Ye, H. Hu, D. C. Zhan, et al. “Few-shot Learning via Embedding Adaptation with Set-to-set Functions”. In: IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020, 8808–8817.
  8. [8] C. Zhang, Y. Cai, G. Lin, and C. Shen. “DeepEMD: Few-shot Image Classification with Differentiable Earth Mover’s Distance and Structured Classifiers”. In: IEEE Conf. Comput. Vis. Pattern Recog. (CVPR). 2020, 12203–12213.
  9. [9] W. Li, L. Wang, J. Xu, J. Huo, Y. Gao, and J. Luo. “Revisiting Local Descriptor Based Image-to-Class Measure for Few-shot Learning”. In: IEEE Conf. Comput. Vis. Pattern Recog. (CVPR). 2019, 7260–7268.
  10. [10] F. Sung, Y. Yang, L. Zhang, T. Xiang, P. H. Torr, and T. M. Hospedales. “Learning to Compare: Relation Network for Few-shot Learning”. In: IEEE Conf. Comput. Vis. Pattern Recog. (CVPR). 2018, 1199–1208.
  11. [11] X. Luo, H. Wu, J. Zhang, L. Gao, J. Xu, and J. Song. “A closer look at few-shot classification again”. In: International Conference on Machine Learning. 2023, 23103–23123.
  12. [12] Y. B. Liu, L. C. Liu, X. H. Wang, M. Yamada, and Y. Yang, (2023) “Bilaterally Normalized Scale-consistent Sinkhorn Distance for Few-shot Image Classification" IEEE Trans. Neural Netw. Learn. Syst. 12203–12213.
  13. [13] A. Afrasiyabi, H. Larochelle, J.-F. Lalonde, and C. Gagné. “Matching Feature Sets for Few-shot Image Classification”. In: Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit. 2022, 9014–9024.
  14. [14] C. Finn, P. Abbeel, and S. Levine. “Model Agnostic Meta-Learning for Fast Adaptation of Deep Networks”. In: ICML. 2017, 1126–1135.
  15. [15] K. Lee, S. Maji, A. Ravichandran, and S. Soatto. “Meta-Learning with Differentiable Convex Optimization”. In: CVPR. 2019, 10657–10665.
  16. [16] A. Nichol, J. Achiam, and J. Schulman, (2018) “On First-Order Meta-Learning Algorithms" CoRR abs/1803.02999:
  17. [17] Y. Tian, Y. Wang, D. Krishnan, J. B. Tenenbaum, and P. Isola. “Rethinking Few-shot Image Classification: A Good Embedding is All You Need?” In: Computer Vision – ECCV. 2020.
  18. [18] N. Lai, M. Kan, C. Han, X. Song, and S. Shan, (2021) “Learning to Learn Adaptive Classifier-Predictor for FewShot Learning" IEEE Trans Neural Netw Learn Syst: 3458–3470. DOI: 10.1109/TNNLS.2020.3011526.
  19. [19] Y. Yu, D. Zhang, and Z. Ji. “Masked Feature Generation Network for Few-Shot Learning”. In: IJCAI. 2022, 7005–7014.
  20. [20] Y. Jian and L. Torresani. “Label Hallucination for Few-Shot Classification”. In: AAAI. 2022, 3695–3701. DOI: 10.1609/aaai.v36i6.20659.
  21. [21] A. Antoniou, A. Storkey, and H. Edwards. “Data Augmentation Generative Adversarial Networks”. In: AAAI. 2018, 1050.
  22. [22] R. Kwitt, S. Hegenbart, and M. Niethammer. “Oneshot Learning of Scene Locations via Feature Trajectory Transfer”. In: CVPR. 2016, 78–86.
  23. [23] L. X, W. J, S. Z, et al., (2020) “BSNet: Bi-similarity Network for Few-shot Fine-grained Image Classification" IEEE Transactions on Image Processing: 1318–1331. DOI: 10.1109/TIP.2020.3043128.
  24. [24] C. Dong, W. Li, J. Huo, et al. “Learning Task-aware Local Representations for Few-shot Learning”. In: IJCAI. 2021, 716–722.
  25. [25] B. Q, R. I, and A. R. “Improving Few-Shot Learning Through Multi-task Representation Learning Theory”. In: Computer Vision–ECCV. 2022, 435–452.
  26. [26] H. Huang, J. Zhang, L. Yu, J. Zhang, Q. Wu, and C. Xu, (2021) “TOAN: Target-oriented alignment network for fine-grained image categorization with few labeled samples" IEEE Transactions on Circuits and Systems for Video Technology 32(2): 853–866.


    



 

2.1
2023CiteScore
 
 
69th percentile
Powered by  Scopus

SCImago Journal & Country Rank

Enter your name and email below to receive latest published articles in Journal of Applied Science and Engineering.