Skip to main content
Log in

A conditioned feature reconstruction network for few-shot classification

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Few-shot classification is one of the most daunting challenges in deep learning. The complexities of this task arise from the fact that category targets are often embedded within intricate and diverse background pixels, resulting in inconspicuous category features. Moreover, obtaining common category characteristics from a limited number of samples is difficult. Compounding the issue, models encounters categories that they have never seen before, rendering the prior guarantee of interclass variance infeasible. To address these dilemmas, this paper leverages the apriori conditioned information of few-shot tasks and introduces a Conditioned Feature Reconstruction Network (CFRN). The CFRN employs prototype reconstruction to minimize the prototype similarity among different classes and query reconstruction to maximize the similarity of (query, prototype) feature pairs. This approach increases the interclass variance while decreasing the intraclass variance, thereby enhancing separability and improving the saliency of the target features. An experimental validation demonstrates the effectiveness of the CFRN, which obtains state-of-the-art results on the mini-ImageNet, tiered-ImageNet, and CUB datasets.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Algorithm 1
Fig. 3

Similar content being viewed by others

Data Availability

The data will be made available upon reasonable request.

References

  1. Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 10012–10022

  2. Hudson DA, Zitnick L (2021) Generative adversarial transformers. In: International conference on machine learning, PMLR, pp 4487–4499

  3. Liu H, Zhang C, Deng Y, Xie B, Liu T, Zhang Z, Li Y-F (2023) Transifc: invariant cues-aware feature concentration learning for efficient fine-grained bird image classification. IEEE Trans Multimed

  4. Liu H, Liu T, Chen Y, Zhang Z, Li Y-F (2022) Ehpe: Skeleton cues-based gaussian coordinate encoding for efficient human pose estimation. IEEE Trans Multimed

  5. Liu H, Liu T, Zhang Z, Sangaiah AK, Yang B, Li Y (2022) Arhpe: Asymmetric relation-aware representation learning for head pose estimation in industrial human-computer interaction. IEEE Trans Industr Inf 18(10):7107–7117

    Article  Google Scholar 

  6. Ning X, Yu Z, Li L, Li W, Tiwari P (2024) Dilf: Differentiable rendering-based multi-view image-language fusion for zero-shot 3d shape understanding. Inform Fusion 102:102033

  7. Hayashi T, Cimr D, Fujita H, Cimler R (2023) Image entropy equalization: A novel preprocessing technique for image recognition tasks. Inf Sci 647:119539

    Article  Google Scholar 

  8. Hayashi T, Cimr D, Studnička F, Fujita H, Bušovskỳ D, Cimler R, Selamat A (2024) Distance-based one-class time-series classification approach using local cluster balance. Expert Syst Appl 235:121201

    Article  Google Scholar 

  9. Huang C, Guan H, Jiang A, Zhang Y, Spratling M, Wang Y-F (2022) Registration based few-shot anomaly detection. In: European conference on computer vision, Springer, pp 303–319

  10. Dinh P-H (2021) Multi-modal medical image fusion based on equilibrium optimizer algorithm and local energy functions. Appl Intell 51(11):8416–8431

    Article  Google Scholar 

  11. Zhang T-T, Shu H, Lam K-Y, Chow C-Y, Li A (2023) Feature decomposition and enhancement for unsupervised medical ultrasound image denoising and instance segmentation. Appl Intell 53(8):9548–9561

    Article  Google Scholar 

  12. Zhang A, Zhang B, Bi W, Mao Z (2022) Attention based trajectory prediction method under the air combat environment. Appl Intell 52(15):17341–17355

    Article  Google Scholar 

  13. Yu Z (2023) An information fusion method for meta-tracker about online aerospace object tracking. Journal of Intelligent & Fuzzy Systems. 45(4):6063–6075. https://doi.org/10.3233/JIFS-230265

    Article  Google Scholar 

  14. Zheng X, Chen J, Wang H, Zheng S, Kong Y (2021) A deep learning-based approach for the automated surface inspection of copper clad laminate images. Appl Intell 51:1262–1279

    Article  Google Scholar 

  15. Tian S, Li L, Li W, Ran H, Ning X, Tiwari P (2024) A survey on few-shot class-incremental learning. Neural Netw 169:307–324

    Article  Google Scholar 

  16. Hayashi T, Cimr D, Studnička F, Fujita H, Bušovskỳ D, Cimler R (2024) Patient deterioration detection using one-class classification via cluster period estimation subtask. Inf Sci 657:119975

    Article  Google Scholar 

  17. Guo Y, Codella NC, Karlinsky L, Codella JV, Smith JR, Saenko K, Rosing T, Feris R (2020) A broader study of cross-domain few-shot learning. In: Computer vision–ECCV 2020: 16th European conference, Glasgow, UK, Proceedings, Part XXVII 16, Springer, pp 124–141. Accessed 23–28 Aug 2020

  18. Wang J, Liu K, Zhang Y, Leng B, Lu J (2023) Recent advances of few-shot learning methods and applications. SCIENCE CHINA Technol Sci 66(4):920–944

    Article  Google Scholar 

  19. Liu H, Fang S, Zhang Z, Li D, Lin K, Wang J (2021) Mfdnet: Collaborative poses perception and matrix fisher distribution for head pose estimation. IEEE Trans Multimedia 24:2449–2460

    Article  Google Scholar 

  20. Liu T, Wang J, Yang B, Wang X (2021) Ngdnet: Nonuniform gaussian-label distribution learning for infrared head pose estimation and on-task behavior understanding in the classroom. Neurocomputing 436:210–220

    Article  Google Scholar 

  21. Zhang C, Liu H, Deng Y, Xie B, Li Y (2023) Tokenhpe: Learning orientation tokens for efficient head pose estimation via transformers. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8897–8906

  22. Simon C, Koniusz P, Nock R, Harandi M (2020) Adaptive subspaces for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 4136–4145

  23. Hou R, Chang H, Ma B, Shan S, Chen X (2019) Cross attention network for few-shot classification. Adv Neural Inform Process Syst 32

  24. Zhang C, Cai Y, Lin G, Shen C (2022) Deepemd: Differentiable earth mover’s distance for few-shot learning. IEEE Trans Pattern Anal Mach Intell 45(5):5632–5648

    Google Scholar 

  25. Wertheimer D, Tang L, Hariharan B (2021) Few-shot classification with feature map reconstruction networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8012–8021

  26. Xie J, Long F, Lv J, Wang Q, Li P (2022) Joint distribution matters: Deep brownian distance covariance for few-shot classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7972–7981

  27. Li X, Li Y, Zheng Y, Zhu R, Ma Z, Xue J-H, Cao J (2023) Renap: Relation network with adaptiveprototypical learning for few-shot classification. Neurocomputing 520:356–364

    Article  Google Scholar 

  28. Vinyals O, Blundell C, Lillicrap T, Wierstra D et al (2016) Matching networks for one shot learning. Adv Neural Inform Process Syst 29

  29. Snell J, Swersky K, Zemel R (2017) Prototypical networks for few-shot learning. Adv Neural Inform Process Syst 30

  30. Fei N, Lu Z, Xiang T, Huang S (2020) Melr: Meta-learning via modeling episode-level relationships for few-shot learning. In: International conference on learning representations

  31. Ye H-J, Ming L, Zhan D-C, Chao W-L (2022) Few-shot learning with a strong teacher. IEEE Trans Pattern Anal Mach Intell

  32. Shao Y, Wu W, You X, Gao C, Sang N (2022) Improving the generalization of maml in few-shot classification via bi-level constraint. IEEE Trans Circ Syst Vid Technol

  33. Zhu X, Li S (2022) Mgml: Momentum group meta-learning for few-shot image classification. Neurocomputing 514:351–361

    Article  Google Scholar 

  34. Finn C, Abbeel P, Levine S (2017) Model-agnostic meta-learning for fast adaptation of deep networks. In: International conference on machine learning, PMLR, pp 1126–1135

  35. Dong Z, Lin B, Xie F (2024) Optimizing distortion magnitude for data augmentation in few-shot remote sensing scene classification. Int J Remote Sens 45(4):1134–1147

    Article  Google Scholar 

  36. Zhang R, Yang Y, Li Y, Wang J, Li H, Miao Z (2023) Multi-task few-shot learning with composed data augmentation for image classification. IET Comput Vision 17(2):211–221

    Article  Google Scholar 

  37. Mangla P, Kumari N, Sinha A, Singh M, Krishnamurthy B, Balasubramanian VN (2020) Charting the right manifold: Manifold mixup for few-shot learning. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp 2218–2227

  38. Chen W-Y, Liu Y-C, Kira Z, Wang Y-CF, Huang J-B (2019) A closer look at few-shot classification. In: International conference on learning representations

  39. Tian Y, Wang Y, Krishnan D, Tenenbaum JB, Isola P (2020) Rethinking few-shot image classification: a good embedding is all you need? In: Computer vision–ECCV 2020: 16th European conference, Glasgow, UK, Proceedings, Part XIV 16, Springer, pp 266–282. Accessed 23–28 Aug 2020

  40. Yang S, Liu L, Xu M (2021) Free lunch for few-shot learning: Distribution calibration. In: International conference on learning representations

  41. Li W, Wang Z, Yang X, Dong C, Tian P, Qin T, Huo J, Shi Y, Wang L, Gao Y et al (2023) Libfewshot: A comprehensive library for few-shot learning. IEEE Trans Pattern Anal Mach Intell

  42. Hu SX, Li D, Stühmer J, Kim M, Hospedales TM (2022) Pushing the limits of simple pipelines for few-shot learning: External data and fine-tuning make a difference. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9068–9077

  43. Xi B, Li J, Li Y, Song R, Hong D, Chanussot J (2022) Few-shot learning with class-covariance metric for hyperspectral image classification. IEEE Trans Image Process 31:5079–5092

    Article  Google Scholar 

  44. Devos A, Grossglauser M (2020) Regression networks for meta-learning few-shot classification. In: 7th ICML Workshop on automated machine learning (AutoML 2020)

  45. Ren M, Triantafillou E, Ravi S, Snell J, Swersky K, Tenenbaum JB, Larochelle H, Zemel RS (2018) Meta-learning for semi-supervised few-shot classification. In: Proceedings of 6th international conference on learning representations ICLR

  46. Wah C, Branson S, Welinder P, Perona P, Belongie S (2011) The caltech-ucsd birds-200-2011 dataset

  47. Li Y, Qing L, He X, Chen H, Liu Q (2023) Image classification based on self-distillation. Appl Intell 53(8):9396–9408

    Article  Google Scholar 

  48. Wang L, He K, Liu Z (2024) Mcs: a metric confidence selection framework for few shot image classification. Multimed Tool Appl 83(4):10865–10880

  49. Snell J, Zemel R (2020) Bayesian few-shot classification with one-vs-each pólya-gamma augmented gaussian processes. In: International conference on learning representations

  50. Oh J, Yoo H, Kim C, Yun S (2021) Boil: Towards representation change for few-shot learning. In: The Ninth International Conference on Learning Representations (ICLR). Int Conf Learn Representations (ICLR)

  51. Liu B, Cao Y, Lin Y, Li Q, Zhang Z, Long M, Hu H (2020) Negative margin matters: Understanding margin in few-shot classification. In: Computer vision–ECCV 2020: 16th European conference, Glasgow, UK, Proceedings, Part IV 16, Springer, pp 438–455. Accessed 23–28 Aug 2020

  52. Chen Y, Liu Z, Xu H, Darrell T, Wang X (2021) Meta-baseline: Exploring simple meta-learning for few-shot learning. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 9062–9071

  53. Chen Z, Ge J, Zhan H, Huang S, Wang D (2021) Pareto self-supervised training for few-shot learning. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13663–13672

  54. Cheng J, Hao F, Liu L, Tao D (2022) Imposing semantic consistency of local descriptors for few-shot learning. IEEE Trans Image Process 31:1587–1600

    Article  Google Scholar 

  55. Lu Y, Wen L, Liu J, Liu Y, Tian X (2022) Self-supervision can be a good few-shot learner. In: European conference on computer vision, Springer, pp 740–758

  56. Zhang M, Huang S, Li W, Wang D (2022) Tree structure-aware few-shot image classification via hierarchical aggregation. In: European Conference on Computer Vision, pp. 453–470 . Springer

  57. Li W, Xie L, Gan P, Zhao Y (2023) Self-supervised pairwise-sample resistance model for few-shot classification. Appl Intell pp 1–14

  58. Afrasiyabi A, Larochelle H, Lalonde J-F, Gagné C (2022) Matching feature sets for few-shot image classification. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 9014–9024

  59. Ye H-J, Hu H, Zhan D-C, Sha F (2020) Few-shot learning via embedding adaptation with set-to-set functions. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8808–8817

  60. Hao F, He F, Cheng J, Tao D (2021) Global-local interplay in semantic alignment for few-shot learning. IEEE Trans Circuits Syst Video Technol 32(7):4351–4363

    Article  Google Scholar 

  61. He X, Lin J (2022) Weakly-supervised object localization based fine-grained few-shot learning. J Image Graph (007):027

  62. Zhang J, Zhang X, Wang Z (2022) Task encoding with distribution calibration for few-shot learning. IEEE Trans Circuits Syst Video Technol 32(9):6240–6252

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Contributions

Bin Song: design, implementation, formal analysis and writing. Hong Zhu: guidance, review and editing. Yuandong Bi: validation.

Corresponding author

Correspondence to Hong Zhu.

Ethics declarations

Ethical and Informed Consent for Data Used

No ethical approval or informed consent was necessary for this study, as the data were already publicly available and did not involve human or animal subjects.

Competing Interests

The paper is original in terms of its contents and is not under consideration for publication in any other journals/proceedings. The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Song, B., Zhu, H. & Bi, Y. A conditioned feature reconstruction network for few-shot classification. Appl Intell (2024). https://doi.org/10.1007/s10489-024-05516-9

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-024-05516-9

Keywords

Navigation