Abstract
In the paper, a method for image search is proposed. It allows for finding images starting from a text query that contains names of object classes and their expected spatial relations. The input query containing the description of desired objects and their spatial relations is used to select and score relevant images. Next, the score is used to sort the output images putting more relevant ones on the top of the list. The proposed approach is based on object detection methods and a fuzzy approach describing the position of detected objects in relation to other ones. The method may be used to retrieve images from databases containing images with metadata produced by object detectors.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Bloch, I.: Fuzzy spatial relationships for image processing and interpretation: a review. Image Vis. Comput. 23(2), 89–110 (2005)
Bloch, I.: Fuzzy Models of Spatial Relations, Application to Spatial Reasoning, pp. 51–58. Springer, Berlin (2013)
Cohn, A., Hazarika, S.: Qualitative spatial representation and reasoning: An overview. Fundam. Inform. 46(4), 1–29 (2001)
Girshick, R., Donahue, J., Darrell, T., Malik, J.: Rich feature hierarchies for accurate object detection and semantic segmentation. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 580–587 (2014)
Iwanowski, M., Bartosiewicz, M.: Describing images using fuzzy mutual position matrix and saliency-based ordering of predicates. In: 2021 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1–8 (2021)
Iwanowski, M., Grzabka, M.: Similarity and symmetry measures based on fuzzy descriptors of image objects’ composition. CoRR abs/2107.13651 (2021). https://arxiv.org/abs/2107.13651
Khan, A., Sohail, A., Zahoora, U., Qureshi, A.S.: A survey of the recent architectures of deep convolutional neural networks. Artif. Intell. Rev. 53(8), 5455–5516 (2020)
LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436–444 (2015)
Lin, T.Y., Maire, M., Belongie, S., Bourdev, L., Girshick, R., Hays, J., Perona, P., Ramanan, D., Zitnick, C.L., Dollr, P.: Microsoft coco: Common objects in context (2014). https://arxiv.org/abs/1405.0312
Naeem, M., Matsakis, P.: Relative position descriptors a review. In: Proceedings of ICPRAM 2015, vol. 1, pp. 286–295 (2015), cited By 7
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779–788 (2016)
Redmon, J., Farhadi, A.: Yolov3: An Incremental Improvement (2018)
Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137–1149 (2017)
Smeulders, A.W.M., Worring, M., Santini, S., Gupta, A., Jain, R.: Content-based image retrieval at the end of the early years. IEEE Trans. Pattern Anal. Mach. Intell. 22(12), 1349–1380 (2000)
Zadeh, L.: Fuzzy sets. Inf. Control 8(3), 338–353 (1965)
Zhang, W., Sugeno, M.: A fuzzy approach to scene understanding. In: [Proceedings 1993] Second IEEE International Conference on Fuzzy Systems, vol. 1, pp. 564–569 (1993)
Zhao, Z., Zheng, P., Xu, S., Wu, X.: Object detection with deep learning: A review. IEEE Trans. Neural Networks Learn. Syst. 30(11), 3212–3232 (2019)
Acknowledgements
Research work was carried out as part of the project “Powiedz mi, co widzisz” (“Tell me, what you see”) founded by the POB SzIR of the Warsaw University of Technology.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Iwanowski, M., Haidukievich, A., Leszczynski, M., Wnorowski, B. (2023). Fuzzy Approach to Object-Detection-Based Image Retrieval. In: Chmielewski, L.J., Orłowski, A. (eds) Computer Vision and Graphics. ICCVG 2022. Lecture Notes in Networks and Systems, vol 598. Springer, Cham. https://doi.org/10.1007/978-3-031-22025-8_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-22025-8_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-22024-1
Online ISBN: 978-3-031-22025-8
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)