Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Feb 22;25(5):1350.
doi: 10.3390/s25051350.

Content-Based Histopathological Image Retrieval

Affiliations

Content-Based Histopathological Image Retrieval

Camilo Nuñez-Fernández et al. Sensors (Basel). .

Abstract

Feature descriptors in histopathological images are an important challenge for the implementation of Content-Based Image Retrieval (CBIR) systems, an essential tool to support pathologists. Deep learning models like Convolutional Neural Networks and Vision Transformers improve the extraction of these feature descriptors. These models typically generate embeddings by leveraging deeper single-scale linear layers or advanced pooling layers. However, these embeddings, by focusing on local spatial details at a single scale, miss out on the richer spatial context from earlier layers. This gap suggests the development of methods that incorporate multi-scale information to enhance the depth and utility of feature descriptors in histopathological image analysis. In this work, we propose the Local-Global Feature Fusion Embedding Model. This proposal is composed of three elements: (1) a pre-trained backbone for feature extraction from multi-scales, (2) a neck branch for local-global feature fusion, and (3) a Generalized Mean (GeM)-based pooling head for feature descriptors. Based on our experiments, the model's neck and head were trained on ImageNet-1k and PanNuke datasets employing the Sub-center ArcFace loss and compared with the state-of-the-art Kimia Path24C dataset for histopathological image retrieval, achieving a Recall@1 of 99.40% for test patches.

Keywords: content-based image retrieval; feature embedding; feature fusion; histopathological image; transfer learning.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflicts of interest.

Figures

Figure 1
Figure 1
The LGFFEM architecture comprises a pre-trained backbone as a feature extractor from multi-scale stages, a trainable neck consisting of layers for local–global feature fusion, and a pooling head composed of trainable GeM mini-heads for each multi-scale fused feature from the neck.
Figure 2
Figure 2
Illustration of the bottleneck operation for the Local and Global Aggregators and the pooling GeM mini-head. (a) Detailed schematic of the Global Feature Aggregator Unit. (b) Detailed schematic of the Local Feature Aggregator Unit. (c) Detailed schematic of the mini-head unit from the GeM head.
Figure 3
Figure 3
Query image selected for the class set S0 and their first retrieve image from the Kimia Patch24C dataset. (a) Query image ID S0-1. (b) First image retrieved ID S0-2.
Figure 4
Figure 4
Grad-CAM applied to the first layer of the neck used in strategy C for the first image’s retrieved ID S0-2. (a) Grad-CAM applied to the outer aggregation fusion node P1_2 in Layer 1. (b) Grad-CAM applied to the outer aggregation fusion node P2_2 in Layer 1. (c) Grad-CAM applied to the outer aggregation fusion node P3_2 in Layer 1. (d) Grad-CAM applied to the outer aggregation fusion node P4_2 in Layer 1. (e) Grad-CAM applied to the collapse of all outer aggregation fusion nodes in Layer 1.
Figure 5
Figure 5
Grad-CAM applied to the second layer of the neck used in strategy C for the first image’s retrieved ID S0-2. (a) Grad-CAM applied to the outer aggregation fusion node P1_2 in Layer 2. (b) Grad-CAM applied to the outer aggregation fusion node P2_2 in Layer 2. (c) Grad-CAM applied to the outer aggregation fusion node P3_2 in Layer 2. (d) Grad-CAM applied to the outer aggregation fusion node P4_2 in Layer 2. (e) Grad-CAM applied to the collapse of all outer aggregation fusion nodes in Layer 2.
Figure 6
Figure 6
Grad-CAM applied to the third layer of the neck used in strategy C for the first image’s retrieved ID S0-2. (a) Grad-CAM applied to the outer aggregation fusion node P1_2 in Layer 3. (b) Grad-CAM applied to the outer aggregation fusion node P2_2 in Layer 3. (c) Grad-CAM applied to the outer aggregation fusion node P3_2 in Layer 3. (d) Grad-CAM applied to the outer aggregation fusion node P4_2 in Layer 3. (e) Grad-CAM applied to the collapse of all outer aggregation fusion nodes in Layer 3.
Figure 7
Figure 7
Visualization of the 2D projection of embeddings from the Kimia Patch24C test dataset using strategy C. Each dot represents an image, and each color represents a class of tissue.

Similar articles

References

    1. Solar M., Castañeda V., Ñanculef R., Dombrovskaia L., Araya M. A Data Ingestion Procedure towards a Medical Images Repository. Sensors. 2024;24:4985. doi: 10.3390/s24154985. - DOI - PMC - PubMed
    1. Rahaman M.M., Li C., Wu X., Yao Y., Hu Z., Jiang T., Li X., Qi S. A Survey for Cervical Cytopathology Image Analysis Using Deep Learning. IEEE Access. 2020;8:61687–61710. doi: 10.1109/ACCESS.2020.2983186. - DOI
    1. Solar M., Aguirre P. Deep learning techniques to process 3D chest CT. J. Univ. Comput. Sci. 2024;30:758. doi: 10.3897/jucs.112977. - DOI
    1. Hegde N., Hipp J.D., Liu Y., Emmert-Buck M., Reif E., Smilkov D., Terry M., Cai C.J., Amin M.B., Mermel C.H., et al. Similar image search for histopathology: SMILY. Npj Digit. Med. 2019;2:56. doi: 10.1038/s41746-019-0131-z. - DOI - PMC - PubMed
    1. Hashimoto N., Takagi Y., Masuda H., Miyoshi H., Kohno K., Nagaishi M., Sato K., Takeuchi M., Furuta T., Kawamoto K., et al. Case-based similar image retrieval for weakly annotated large histopathological images of malignant lymphoma using deep metric learning. Med. Image Anal. 2023;85:102752. doi: 10.1016/j.media.2023.102752. - DOI - PubMed

Grants and funding

LinkOut - more resources