Hostname: page-component-89b8bd64d-dvtzq Total loading time: 0 Render date: 2026-05-07T06:38:42.577Z Has data issue: false hasContentIssue false

Content-based image retrieval for industrial material images with deep learning and encoded physical properties

Published online by Cambridge University Press:  21 September 2023

Myung Seok Shim*
Affiliation:
Shell Global Solutions (US), Inc., Houston, TX, USA
Christopher Thiele
Affiliation:
ERGO Group AG, Düsseldorf, Germany
Jeremy Vila
Affiliation:
Shell International E&P, Inc., Houston, TX, USA
Nishank Saxena
Affiliation:
Shell International E&P, Inc., Houston, TX, USA
Detlef Hohl
Affiliation:
Shell International E&P, Inc., Houston, TX, USA
*
Corresponding author: Myung Seok Shim; Email: Myung-Seok.Shim@shell.com

Abstract

Industrial materials images are an important application domain for content-based image retrieval. Users need to quickly search databases for images that exhibit similar appearance, properties, and/or features to reduce analysis turnaround time and cost. The images in this study are 2D images of millimeter-scale rock samples acquired at micrometer resolution with light microscopy or extracted from 3D micro-CT scans. Labeled rock images are expensive and time-consuming to acquire and thus are typically only available in the tens of thousands. Training a high-capacity deep learning (DL) model from scratch is therefore not practicable due to data paucity. To overcome this “few-shot learning” challenge, we propose leveraging pretrained common DL models in conjunction with transfer learning. The “similarity” of industrial materials images is subjective and assessed by human experts based on both visual appearance and physical qualities. We have emulated this human-driven assessment process via a physics-informed neural network including metadata and physical measurements in the loss function. We present a novel DL architecture that combines Siamese neural networks with a loss function that integrates classification and regression terms. The networks are trained with both image and metadata similarity (classification), and with metadata prediction (regression). For efficient inference, we use a highly compressed image feature representation, computed offline once, to search the database for images similar to a query image. Numerical experiments demonstrate superior retrieval performance of our new architecture compared with other DL and custom-feature-based approaches.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press
Figure 0

Figure 1. Proposed Double Siamese Neural Network (DSNN) architecture. The blue Siamese part of the network is trained with image similarity and the gray Siamese part is trained with image and metadata similarity in classification. The dark gray part is trained with metadata for regression. “x,” “+,” “-“designate multiplication, addition, and subtraction. The green and yellow images are pairs of images, with binary classification label “similar” or “different.” After the training phase the networks are used in the inference phase for image retrieval and metadata regression as shown in Figure 3. See text for the mathematical details and symbols.

Figure 1

Figure 2. Examples of 2D micro-CT and thin-section rock images. Micro-CT images are gray scale, while thin-section images are RGB.

Figure 2

Figure 3. Image retrieval system with the trained DSNN. Note that the feature encoding for the images in the data base (top) is computed offline and not during inference for a new query image (bottom).

Figure 3

Table 1. Image retrieval performance comparison.

Figure 4

Figure 4. Regression coefficient of determination $ {R}^2 $ for porosity and permeability in the 2D micro-CT dataset.

Figure 5

Figure 5. Regression coefficient of determination $ {R}^2 $ for porosity in the thin-section dataset.

Figure 6

Figure 6. Cross plot of porosity and permeability of query images and their top three matching retrieval images for the 2D micro-CT dataset. (a and b) are for porosity and (c and d) are for permeability. Our DSNN (right panels) reliably retrieves rock samples with similar rock properties; most of the retrievals are on the diagonal line. The top, second-, and third-best matches are colored blue, orange, and green, respectively.

Figure 7

Figure 7. (Porosity, permeability) of query and (porosity, permeability) of top three retrievals compared for the 2D micro-CT dataset. (a and b) are for porosity and (c and d) are for permeability. The error histogram for our DSNN shows a narrower spread around zero (perfect match) than the vanilla ResNet-34.

Figure 8

Figure 8. Comparison of the porosity of a query image and the top three ranked images retrieved in the thin-section dataset. (a) and (b) are cross plots of porosity of query images and the top3 matching retrieval images. (c) and (d) are histograms to display the cross plots in terms of distributions. Similar to the results in the 2D micro-CT dataset, our DSNN in (d) has a higher peak at zero.

Figure 9

Figure 9. Qualitative retrieval results with reservoir E2 on 2D micro-CT dataset. Vanilla ResNet-34 (yellow) vs our DSNN (purple). $ por $ and $ perm $ are lab-measured porosity and permeability. $ po{r}^{\prime } $ and $ per{m}^{\prime } $ represent predicted porosity and permeability from DSNN.

Figure 10

Figure 10. Qualitative retrieval results with Reservoir K on 2D micro-CT dataset. Vanilla ResNet-34 (yellow) vs our DSNN (purple).

Figure 11

Figure 11. Qualitative retrieval results with reservoir U on the 2D thin-section dataset. Vanilla ResNet-34 (yellow) vs our DSNN (purple). $ por $ is a lab-measured porosity and $ po{r}^{\prime } $ represents porosity predicted from DSNN.

Figure 12

Figure 12. Qualitative retrieval results for reservoir D on the 2D thin-section dataset. Vanilla ResNet-34 (yellow) vs our DSNN (purple).

Submit a response

Comments

No Comments have been published for this article.