Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Jul 20;24(14):4707.
doi: 10.3390/s24144707.

Deformation Estimation of Textureless Objects from a Single Image

Affiliations

Deformation Estimation of Textureless Objects from a Single Image

Sahand Eivazi Adli et al. Sensors (Basel). .

Abstract

Deformations introduced during the production of plastic components degrade the accuracy of their 3D geometric information, a critical aspect of object inspection processes. This phenomenon is prevalent among primary plastic products from manufacturers. This work proposes a solution for the deformation estimation of textureless plastic objects using only a single RGB image. This solution encompasses a unique image dataset of five deformed parts, a novel method for generating mesh labels, sequential deformation, and a training model based on graph convolution. The proposed sequential deformation method outperforms the prevalent chamfer distance algorithm in generating precise mesh labels. The training model projects object vertices into features extracted from the input image, and then, predicts vertex location offsets based on the projected features. The predicted meshes using these offsets achieve a sub-millimeter accuracy on synthetic images and approximately 2.0 mm on real images.

Keywords: deformation estimation; graph convolution; image dataset; label generation; single image; textureless deformed object.

PubMed Disclaimer

Conflict of interest statement

Authors Joshua K. Pickard and Ganyun Sun were employed by the company Eigen Innovations. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

Figure 1
Figure 1
Images of the five textureless plastic objects studied in this paper: (a) skateboard, (b) bracket, (c) round flat receptacle lid (rfrl), (d) oval flange, and (e) vent cover.
Figure 2
Figure 2
Images of the deformed versions of the skateboard; the wireframe illustrates the undeformed model: (a) F=22.5 N, MD=10.08 mm; (b) F=44.5 N, MD=19.93 mm; (c) F=22.5 N, MD=10.08 mm; (d) F=44.5 N, MD=19.93 mm.
Figure 3
Figure 3
Images of the deformed versions of the bracket; the wireframe illustrates the undeformed model: (a) F=24 N, MD=5.14 mm; (b) F=66 N, MD=14.15 mm; (c) F=24 N, MD=5.14 mm; (d) F=66 N, MD=14.15 mm.
Figure 4
Figure 4
Images of the deformed versions of the round flat receptacle lid; the wireframe illustrates the undeformed model: (a) F=1050 N,MD=6.09 mm; (b) F=2250 N,MD=13.06 mm; (c) F=3450 N, MD=20.03 mm; (d) F=4300 N,MD=4300 mm.
Figure 5
Figure 5
Images of the deformed versions of the oval flange; the wireframe illustrates the undeformed model: (a) F=1300 N, MD=8.93 mm; (b) F=2650 N, MD=18.20 mm; (c) F=1300 N, MD=8.93 mm; (d) F=2650 N,MD=18.20 mm.
Figure 6
Figure 6
Images of the deformed versions of the vent cover; the wireframe illustrates the undeformed model: (a) F=150 N,MD=3.88 mm; (b) F=540 N, MD=13.97 mm; (c) F=150 N, MD=3.88 mm; (d) F=540 N, MD=13.97 mm.
Figure 7
Figure 7
This schematic depicts the camera position relative to the deformed vent cover, located at the origin (0,0,0) of Blender’s global coordinate system. The red, green, and blue axes represent the X, Y, and Z axes, respectively.
Figure 8
Figure 8
Images of the deformed vent cover with F=540.0 N, MD=13.97 mm, and RZO=0°. (a) TXL=0 mm, (b) TXL=25 mm, (c) TXL=35 mm, (d) TXL=45 mm, (e) RXC=0°, (f) RXC=10°, (g) RXC=20°, and (h) RXC=30°. If not mentioned, RXC=60° and TXL=0 mm.
Figure 9
Figure 9
Images of the 3D-printed deformed models captured by a smartphone camera. (a) Deformed model of the bracket with F=66.0 N, MD=14.15 mm. (b) Deformed bracket component with F=66.0 N, MD=14.15 mm. (c) Deformed vent cover component with F=270 N, MD=6.98 mm. (d) Deformed model of the vent cover object with F=500.0 N, MD=12.93 mm.
Figure 10
Figure 10
Setup devised to capture pictures from the printed deformed models.
Figure 11
Figure 11
Real-world images of a 3D-printed, deformed vent cover (F=500.0 N, MD=12.93 mm) used for training the machine learning model. (a) RZO=0.70° and RXC=85.36°. (b) RZO=36.46° and RXC=82.09°. (c) RZO=34.40°, and RXC=72.71°. (d) RZO=3.49° and RXC=76.82°.
Figure 12
Figure 12
Mesh comparison: (a) Deformed mesh model, Ansys output with 6268 vertices and 12,544 faces. (b) Initial undeformed mesh model (training model input) with 1145 vertices and 2298 faces.
Figure 13
Figure 13
The chamfer distance algorithm applied on two 2D distributions. False correspondences (red dashed line) for middle blue point, and true correspondence (green dashed line).
Figure 14
Figure 14
Sequential deformation (lower path) vs. direct application of the chamfer distance (upper path) on a vent cover object with significant geometric variation (e). (a) Initial undeformed model. (b) Least deformed model with MD=0.25 mm,F=10 N. (c) Average deformed model with MD=7.76 mm,F=300 N. (d) Deformed model with MD=14.75 mm,F=570 N. (e) Highest deformed model with MD=15.00 mm,F=580 N. Labels were generated using (f) direct chamfer distance and (g) sequential deformation algorithm.
Figure 15
Figure 15
Sequential deformation (SD) vs. chamfer distance (CD) on skateboard object with MD=19.93 mm (first row) and MD=19.93 mm (second row). (a,e) Initial undeformed model. (b,f) Deformed models (Ansys output). (c,g) CD algorithm outputs. (d,h) SD algorithm outputs.
Figure 16
Figure 16
Sequential deformation (SD) vs. chamfer distance (CD) on bracket object with MD=14.15 mm (first row) and MD=14.15 mm (second row). (a,e) Initial undeformed model. (b,f) Deformed models (Ansys output). (c,g) CD algorithm outputs. (d,h) SD algorithm outputs.
Figure 17
Figure 17
Sequential deformation (SD) vs. chamfer distance (CD) on RFRL object with MD=24.97 mm. (a) Initial undeformed model. (b) Deformed model (Ansys output). (c) CD algorithm output. (d) SD algorithm output.
Figure 18
Figure 18
Sequential deformation (SD) vs. chamfer distance (CD) on oval flange object with MD=18.20 mm (first row) and MD=18.20 mm (second row). (a,e) Initial undeformed model. (b,f) Deformed models (Ansys output). (c,g) CD algorithm outputs. (d,h) SD algorithm outputs.
Figure 19
Figure 19
Sequential deformation (SD) vs. chamfer distance (CD) on vent cover object with MD=15.00 mm. (a) Initial undeformed model. (b) Deformed model (Ansys output). (c) CD algorithm output. (d) SD algorithm output.
Figure 20
Figure 20
Training model pipeline: Blue cubes represent 2D convolutional layers, and yellow cubes denote max-pooling layers. The “cam” block represents the camera model’s intrinsic and extrinsic parameters. Green rectangles symbolize graph convolutional layers, while magenta rectangles represent dense layers.
Figure 21
Figure 21
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted skateboard [F=33.5 N,MD=15.0 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed skateboard fed to the training model. (ce) Predicted mesh of the deformed skateboard from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 22
Figure 22
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted skateboard [F=33.5 N,MD=15.0 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed skateboard fed to the training model. (ce) Predicted mesh of the deformed skateboard from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 23
Figure 23
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted bracket [F=28 N,MD=6.00 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed bracket fed to the training model. (ce) Predicted mesh of the deformed bracket from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 24
Figure 24
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted bracket [F=46 N,MD=9.86 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed bracket fed to the training model. (ce) Predicted mesh of the deformed bracket from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 25
Figure 25
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted RFRL [F=2750 N,MD=15.96 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed RFRL fed to the training model. (ce) Predicted mesh of the deformed RFRL from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 26
Figure 26
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted RFRL [F=3950 N,MD=22.93 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed RFRL fed to the training model. (ce) Predicted mesh of the deformed RFRL from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 27
Figure 27
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted oval flange [F=1750 N,MD=12.02 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed oval flange fed to the training model. (ce) Predicted mesh of the deformed oval flange from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 28
Figure 28
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted oval flange [F=2200 N,MD=15.11 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed oval flange fed to the training model. (ce) Predicted mesh of the deformed oval flange from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 29
Figure 29
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted vent cover [F=310 N,MD=8.02 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed vent cover fed to the training model. (ce) Predicted mesh of the deformed vent cover from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 30
Figure 30
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted vent cover [F=390 N,MD=10.09 mm]. (a) Input mesh of the training network (undeformed). (b) Input image of the deformed vent cover fed to the training model. (ce) Predicted mesh of the deformed vent cover from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 31
Figure 31
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted bracket [F=28 N,MD=6.00 mm]. (a) Input mesh of the training network (undeformed). (b) Real image of the actual deformed bracket fed to the training model. (ce) Predicted mesh of the deformed bracket from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).
Figure 32
Figure 32
Visualization of actual Euclidean distance error between the predicted mesh and the ground truth for each vertex of the predicted vent cover [F=500 N,MD=10.09 mm]. (a) Input mesh of the training network (undeformed). (b) Real image of the 3D-printed deformed vent cover fed to the training model. (ce) Predicted mesh of the deformed vent cover from different viewpoints. (f) Color bar representing the magnitude of the Euclidean distance error in millimeters (mm).

References

    1. Lowe D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004;60:91–110. doi: 10.1023/B:VISI.0000029664.99615.94. - DOI
    1. Bay H., Tuytelaars T., Van Gool L. SURF: Speeded Up Robust Features. In: Leonardis A., Bischof H., Pinz A., editors. Computer Vision—ECCV 2006. Volume 3951. Springer; Berlin/Heidelberg, Germany: 2006. pp. 404–417. Lecture Notes in Computer Science. - DOI
    1. Wang N., Zhang Y., Li Z., Fu Y., Yu H., Liu W., Xue X., Jiang Y.-G. Pixel2mesh: 3D mesh model generation via image guided deformation. IEEE Trans. Pattern Anal. Mach. Intell. 2021;43:3600–3613. doi: 10.1109/TPAMI.2020.2984232. - DOI - PubMed
    1. Moemen M.Y., Elghamrawy H., Givigi S.N., Noureldin A. 3-D reconstruction and measurement system based on multimobile robot machine vision. IEEE Trans. Instrum. Meas. 2021;70:5003109. doi: 10.1109/TIM.2020.3026719. - DOI
    1. Wang Y., James S., Stathopoulou E.K., Beltrán-González C., Konishi Y., Del Bue A. Autonomous 3-D reconstruction, mapping, and exploration of indoor environments with a robotic arm. IEEE Trans. Robot. Autom. Lett. 2019;4:3340–3347. doi: 10.1109/LRA.2019.2926676. - DOI

Grants and funding

LinkOut - more resources