Journal article
Journal of Manufacturing and Materials Processing, 2022
Additive Manufacturing R&D
Electrical & Computer Eng.
Western University
APA
Click to copy
Petsiuk, A. L., Singh, H., Dadhwal, H., & Pearce, J. M. (2022). Synthetic-to-real Composite Semantic Segmentation in Additive Manufacturing. Journal of Manufacturing and Materials Processing.
Chicago/Turabian
Click to copy
Petsiuk, Aliaksei L., Harnoor Singh, Himanshu Dadhwal, and Joshua M. Pearce. “Synthetic-to-Real Composite Semantic Segmentation in Additive Manufacturing.” Journal of Manufacturing and Materials Processing (2022).
MLA
Click to copy
Petsiuk, Aliaksei L., et al. “Synthetic-to-Real Composite Semantic Segmentation in Additive Manufacturing.” Journal of Manufacturing and Materials Processing, 2022.
BibTeX Click to copy
@article{aliaksei2022a,
title = {Synthetic-to-real Composite Semantic Segmentation in Additive Manufacturing},
year = {2022},
journal = {Journal of Manufacturing and Materials Processing},
author = {Petsiuk, Aliaksei L. and Singh, Harnoor and Dadhwal, Himanshu and Pearce, Joshua M.}
}
The application of computer vision and machine learning methods for semantic segmentation of the structural elements of 3D-printed products in the field of additive manufacturing (AM) can improve real-time failure analysis systems and potentially reduce the number of defects by providing additional tools for in situ corrections. This work demonstrates the possibilities of using physics-based rendering for labeled image dataset generation, as well as image-to-image style transfer capabilities to improve the accuracy of real image segmentation for AM systems. Multi-class semantic segmentation experiments were carried out based on the U-Net model and the cycle generative adversarial network. The test results demonstrated the capacity of this method to detect such structural elements of 3D-printed parts as a top (last printed) layer, infill, shell, and support. A basis for further segmentation system enhancement by utilizing image-to-image style transfer and domain adaptation technologies was also considered. The results indicate that using style transfer as a precursor to domain adaptation can improve real 3D printing image segmentation in situations where a model trained on synthetic data is the only tool available. The mean intersection over union (mIoU) scores for synthetic test datasets included 94.90% for the entire 3D-printed part, 73.33% for the top layer, 78.93% for the infill, 55.31% for the shell, and 69.45% for supports.