Abstract
Generative models, particularly diffusion-based approaches, have gained significant attention recently due to their ability to create realistic outputs. Despite their potential, the application of these models in manufacturing remains largely unexplored. This work presents a framework that addresses this gap by generating machined surface images guided by multiple sensor inputs in manufacturing. The proposed model integrates information from multiple sensors with varying sampling rates using multimodal embedding and employs a latent diffusion model to translate the fused sensor embedding into an image embedding, which is then converted into a machined surface image. The effectiveness of the framework is validated using real-world time-series data, including force, torque, acceleration, and sound, collected from various industrial processes, such as a carbon-fiber-reinforced plastic drilling process. The results demonstrate the model’s ability to predict defects from the generated machined surface images. The proposed approach can potentially revolutionize prognostics and health management (PHM) in smart manufacturing by enabling sensor-guided visual inspection, defect detection, process monitoring, and predictive maintenance.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.