Abstract

This paper introduces a novel approach aimed at efficiently extracting dominant colors from online fashion images. The method addresses challenges related to detecting overlapping objects and computationally expensive methods by combining K-means clustering and graph-cut techniques into a framework. This framework incorporates an adaptive weighting strategy to enhance color extraction accuracy. Additionally, it introduces a two-phase fashion apparel detection method called YOLOv4, which utilizes U-Net architecture for clothing segmentation to precisely separate clothing items from the background or other elements. Experimental results show that K-means with YOLOv4 outperforms K-means with the U-Net model. These findings suggest that the U-Net architecture and YOLOv4 models can be effective methods for complex image segmentation tasks in online fashion retrieval and image processing, particularly in the rapidly evolving e-commerce environment.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.