Abstract

In recent years, the representation of social images has achieved significant results in multilabel classification and cross-modal retrieval. However, simply modeling content information may lead to less accurate embeddings, as social images contain both content and social relationships between images in a variety of modalities. Learning low-dimensional, dense, and sequential representations of social networked multimodal data is the goal of social image representation learning. Many practical applications can also be facilitated by these methods. Many tasks are aided by multiview representations learning, such as cross-view classification. Since social images often contain link information in addition to multimodal content, simply adopting data content may lead to suboptimal multiview representations of social images. Social images are often rich in metadata, and there are ways to turn images and their textual descriptions into links between content, which are ultimately handled as multimodal content.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.