Abstract

Nowadays, modular domain ontology, where each module represents a subdomain of the ontology domain, facilitates the reuse of information and provides users with domain-specific knowledge. In this paper, we focus on modular taxonomy learning from text, where each module collects terms with the same topic insights, and in parallel we manage to discover hypernym and 'related' relations among those collected terms.However, it is difficult to automatically fit terms into modules and discover relations.We propose to employ twice trainedLDAto partition termsof each subdomain, and relate subdomains into modules of ontology. Meanwhile, we apply core concept replacement and subdomain knowledge supplementation as supportive information embedding technique over the corpus. This shows that the twice trained LDA strategy can effectively identify topic-relevant terms into subdomains, with nearly two-fold precision comparing to that of normal LDA training. The combination of core concept replacement and subdomain knowledge supplementation contributes to significant improvements in modular taxonomy learning.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.