Abstract
Subject matter experts can sometimes provide incorrect and/or incomplete knowledge in the process of building intelligent systems. Other times, the expert articulates correct knowledge only to be misinterpreted by the knowledge engineer. In yet other cases, changes in the domain can lead to outdated knowledge in the system. This paper describes a technique that improves a flawed tactical agent by revising its knowledge through practice in a simulated version of its operational environment. This form of theory revision repairs agents originally built through interaction with subject matter experts. It is advantageous because such systems can now cease to be completely dependent on human expertise to provide correct and complete domain knowledge. After an agent has been built in consultation with experts, and before it is allowed to become operational, our method permits its improvement by subjecting it to several practice sessions in a simulation of its mission environment. Our method uses reinforcement learning to correct such errors and fill in gaps in the knowledge of a context-based tactical agent. The method was implemented and evaluated by comparing the performance of an agent improved by our method, to the original hand-built agent whose knowledge was purposely seeded with known errors and/or gaps. The results show that the improved agent did in fact correct the seeded errors and did gain the missing knowledge to permit it to perform better than the original, flawed agent.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.