Abstract
To develop new computer vision capabilities leveraging artificial intelligence, we will increasingly need to use operationally realistic training and validation datasets. Although operational full motion video and imagery datasets present information regarding their provenance and classification level, these designations are often not indicative of the presence of potentially offensive or traumatizing content. As machine learning and data scientists increasingly need to work with operational unsanitized operational video and imagery data, they will have a higher risk of being exposed to sensitive and traumatic content. In this paper, we first raise awareness about this risk within the Defense community. Then, we propose several approaches for mitigating machine learning practitioner's exposure to offensive and traumatizing media, including dataset preprocessing procedures and viewing tool design considerations.
Published Version (Free)
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have