A 2018 study by HfS Research, conducted in partnership with IPsoft, found that C-Level executives are not only optimistic about securing Artificial Intelligence (AI) tools, but they’re also intrigued by the role AI can play in actually improving overall data security. In that study, 59% of respondents said they were pleased with the security benefits gained by implementing cognitive tools. As their satisfaction with AI security and the safeguards around it grows, organizations are starting to examine the inherent biases in AI tools when programmed by homogeneous teams of developers.
Fast Company explores this topic in great detail in a recent article. The author discusses AI bias with developers from the world’s leading AI companies, including IPsoft’s Tracey Robinson, Director of Cognitive Implementation for Amelia. To demonstrate how seriously IPsoft takes the possibility of AI bias, Tracey discusses the work we do with linguists and designers to program our solutions, and the lengths at which we go to best diversify our implementation teams.
“Human nature makes the complete elimination of biases impossible, which is why it is an organizational imperative to employ as diverse an AI training group as possible, be it culturally, geographically, gender, experience and skill set,” Tracey tells Fast Company.
For more on the ways in which IPsoft attempts to root out bias in our AI solutions, as well as additional thoughts from Tracey on the dramatic potential for AI to improve society, be sure to read the Fast Company article.