As we are talking about data-driven AI decisions in all areas of the enterprise, we need to consider the risk of relying on AI models for optimal decision making. Think of an AI model like a prism that focuses your sight on the decision relevant facts and scenarios for your own individual job tasks:
What happens if malicious hackers inject a biased version of "reality" into your corporate AI platforms?
Can maliciously manipulated AI models endanger an organization?
How do you find out about internal or external AI services being compromised?
Read my blog post "Hijacking Corporate Reality through AI" over on my LinkedIn blog.
By definition, artificial intelligence (AI) has the capability to introduce a bias, or false sense of reality, into the enterprise. Today's belief often is that data-driven decision making must lead to the right conclusions and that the more data points we gather, the better founded in reality the decisions are. This is true as long as long as we carefully create, secure, and manage our AI models in order to directly tie them to reality. Here are my six rules to prevent AI bias in a nutshell. For examples and more detail, please read my article on this topic at TechTarget's SearchSoftwareQuality site.
1. AI Model Transparency: Clearly document the what goes into the AI model, how the decision process works and what the model's limitations are.
2. Validate the Training Data: Always talk to subject matter experts to fully understand the business background of input variables. This will also help tune and test the model in the end.
3. Carefully evaluate commercial data sets: While this rule applies to all data sets, it is even more important for commercially purchased data sets. Always carefully keep in mind any potential bias that could have been introduced through "cutting corners" or through commercial interests of the vendor.
4. Dictionaries: Dictionaires are the "connector" between the real world and your AI model. If dictionaries are incomplete, biased, or inaccurate, you model will not be able to recognize the relevant input variables and arrive at invalid conclusions.
5. Transfer Learning: When using an existing AI model to solve a related but different problem, it is crucial to carefully test your assumptions about the original model's ability to "grasp" the new task.
6. Feedback Loop: Modern reinforcement learning requires well designed feedback loops to continuously tune the AI model based on the results of its output. Often identifying these results is tricky, as the environment can contain an infinite number of confounding variables.
Read my piece at TechTarget on the same topic: Prevent AI bias in predictive applications before it starts
This Blog is all about demystifying artificial intelligence and machine learning (AI/ML) for enterprise use. The EMA team and outside experts will offer pragmatic advice to help you plan, prepare, and execute your AI/ML projects. Without becoming overly technical, this blog will provide perspective and a clear understanding of how ML/AI works and what results we can and cannot expect today.