Even though AI has been around for years, the AI and ethics conversation is just getting started, and it has a long, long way to go. As AI is used in new ways, new ethical concerns will arise. There is much that still needs to be done, but increasing awareness and education, as well as broadening the types of participants involved in the conversation, are foundational first steps.
As AI becomes more widely used, many companies are unclear about how to best navigate the murky waters of AI and ethics. No organization wants to risk the public relations fiasco that would ensue, should it be determined that the AI algorithms it used yielded results that were biased against a specific demographic, or were being applied a way that was not in line with corporate ethics policies.
Ethical concerns over AI are not new
But changing regulations and privacy laws, a lack of definitions and standards, concerns over unintentional bias in training data, the ability for bias to creep into models over time, the lack of transparency in machine learning models, and the dearth of experience with use cases leads to numerous challenges. The preeminent ethical concerns used to be the loss of jobs due to AI, bias, or applications of facial recognition; but the ethical debate has become more complicated.
For instance, using location data that is linked to a specific group can be problematic. AI analysis of unstructured data from social media apps can lead to false or undesirable assumptions about individuals. And it’s possible that AI algorithms don’t take into account the cultural norms of a sub-segment of society (for example an algorithm that evaluates creditworthiness based on individual savings doesn’t apply to a society that values the distribution of extra wealth over accumulated savings)? The list goes on and on.
No easy answers on ethics
There are no easy answers, but initiatives to help support AI and Ethics are underway. AI platforms have started to offer ‘model cards’ to support greater transparency of machine learning models and their findings. The model cards explain how an algorithm works and identifies the degree to which various factors impact a model’s findings. Users can then tweak or remove inputs as needed. The platforms have also started offering monitoring capabilities that can flag algorithms that begin to drift and stop performing as expected. Data scientists can then intervene as necessary to mitigate the impact.
Additionally, organizations have started expanding the teams involved in project deployments to ensure a more multi-disciplinary perspective, including employees from finance, legal, human resources, and other departments beyond IT. The hope is that additional voices and expertise can identify potential ethical concerns early in the project development process, and guide course corrections as needed.
How well do you really know your competitors?
Access the most comprehensive Company Profiles on the market, powered by GlobalData. Save hours of research. Gain competitive edge.
Your download email will arrive shortly
Not ready to buy yet? Download a free sample
We are confident about the unique quality of our Company Profiles. However, we want you to make the most beneficial decision for your business, so we offer a free sample that you can download by submitting the below formBy GlobalData
Similarly, providers of AI platforms have started withholding certain features, such as facial analysis and recognition, or put in place restrictions to prevent its use in applications that they deem unethical. They have also crafted Ethical Use policies that customers must adhere to when using their platform (although the degree to which they can monitor and therefore enforce this is questionable). They have also established internal teams that review new AI-enabled capabilities and internal use of the technology. Salesforce is creating battle cards for its sales teams that help them explain to customers how to use data more ethically and consulting organizations and IT services providers offer guidance on ethical adoption of AI to their customers.
Regulatory agencies, such as the NIST, are also involved in the conversation. NIST is developing taxonomies, terminology and testbeds for measuring AI risks, and is developing benchmarks, and qualitative and quantitative metrics to evaluate AI technologies. Furthermore, AI is inextricably linked to data. Bad data yields bad AI results and biased data yields biased model results. It therefore stands to reason that conversations related to the more ethical application of AI will need to broaden to address ethical issues around data management – particularly data that can be linked to individuals.