by Jack Berkowitz

As organizations develop their own internal ethical practices and countries continue to develop legal requirements, we are at the beginning of determining standards for the ethical use of data and artificial intelligence (AI).

In the past 20 years, our ability to collect, store, and process data has dramatically increased. There are exciting new tools that can help us automate processes, learn things we couldn’t see before, recognize patterns, and predict what is likely to happen. Since our capacity to do new things has developed quickly, the focus in tech has been primarily on what we can do. Today, organizations are starting to ask what’s the right thing to do.

This is partly a global legal question as countries implement new requirements for the use and protection of data, especially information directly or indirectly connected to individuals. It’s also an ethical question as we address concerns about bias and discrimination, and explore concerns about privacy and a person’s rights to understand how data about them is being used.

What is AI and Data Ethics?

Ethical use of data and algorithms means working to do the right thing in the design, functionality, and use of data in Artificial Intelligence (AI).

It’s evaluating how data is used and what it’s used for, considering who does and should have access, and anticipating how data could be misused. It means thinking through what data should and should not be connected with other data and how to securely store, move, and use it. Ethical use considerations include privacy, bias, access, personally identifiable information, encryption, legal requirements and restrictions, and what might go wrong.

Data Ethics also means asking hard questions about the possible risks and consequences to people whom the data is about and the organizations that use that data. These considerations include how to be more transparent about what data organizations have and what they do with it. It also means being able to explain how the technology works, so people can make informed choices on how data about them is used and shared.

Why is Ethics Important in HR Technology?

Technology is evolving fast. We can create algorithms that connect and compare information, see patterns and correlations, and offer predictions. Tools based on data and AI are changing organizations, the way we work, and what we work on. But we also need to be careful about arriving at incorrect conclusions from data, amplifying bias, or relying on AI opinions or predictions without thoroughly understanding what they are based on.

We want to think through what data goes into workplace decisions, how AI and technology affect those decisions, and then come up with fair principles for how we use data and AI.

What Are Data Ethics Principles?

Ethics is about acknowledging competing interests and considering what is fair. Ethics asks questions like: What matters? What is required? What is just? What could possibly go wrong? Should we do this?

In trying to answer these questions, there are some common principles for using data and AI ethically.

  1. Transparency – This includes disclosing what data is being collected, being compliant with what can be collected and how it is allowed to be used, what decisions are made with the assistance of AI, and whether a user is dealing with bots or humans. It also means being able to explain how algorithms work and what their outputs are based on. That way, we can evaluate the information they give us against the problems we’re trying to solve. Transparency also includes how we let people know what data an organization has about them and how it is used. Sometimes, this includes giving people an opportunity to have the information corrected or deleted.
  2. Fairness – AI doesn’t just offer information. Sometimes it offers opinions. This means we have to think through how these tools and the information they give us are used. Since data comes from and concerns humans, it’s essential to look for biases in what data is collected, what rules are applied, and what questions are asked of the data. For example, if you want to increase diversity in hiring, you don’t want to only rely on tools that tell you who has been successful in your organization in the past. This information alone would likely give you more of the same rather than more diversity. While there is no way to completely eliminate bias in tools created by and about people, we need to understand how the tools are biased so we can reduce and manage the bias and correct for it in our decision making.
  3. Accuracy – The data used in AI should be up to date and accurate. And there needs to be ways to correct it. Data should also be handled, cleaned, sorted, connected, and shared with care to retain its accuracy. Sometimes taking data out of context can make it appear misleading or untrue. So accuracy depends partly on whether the data is true, and partly on whether it makes sense and is useful based on what we are trying to do or learn.
  4. Privacy – Some cultures believe that privacy is part of fundamental human rights and dignity. Canada’s data privacy laws vary by region. An increasing number of privacy laws around the globe recognize privacy rights in our names and likeness, financial and medical records, personal relationships, homes, and property. We are still working out how to balance privacy and the need to use so much personal data. Law makers have been more comfortable allowing broader uses of anonymized data than data where you know, or can easily discover, who it’s about. But as more data is collected and connected, questions arise about how to maintain that anonymity. Other privacy issues include the security of the information and what people should know about who has data about them and how it’s used.
  5. Accountability – This is not just compliance with global laws and regulations. Accountability is also about the accuracy and integrity of data sources, understanding and evaluating risks and potential consequences of developing and using data and AI, and implementing processes to make sure that new tools and technologies are created ethically.

As organizations develop their own internal ethical practices and countries continue to develop legal requirements, we are at the beginning of determining standards for the ethical use of data and AI.

ADP is already working on its AI and data ethics, through establishing an AI and Data Ethics Board and developing ethical principles that are customized to ADP’s data, products and services. Next in our series on AI and Ethics, we will be talking to each of ADP’s AI and Data Ethics Board members about ADP’s guiding ethical principles and how ADP applies those principles to its design, processes, and products.

Learn more

We invite you to explore this webpage to learn more about Privacy at ADP, including understanding the steps that we’ve taken to protect personal data globally. Within this webpage, you may find our Privacy Statements, an overview of our Binding Corporate Rules, and of our Privacy Program, as well as brochures and other privacy-related materials.

Check out this article: Algorithms and Ethics: What’s in your A.I.?

This article originally appeared on SPARK powered by ADP.

Previous articleTaking a Data-Based Approach to Diversity and Inclusion
Next articleHow Employers Can Show Their Character and Stand Out in Trying Times