Creating A Shared Code Of Ethics To Guide Ethical and Responsible Use of Data with Natalie Evans Harris
Natalie Michelle Evans Harris, a leader on ethical and responsible use of data, explains on the Women in Data Science (WiDS) Podcast how building trust through a shared vision and data “code of ethics” is essential to promote both innovation and privacy.
During her career at the National Security Agency, Capitol Hill and the White House, Natalie Evans Harris saw that while we collected troves of data, we didn’t have strong frameworks and governance in place to protect people in a data driven world. “Data has been used to intrude in our lives. Things are happening based upon data that nobody communicated to the public was actually happening,” she explained during a conversation with Stanford’s Margot Gerritsen, Stanford professor and host of the WiDS Podcast.
Data ethics and responsible use of data are essentially about building trust. There’s this gap in understanding what sharing data means. Two things have to happen if we’re going to build a relationship where people allow their data to be used by a company. Individuals have to trust that what the company is doing with that data is something they’re okay with. And the company has to be able to prove that they’re being responsible with the use of the data. A company could have the best products out there, but if people don’t trust you or understand what you’re doing with the data, then they’re not going to trust you to use the data. And then innovation stops.
She believes the biggest problem is we do not have a shared vision of what ethical practices mean. We don’t want to put broad impact laws in place to govern responsible use of data when we’re still trying to define that vision. To change business practices, we have to change company expectations so that they’re not only incentivized to be ethical and responsible in their business models, but they’re also penalized when they violate.
Harris has been advocating for a data science “code of ethics” to create a shared vision to guide our behaviors, and then start to develop best practices around. Companies are now taking this code of ethics and personalizing it to their businesses around principles like informed consent, transparency, fairness and diversity. Companies then publicize the practices that they’re putting in place to align with those principles. That’s how you start to create that shared vision.
She sees that there’s this transformation happening with the relationship between technology and people. For so long, technology has been this very passive thing in our lives, and now with AI and machine learning and all of these uses of data and technology, there’s this tension around what technology can do and what humans should do. Until people know and understand what is happening with their data, and until companies can thoughtfully express what they’re doing with the data in a very transparent fashion, we will continue to have this tension. She is hoping that this code of ethics can start to ease that tension.