The rising interest in Ethical AI has called organizations to be more responsible with their technology. On the latest episode of the Decisions Now podcast, we are joined by, the CEO and Founder of Emergent Line, Christopher Sanchez who shares all about the Global AI Bill of Rights, which he wrote.
In this episode, co-hosts Rigvinath Chevala, Evalueserve’s chief technology officer, and Erin Pearson, our vice president of marketing, and Sanchez discuss the bill, the need for responsible tech and its effects, privacy, and why explainability is important.
Global AI Bill of Rights
The bill was developed as an idea from conversations Sanchez had with his client and teams with the motive of being human-centric, ethical and designing products that empowered not only end users but even workers.
He founded dataoath.org and wrote The Absurd Dinner Bill, all of which led to conversations and buzz about the need for defined user rights.
“AI is inherently global. And then what would that algorithm look like that applies to those rights. And then what are the data practices? Because the data practices feed the algorithms that would ensure those rights,” Sanchez said. “So, if there’s one thing that I’ll touch on, I firmly believe everybody wants to do the right thing. Everybody inherently is a good person, but you also have to make it simple for them to do the right thing.”
The bill touches upon topics like facial recognition rights, biometrics, and other identifying data rights, ESG (Environmental Social Governance), emotional and mental state identification, data rights of users, system bias, and equality of outcomes.
Making Rights Accessible
Due to the abundance of information and different AI frameworks, businesses may find it challenging to apply ethics to their operations, making it essential to simplify them, Sanchez mentioned.
A key aspect in algorithmic due process is explainability.
When looking at human impact AI, that’s where you need explainability with what’s going on. Not always in the case of when my package will be delivered but more impactful situations that determine someone getting a job interview or their insurance premiums etc., he said.
“I believe as AI practitioners, we have a responsibility, not only to create amazing products that improve people’s lives, but to think about everybody who will be impacted, even if we never see them, even if they never use our products,” Sanchez mentioned.
When making explainability work, here are a few questions teams can ask themselves:
- How are you thinking when creating these products?
- How are you collecting peoples’ data?
- Who is the data representing?
- Do you have a diverse team?
- Am I testing the outputs and are people being treated correctly?
Make a list of people that exist in society and ask if the products will work for everyone.
“Remember the problem with AI is problems happen at scale, it doesn’t impact one person. It impacts tens of thousands of people at once and very fast,” he added.
Issues with unethical technology effects the everyday worker, what can people do to understand this and educate themselves? Pearson asked Sanchez.
“Because even if businesses make it available to them, it doesn’t necessarily mean that people are going to seek out that information. So, do you think that’s an element of it? There has to be a shift in culture for them to understand it better or understand how it’s actually impacting them,” Pearson asked.
Sanchez said this was a three-part issue, where companies must be clear about the AI their consumers are interacting with. And secondly, there needs to be some personal responsibility as well. Lastly, this education must be included in school curriculums.
In 2016, his team launched a site called Wandering Alpha, with a mission to educate folks on AI, NFTs, blockchain and more.
The whole point was to make it simple and concise for someone to understand, Sanchez said.
Often people still believe AI is limited to robots, education will help bridge the disconnect between reality and outdated views and help people understand the numerous AI interactions they have every day on their phones and computers.
“There is a bigger chance that you will be mistreated by poorly designed AI systems than you will be by super intelligence,” Sanchez said. “Where they start developing populations that understand this from a very young age, they’re more sophisticated about what they’re dealing with and the data that they’re leaving and also the technologies they have to master in order to thrive in the world that we’re all going to be going into.”