BrahMos WORLD INDIA MADHYA PRADESH BHOPAL WTN SPECIAL Astrology GOSSIP CORNER SPORTS BUSINESS FUN FACTS ENTERTAINMENT LIFESTYLE TRAVEL ART & LITERATURE SCIENCE & TECHNOLOGY HEALTH EDUCATION DIASPORA OPINION & INTERVIEW RECIPES DRINKS FUNNY VIDEOS VIRAL ON WEB PICTURE STORIES
WTN HINDI ABOUT US PRIVACY POLICY SITEMAP CONTACT US
logo
Breaking News

IBM bats for regulating AI based on accountability,security

Wednesday - January 22, 2020 7:20 pm , Category : SCIENCE & TECHNOLOGY
San Francisco, Jan 22 (IANS) IT major IBM has released a regulatory framework for organisations involved in developing or using Artificial Intelligence based on accountability, transparency, fairness and security.
These IBM recommendations come as the new European Commission has indicated that it will legislate on AI within the first 100 days of 2020 and the White House has released new guidelines for regulation of AI.
The "Precision Regulation for Artificial Intelligence" released by The IBM Policy Lab builds upon IBM's calls for a "precision regulation" approach to facial recognition and illegal online content - laws tailored to hold companies more accountable, without becoming over-broad in a way that hinders innovation or the larger digital economy.
Specifically, IBM's new policy paper outlines five policy imperatives for companies, whether they are providers or owners of AI systems that can be reinforced by regulation.
To ensure compliance with these expectations, providers and owners should designate a person responsible for trustworthy AI, such as a lead AI ethics official.
All entities providing or owning an AI system should conduct an initial high-level assessment of the technology's potential for harm. And regulation should treat different use cases differently based on the possible inherent risk.
The best way to promote transparency is through disclosure, making the purpose of an AI system clear to consumers and businesses, according to the regulatory framework.
No one should be tricked into interacting with AI, it added.
Any AI system on the market that is making determinations or recommendations with potentially significant implications for individuals should be able to explain and contextualise how and why it arrived at a particular conclusion.
All organisations in the AI developmental lifecycle have some level of shared responsibility in ensuring the AI systems they design and deploy are fair and secure.
This requires testing for fairness, bias, robustness and security, and taking remedial actions as needed, both before sale or deployment and after it is operationalised.
This should be reinforced through "co-regulation", where companies implement testing and government conducts spot checks for compliance, IBM said.

--IANS gb/bg