{"id":20913,"date":"2023-08-22T06:28:27","date_gmt":"2023-08-22T06:28:27","guid":{"rendered":"https:\/\/web3unplugged.io\/blog\/?p=20913"},"modified":"2023-08-22T06:28:29","modified_gmt":"2023-08-22T06:28:29","slug":"white-house-science-adviser-of-indian-origin-calls-for-more-safeguards-against-artificial-intelligence-risks","status":"publish","type":"post","link":"https:\/\/web3unplugged.io\/blog\/white-house-science-adviser-of-indian-origin-calls-for-more-safeguards-against-artificial-intelligence-risks\/","title":{"rendered":"White House science adviser of Indian-origin calls for more safeguards against artificial intelligence risks"},"content":{"rendered":"\n<p>Arati Prabhakar is helping to guide the U.S. approach to safeguarding AI technology<\/p>\n\n\n\n<p>When President Joe Biden has questions about artificial intelligence, one expert he turns to is his science adviser Arati Prabhakar, director of the White House Office of Science and Technology Policy.<\/p>\n\n\n\n<p>Prabhakar is helping to guide the U.S. approach to safeguarding AI technology, relying in part on cooperation from big American tech firms like Amazon, Google, Microsoft and Meta.<\/p>\n\n\n\n<p>The India-born, Texas-raised engineer and applied physicist is coming at the problem from a career bridging work in government \u2014 including leading the Defense Department\u2019s advanced technology research arm \u2014 and the private sector as a former Silicon Valley executive and venture capitalist.<\/p>\n\n\n\n<p>She spoke with The Associated Press earlier this month ahead of a White House-organized test of AI systems at the DefCon hacker convention in Las Vegas. The interview has been edited for length and clarity.<\/p>\n\n\n\n<p>Q: Does the president come to ask you about AI?<br>A: I\u2019ve had the great privilege of talking with him several times about artificial intelligence. Those are great conversations because he\u2019s laser-focused on understanding what it is and how people are using it. Then immediately he just goes to the consequences and deep implications. Those have been some very good conversations. Very exploratory, but also very focused on action.<\/p>\n\n\n\n<p>Q: Senate Majority Leader Chuck Schumer (who\u2019s pushing for AI regulations) says making AI models explainable is a priority. How realistic is that?<br>A: It\u2019s a technical feature of these deep-learning, machine-learning systems, that they are opaque. They are black box in nature. But most of the risks we deal with as human beings come from things that are not explainable. As an example, I take a medicine every single day. While I can\u2019t actually predict exactly how it\u2019s going to interact with the cells in my body, we have found ways to make pharmaceuticals safe enough. Think about drugs before we had clinical trials. You could hawk some powder or syrup and it might make you better or it might kill you. But when we have clinical trials and a process in place, we started having the technical means to know enough to start harnessing the value of pharmaceuticals. This is the journey we have to be on now for artificial intelligence. We\u2019re not going to have perfect measures, but I think we can get to the point where we know enough about the safety and effectiveness of these systems to really use them and to get the value that they can offer.<\/p>\n\n\n\n<p>Q: What are some specific AI applications you&#8217;re concerned about?<br>A: Some of the things we see are big and obvious. If you break the guardrails of a chatbot, which people do routinely, and coax it to tell you how to build a weapon, well, clearly that\u2019s concerning. Some of the harms are much more subtle. When these systems are trained off human data, they sometimes distill the bias that\u2019s in that data. There\u2019s now a fairly substantial, distressing history of facial recognition systems being used inappropriately and leading to wrongful arrests of Black people. And then privacy concerns. All of our data that\u2019s out in the world, each individual piece may not reveal much about us, but when you put it all together it tells you rather a lot about each of us.<\/p>\n\n\n\n<p>Q: Seven companies (including Google, Microsoft and ChatGPT-maker OpenAI) agreed in July to meet voluntary AI safety standards set by the White House. Were any of those commitments harder to get? Where&#8217;s the friction?<br>A: I want to start by just acknowledging how fortunate we are that so many of the companies that are driving AI technology today are American companies. It reflects a long history of valuing innovation in this country. That\u2019s a tremendous advantage. We also just have to be very, very clear that with every good intention in the world, the realities of operating in the market are, by definition, a constraint on how far these individual companies can go. We hope many more will join them and voluntary commitments will grow. We just have to be clear that\u2019s only one part of it. That\u2019s companies stepping up to their responsibilities. But we in government need to step up to ours, both in the executive branch and for the legislative branch.<\/p>\n\n\n\n<p>Q: Do you have a timeline for future actions (such as a planned Biden executive order)? Will it include enforceable accountability measures for AI developers?<br>A: Many measures are under consideration. I don\u2019t have a timeline for you. I will just say fast. And that comes directly from the top. The president has been clear that this is an urgent issue.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Arati Prabhakar is helping to guide the U.S. approach to safeguarding AI technology When President Joe Biden has questions about artificial intelligence, one expert he turns to is his science adviser Arati Prabhakar, director of the White House Office of Science and Technology Policy. Prabhakar is helping to guide the U.S. approach to safeguarding AI [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":20915,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_seopress_robots_primary_cat":"none","_seopress_titles_title":"","_seopress_titles_desc":"","_seopress_robots_index":"","footnotes":""},"categories":[2],"tags":[],"class_list":["post-20913","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news"],"rttpg_featured_image_url":{"full":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",549,309,false],"landscape":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",549,309,false],"portraits":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",549,309,false],"thumbnail":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23-150x150.jpg",150,150,true],"medium":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23-300x169.jpg",300,169,true],"large":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",549,309,false],"1536x1536":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",549,309,false],"2048x2048":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",549,309,false],"post-thumbnail":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",549,309,false],"graptor-sq-xs":["https:\/\/web3unplugged.io\/blog\/wp-content\/uploads\/2023\/08\/Untitled-23.jpg",100,56,false]},"rttpg_author":{"display_name":"Admin CG","author_link":"https:\/\/web3unplugged.io\/blog\/author\/admin-cg\/"},"rttpg_comment":0,"rttpg_category":"<a href=\"https:\/\/web3unplugged.io\/blog\/category\/news\/\" rel=\"category tag\">news<\/a>","rttpg_excerpt":"Arati Prabhakar is helping to guide the U.S. approach to safeguarding AI technology When President Joe Biden has questions about artificial intelligence, one expert he turns to is his science adviser Arati Prabhakar, director of the White House Office of Science and Technology Policy. Prabhakar is helping to guide the U.S. approach to safeguarding AI&hellip;","_links":{"self":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts\/20913","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/comments?post=20913"}],"version-history":[{"count":1,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts\/20913\/revisions"}],"predecessor-version":[{"id":20916,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/posts\/20913\/revisions\/20916"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/media\/20915"}],"wp:attachment":[{"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/media?parent=20913"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/categories?post=20913"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/web3unplugged.io\/blog\/wp-json\/wp\/v2\/tags?post=20913"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}