AUTHORITYID | CHAMBER | TYPE | COMMITTEENAME |
---|---|---|---|
sscm00 | S | S | Committee on Commerce, Science, and Transportation |
[Senate Hearing 115-649] [From the U.S. Government Publishing Office] S. Hrg. 115-649 DIGITAL DECISION-MAKING: THE BUILDING BLOCKS OF MACHINE LEARNING AND ARTIFICIAL INTELLIGENCE ======================================================================= HEARING BEFORE THE SUBCOMMITTEE ON COMMUNICATIONS, TECHNOLOGY, INNOVATION, AND THE INTERNET OF THE COMMITTEE ON COMMERCE, SCIENCE, AND TRANSPORTATION UNITED STATES SENATE ONE HUNDRED FIFTEENTH CONGRESS FIRST SESSION __________ DECEMBER 12, 2017 __________ Printed for the use of the Committee on Commerce, Science, and Transportation [GRAPHIC NOT AVAILABLE IN TIFF FORMAT] Available online: http://www.govinfo.gov ___________ U.S. GOVERNMENT PUBLISHING OFFICE 37-295 PDF WASHINGTON : 2019 SENATE COMMITTEE ON COMMERCE, SCIENCE, AND TRANSPORTATION ONE HUNDRED FIFTEENTH CONGRESS FIRST SESSION JOHN THUNE, South Dakota, Chairman ROGER F. WICKER, Mississippi BILL NELSON, Florida, Ranking ROY BLUNT, Missouri MARIA CANTWELL, Washington TED CRUZ, Texas AMY KLOBUCHAR, Minnesota DEB FISCHER, Nebraska RICHARD BLUMENTHAL, Connecticut JERRY MORAN, Kansas BRIAN SCHATZ, Hawaii DAN SULLIVAN, Alaska EDWARD MARKEY, Massachusetts DEAN HELLER, Nevada CORY BOOKER, New Jersey JAMES INHOFE, Oklahoma TOM UDALL, New Mexico MIKE LEE, Utah GARY PETERS, Michigan RON JOHNSON, Wisconsin TAMMY BALDWIN, Wisconsin SHELLEY MOORE CAPITO, West Virginia TAMMY DUCKWORTH, Illinois CORY GARDNER, Colorado MAGGIE HASSAN, New Hampshire TODD YOUNG, Indiana CATHERINE CORTEZ MASTO, Nevada Nick Rossi, Staff Director Adrian Arnakis, Deputy Staff Director Jason Van Beek, General Counsel Kim Lipsky, Democratic Staff Director Chris Day, Democratic Deputy Staff Director Renae Black, Senior Counsel ------ SUBCOMMITTEE ON COMMUNICATIONS, TECHNOLOGY, INNOVATION, AND THE INTERNET ROGER F. WICKER, Mississippi, BRIAN SCHATZ, Hawaii, Ranking Chairman MARIA CANTWELL, Washington ROY BLUNT, Missouri AMY KLOBUCHAR, Minnesota TED CRUZ, Texas RICHARD BLUMENTHAL, Connecticut DEB FISCHER, Nebraska EDWARD MARKEY, Massachusetts JERRY MORAN, Kansas CORY BOOKER, New Jersey DAN SULLIVAN, Alaska TOM UDALL, New Mexico DEAN HELLER, Nevada GARY PETERS, Michigan JAMES INHOFE, Oklahoma TAMMY BALDWIN, Wisconsin MIKE LEE, Utah TAMMY DUCKWORTH, Illinois RON JOHNSON, Wisconsin MAGGIE HASSAN, New Hampshire SHELLEY CAPITO, West Virginia CATHERINE CORTEZ MASTO, Nevada CORY GARDNER, Colorado TODD YOUNG, Indiana C O N T E N T S ---------- Page Hearing held on December 12, 2017................................ 1 Statement of Senator Wicker...................................... 1 Letter dated December 11, 2017 to Hon. Roger Wicker and Hon. Brian Schatz from Dean Garfield, President and CEO, Information Technology Industry Council (ITI).............. 83 Letter dated December 12, 2017 to Hon. John Thune and Hon. Bill Nelson from Marc Rotenberg, President, EPIC; Caitriona Fitzgerald, Policy Director, EPIC; and Christine Bannan, Policy Fellow, EPIC........................................ 84 Statement of Senator Schatz...................................... 2 Statement of Senator Moran....................................... 41 Statement of Senator Peters...................................... 44 Statement of Senator Udall....................................... 46 Statement of Senator Young....................................... 48 Statement of Senator Cantwell.................................... 50 Statement of Senator Markey...................................... 55 Statement of Senator Cruz........................................ 56 Statement of Senator Cortez Masto................................ 58 Statement of Senator Blumenthal.................................. 60 Witnesses Dr. Cindy L. Bethel, Associate Professor, Department of Computer Science and Engineering, Mississippi State University.......... 4 Prepared statement........................................... 5 Daniel Castro, Vice President, Information Technology and Innovation Foundation (ITIF)................................... 8 Prepared statement........................................... 10 Victoria Espinel, President and CEO, BSAThe Software Alliance....................................................... 17 Prepared statement........................................... 18 Report entitled ``The $1 Trillion Economic Impact of Software 63 Dr. Dario Gil, Ph.D., Vice President, AI and IBM Q............... 26 Prepared statement........................................... 27 Dr. Edward W. Felten, Ph.D., Robert E. Kahn Professor of Computer Science and Public Affairs, Princeton University............... 32 Prepared statement........................................... 34 Appendix Response to written questions submitted to Dr. Cindy L. Bethel by: 2Hon. Amy Klobuchar.......................................... 87 Hon. Tom Udall............................................... 87 Hon. Maggie Hassan........................................... 88 Response to written questions submitted to Daniel Castro by: Hon. Tom Udall............................................... 90 Hon. Gary Peters............................................. 90 Hon. Maggie Hassan........................................... 91 Response to written questions submitted to Victoria Espinel by: Hon. Gary Peters............................................. 92 Hon. Maggie Hassan........................................... 94 Response to written questions submitted to Dr. Dario Gil, Ph.D. by: Hon. Amy Klobuchar........................................... 96 Hon. Tom Udall............................................... 97 Hon. Gary Peters............................................. 98 Hon. Maggie Hassan........................................... 99 Response to written questions submitted to Dr. Edward W. Felten, Ph.D. by: Hon. Amy Klobuchar........................................... 102 Hon. Tom Udall............................................... 103 Hon. Gary Peters............................................. 103 Hon. Maggie Hassan........................................... 104 DIGITAL DECISION-MAKING: THE BUILDING BLOCKS OF MACHINE LEARNING AND ARTIFICIAL INTELLIGENCE ---------- TUESDAY, DECEMBER 12, 2017 U.S. Senate, Subcommittee on Communications, Technology, Innovation, and the Internet, Committee on Commerce, Science, and Transportation, Washington, DC. The Subcommittee met, pursuant to notice, at 10 a.m. in room SR-253, Russell Senate Office Building, Hon. Roger Wicker, Chairman of the Subcommittee, presiding. Present: Senators Wicker [presiding], Schatz, Blunt, Cruz, Fischer, Moran, Sullivan, Heller, Inhofe, Capito, Young, Cantwell, Klobuchar, Blumenthal, Markey, Booker, Udall, Peters, Hassan, and Cortez Masto. OPENING STATEMENT OF HON. ROGER F. WICKER, U.S. SENATOR FROM MISSISSIPPI Senator Wicker. This hearing will come to order. Senator Schatz will be here in a few moments and has sent word that we should go ahead and proceed. Today's Subcommittee meets to examine the commercial applications of artificial intelligence and machine learning for the U.S. economy. We are also gathered to discuss how the responsible design and deployment of intelligent systems can foster innovation and investment, propelling the United States as a leader in artificial intelligence. I'm glad to convene this hearing, and as I mentioned, my colleague and friend Senator Schatz will be here in a moment. Artificial intelligence refers to technology that is capable of taking on human-like intelligence. Through data inputs and algorithms, AI systems have the potential to learn, to reason, to plan, perceive, process, make decisions, and even act for themselves. Although AI applications have been around for decades, recent advancements, particularly in machine learning, have accelerated in their capabilities because of the massive growth in data gathered from billions of connected devices and the digitization of everything. Developments in computer processing technologies and better algorithms are also enabling AI systems to become smarter and perform more unique tasks. Every day consumers use technology that employ some degree of AI, smartphone mapping apps that suggest faster driving routes, for example. Online search platforms are learning from past queries to generate increasingly customized results for users. We all know that when we click on a site on the Internet, news suggestions and advertisements on social media, and semi-autonomous vehicles are just a few examples of how machines and computer programs are taking on increasingly cognitive tasks. The excitement surrounding this technology is deserved. AI has the potential to transform our economy, and so let's talk about that today. AI's ability to process and sort through troves of data can greatly inform human decisionmaking and processes across industries, including agriculture, health care, and transportation. In turn, businesses can be more productive, profitable, and efficient in their operations. As AI systems mature and become more accurate in their descriptive, predictive, and prescriptive capabilities, there are issues that should be addressed to ensure the responsible development and use of this technology. Some of these issues include: understanding how data is gathered, what data is provided for an intelligent machine to analyze, and how algorithms are programmed by humans to make certain predictions. Moreover, understanding how the human end user interacts with or responds to the digital decision, and how humans interpret or explain decisions of the AI system over time will also need to be addressed. These are important considerations to ensure that the decisions made by AI systems are based on representative data that does not unintentionally harm vulnerable populations or act in an unsafe, anticompetitive, or biased way. So there's a lot to think about. In addition to these issues, other considerations, such as data privacy and cybersecurity, AI's impact on the workforce, and human control and oversight over intelligence systems should also be addressed as the technology develops. Fundamental to the success of machine learning and AI in enhancing U.S. productivity and empowering human decisionmaking is consumer confidence and trust in these systems. To build consumer confidence and trust, it is critical that the integration of AI into our commercial and government processes be done responsibly. To that end, I look forward to learning from today's witnesses about how AI is advancing in our economy and what best practices our industry and AI researchers are considering to achieve all the promised economic and societal benefits of this technology. Senator Schatz, do you have anything to add to my comprehensive opening statement? STATEMENT OF HON. BRIAN SCHATZ, U.S. SENATOR FROM HAWAII Senator Schatz. I think everything has been said, but not everybody has said it. So good morning. Thank you very much, Mr. Chairman. AI is advancing fast. Each year, processing power gets better, hardware gets cheaper, and algorithms are easier to train thanks to bigger and better datasets. These advances are, for companies and economies, great opportunities. Technologists, historians, and economists say that we're at the cusp of the next industrial revolution, but there are concerns. We've seen that AI can be a black box. It can make decisions and come to conclusions without showing its reasoning. There are also known cases of algorithms that discriminate against minority groups, and when you start to apply these systems to criminal justice, health care, or defense, the lack of transparency and accountability is worrisome. Given the many concerns in a field that's advancing so quickly and is so revolutionary, it's hard to believe that there is no AI policy at the Federal level, and that needs to change. To start, the government should not purchase or use AI systems if we can't explain what it does, especially if these systems are making decisions about our citizens' lives. There also needs to be more transparency and consumer control on data collection. Too many consumers still do not know what data is being collected, how it's being collected, or who owns it. Some of our current laws and regulations work, but some of them are too old and outdated to be used as a strong foundation for AI. For example, companies often use data scraping to build their AI models. This falls under the Computer Fraud and Abuse Act, a 1986 law that was written before the Web was really in operation. AI is now used to write news articles, edit photographs, artificially reconstruct movies, all actions that fall under the Digital Millennium Copyright Act, which was passed in 1998, 10 years before the iPhone. Our laws used to apply to actions in the physical world, but now they apply to software systems that ultimately do the same thing. I'm glad to see that industry and academia are being proactive by coming up with policy, principles, and professional ethics codes, but with this kind of patchwork, the system is only as strong as its weakest link. From the private sector to academia to government, everyone has to wrestle with the ethical and policy questions that AI raises. And that's why I intend to introduce a bill creating an independent Federal commission to ensure that AI is adopted in the best interest of the public. If created, the commission would serve as a resource and coordinating body to the 16-plus Federal agencies that govern the use of AI. Otherwise, we risk bureaucratic turf wars and outdated inconsistent rules. The commission would also be tasked with asking the tough ethical questions around AI. For instance, employers may not legally ask interviewees about their religion, marital status, or race, but they use software that mines social media data that may make the same inferences. Judges cannot base their sentencing decisions on whether the defendant has family members in jail, yet these facts contribute to a risk score calculated by machine learning algorithms. In these cases, it's not clear yet where we draw the line on what is legal and what is not. In some instances, existing statutes will suffice. A lot of our laws actually work just fine when it comes to AI, and great laws can survive the test of time. But there are a few things that need to be wrestled with today in Congress and with our agencies, and that's why this hearing is so important. Thank you very much. I look forward to hearing the testimony. Senator Wicker. Thank you, Senator Schatz. Our witnesses today are Dr. Cindy Bethel, Associate Professor, Department of Computer Science and Engineering at Mississippi State University; Mr. Daniel Castro, Vice President, Information Technology and Innovation Foundation; Ms. Victoria Espinel, Chief Executive Officer, BSA-The Software Alliance; Dr. Dario Gil, Vice President, IBM Research, AI, and IBM Q; and Dr. Edward Felten, Robert E. Kahn Professor of Computer Science and Public Affairs, Princeton University. Friends, we will begin on my left with Dr. Bethel and proceed with 5-minute opening statements down the table. Thank you very much. Dr. Bethel. STATEMENT OF DR. CINDY L. BETHEL, ASSOCIATE PROFESSOR, DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING, MISSISSIPPI STATE UNIVERSITY Dr. Bethel. Good morning, Chairman Wicker, Ranking Member Schatz, and members of the Committee. Thank you for the opportunity to appear before you today. I'm Dr. Cindy Bethel. I'm Associate Professor of Computer Science and Engineering at Mississippi State University, or MSU. It is an honor to speak with the Committee today about digital decisionmaking associated with artificial intelligence, known as AI, from the academic perspective, and about applications of AI being developed at MSU. Today I will address three primary areas: first, a brief introduction to AI; next, three AI projects being developed at MSU; and last I will discuss some key points associated with AI. A critical aspect associated with the advancement of science and technology is the development of algorithms with AI including machine learning for digital decisionmaking. In order for a system to make a decision, it must first acquire information, process and learn from that information, and then use that information to make decisions. The gold standard would be the ability for a system to make a decision in a manner similar to a human expert in that area. This is a relatively new field of science that will be--provide ongoing research opportunities, including efforts to enhance existing algorithms and develop new and more efficient and effective algorithms. This is critical to the overall advancement of science in various disciplines, such as robotics, medicine, economics, and others. There are countless application areas for which AI is beneficial and may make a significant impact on society. At MSU, we are on the forefront of AI developments with our research efforts, and today I will focus on three of those-- excuse me--projects. First is the integration of robots into high-risk law enforcement operations, such as SWAT teams, who I've trained with monthly for the last 6 years. This is an application for highly dynamic situations that require officers to make life- critical decisions. The more information they have prior to entering these dangerous and unknown environments, the better decisions the officers can make. We are developing algorithms using robots into an environment prior to entry to provide critical audio and video intelligence to allow officers to make more informed decisions. This information continues the dynamics of how they make entry or process a scene. Our research through intelligent interfaces will help inform how this information is delivered to the officers and maximize safety and performance. Second, Mississippi State University is researching the development of autonomous cargo transport systems used in the fast-paced dynamic environment of a top 100 logistics company. This involves choosing the proper sensors to ensure that data is able to make--help make informed decisions on the vehicle's path. Further, we are researching a variety of human factors because controlled vehicles will transfer from autonomous to human driver control and back again. Our work shows that if a human is not actively engaged in driving, they may--there may be insufficient situational awareness to control--take control on short notice. So how the human driver will be notified of the transfer of control is another critical aspect of our work. To ensure maximum safety, researchers at MSU are exploring situations in which humans are operating in close proximity to the autonomous vehicles, such as a vehicle that's docking to deliver cargo, and a worker is unloading and loading this cargo. Finally, another one of MSU's AI research projects involves a robotic therapy support system, known as Therabot that is in the form of a stuffed robotic dog. Therabot is an alternative to animal-assisted therapy for people who may be allergic to animals or may not be able to care for a live animal. Therabot will be used to provide support during clinical therapy sessions and home therapy practice with children or adults who are dealing with post-traumatic stress disorders or other mental health concerns. The algorithms being developed modify the autonomous behaviors of the robot based on the interactions with the human to accommodate his or her preferences and in response to different levels of stress detected by the robot, and to provide improved comfort and support. AI will only be as good as the data the system receives and its ability to process that information to make a decision. The better the quality and quantity of information available to the system, the better the results will be from the machine learning process, which results in better final decisions from the system. Otherwise, decisionmaking capabilities can be limited or inaccurate. The potential applications of AI are almost limitless. The United States can and should remain at the forefront of AI research, application development, and innovation provided that government proceeds with a light regulatory touch that doesn't stifle this potential. Thank you so much for the opportunity to testify today on these important topics. I appreciate your time and attention to the advancement and impact of digital decisionmaking and AI. [The prepared statement of Dr. Bethel follows:] Prepared Statement of Dr. Cindy L. Bethel, Associate Professor of Computer Science and Engineering; Director of the Social, Therapeutic, and Robotic Systems (STaRS) Lab; Billie J. Ball Endowed Professor in Engineering; Mississippi State University Chairman Wicker, Ranking Member Schatz, and Members of the Committee, thank you for the opportunity to appear before you today. I am Dr. Cindy Bethel, an Associate Professor of Computer Science and Engineering, the Billie J. Ball Endowed Professor in Engineering, and the Director of the Social, Therapeutic, and Robotic Systems (STaRS) Lab at Mississippi State University. It is an honor to speak with the committee today about digital decision-making associated with artificial intelligence (AI) from the academic perspective and about applications of AI being developed at Mississippi State University. A critical aspect associated with the advancement of science and technology is the development of algorithms associated with AI including machine learning for digital decision-making. In order for a system to make a decision it must acquire information, have a means of processing and learning from that information, and have the ability to use that information to make an informed decision. The gold standard would be the ability for a machine or system to make a decision in a manner similar to that of a human who is an expert in that area. We have made considerable progress toward this goal since the inception of what is considered artificial intelligence, which began in 1943 with the research performed by McCulloch and Pitts. Today many machines and systems rely upon the use of artificial intelligence for digital decision-making. This is a relatively new field of science that will provide many lifetimes of research including efforts to enhance existing algorithms and develop new, more efficient, and effective algorithms. This is critical to the overall advancements of science in many disciplines, such as robotics, medicine, economics, and many others. There are often disagreements in the field as to what is considered AI and what algorithms and techniques used for learning are considered the best. There are many application areas for which artificial intelligence is beneficial and may make a significant impact on society. At Mississippi State University, we are actively conducting AI research with several research projects that use AI and machine learning techniques, but I will focus on three primary projects. The first project is the integration of robots into law enforcement operations, especially high risk, life critical incident responses such as those used with special weapons and tactics (SWAT) teams. I have been training monthly with rural regional SWAT team members since 2011. This is an example of an application in which high risk, dynamic situations are encountered, that require often officers to make life- critical decisions. The more information or intelligence they have prior to entering these dangerous and unknown environments, the better decisions the officers can make. We are investigating and developing algorithms related to computer vision, sensor fusion, and scene understanding to send a robot in prior to entry to provide audio and video feedback to officers during a response, highlighting what is critical information for them to attend to so that they are not overwhelmed with information when under high stress. The algorithms identify what is important to the officers in the environment, such as children, weapons, and other possible threats. This information can change the dynamics of how they make entry or process the scene. We are also researching in what ways this information needs to be provided to the officers, so that they can use it to their advantage to keep them and others safer in the performance of their duties. For example if officers are conducting a slow and methodical search of a building for a suspect, and the environment is quiet, dark and threat is nearby, they would not want to receive the information in a openly visual manner such as a video stream on a mobile phone that would highlight them in the environment and make them more likely to be the target of harm. In this case, they may want a verbal description of the scene that comes across on their radio earpiece. If they are in a gunfight or if there is an alarm that is sounding in the environment, but they are in a relatively ``safe'' location, then they may want to receive this information in a visual form, because audio transmission would be difficult to hear. We are researching the development of intelligent interface switching in which the manner that information is delivered to the officers may change depending on what is happening in the environment they are operating in. The officers are excited and ready to start deploying some of these artificial intelligence and machine learning applications in real-world responses. A second project that we are working on at MSU is the development of autonomous cargo transport systems to be used in the fast-paced, dynamic environment of a top 100 logistics company. A primary factor that needs to be considered is what sensors need to be used to make informed decisions on the path the vehicles must travel. We also need to consider humans, who are sitting in the driver's seat of these vehicles, because control of the vehicles will change between fully autonomous to human driver operated. Research has shown that if the human is not actively involved in the activity of driving, that there may not be adequate situation awareness to be able to take back control of the vehicle with relatively short notice if needed. This has potential for life critical decision-making. We are investigating how the system needs to be able to alert the driver that control is being transferred, either to the vehicle or back to the human driver. It is also important to consider the types of notifications that need to occur to ensure safety and situational awareness of what is happening around the vehicle. Also there are situations in which humans are operating in close proximity to the autonomous vehicle. As human drivers, we observe consciously or unconsciously behaviors of other drivers to infer what nearby vehicles will do next, but if there are not those cues, then how do humans in the environment understand what the vehicle or system will do next? This is a major issue of concern. This occurs also when the vehicle is docking to deliver cargo and a human is involved in unloading and loading this cargo. We are exploring methods of notification to the person about what the vehicle will do next. This is of significant concern, and if the incorrect decision is made, the vehicle could cause harm to the human. The third project involves a robotic therapy support system, known as TherabotTM that is in the form of a stuffed robotic dog. TherabotTM is an alternative to animal-assisted therapy, for people who may be allergic to animals or may not be able to care for a live animal. TherabotTM will be used to provide support during clinical therapy sessions and for home therapy practice with children or adults who are dealing with post-traumatic stress disorders and other mental health concerns. The algorithms being developed for this project modify the autonomous behaviors and responses of the robot based on the interactions with the human to accommodate his or her preferences and in response to different levels of stress detected. Machine learning is being used to understand what behaviors the user prefers and to provide better support. It allows TherabotTM to be customizable to each individual user. It will learn and respond to each user as a dog would each person it encounters. Currently, TherabotTM can detect different types of touch, such as petting, patting, and hugging and will respond in different ways. If a person is under stress during the interaction he or she may squeeze the robot and the robot will adapt its behaviors to provide more comfort and support. Algorithms used in the research and development of systems that are capable of digital decision-making are being developed and enhanced by researchers all across the world. Mississippi State University is at the forefront of these research developments and is continually contributing through publications and sharing of knowledge, algorithms, and software developments. Research as a whole involves exploring what others have performed and then determining if there are modifications that can be made to improve upon those algorithms or the development of new algorithms to meet the needs of the application of use. For example, algorithms are being developed to learn a user's preference for how close a robot stands to them and still feel comfortable, or which friend they like most on a social media account. There are many methods for solving a problem. I typically tell my students who are interested in pursuing research that ``you need to be first or best at something and it is always best to be first.'' Artificial intelligence will only be as good as the data the system receives and its ability to process that information to make a decision. This is a critical aspect to the advancement of this field that can impact almost any other discipline. The system or machine must have the ability to perceive information and that typically comes from sensors and other forms of data. There are many types of sensing systems such as a camera, streaming video, thermal images based on the heat signature of items in the environment, radar, and many others. There are also methods of gathering information from sources such as social media, mobile device location history, purchase history, product preferences, websites visited, etc., that can assist in the decision- making process. The better the quality and quantity of information available to the system, the better the results will be from the machine learning process, which results in a better final decision from the system. Algorithms are programmed to receive data as an input, process that data, learn from large amounts of data, and then use that information to make a digital decision. If there is not sufficient amounts of data available to train the machine learning algorithms or enough diversity in the data to allow the learning algorithms to adapt to different aspects, then the decision-making capabilities can be limited or inaccurate. Another major issue of concern is the processing power necessary to handle large amounts of information received and come to a decision. This is especially a concern on smaller systems such as robots that have limited onboard processing capabilities. Many of the AI problems that are being addressed in the research community are performed on high powered computing resources and simulations are performed to validate the results. This is fine for many scientific applications, but in order for AI and machine learning to be beneficial in real-world applications, it will be necessary to perform the decision-making processes in real-time. The results need to be made available in an instant and not have to wait for processing time to provide a result. This is improving, but sensing and processing are currently significant limitations to the application and use of AI in digital decision- making. The level of human engagement necessary for digital decision-making depends on the state of the AI system. There are different levels of autonomous decision making. There is full autonomy, where the system receives the input and then processes the information from that data and makes a decision with no input from a human. There is supervised autonomy, in which the system receives information, processes the data, and comes up with possible results, and the human may have the ability to override or make the final decision. The more common level of autonomy is supervised autonomy. The level of human engagement also needs to consider the ramifications of the decision-making process. If it is a life-critical decision, then most people are more comfortable with a human remaining involved in the process to ensure an ethical and ``good'' decision is the final result. There are many ethical hurdles that will need to be decided at some point as to who is responsible if an AI system makes an incorrect decision, especially if the decision could result in harm to humans. There have been discussions in the field among researchers regarding who is responsible for this decision, such as the programmer, the company that made the system, or others. The current state often requires a human to be involved at some level of the final decision-making process unless it is low risk or well validated that the system will always make a ``right'' decision. The fields of artificial intelligence and machine learning are at such an early stage of scientific development that standards and best practices are discussed among researchers; however, there is not a single standard or set of best practices that I am aware of that all researchers and scientists follow at this point. The biggest concern is providing the best possible answers that will not result in harm and provides benefits to the users. The algorithms developed for machine learning and artificial intelligence can be used in almost any area of research, development, and discipline. It can be used to improve the decision-making process of humans. The processing of some information by a computer can be faster than what can be achieved by a human brain. Almost any aspect of society can benefit from the use of high quality artificial intelligence capabilities. A critical aspect in the development of artificial intelligence using machine learning and other techniques is the impact on the humans that are involved. A current limitation to the advancement of artificial intelligence is the quality and cost effectiveness of sensing capabilities to provide high quality information or data to the system to make those digital decisions. Another critical limitation to current artificial intelligence capabilities is onboard processing capabilities and the cost effectiveness of those systems. We have come a long way in the advancement of artificial intelligence; however we still have a long way to go! The potential applications of AI are almost limitless. The United States can and should remain at the forefront of AI research, application development, and innovation provided that the government proceeds with a light regulatory touch that doesn't stifle this potential. Thank you so much for the opportunity to testify today on these important topics. I appreciate your time and attention to the advancement and impacts of digital decision-making and artificial intelligence. Senator Wicker. Thank you, Dr. Bethel. Precisely 5 minutes. Dr. Bethel. Thank you. [Laughter.] Senator Wicker. Mr. Castro, we're delighted to have you. STATEMENT OF DANIEL CASTRO, VICE PRESIDENT, INFORMATION TECHNOLOGY AND INNOVATION FOUNDATION (ITIF) Mr. Castro. Thank you. Chairman Wicker, Ranking Member Schatz, and members of the Committee, I appreciate the invitation to be here today. AI has the potential to create a substantial and lasting impact on the economy by increasing the level of automation in virtually every sector, leading to more efficient processes and higher quality outputs, and boosting productivity in per capita incomes. In the coming years, AI is expected to generate trillions of dollars of economic value and help businesses make smarter decisions, develop innovative products and services, and boost productivity. For example, manufacturers are using AI to invent new metal alloys for 3D printing, pharmaceutical companies are using AI to discover new lifesaving drugs, and agricultural businesses are using AI to increase automation on farms. Companies that use AI will have an enormous advantage compared to their peers that do not; therefore, the United States should prioritize policy initiatives that promote AI adoption in its traded sectors where U.S. firms will face international competition. Many other countries already see the strategic importance of becoming lead adopters of AI, and they have begun implementing policies to pursue this goal. For example, this past March, Canada launched the Pan-Canadian AI Strategy, which is intended to help Canada become an international leader in AI research. The U.K.'s new budget, which was published last month, includes several provisions that have the goal of making the U.K. a world leader in AI, including establishing a new research center and funding about 500 Ph.D.'s. Japan has created an AI technology strategy designed to develop and commercialize AI in a number of fields, including transportation and health care. And China has declared its intent to be the world's premier AI innovation center by 2030. However, to date, the U.S. Government has not declared its intent to be a global leader in this field, nor has it begun the even harder task of developing a strategy to achieve that vision. Moreover, China, which has launched this ambitious program to dominate the field, has already surpassed the United States in terms of the total number of papers published and cited in some AI disciplines, such as deep learning. The U.S. should not cede its existing advantages in AI. Instead, it should pursue a multipronged national strategy to remain competitive in this field. First, the Federal Government should continue to expand its funding to support strategic areas of AI, especially in areas industry is unlikely to invest in, as well as better plan and coordinate Federal funding for AI R&D across different agencies. Second, the Federal Government should support educational efforts to ensure a strong pipeline of talent to create the next generation of AI researchers and developers, including the retraining and diversity programs as well as pursue integration policies that allow U.S. businesses to recruit and retain highly skilled computer scientists. Third, Federal and state regulators should conduct regulatory reviews to identify regulatory barriers to commercial use of AI in various industries, such as transportation, health care, education, and finance. Fourth, the Federal Government should continue to supply high-value datasets that enable advances in AI, such as providing open access to standardized reference datasets for text analysis and facial recognition. Federal agencies should also facilitate data-sharing between industry stakeholders just as the Department of Transportation has done on safety for autonomous vehicles. And, fifth, the Federal Government should assess what types of economic data it needs to gather from businesses to monitor and evaluate AI adoption, much like it tracked rural electrification or broadband connectivity as key economic indicators. Now, as with any technology, there will be some risks and challenges associated with AI that require government oversight, but the U.S. should not replicate the European approach to AI, where rules creating a right to explanation and a right to human review for automated decisions risk severely curtailing the uses of AI. Instead, the U.S. should create its own innovation-friendly approach to providing oversight of the emerging algorithmic economy just as it has for the Internet economy. Such an approach should prioritize sector-specific policies over comprehensive regulation, outcomes over transparency, and enforcement actions against firms that cause tangible harm over those that merely make missteps without injury. In many cases, regulators will not need to intervene because the private sector will address problems about AI, such as bias or discrimination, on its own. Moreover, given that U.S. companies are at the forefront of efforts to build AI that is safe and ethical, maintaining U.S. leadership in this field will be important to ensure these values remain embedded in this technology. AI is a transformational technology that has the potential to significantly increase efficiency and innovation across the U.S. economy, creating higher living standards and improve quality of life. But while the United States has an early advantage in AI, many other countries are trying to be number one--they're trying to be number one. We need more leadership on this issue. And I look forward to working with any member of the Committee on their proposed legislation and new ideas in this space. And I commend you all for holding this hearing. Thank you for the opportunity to be here today. And I look forward to the questions. [The prepared statement of Mr. Castro follows:] Prepared Statement of Daniel Castro, Vice President, Information Technology and Innovation Foundation (ITIF) Introduction Chairman Wicker, Ranking Member Schatz and members of the subcommittee, I appreciate the opportunity to appear before you to discuss the importance of artificial intelligence (AI) to the U.S. economy and how best to govern this important technology. My name is Daniel Castro, and I am vice president of the Information Technology and Innovation Foundation (ITIF), a non-profit, nonpartisan think tank whose mission is to formulate and promote public policies to advance technological innovation and productivity, and director of ITIF's Center for Data Innovation. What is Artificial Intelligence? AI is a field of computer science devoted to creating computer systems that perform tasks much like a human would, particularly tasks involving learning and decision-making.\1\ AI has many functions, including, but not limited to: --------------------------------------------------------------------------- \1\ Daniel Castro and Joshua New, ``The Promise of Artificial Intelligence,'' Center for Data Innovation, October 2016, http:// www2.datainnovation.org/2016-promise-of-ai.pdf. Learning, which includes several approaches such as deep learning (for perceptual tasks), transfer learning, --------------------------------------------------------------------------- reinforcement learning, and combinations thereof; Understanding, or deep knowledge representation required for domain-specific tasks, such as medicine, accounting, and law; Reasoning, which comes in several varieties, such as deductive, inductive, temporal, probabilistic, and quantitative; and Interacting, with people or other machines to collaboratively perform tasks, and for interacting with the environment. The cause of many misconceptions about AI, particularly its potential harms, is that some people conflate two very distinct types of AI: narrow AI and strong AI. Narrow AI describes computer systems adept at performing specific tasks, but only those specific types of tasks--somewhat like a technological savant.\2\ For example, Apple's Siri virtual assistant is capable of interpreting voice commands, but the algorithms that power Siri cannot drive a car, predict weather patterns, or analyze medical records. While other algorithms exist that can accomplish those tasks, they too are narrowly constrained--the AI used for an autonomous vehicle will not be able predict a hurricane's trajectory or help doctors diagnose a patient with cancer. --------------------------------------------------------------------------- \2\ Irving Wladawksy-Berger, `` `Soft' Artificial Intelligence Is Suddenly Everywhere,'' The Wall Street Journal, January 16, 2016, http://blogs.wsj.com/cio/2015/01/16/soft-artificial-intelligence-is- suddenly-everywhere/. --------------------------------------------------------------------------- In contrast, strong AI, also referred to as artificial general intelligence (AGI), is a hypothetical type of AI that can meet or exceed human-level intelligence and apply this problem-solving ability to any type of problem, just as the human brain can easily learn how to drive a car, cook food, and write code.\3\ Many of the dystopian fears about AI--that it will eliminate most jobs or go out of control and wipe out humanity, for example--stem from the notion that AGI is feasible, imminent, and uncontrollable.\4\ However, at least for the foreseeable future, computer systems that can fully mimic the human brain are only going to be found in scripts in Hollywood, and not labs in Silicon Valley. --------------------------------------------------------------------------- \3\ Ibid. \4\ Robert D. Atkinson, ``'It's Going to Kill Us!' and Other Myths About the Future of Artificial Intelligence,'' (Information Technology and Innovation Foundation, June 2016), http://www2.itif.org/2016-myths- machine-learning.pdf?_ga=1.201838291.334601971.1460947053. --------------------------------------------------------------------------- The application of AI has seen a surge in recent years because of the development of machine learning--a branch of AI that focuses on designing algorithms that can automatically and iteratively build analytical models from data without needing a human to explicitly program the solution. Before machine learning, computer scientists had to manually code a wide array of functions into a system for it to mimic intelligence. But now developers can achieve the same, or better, results more quickly and at a lower cost using machine learning techniques. For example, Google uses machine learning to automatically translate content into different languages based on translated documents found online, a technique that has proven to be much more effective than prior attempts at language translation.\5\ --------------------------------------------------------------------------- \5\ Pedro Domingos, The Master Algorithm: How the Quest for the Ultimate Learning Machine Will Remake Our World (New York: Basic Books, 2015). --------------------------------------------------------------------------- What Are the Potential Benefits of AI? AI will have a substantial and lasting impact on the economy by increasing the level of automation in virtually every sector, leading to more efficient processes and higher-quality outputs, and boosting productivity and per-capita incomes. For example, the McKinsey Global Institute estimates that by 2025 automating knowledge work with AI will generate between $5.2 trillion and $6.7 trillion of global economic value, advanced robotics relying on AI will generate between $1.7 trillion and $4.5 trillion, and autonomous and semi-autonomous vehicles will generate between $0.2 trillion and $1.9 trillion.\6\ Deloitte estimates that the Federal Government could save as much as $41.1 billion annually by using AI to automate tasks.\7\ And Accenture predicts that by 2035, AI could increase the annual growth rate of the U.S. economy by 2 percentage points, the Japanese economy by 1.9, and the German economy by 1.6.\8\ The report also found that, for the 12 countries surveyed, AI would boost labor productivity rates by 11 to 37 percent.\9\ --------------------------------------------------------------------------- \6\ James Manyika et al., Disruptive Technologies: Advances That Will Transform Life, Business, and the Global Economy,'' (McKinsey Global Institute, May 2013), http://www.mckin sey.com/business-functions/business-technology/our-insights/disruptive- technologies. \7\ Peter Viechnicki and William D. Eggers, ``How much time and money can AI save government?'' (Deloitte, April 26, 2017), https:// dupress.deloitte.com/dup-us-en/focus/cognitive-technologies/artificial- intelligence-government-analysis.html. \8\ Mark Purdy and Paul Daugherty, ``Why Artificial Intelligence Is the Future of Growth,'' (Accenture, September 28, 2016), https:// www.accenture.com/us-en/_acnmedia/PDF-33/Accen ture-Why-AI-is-the-Future-of-Growth.pdf. \9\ Ibid. --------------------------------------------------------------------------- There are a vast and diverse array of uses for AI, and many U.S. businesses are already using the technology today. Manufacturers are using AI to invent new metal alloys for 3D printing; pharmaceutical companies are using AI to discover new lifesaving drugs; mining companies are using AI to predict the location of mineral deposits; and agricultural businesses are using AI to increase automation on farms. The International Data Corporation (IDC) estimates that the market for AI technologies that analyze unstructured data will reach $40 billion by 2020.\10\ And AI startups have attracted significant investment, with U.S. investors putting $757 million in venture capital in AI start-ups in 2013, $2.18 billion in 2014, and $2.39 billion in 2015.\11\ --------------------------------------------------------------------------- \10\ ``Cognitive Systems Accelerate Competitive Advantage,'' IDC, accessed September 29, 2016, http://www.idc.com/promo/thirdplatform/ innovationaccelerators/cognitive. \11\ ``Artificial Intelligence Explodes: New Deal Activity Record for AI Startups,'' CB Insights, June 20, 2016, https:// www.cbinsights.com/blog/artificial-intelligence-funding-trends/. --------------------------------------------------------------------------- In some cases, the principle benefit of AI is that it automates work that would otherwise need to be performed by a human, thereby boosting efficiency. Sometimes AI can complete tasks that it is not always worth paying a human to do but still creates value, such as writing newspaper articles to summarize Little League games.\12\ In other cases, AI adds a layer of analytics that uncovers insights human workers would be incapable of providing on their own, thereby boosting quality. In some cases, it does both. For example, researchers at Stanford have used machine learning techniques to develop software that can analyze lung tissue biopsies with significantly more accuracy than a top human pathologist and at a much faster rate.\13\ By analyzing large volumes of data, researchers can train their computer models to reliably recognize known indicators of specific cancer types as well as discover new predictors. --------------------------------------------------------------------------- \12\ Steven Levy, ``Can an Algorithm Write a Better News Story Than a Human Reporter?'' Wired, April 24, 2012, https://www.wired.com/2012/ 04/can-an-algorithm-write-a-better-news-story-than-a-human-reporter/. \13\ Kun-Hsing Yu et al., ``Predicting non-small cell lung cancer prognosis by fully automated microscopic pathology image features,'' Nature, August 16, 2017, https://www.nature.com/articles/ncomms12474. --------------------------------------------------------------------------- AI is also delivering valuable social benefits, such as by helping authorities rapidly analyze the deep web to crack down on human trafficking, fighting bullying and harassment online, helping development organizations better target impoverished areas, reducing the influence of gender bias in hiring decisions, and more.\14\ Just as AI can help businesses make smarter decisions, develop innovative new products and services, and boost productivity to drive economic value, it can achieve similar results for organizations generating social value, and many of these solutions have the potential to scale globally. --------------------------------------------------------------------------- \14\ Larry Greenemeier, ``Human Traffickers Caught on Hidden Internet,'' Scientific American, February 8, 2015 http:// www.scientificamerican.com/article/human-traffickers-caught-on-hidden- internet/; Davey Alba, ``Weeding Out Online Bullying Is Tough, So Let Machines Do It,'' Wired, July 10, 2015, https://www.wired.com/2015/07/ weeding-online-bullying-tough-let-machines/; Michelle Horton, ``Stanford Scientists Combine Satellite Data, Machine Learning to Map Poverty,'' Stanford News, August 18, 2016 http://news.stanford.edu/ 2016/08/18/combining-satellite-data-machine-learning-to-map-poverty/; Sean Captain, ``How Artificial Intelligence is Finding Gender Bias at Work,'' Fast Company, October 10, 2015, https://www.fastcompany.com/ 3052053/elasticity/how-artificial-intelligence-is-finding-gender-bias- at-work. --------------------------------------------------------------------------- Finally, AI will be an increasingly important technology for defense and national security. AI can address many goals, such as improving logistics, detecting and responding to cybersecurity incidents, and analyzing the enormous volume of data produced on the battlefield. Moreover, AI will be a core enabler of the Pentagon's ``Third Offset Strategy,'' a policy designed to keep the United States ahead of adversaries, especially ones capable of fielding numerically superior forces, through technological superiority.\15\ Indeed, one top Pentagon general has suggested that the Defense Department should never buy another weapons system that does not have AI built into it.\16\ --------------------------------------------------------------------------- \15\ Sydney Freedberg, ``Faster Than Thought: DARPA, Artificial Intelligence, & The Third Offset Strategy,'' Breaking Defense, February 11, 2016, https://breakingdefense.com/2016/02/faster-than-thought- darpa-artificial-intelligence-the-third-offset-strategy/. \16\ Jack Corrigan, ``Three-Star General Wants Artificial Intelligence in Every New Weapon System,'' Nextgov, November 2, 2017, http://www.nextgov.com/cio-briefing/2017/11/three-star-general-wants- artificial-intelligence-every-new-weapon-system/142225/. --------------------------------------------------------------------------- How Should Policymakers Support the Adoption and Use of AI? Given the potential economic impact of AI in raising productivity, policymakers should develop a national strategy to support the development and adoption of AI in U.S. businesses. In particular, given the enormous advantage that AI-enabled firms will have compared to their non-AI-enabled peers, the United States should focus on AI adoption in its traded sectors where U.S. firms will face international competition. Many other countries see the strategic importance of becoming lead adopters of AI, and they have begun implementing policies to pursue this goal. These include: Canada: In March 2017, Canada launched the Pan-Canadian Artificial Intelligence Strategy which sets a goal of establishing Canada as an international leader in AI research. The strategy has four goals, which include increasing the number of AI researchers and graduates; establishing three major AI research centers; developing global thought leadership on the economic, ethical, policy and legal implications of advances in AI; and supporting the national AI research community.\17\ --------------------------------------------------------------------------- \17\ ``Pan-Canadian Artificial Intelligence Strategy Overview,'' Canadian Institute for Advanced Research, March 3, 2017, https:// www.cifar.ca/assets/pan-canadian-artificial-intelligence-strategy- overview/. China: China's State Council issued a development plan for AI in July 2017 with the goal of making China a leader in the field by 2030. China's goal is to be equal to countries currently leading in AI by 2020. Then, over the subsequent five years, China will focus on developing breakthroughs in areas of AI that will be a ``a key impetus for economic transformation.'' \18\ Finally, by 2030, China intends to be the world's ``premier artificial intelligence innovation center.'' \19\ China's plan also signals its intent to require high school students to take classes in AI, which is one of the most ambitious efforts to develop human capital for the AI economy of any nation. --------------------------------------------------------------------------- \18\ Graham Webster et al., ``China's Plan to `Lead' in AI: Purpose, Prospects, and Problems,'' New America Foundation, August 1, 2017, https://www.newamerica.org/cybersecurity-initiative/blog/chinas- plan-lead-ai-purpose-prospects-and-problems/. \19\ Ibid. Japan: Prime Minister Abe launched the Artificial Intelligence Technology Strategy Council in April 2016 to develop a roadmap for the development and commercialization of AI.\20\ Published in May 2017, the roadmap outlines priority areas for research and development (R&D), focusing on the themes of productivity, mobility, and health. The strategy also encourages collaboration between industry, government, and academia to advance AI research, as well as stresses the need for Japan to develop the necessary human capital to work with AI. Japan also launched its Japan Revitalization Strategy 2017, which details how the government will work to support growth in certain areas of the economy. The 2017 strategy includes a push to promote the development of AI for telemedicine and self- driving vehicles to address the shortage of workers in Japan. --------------------------------------------------------------------------- \20\ Josh New, ``How Governments Are Preparing for Artificial Intelligence,'' August 8, 2017, https://www.datainnovation.org/2017/08/ how-governments-are-preparing-for-artificial-intelligence/. UK: The United Kingdom has taken several steps to promote AI. The UK Digital Strategy, published in March 2017, recognizes AI as a key field that can help grow the United Kingdom's digital economy.\21\ The UK's new budget, published in November 2017, includes several provisions that have the goal of establishing the UK as a world leader in AI, such as by establishing a ``Centre for Data Ethics and Innovation'' to promote the growth of AI, facilitating data access for AI through ``data trusts,'' and funding 450 PhD researchers working on AI.\22\ --------------------------------------------------------------------------- \21\ Department of Digital, Culture, Media, and Sport, UK Digital Strategy, (United Kingdom: Department for Digital, Culture, Media, and Sport, 2017), https://www.gov.uk/government/publications/uk-digital- strategy. \22\ Her Majesty's Treasury (HM Treasury), Autumn Budget 2017 (United Kingdom: HM Treasury, 2017), https://www.gov.uk/government/ publications/autumn-budget-2017-documents/autumn-budget-2017. While the U.S. Government has put significant funding behind AI R&D--approximately $1.1 billion in 2015--it has not done enough to maintain U.S. leadership.\23\ The most ambitious AI program comes from China, which as of 2014 surpassed the United States in terms of total number of papers published and cited in AI fields, such as deep learning.\24\ For both economic and national security reasons, the United States cannot afford to cede its existing advantages in AI, and should instead look to capitalize on its head start by developing a strategy to support AI development and adoption. Such a strategy should include policies to address the following: --------------------------------------------------------------------------- \23\ Ibid. \24\ ``National Artificial Intelligence Research and Development Strategic Plan,'' (National Science and Technology Council, October 2016), https://www.nitrd.gov/PUBS/national_ai _rd_strategic_plan.pdf. Funding: The government should continue to expand its funding to support the ``National Artificial Intelligence Research and Development Strategic Plan,'' a set of R&D priorities identified by the Networking and Information Technology Research and Development (NITRD) program that addresses strategic areas of AI in which industry is unlikely to invest, as well as better plan and coordinate Federal funding for AI R&D across different agencies.\25\ --------------------------------------------------------------------------- \25\ Ibid. Skills: The Federal Government should support educational efforts to ensure a strong pipeline of talent to create the next generation of AI researchers and developers, including through retraining and diversity programs, as well as pursue immigration policies that allow U.S. businesses to recruit and --------------------------------------------------------------------------- retain highly skilled computer scientists. AI-Friendly Regulations: Federal and state regulators should conduct regulatory reviews to identify regulatory barriers to commercial use of AI in various industries, such as transportation, health care, education, and finance. Data Sharing: Some advances in AI are made possible when large volumes of accurate and representative data are made part of a data commons. The government should continue to supply high-value datasets that enable advances in AI, such as its efforts to produce standardize reference datasets for text analysis and facial recognition. Similarly, Federal agencies should facilitate data sharing between industry stakeholders, such as the Department of Transportation's draft ``Guiding Principles on Data Exchanges to Accelerate Safe Deployment of Automated Vehicles.'' \26\ --------------------------------------------------------------------------- \26\ ``Draft U.S. DOT Guiding Principles on Voluntary Data Exchanges to Accelerate Safe Deployment of Automated Vehicles,'' (U.S. Department of Transportation, December 1, 2017) https:// www.transportation.gov/av/data. Economic Indicators: Understanding the degree to which U.S. firms have automated processes using AI will be a key metric to assessing the effectiveness of various policies. The Census Bureau should assess what type of economic data it should gather from businesses to monitor and evaluate AI adoption, much like it has tracked rural electrification or broadband connectivity as key economic indicators. How Should Policymakers Address Concerns About Workforce Disruption? One of the most common fears about AI is that it will lead to significant disruptions in the workforce.\27\ This fear is not new-- concerns about technology-driven automation have been a perennial policy concern since at least the 1930s when Congress debated legislation that would direct the Secretary of Labor to make a list of all labor-saving devices and estimate how many people could be employed if these devices were eliminated.\28\ This concern has been exacerbated by a frequently-cited study by two Oxford academics which predicted that 47 percent of U.S. jobs could be eliminated over the next 20 years.\29\ --------------------------------------------------------------------------- \27\ For a thorough rebuttal of this concern, see Robert D. Atkinson, ``'It's Going to Kill Us!' And Other Myths of Artificial Intelligence,'' (Information Technology and Innovation Foundation, June 2016), http://www2.itif.org/2016-myths-machine-learning.pdf. \28\ John Scoville, ``Technology and the Volume of Employment,'' Proceedings of the Academy of Political Science 18, no. 1 (May 1938): 84-99. \29\ Carl B. Frey and Michael A. Osbourne, ``The Future of Employment: How Susceptible Are Jobs to Computerisation?'' (Oxford Martin School, University of Oxford, Oxford, September 17, 2013), http://www.oxfordmartin.ox.ac.uk/downloads/academic/ The_Future_of_Employment. pdf. --------------------------------------------------------------------------- This study's predictions are misleading and unlikely for at least three reasons. First, the estimate includes a number of occupations that have little chance of automation, such as fashion models and barbers. Second, while this rate of productivity seems high and even threatening, it is only slightly higher than rates enjoyed in the mid- 1990s when U.S. job creation was robust and unemployment rates low. Third, it succumbs to what economists call the ``lump of labor'' fallacy which holds that once a job is gone, there are no other jobs to replace it. The reality is that AI-driven productivity enables organizations to either raise wages or reduce prices. These changes lead to increases in spending, which in turn creates more jobs. And given that consumers' wants are far from satisfied, there is no reason to believe that this dynamic will change anytime soon. But while predictions about massive AI-driven unemployment are vastly overstated--indeed, by historical standards occupational churn, the rate at which some jobs expand while others contract, is at its lowest levels in 165 years--there will still be some worker displacement as AI creates higher levels of productivity.\30\ So policymakers can and should do more to help workers make transitions between jobs and occupations, such as by providing strong social safety net programs, reforming unemployment insurance, and offering worker retraining. The failure to give workers training and assistance to move into new jobs or occupations not only contributes to higher structural unemployment, but also increases resistance to innovation and automation.\31\ --------------------------------------------------------------------------- \30\ Robert D. Atkinson and John Wu, ``False Alarmism: Technological Disruption and the U.S. Labor Market, 1850-2015,'' (Information Technology and Innovation Foundation, May 2017), http:// www2.itif.org/2017-false-alarmism-technological-disruption.pdf. \31\ See forthcoming report: ``Technological Innovation, Employment, and Workforce Adjustment Policies,'' (Information Technology and Innovation Foundation, January 2018). --------------------------------------------------------------------------- How Should Policymakers Provide Oversight of AI? When it comes to AI, the primary goal of the United States should be to accelerate the development and adoption of the technology. But as with any technology, there will be some risks and challenges that require government oversight. The presence of risk, however, does not mean that the United States should embrace the precautionary principle, which holds that new technology must first be proven safe before it can be used. Instead, policymakers should rely on the innovation principle, which says that policymakers should address risks as they arise, or allow market forces to address them, and not hold back progress because of speculative concerns. The innovation principle is especially useful when fears about a new technology exceed public awareness and understanding about how the technology works and how potential problems will be mitigated.\32\ --------------------------------------------------------------------------- \32\ Daniel Castro and Alan McQuinn, ``The Privacy Panic Cycle: A Guide to Public Fears About New Technologies,'' (Information Technology and Innovation Foundation, September 2015), http://www2.itif.org/2015- privacy-panic.pdf. --------------------------------------------------------------------------- To understand why this is important, consider the differences between the United States and the European Union in the Internet economy. Compared to Europe, the United States has had more success in the Internet economy, at least in part, because of its vastly more simplified data protection regulations. Yet even as the United States continues to produce the majority of the major global Internet companies, the European Union has decided to double down on its onerous data protection rules in the forthcoming General Data Protection Regulation (GDPR), a far-reaching set of policies that will substantially raise the costs, and in some cases, limit the feasibility of using AI in Europe. For example, the GDPR creates both a right to explanation and a right to human review for automated decisions, two requirements that will make it difficult for companies to construct business models that rely extensively on complex algorithms to automate consumer-facing decisions. The GDPR also requires organizations to only use data for the purposes for which they originally collected it, a rule that strictly limits the application of AI to existing data.\33\ If the United States wants to compete for global leadership in AI, it should be careful not to follow Europe down this path. --------------------------------------------------------------------------- \33\ Nick Wallace, ``UK Regulations Need an Update to Make Way for Medical AI,'' Center for Data Innovation, August 12, 2017, http:// datainnovation.org/2017/08/uk-regulations-need-an-update-to-make-way- for-medical-ai/. --------------------------------------------------------------------------- While the United States should not replicate the European model, it should create its own innovation-friendly approach to providing oversight of the emerging algorithmic economy just as it has for the Internet economy. Such an approach should prioritize sector-specific policies over comprehensive regulations, outcomes over transparency, and enforcement actions against firms that cause tangible harm over those that merely make missteps without injury. For example, rather than industry-wide rules requiring ``algorithmic transparency'' or ``AI ethics''--proposals that focus on means, rather than ends--policymakers should look to address specific problems, such as ensuring financial regulators have the skills necessary to provide oversight of fintech companies relying heavily on AI to make lending decisions or provide automated financial advisors. In many cases, regulators will not need to intervene because the private sector will address problems about AI, such as bias or discrimination, on its own--even if to outsiders an algorithm appears to be a ``black box.'' After all, one company's hidden biases are another company's business opportunities. For example, if certain lenders were to use algorithms that consistently denied loans to ethnic or religious minorities who have good credit, then their competitors would have an incentive to target these individuals to gain new customers. Moreover, the private sector is actively seeking out solutions to eliminate problems like unintentional bias in AI that may skew its results.\34\ For example, a group of leading AI companies in the United States have formed an association to develop and share best practices to ensure that AI is fair, safe, and reliable, while another technology trade association has publicly committed itself to ensuring that the private sector designs and uses AI responsibly.\35\ Indeed, given that U.S. companies are at the forefront of efforts to build AI that is safe and ethical, maintaining U.S. leadership in this field will be important to ensure these values remain embedded in the technology. --------------------------------------------------------------------------- \34\ Cliff Kuang, ``Can A.I. Be Taught to Explain Itself?'' New York Times, November 21, 2017, https://www.nytimes.com/2017/11/21/ magazine/can-ai-be-taught-to-explain-itself.html. \35\ See ``Partnership on AI,'' https://www.partnershiponai.org/and ``AI Policy Principles,'' Information Technology Industry Council, https://www.itic.org/resources/AI-Policy-Principles-FullReport2.pdf. --------------------------------------------------------------------------- But policymakers should be careful not to misclassify certain concerns as ``AI problems'' that would be best dealt with on a technology-neutral basis. For example, discrimination in areas such as access to financial services and housing are best addressed through existing legal mechanisms. No new laws and regulations are needed simply because a company uses AI, instead of human workers, to make certain decisions.\36\ Companies cannot use AI to circumvent laws outlawing discrimination. --------------------------------------------------------------------------- \36\ Travis Korte and Daniel Castro, ``Disparate Impact Analysis is Key to Ensuring Fairness in the Age of the Algorithm,'' Center for Data Innovation, January 20, 2015, http://datainnovation.org/2015/01/ disparate-impact-analysis-is-key-to-ensuring-fairness-in-the-age-of- the-algorithm/. --------------------------------------------------------------------------- Finally, certain problems, such as sexism in hiring practices, are not necessarily made worse by AI. On the contrary, using AI can actually reduce human biases. For example, companies can use AI to police undesirable behaviors, like automatically flagging job advertisements that use gender-specific terminology, such as ``waitress'' instead of ``wait staff,'' or stereotypical images, such as a female nurse.\37\ And unlike human processes, where it may take years or decades to change social norms and company culture, businesses can refine and tweak code over a period of days or weeks. For example, Google changes its search engine 500 to 600 times per year.\38\ Thus companies will likely have more success eliminating bias when it appears in AI, than when it appears elsewhere in society. --------------------------------------------------------------------------- \37\ Amber Laxton, ``Critics of `Sexist Algorithms' Mistake Symptoms for Illness,'' Center for Data Innovation, August 3, 2015, http://datainnovation.org/2015/08/critics-of-sexist-algorithms-mistake- symptoms-for-illness/. \38\ Daniel Castro, ``Data Detractors Are Wrong: The Rise of Algorithms Is a Cause for Hope and Optimism,'' Center for Data Innovation, October 25, 2016, http://datainnovation.org/2016/10/data- detractors-are-wrong-the-rise-of-algorithms-is-a-cause-for-hope-and- optimism/. --------------------------------------------------------------------------- Conclusion AI is a transformational technology that has the potential to significantly increase efficiency and innovation across the U.S. economy, creating higher living standards and improved quality of life. But while the United States has an early advantage in AI given its top talent in computer science and deep bench of companies, large and small, investing in the field, many other countries are actively vying to challenge U.S. leadership in this domain. In particular, China, with its highly skilled computer science workforce and significant funding for AI R&D, could easily catch and surpass the United States, leading to it gaining economic and military advantages. Unfortunately, U.S. policy debates about AI too often overemphasize the potential impact on worker displacement from automation or bias from algorithms and ignore the much more pressing concern about the potential loss of competitiveness and defense superiority if the United States falls behind in developing and adopting this key technology. Yet, when it comes to AI, successfully integrating this technology into U.S. industries should be the primary goal of policymakers, and given the rapid pace at which other countries are pursuing this goal, the United States cannot afford to rest on its laurels. To date, the U.S. Government has not declared its intent to remain globally dominant in this field, nor has it begun the even harder task of developing a strategy to achieve that vision. Some may think this is unnecessary, believing that the United States will automatically prevail in this technology simply because it has a unique culture of innovation and has prevailed on past technologies.\39\ Such views are naive and dangerous, and if followed, likely will lead to the United States being surpassed as the global leader in AI with significant negative consequences for the U.S. economy and society. However, it is not too late to begin to ensure continued U.S. leadership, and I commend you for holding this hearing so that we can have this conversation. --------------------------------------------------------------------------- \39\ Patrick Tucker, ``What the CIA's Tech Director Wants from AI,'' Defense One, September 6, 2017, http://www.defenseone.com/ technology/2017/09/cia-technology-director-artificial-intelligence/ 140801/. Senator Wicker. Thank you, Mr. Castro. Ms. Espinel. STATEMENT OF VICTORIA ESPINEL, PRESIDENT AND CEO, BSA THE SOFTWARE ALLIANCE Ms. Espinel. Good morning, Chairman Wicker, Ranking Member Schatz, and members of the Subcommittee. My name is Victoria Espinel, and I am the President and CEO of BSA The Software Alliance. BSA is the advocate for the global software industry in the United States and around the world. Our members are at the forefront of developing artificial intelligence and related software services. I commend the Subcommittee for a hearing on this important topic, and I thank you for the opportunity to testify. At the outset, I think it's important to answer a key question: What is AI? So let me provide you with a brief anecdote. A 60-year-old woman was initially diagnosed with a conventional form of leukemia. She went through chemotherapy to treat the disease, but her recovery was unusually slow. Conventional tests failed to reveal a problem, but her doctor suspected that something was still wrong. After several frustrating months, they turned to an AI-powered, cloud-based system capable of cross-referencing the patient's genetic data with insights gleaned from tens of millions of studies from around the world. Within minutes, the doctors learned that the patient might be suffering from an extremely rare form of leukemia that required a unique course of treatment. The doctors were able to quickly update her treatment plan and watch her condition improve significantly. This is AI: it's innovative; it's powerful; it's lifesaving. AI is not the image that we see in science fiction movies of robots demolishing tall buildings; instead, the AI provided by BSA members today is a tool that uses data to help people solve complex problems, simplify our daily lives, improve business operations, and enhance government services. AI is powered by software, which is itself a major engine of economic growth. The software industry contributed more than $1.14 trillion to the U.S. GDP in 2016--a $70 billion increase in just 2 years. The software industry is a powerful job creator supporting over 10.5 million jobs with a significant positive impact on jobs and economic growth in every one of the 50 states. For example, in Mississippi, software is contributing over $800 million to its GDP and over 7,000 jobs, a 25 percent increase in jobs in just 2 years. Over 4,000 miles away in Hawaii, software is contributing over $1 billion to its GDP and over 16,000 jobs. Across every single state in the country, the economic impact of software is on the rise. AI is helping all industry sectors. Whether it is securing networks, improving health, or helping American farmers save money, the impact of AI is already visible in every industry, in every state, and across the globe. We should also be prepared to address important issues that may arise as AI-enabled services are used. Let me focus on two. First, AI will change the skill sets needed for certain jobs, and while new AI-related jobs will be created, there will be shifts in the economy. BSA members are already helping launching groundbreaking initiatives to provide free training, including to youth and military veterans, to ensure that both the current workforce and the next generation are prepared for the future. We are dedicated to this work, and we look forward to collaborating with all of you on this effort. Second, we are mindful of the need to ensure that AI is both trained and used fairly and responsibly. At the same time, we recognize the potential of AI to make human decisions more accurate and less biased and the need to push toward that outcome. As our companies seek to ensure responsible AI deployment, there are several steps that Congress and the administration can take. First, as I highlighted earlier, AI depends on data, so we urge Congress to pass the Open Government Data Act, which would make non-sensitive government data more open, more available, and more usable for the general public. Ranking Member Schatz, thank you for your great work as sponsor of the Open Government Data Act. We hope that Congress will act soon to send it to the President's desk. We also encourage Congress and the administration to be leaders on digital trade to encourage global data flows. Second, we encourage increased investment in government research, including on how AI can contribute to both positive economic and social outcomes and policies that incentivize private sector research and development. And, third, we need to prioritize education and workforce development so that our young people and our current workforce are prepared for the future. As part of all of this, we need to have a meaningful dialogue with all stakeholders about how to address any challenges that lie ahead. The legislation introduced by Senators Cantwell, Young, and Markey is a good step, and we thank you for that. Thanks to Senator Schatz as well for the legislation that you are currently working on. In closing, we look forward to working with all of you towards a clear understanding of AI and to address the challenges and embrace the opportunities ahead. BSA members are part of the solution to these challenges, and we are eager to work with you as we chart a responsible path forward. Thank you, and I look forward to your questions. [The prepared statement of Ms. Espinel follows:] Prepared Statement of Victoria Espinel, President and CEO, BSA--The Software Alliance Good morning Chairman Wicker, Ranking Member Schatz, and members of the Subcommittee. My name is Victoria Espinel, and I am the President and CEO of BSA The Software Alliance. BSA is the leading advocate for the global software industry in the United States and around the world.\1\ Our members are at the forefront of developing cutting-edge artificial intelligence (AI) and related software-enabled technologies and services that are having a significant impact on the U.S. and global economy. I commend the Subcommittee for holding a hearing on this important topic, and I thank you for the opportunity to testify on behalf of BSA. --------------------------------------------------------------------------- \1\ BSA's members include: Adobe, ANSYS, Apple, Autodesk, Bentley Systems, CA Technologies, CNC/Mastercam, DataStax, DocuSign, IBM, Microsoft, Oracle, salesforce.com, SAS Institute, Siemens PLM Software, Splunk, Symantec, Trimble Solutions Corporation, The MathWorks, Trend Micro and Workday. --------------------------------------------------------------------------- I. AI: Defining the Landscape The term ``artificial intelligence'' often conjures images of all- knowing robots with physical and cognitive abilities far superior to those of their human creators. The actual AI services that are in the market today--and that BSA members provide--bear no resemblance to the sinister images of the future that consumers often see in the movies, with robots taking over big cities and small towns. Instead, they increasingly are becoming a foundational technology that drives many products and services that people use every day. Whether it is a personal digital assistant that helps consumers locate the nearest restaurant, a fraud detection monitoring service that prevents criminals from placing charges on credit cards, or a tool that helps teachers identify students with additional needs and develop personalized lesson plans, we increasingly rely on a diverse range of AI-enabled services every day. But what is ``AI''? Although definitions of AI vary, one common description of AI is that it refers to machines that act intelligently in pursuit of human- defined objectives. At its core, AI is simply a tool. It includes a broad range of technologies, but the AI systems that BSA members largely provide assist in the analysis of enormous volumes of data to find connections that improve the quality and accuracy of human decision-making. Although some AI systems have a limited degree of autonomy, such as submarines that map the ocean bed and measure ocean currents, and others are minutely supervised, such as robot surgical tools assisting doctors with hip replacement surgeries, the vast majority provide advice and recommendations to humans rather than acting independently. AI makes possible important tasks that would otherwise be economically or physically infeasible, such as inspecting wind turbine blades or the interior of oil pipelines. AI systems, like other software systems, use sophisticated algorithms. An algorithm is a set of instructions that processes various inputs and provides an output in a systematized way. The algorithms used in AI are particularly well-suited to analyzing massive volumes of data from many different sources, and in identifying patterns across the enormous number of variables in such data that may interact in complex and unexpected ways. Through this analysis, AI systems can enhance perception, learning, reasoning, and decision- making, and improve the ability of people to solve complex and challenging problems. The use of systems, including software, to help people solve complex problems is not new. Research into AI dates back many decades, but we have witnessed tremendous advances in AI capabilities over the past five to ten years. These advances have been fueled by a number of related developments, including the proliferation of technologies that generate vast amounts of data, the affordability of data storage, and ever-growing data processing capabilities. BSA members have made significant investments in enhancing these data-driven technologies to develop innovative AI solutions for use across a broad range of applications in a wide variety of contexts. II. AI Services Provide Substantial Benefits Advances in AI and software-enabled data analytics are fueling job and economic growth in the United States and around the world, improving how businesses in every sector operate, and producing real societal gains. We must recognize that AI will change the skill sets needed for certain jobs. And while new, AI-related jobs will be created, there will be shifts in the labor market. And although we should be mindful of the need to ensure that AI is deployed fairly and responsibly, we should also recognize the potential of AI to make human decisions more accurate and less biased, and thereby to promote fairness and inclusiveness across all segments of society. A. AI and Related Software Services Are Creating Jobs and Economic Growth In high-tech and low-tech industries alike, the analysis of data has made businesses more agile, responsive, and competitive, boosting the underlying productivity of many key pillars of our economy. The economic implications of the data revolution--and AI and related software solutions that leverage that data--are enormous. Economists predict that making better use of data could lead to a ``data dividend'' of $1.6 trillion in the next four years, and that data-enabled efficiency gains could add almost $15 trillion to global GDP by 2030.\2\ In addition, experts predict that applications of AI technologies could grow the global economy by $7.1 to $13.17 trillion over the next eight years.\3\ --------------------------------------------------------------------------- \2\ See BSA, What's the Big Deal With Data? 14 (Oct. 2015), available at http://data.bsa.org/wp-content/uploads/2015/12/ bsadatastudy_en.pdf. The potential of digital data to improve the healthcare system is substantial: some estimates predict that if the healthcare sector were to use data more effectively to drive efficiency and quality, the sector could save more than $300 billion every year. See James Manyika et al., Big Data: The Next Frontier for Innovation, Competition, and Productivity, McKinsey Global Institute (May 2011), available at http://www.mckinsey.com/insights/business_technology/ big_data_the_next_frontier_for_innovation. \3\ See Disruptive technologies: Advances that will transform life, business, and the global economy, McKinsey Global Institute (May 2013), available at http://www.mckinsey.com/business-functions/digital- mckinsey/our-insights/disruptive-technologies. --------------------------------------------------------------------------- AI systems are powered by software, which itself is a major engine of economic growth. In September, Software.org: the BSA Foundation released a study with data from the Economist Intelligence Unit (EIU) showing that the software industry alone contributed more than $1.14 trillion to U.S. GDP in 2016--a $70 billion increase in just two years.\4\ The study also showed that the software industry is a powerful job creator, supporting over 10.5 million jobs, with a significant impact on job and economic growth in each of the 50 states.\5\ --------------------------------------------------------------------------- \4\ Software.org: The BSA Foundation, The Growing $1 Trillion Economic Impact of Software 5 (Sept. 2017), available at https:// software.org/wp-content/uploads/2017_Software_Economic _Impact_Report.pdf. \5\ Id. --------------------------------------------------------------------------- B. AI and Related Software Services Are Improving Every Industry The benefits of AI are not limited to the software sector. In fact, AI innovation is stimulating growth across all industry sectors as businesses, big and small, use AI and related software services to improve supply chains, secure their networks, and evaluate how to improve their products and services. There are numerous examples of this positive impact across a wide swath of industries, for instance: Cybersecurity. AI tools are revolutionizing how we monitor network security, helping analysts parse through hundreds of thousands of security incidents per day to weed out false positives and identify threats that warrant further attention by network administrators. By automating responses to routine incidents and enabling security professionals to focus on truly significant threats, AI-enabled cyber tools are helping enterprises stay ahead of their malicious adversaries.\6\ --------------------------------------------------------------------------- \6\ For example, IBM's Watson for Cyber Security is a cybersecurity tool that can analyze 15,000 security documents per day--a rate essentially impossible for any individual to achieve. Watson's data processing capabilities enable analysts to more quickly identify incidents that require human attention. See IBM, IBM Delivers Watson for Cyber Security to Power Cognitive Security Operations Centers (Feb. 13, 2017), https://www-03.ibm.com/press/us/en/press release/51577.wss; Jason Corbin, Bringing the Power of Watson and Cognitive Computing to the Security Operations Center, Security Intelligence (Feb. 13, 2017), https://securityintelligence .com/bringing-the-power-of-watson-and-cognitive-into-the-security- operations-center/?cm_mc_uid =70595459933115020631816&cm_mc_sid_50200000=1503364089&cm_mc_sid_5264000 0=150336 5578. Splunk uses a similar model, with machine-learning algorithms conducting real-time analysis and processing of massive volumes of data from all sensors on a network to identify anomalies, feeding visualization tools that help network administrators efficiently triage security incidents. See David Braue, Machine learning key to building a proactive security response: Splunk, CSO Online (Aug. 20, 2015), https://www.cso.com.au/article/582483/machine-learning-key-building- proactive-security-response-splunk/. Microsoft's Windows 10 Anniversary Edition introduced AI-driven capabilities for automatically isolating suspicious network traffic pending adjudication by network administrators. See Chris Hallum, Defense Windows clients from modern threats and attacks with Windows 10, Channel 9 video content (Oct. 6, 2016), available at https://channel9.msdn.com/events/Ignite/2016/ BRK2135-TS); ``Intelligent Security: Using Machine Learning to Help Detect Advanced Cyber Attacks,'' https://www.microsoft.com/en-us/ security/intelligence. Financial Services. AI is improving fraud detection by providing companies with real-time information that helps them identify and investigate different types of fraud, reducing the losses attributed to fraudsters by billions of dollars. In a matter of seconds, machine learning algorithms can generate a risk score for a transaction by parsing through large volumes of data about the vendor and the purchaser to determine the likelihood of fraud.\7\ These tools are protecting consumers from the risk of fraudulent charges and from the frustration associated with ``false declines.'' --------------------------------------------------------------------------- \7\ See generally Pablo Hernandez, CA Technologies Uses AI Tech to Combat Online Fraud, eSecurityPlanet, May 4, 2017, available at https:/ /www.esecurityplanet.com/network-security/ca-technologies-uses-ai-tech- to-combat-online-fraud.html. Agriculture. AI is helping farmers tackle some of the biggest issues they face, including declining crop yields and changing weather patterns, through precision farming, better data analysis, and improved operational efficiency. For instance, tools like computer vision and deep-learning algorithms are enabling farmers more effectively to process data for purposes of monitoring crop and soil health.\8\ --------------------------------------------------------------------------- \8\ See Kumba Sennaar, AI in Agriculture--Present Applications and Impact, techemergence (Nov. 17, 2017), https://www.techemergence.com/ ai-agriculture-present-applications-impact/. Manufacturing. AI-enabled tools are also helping factory owners streamline their manufacturing processes and resolve problems common to most factories, such as inaccurate demand forecasting and capacity planning, unexpected equipment failures and downtimes, and supply chain bottlenecks. Predictive maintenance, for instance, allows manufacturers to achieve 60 percent or more reduction in unscheduled system downtime. Cameras powered by computer vision algorithms can fix product defects immediately and identify root causes of failure. AI thus enables manufacturers to reduce waste, shorten production periods, increase yields on production inputs, and improve both revenue and workplace safety.\9\ --------------------------------------------------------------------------- \9\ See Mariya Yao, Factories Of The Future Need AI To Survive And Compete, Forbes.com (Aug. 8, 2017), https://www.forbes.com/sites/ mariyayao/2017/08/08/industrial-ai-factories-of-future/#2d7ab2fd128e. Healthcare. AI technologies are already providing solutions that help save lives. A 2016 Frost & Sullivan report predicts that AI has the potential to improve health outcomes by 30 to 40 percent.\10\ AI is helping fuel these improved health outcomes not by replacing the decision-making of healthcare professionals, but by giving these professionals new insights and new ways of analyzing and understanding the health data to which they have access. For example, AI tools are powering machine-assisted diagnosis and surgical applications are being used to improve treatment options and outcomes. Image recognition algorithms are helping pathologists more effectively interpret patient data, thereby helping physicians form a better picture of patients' prognosis.\11\ The ability of AI to process and find patterns in vast amounts of data from disparate sources is also driving important progress in biomedical and epidemiological research.\12\ --------------------------------------------------------------------------- \10\ See From $600 M to $6 Billion, Artificial Intelligence Systems Poised for Dramatic Market Expansion in Healthcare, Frost & Sullivan (Jan. 5, 2016), https://ww2.frost.com/news/press-releases/600-m-6- billion-artificial-intelligence-systems-poised-dramatic-market- expansion-healthcare. \11\ See e.g., Meg Tirrell, From coding to cancer: How AI is changing medicine, cnbc.com (May 11, 2017), https://www.cnbc.com/2017/ 05/11/from-coding-to-cancer-how-ai-is-changing-medicine.html. \12\ For instance, AI is helping biologists who are aiming to treat 100 molecular genetic diseases by 2025. See Splunk, Machine Learning Helps Recursion Pharmaceuticals Treat Genetic Diseases (Nov. 7, 2017), https://www.splunk.com/en_us/newsroom/press-releases/2017/splunk- machine-learning-helps-recursion-pharmaceuticals-treat-genetic- diseases.html. In another example, Microsoft researchers are also using AI and related technologies to better understand the behavior of cells and their interaction, which could ultimately help ``debug'' an individual's specific form of cancer and allow doctors to provide personalized cancer treatment. See generally, Microsoft, Biological Computation, https://www.microsoft.com/en-us/research/group/biological- computation/. Education. AI technologies offer tools for students, teachers, and administrators to help students learn more effectively both within and outside of the classroom. AI programs can, for example, analyze a student's performance in a particular skill across subjects over the course of a year and automatically provide new content or specified learning parameters, offering students continual, individualized practice and feedback. They can also help teachers better understand student performance, quickly identify students that need particular attention, and develop lesson plans that customize instruction, content, pace, and testing to individual students' strengths and interests.\13\ AI solutions also are helping administrators track attendance patterns and gain insights on student performance more broadly.\14\ --------------------------------------------------------------------------- \13\ See Software.org: The BSA Foundation, The Growing $1 Trillion Economic Impact of Software, supra note 4, at 7; see also Daniel Faggella, Examples of Artificial Intelligence in Education, TechEmergence (Mar. 7, 2017), https://www.techemergence.com/examples- of-artificial-intelligence-in-education/. \14\ Benjamin Herold, Are schools ready for the power and problems of big data?, Education Week (Jan. 11, 2016), available at http:// www.edweek.org/ew/articles/2016/01/13/the-future-of-big-data-and- analytics.html. --------------------------------------------------------------------------- C. AI Services Provide Tremendous Societal Benefits The range of potential societal benefits from the use of AI services is equally vast. For example, AI solutions are at the heart of new devices and applications that improve the lives of people with disabilities, including helping people with vision-related impairments interpret and understand photos and other visual content.\15\ This technology opens new possibilities for people with vision impairments to navigate their physical surroundings, giving them increased independence and greater ability to engage with their communities. --------------------------------------------------------------------------- \15\ For instance, Microsoft recently released an intelligent camera app that uses a smartphone's built-in camera functionality to describe to low-vision individuals the objects that are around them. See Microsoft, Seeing AI, https://www.microsoft.com/en-us/seeing-ai/. --------------------------------------------------------------------------- AI is also helping governments improve constituent services in ways that save time, money, and lives. For example, cities are optimizing medical emergency response processes using AI-based systems, enabling them to more strategically position personnel and reduce both response times and the overall number of emergency trips.\16\ AI is also helping to leverage data to improve disaster response and relief efforts, including after the 2015 earthquake in Nepal.\17\ --------------------------------------------------------------------------- \16\ See Kevin C. Desouza, Rashmi Krishnamurthy, and Gregory S. Dawson, Learning from public sector experimentation with artificial intelligence, Brookings Institution (June 23, 2017), https:// www.brookings.edu/blog/techtank/2017/06/23/learning-from-public-sector- experimentation-with-artificial-intelligence/. \17\ See Patrick Meier, Virtual Aid to Nepal: Using Artificial Intelligence in Disaster Relief, Foreign Affairs (June 1, 2015), available at https://www.foreignaffairs.com/articles/nepal/2015-06-01/ virtual-aid-nepal. --------------------------------------------------------------------------- * * * * * Whether it is detecting financial fraud, improving health outcomes, making American farmers more competitive, or enhancing government and emergency services, the impact of AI and related software services is already visible in every industry, in every state, and across the globe. III. Fostering Consumer Trust in AI Even as society gains from the substantial benefits that AI offers, we also recognize that there may be legitimate concerns about how AI systems are deployed in practice, which may also affect trust and confidence in AI. In particular, as people increasingly apply AI services in new contexts, questions may arise about how they operate, whether they treat people fairly and are free from improper bias, and their impact on jobs. Like many technologies, AI has an almost infinite range of beneficial uses, but we should also take appropriate steps to ensure that it is deployed responsibly. We recognize that responsible deployment of AI should instill consumer confidence that these important issues will be appropriately addressed. A. Enhancing Understanding of AI Systems Building trust and confidence in AI-enabled systems is an important priority. In some instances, the complexity of these technologies, which are designed to identify patterns and connections that humans could not easily identify on their own, can make it challenging to explain how certain aspects of AI systems work. BSA members understand that, in order to promote trust, companies that build and deploy AI systems will need to provide meaningful information to enhance understanding of how these systems operate. Indeed, ensuring that AI systems operate as intended and treat people fairly is an important priority. We are eager to participate in meaningful dialogues with other stakeholders about how best to accomplish that goal, and we welcome opportunities such as this one to help advance that dialogue. Currently, relevant technical tools and operational processes that could improve understanding and confidence in AI systems are still being developed, and it is an area of robust research. Although more work needs to be done, it is already clear that expectations are highly context-specific--and demands will vary based on this context. As we seek to address these important issues, we will aim to ensure that we remain sufficiently flexible to respond to concerns, and to adapt to the changing landscape as these emerging technologies, and potential solutions to new challenges, continue to evolve. B. Preparing the Workforce for the Jobs of the Future As AI services improve every industry, they will likely have a multi-dimensional impact on employment. The deployment of AI in the workplace will enable employees to focus on tasks that are best suited to uniquely human skillsets, such as creativity, empathy, foresight, judgment, and other social skills. Although there appears to be no consensus on the precise impact AI will have on employment, there is broad recognition that widespread deployment of these technologies will create demand for new types of jobs, and that these jobs often will require skills that many workers today do not yet have. Current estimates indicate the United States will not have enough workers to meet the predicted high demand for computer science-related jobs. For example, by 2020, the U.S. Bureau of Labor Statistics predicts that there will be 1.4 million computing jobs, but just 400,000 computer science students with the skills necessary to fill those jobs.\18\ It is imperative that the United States takes steps now to ensure that we have a sufficient pipeline of workers with the skills needed to perform these new, high-quality jobs. --------------------------------------------------------------------------- \18\ See Allie Bidwell, Tech Companies Work to Combat Computer Science Education Gap, U.S. News & world report, Dec. 27, 2013, available at https://www.usnews.com/news/articles/2013/12/27/tech- companies-work-to-combat-computer-science-education-gap. --------------------------------------------------------------------------- Yet even these estimates do not take into account the extent to which the use of AI may require new skills. Because AI services will likely be integrated across all sectors of the economy, the new jobs AI creates, and the new skills that will be needed, will reach beyond the tech sector, and will also likely extend to workers in both urban and rural areas. Indeed, many of these jobs will ``look nothing like those that exist today,'' and will include ``entire categories of new, uniquely human jobs'' that will require ``skills and training that have no precedents.'' \19\ As a result, one key challenge that lies ahead is determining how to ensure that the U.S. workforce has the skills necessary for the future. --------------------------------------------------------------------------- \19\ H. James Wilson, Paul R. Daugherty, Nicola Morini-Bianzino, The Jobs that Artificial Intelligence will Create, MIT Sloan Management Review (Mar. 23, 2017), available at https://sloanreview.mit.edu/ article/will-ai-create-as-many-jobs-as-it-eliminates/. --------------------------------------------------------------------------- BSA members are working hard to help address this challenge. BSA recognizes that this will require a multi-faceted solution, including cooperation with public and private stakeholders. We seek to identify opportunities and partnerships that focus on retraining the workforce with new skills, creating a pipeline of workers with skills to fill the next generation of jobs, increasing access to those jobs for skilled workers, and increasing deployment of cloud services, which facilitate employment and collaboration in different geographic regions. Notably, BSA members already have begun helping workers and youth acquire new skills that will enable them to leverage AI systems.\20\ BSA members offer several high-tech and business training programs, including at the high school level. Some programs target populations not traditionally associated with tech jobs, such as military veterans.\21\ These initiatives illustrate just some of the ways in which AI-based employment concerns can be meaningfully addressed. --------------------------------------------------------------------------- \20\ See, e.g., Allen Blue, How LinkedIn is Helping Create Economic Opportunity in Colorado and Phoenix (Mar. 17, 2016), https:// blog.linkedin.com/2016/03/17/how-linkedin-is-helping-create-economic- opportunity-in-colorado-and-phoenix; Markel Foundation, Why Microsoft and the Markle Foundation are Working Together to Connect Workers with New Opportunities in the Digital Economy, https://www.markle.org/ microsoft. IBM, for instance, has established Pathways in Technology Early College High Schools (P-TECH Schools). P-TECH schools are innovative public schools that offer students the opportunity to earn a no-cost associates degree within six years in fields such as applied science and engineering--and to acquire the skills and knowledge necessary to pursue further educational opportunities or to step easily into well paying, high-potential informational technology jobs. IBM designed the P-TECH model to be both widely replicable and sustainable as part of an effort to reform career and technical education. See IBM, IBM and P-TECH, https://www-03.ibm.com/press/us/en/presskit/42300.wss. Likewise, Salesforce offers free high-tech and business skills training through Trailhead, its online learning platform, with the goal of preparing them for the estimated 3.3 million jobs created by the Salesforce economy worldwide from 2016 to 2022, nearly 1 million of which are forecasted to be in the United States. See International Data Corporation, The Salesforce Economy Forecast: 3.3 Million New Jobs and $859 Billion New Business Revenue to Be Created from 2016 to 2022 (Oct. 2017), available at http://www.salesforce.com/assets/pdf/misc/idc- study-salesforce-economy.pdf; see also Gavin Mee, How the Salesforce Economy is Driving Growth and Creating Jobs, Oct. 24, 2017, available at https://www.salesforce.com/uk/blog/2017/10/idc-how-the-salesforce- economy-is-driving-growth-and-creating-jo; Gavin Mee, Guest Blog: Gavin Mee, Salesforce--Evolving tech means change in digital skills, TechUK (Apr. 26, 2017), at https://www.techuk .org/insights/opinions/item/10695-guest-blog-gavin-mee-salesforce- evolving-tech-means-change -in-digital-skills. \21\ For example, the Splunk4Good initiative, which partners with non-profits, is helping military veterans and their families, along with youth, train for careers in technology, providing free access to Splunk licenses and its extensive education resources to help them attain marketable skillsets. See Splunk, Splunk Trains Workforce of Tomorrow With Amazon Web Services, NPower, Wounded Warrior Project and Year Up, (Sept. 26, 2017) https://www.splunk.com/en_us/newsroom/press- releases/2017/splunk-trains-workforce-of-tomorrow-with-amazon-web- services-npower-wounded-warrior-project-and-year-up.html. --------------------------------------------------------------------------- IV. Opportunities for Congress and the Administration to Facilitate AI Innovation As innovation in AI and related software services increasingly fuels growth in the global economy, countries around the world are taking steps to invest in education, research, and technological development to become a hub for AI innovation. For example, the UK government recently released an Industrial Strategy, which identifies putting the UK at the forefront of the AI and data revolution as one of four key strategies that will secure its economic future.\22\ In the EU, the European Parliament recently issued a report on civil law rules regarding robotics, which highlights the opportunities robotics and AI offer and encourages investment in such technology so Europe can maintain leadership in this space.\23\ Likewise, in Japan, the government recently issued a new strategy designed to strengthen collaboration between industry, the government, and academia on matters related to robotics, and also issued a report offering the first systematic review of AI networking issues in Japan.\24\ In China, the government has issued a ``Next Generation Artificial Intelligence Development Plan,'' which lays out objectives for AI development in China for the next 13 years and calls on China to become a global AI innovation center by 2030.\25\ --------------------------------------------------------------------------- \22\ See UK Secretary of State for Business, Energy and Industrial Strategy, Industrial Strategy Building a Britain fit for the future (Nov. 2017), available at https://www.gov.uk/government/uploads/system/ uploads/attachment_data/file/662541/industrial-strategy-white-paper- print-version.pdf. \23\ See European Parliament 2014-2019, Resolution of 16 February 2017 with recommendations to the Commission on Civil Law Rules on Robotics, Eur. Parl. Doc. P8_TA (2017)0051, http:// www.europarl.europa.eu/sides/getDoc.do?pubRef=-//EP//NONSGML+TA+P8-TA- 2017-0051+0 +DOC+PDF+V0//EN. \24\ See Fumio Shimpo, Japan's Role in Establishing Standards for Artificial Intelligence Development, Carnegie Endowment for International Peace (Jan 12, 2017), http://carnegie endowment.org/2017/01/12/japan-s-role-in-establishing-standards-for- artificial-intelligence-development-pub-68311. \25\ See Elsa Kania, China's Artificial Intelligence Revolution, The Diplomat (Jul. 27, 2017), available at https://thediplomat.com/ 2017/07/chinas-artificial-intelligence-revolution/. --------------------------------------------------------------------------- In the United States, a flexible policy framework that facilitates responsible AI deployment and increased investment will be key to preserving U.S. global economic competitiveness. An essential part of that effort will be ensuring the ability to access data, and to transfer that data seamlessly across borders, which are vital for AI to flourish. It also will be important to support investment in AI-related education, workforce development, and research. To that end, there are several steps that Congress and the Administration could take to spur AI innovation and continued economic growth. A. Pass OPEN Government Data Act First, Congress should pass the OPEN Government Data Act. This legislation, which the House recently passed as Title II of the Foundations for Evidence-Based Policymaking Act, recognizes that government-generated data is a national resource that can serve as a powerful engine for creating new jobs and a catalyst for economic growth. To that end, the OPEN Government Data Act would require agencies to make non-sensitive government data more open, available, and usable for the general public. Making such data more readily available will improve government transparency, promote government efficiency, and foster innovation of data-driven technologies such as artificial intelligence. We would like to thank Ranking Member Schatz for his tireless work as an original sponsor of the OPEN Government Data Act. We are hopeful that the Senate will act soon to secure its final passage into law. B. Support Efforts to Promote Digital Trade and Facilitate Data Flows We also urge Congress and the Administration to continue supporting efforts to expand digital trade. Indeed, the new digital data economy, which increasingly relies on AI and related software services, will benefit from a globally recognized system for digital trade that facilitates cross-border data flows and establishes clear rules, rights, and protections. There are several opportunities for Congress and the Administration to lead in this area. First, the ongoing NAFTA discussions provide an important opportunity to modernize the trade agreement, which was initially negotiated when digital services were in their infancy. We are encouraged that the Administration has made it an objective to seek to prohibit market access barriers to digital trade, including restrictions on data transfers, data localization mandates, and technology transfer requirements. Second, another key priority is ensuring that transatlantic trade continues to thrive. In particular, we appreciate Congress's and the Administration's leadership on issues relating to the EU-U.S. Privacy Shield, which both protects privacy and facilitates data transfers between the EU and United States. We encourage your continued support as the Administration proceeds with its ongoing successful implementation of the framework. Third, as other countries seek to modernize their trade policies, the Administration should engage key global partners to ensure that new trade initiatives facilitate data-driven innovation and protect against market access barriers for e-commerce and digital trade. C. Invest in AI research, education, and workforce development Unlocking the full promise of AI technologies also requires a long- term strategy of investing in education, workforce development, and research. Because human beings ultimately drive the success of AI, supporting education, training, and research is essential to extracting the maximum level of benefit that AI technologies offer. As an initial matter, Congress and the Administration should ensure that education programs are developing human talent more effectively. Broadly speaking, this means that Congress and the Administration should support science, technology, engineering, and mathematics (STEM) education at all levels. It also means creating and supporting programs that help educate researchers and engineers with expertise in AI, as well as specialists who apply AI methods for specific applications and users who operate those applications in specific settings.\26\ For researchers and engineers, these programs should include training in computer science, statistics, mathematical logic, and information theory, and for specialists, they should focus on software engineering and related applications.\27\ --------------------------------------------------------------------------- \26\ See U.S. Executive Office of the President, Preparing for the Future of Artificial Intelligence, National Science and Technology Council Committee on Technology 26 (Oct. 2016), available at https:// obamawhitehouse.archives.gov/sites/default/files/whitehouse_files/ microsites/ostp/NSTC/preparing_for_the_future_of_ai.pdf. \27\ See id. --------------------------------------------------------------------------- Congress and the Administration should also support the development of new and innovative ways to ensure the U.S. workforce is prepared for the jobs of the future. Because AI will generate new jobs in categories both known and unforeseen, we need to develop thoughtful and effective approaches to equip the U.S. workforce with the skills necessary to seize the opportunities these new technologies create and to optimize the role of AI in modern life. Continued scientific research is essential to fully tapping the potential of AI technology. Congress and the Administration should therefore also promote both public and private sector research to help ensure that the United States remains a leader in this space. The U.S. Government should invest in the types of ``long-term, high-risk research initiatives'' in which the commercial sector may be reluctant to invest. In the past, such R&D investments have led to ``revolutionary technological advances. . .[such as] the Internet, GPS, smartphone speech recognition, heart monitors, solar panels, advanced batteries, cancer therapies, and much, much more.'' \28\ Congress and the Administration should also adopt policies that incentivize private- sector R&D, including by expanding access to financing. --------------------------------------------------------------------------- \28\ See The National Artificial Intelligence Research and Development Strategic Plan (Oct. 2016), available at https:// www.nitrd.gov/PUBS/national_ai_rd_strategic_plan.pdf. --------------------------------------------------------------------------- Passing the OPEN Government Data Act, supporting efforts to promote digital trade and facilitate cross-border data flows, and investing in AI research, education, and workforce development will be critical to maximizing the opportunities AI presents and helping to ensure that the United States maintains leadership in AI innovation and deployment, even as other nations increase their own efforts to take advantage of the possibilities that AI offers. * * * We appreciate Congress's leadership on the important issue of facilitating AI innovation and its responsible deployment. Thank you and I look forward to your questions. Senator Wicker. Thank you very much. Dr. Gil. STATEMENT OF DR. DARIO GIL, Ph.D., VICE PRESIDENT, AI AND IBM Q Dr. Gil. Chairman Wicker, Ranking Member Schatz, members of the Subcommittee, thank you for inviting me here today. My name is Dario Gil, and I am the Vice President of AI and quantum computing at IBM. The idea of creating a thinking machine is not new, and precedes modern computing. Calculating machines were built in antiquity and improved throughout history by many mathematicians. The term ``artificial intelligence'' was first introduced 61 years ago in 1956, and AI, as an academic discipline, took off. Three years later, IBM scientist Arthur Samuel coined the term ``machine learning'' to refer to computer algorithms that learn from and make predictions on data by building a model from sample inputs without following a set of static instructions. One type of machine learning and AI algorithm that has gained tremendous attention over the past several years is an artificial neural network, notably, deep learning. These networks are inspired by the architecture of the human brain, with neurons organized as layers, and different layers may perform different kinds of operations on their inputs. When presented with sample data, neural net can be trained to perform a specific task, such as recognizing speech or images. Over the last decade, the explosion of digital data and the growth in processing speed and power have made it possible to use neural nets in real-world solutions. While many tend to focus on the automation features of AI, we believe its true impact will be felt in humans carrying out complex tasks which they cannot do on their own. My prepared testimony provides detailed examples on the many ways in which IBM's AI platform for enterprise business, Watson, is being used to augment human abilities across many industries, from strengthening cybersecurity to enhancing the customer experience to improving agriculture and optimizing supply chains. AI is playing a bigger and bigger role in all realms of commerce, and its uses will only grow. Now, there is no question the advent of AI will impact the nature of jobs, yet history suggests that even in the face of technological transformation, employment continues to growth with economic expansion and the creation of entirely new jobs despite the disappearance of some occupations. Jobs are made out of tasks. Those tasks that cannot be automated, that cannot be automated by AI, are those in which workers will provide the greatest value, commanding higher wages and incomes as a result. Now, the creation of AI itself will require new job categories associated with how we design and train them, how we secure them, and verify that they work as planned, and how we integrate them into our workflows. The application of AI will change our professions, opening up new categories of work and increasing demand for some existing professions as we combine the capabilities of these AI systems with our own expertise. For example, many more cybersecurity professionals will be needed to engage with AI systems and act decisively upon the threat to information they provide. We must address the problem of shortage of workers with the skills needed for these and many other roles. A useful example in this regard is software programming, which is taught as a critical skill in many high schools and colleges. We should promote a similar movement for AI techniques, such as machine learning. To enjoy the full benefits of AI, we'll also need to have confidence in the recommendations, judgments, and uses of AI systems. In some cases, users will need to justify why an AI system produced its recommendations. For example, doctors and clinicians using AI systems to support medical decisionmaking may require to provide specific explanations for a diagnosis or course of treatment, both for regulatory and liability reasons. IBM is actively innovating in this field. We're deriving best practices for how, when, and where AI algorithms should be used. We're creating new AI algorithms that can be trusted, are explainable, and are more accurate. Of course, no single company can guarantee the safe and responsible use of such a pervasive technology. For this reason, IBM is a founding member of the Partnership on AI, a collaboration with other key industry leaders and many scientific and nonprofit organizations. It's focused on the formulation of best practices, on AI technologies, and advancing the public's understanding of AI. In addition, a recently created MIT-IBM Watson AI Lab has one of its research pillars advancing shared prosperity with AI, exploring how AI can deliver economic and societal benefits to a broader range of people, nations, and enterprises. In a similar way, we look forward to working closely with the Members of Congress to ensure the responsible, ethical, and equitable use of AI as this technology continues to evolve. [The prepared statement of Dr. Gil follows:] Prepared Statement of Dario Gil, Vice President, AI and IBM Q Introduction Chairman Wicker, Ranking Member Schatz, members of the Subcommittee. Thank you for inviting me here today. My name is Dario Gil and I am Vice President, AI and quantum computing at IBM. We have arrived at a remarkable moment in the history of information technology. An explosion of data and computation has given us access to massive amounts of digitized knowledge. With it, we have enormous intelligence and power to see patterns and solve problems we never could have before. Increasingly, the engine we use to tap into this knowledge is artificial intelligence. We can now train algorithms with emerging Artificial Intelligence (AI) technologies to learn directly from data by example. Moreover, we can do this at scale cost through cloud networks to create machines that help humans think. For these reasons, AI is the most important technology in the world today. The rise of AI has been accompanied by both boundless enthusiasm about its ability to transform our lives, and fears it could potentially harm or displace us. At IBM, we take a different approach. We are guided by the use of artificial intelligence to augment human intelligence. We focus on building practical AI applications that assist people with well-defined tasks. We believe people working collaboratively with these learning systems is the future of expertise. In my testimony, I'll provide an overview of AI and describe how work in this field has evolved. Then I'll offer some examples of the rapidly growing commercial applications of IBM Watson, the best-known artificial intelligence platform for enterprise business today. I'll also look at how we're beginning to combine AI with other emerging technologies, such as blockchain, to optimize business and provide trust in transactions. I'll describe how AI will impact the nature of work leading to many new and improved job opportunities. Finally, I'll examine IBM's position on the responsible and ethical use of AI. The Evolution of AI The idea of creating a `thinking' machine is not new and precedes modern computing. The study of formal reasoning dates to ancient philosophers such as Aristotle and Euclid. Calculating machines were built in antiquity and were improved throughout history by many mathematicians. In the 17th century Leibniz, Hobbes and Descartes explored the possibility that all rational thought could be made as systematic as algebra or geometry. In 1950, Alan Turing, in his seminal paper Computing Machinery and Intelligence, laid out several criteria to assess whether a machine could be deemed intelligent. They have since become known as the ``Turing test.'' The term ``artificial intelligence'' was first introduced in 1956, sixty-one years ago, and AI as an academic discipline took off. Three years later, in 1959, IBM scientist Arthur Samuel coined the term ``machine learning'' to refer to computer algorithms that learn from and make predictions on data by building a model from sample inputs, without following a set of static instructions. An algorithm is simply a set of rules to be followed in calculations or other problem-solving operations. It can be as basic as the steps involved in solving an addition problem or as complex as instructing a computer how to perform a specific task. One type of machine learning and AI algorithm that has gained tremendous attention over the past several years is an artificial neural network. It has been essential to the explosive growth of AI systems today. Artificial neural networks are inspired by the architecture of the human brain. They contain many interconnected processing units, called artificial neurons, which are analogous to biological neurons in the brain. Typically, neurons are organized in layers. Different layers may perform different kinds of operations on their inputs. When presented with sample data, an artificial neural network can be trained to perform a specific task, such as recognize speech or images. For example, an algorithm can learn to identify images that contain cars by analyzing numerous images that have been manually labeled as ``car'' or ``no car.'' It can then use those results to identify cars in images that it has not seen before. Even though neural networks and other machine learning algorithms were actively researched more than six decades ago, their practical use was hindered by the lack of digitized data from which to learn from and insufficient computational power. At the time, most data were in analog form and not easily available to a computer. Training of the neural network algorithm was and remains a computationally intensive process. Due to the limitations of processors, it could not be implemented effectively. Over the last decade, the explosion of digital data, the growth in processing speed and power, and the availability of specialized processing devices such as graphical processing units (GPUs) have made it possible to use artificial neural networks in real-world solutions. Today, computation is carried out not only in the cloud and in data centers, but also at the edge of the network, in sensors, wearable devices, smart phones, embedded electronics, factory machines, home devices, or components in a vehicle. These conditions have also allowed researchers and engineers to create incredibly complex neural networks, called deep learning networks. They perform in ways comparable to humans in many tasks. For certain tasks, such as speech and image recognition, game playing, and medical image classification, these networks can outperform people. Today, neural networks are used in a variety of applications, including computer vision, speech recognition, machine translation, social network analysis, playing board and video games, home assistants, conversational devices and chatbots, medical diagnostics, self-driving cars, and operating robots. In addition to machine learning, AI systems deploy a variety of other algorithms and technologies that include knowledge representation, machine reasoning, planning and scheduling, machine perception (speech and vision), and natural language processing and understanding. At IBM, we are actively researching and advancing these and other technologies so that we can continue to enhance AI systems. We are also envisioning and developing the next-generation infrastructure required for increasingly complex AI tasks and workloads. This is the physical hardware required to run AI algorithms: the processors, servers, databases, storage, data centers and cloud infrastructure. When all these pieces are aligned in a way that allows algorithms to analyze data with maximum efficiency, we refer to it as the ``full stack.'' By successfully engineering the full stack, we can build AI-powered solutions that we can apply to a broad array of societal and industry challenges. While many tend to focus on the benefit of automation, we believe that AI's true impact will be felt in assisting people's daily lives, and by helping us carry out extremely complex tasks we cannot do on our own. That includes everything from forecasting the weather, to predicting how traffic will flow, to understanding where crops will grow the best. AI will also help us research the best combinations of compounds for drug development, repurpose chemical structures for new medicines, and optimize vastly intricate supply chains. I'd like to illustrate this further with a look at how IBM Watson is already being used across a range of different industries. AI applications to industries My first example illustrates how AI can assist humans in reacting to a problem when there is very little time to react. The security of data and on-line transactions is fundamental to the growth of commerce. But the simple fact is that most organizations can't keep up with the threats. Security experts are too few and overstretched. Sophisticated attacks, including those using AI tools, are coming at a rate that makes them extremely difficult to stop. Entire networks are compromised in the blink of an eye. Watson for cybersecurity allows us to turn the tables. It sifts through the insights contained within vast amounts of unstructured data, whether it's documented software vulnerabilities or the more than 70,000 security research papers and blogs published each year. It instantly alerts security experts to relevant information, scaling and magnifying human cognition. It also learns from each interaction that has an alert, and works proactively to stop continued intrusion. Security analysts, armed with this collective knowledge, can respond to threats with greater confidence and speed. The second example shows how AI is enhancing customer experience. Tax preparation is an area ripe for AI solutions. H&R Block is using Watson to understand context, interpret intent and draw connections between clients' statements and relevant areas of their tax return. Watson is working alongside H&R Block Tax Pros as they take clients through the tax return process, suggesting key areas where they may qualify for deductions and credits. Clients can follow along and understand how their taxes are being computed and impacted by numerous aspects of the tax code. They can also see the many paths to file a return with the IRS, pinpointing the route that delivers a maximum refund. A third example demonstrates AI's ability to personalize the client experience. 1-800-Flowers launched an AI-powered gift concierge powered by Watson Conversation. It interacts with online customers using natural language. The service can interpret questions, then ask qualifying questions about the occasion, sentiment and who the gift is for to ensure that suggestions are appropriate and tailored to each customer. In this way, the customer can get the right flower for the right occasion. The next example highlights AI's role in enhancing agricultural productivity. A program led by our Research division called IBM PAIRS is bringing the power of AI to improve crop yields. It works by processing and analyzing vast amounts of geospatial data to generate vastly improved long term weather and irrigation forecasts. Using these methods, IBM Research and Gallo Winery co-developed a precision irrigation method and prototype system that allows Gallo to use 20 percent less water for each pound of grapes it produces. A final example shows AI's ability to optimize supply chains. Traditional brick and mortar retailers are under tremendous pressure from e-commerce. They must find new, cost-effective and efficient ways to deliver goods to buyers in order to stay in business. That means offering customers a range of delivery options--pick up in store, ship from the nearest store, or move goods seamlessly between store and e- commerce. Our client, a major American retailer, had to coordinate this effort across a thousand stores in their fulfillment chain. Our predictive models enabled them to determine optimal distribution across their entire chain, factoring in dozens of different variables. Over an eight-day period including Black Friday and Cyber Monday, they processed more than 4 million orders--a company record--at a savings of 19 percent per order compared to the prior year. This led to an overall savings of $7.5 million dollars. From cybersecurity, to customer experience, to personalization, to productivity and optimization, AI is playing a bigger and bigger role in all realms of commerce. And its uses will only grow. AI and blockchain The potential for AI becomes even greater when combined with other emerging technologies such as blockchain. Blockchain stores data in shared, secure, distributed ledgers that allow every participant appropriate access to the entire history of a transaction using a ``permissioned'' network--one that is highly secure and can determine who can see what. Blockchain holds the promise to be the way we may do all transactions in the future. A typical AI process could include data collection, use of the data to train a neural network and then deployment of the pre-trained model to power the application. Blockchain supports the AI process by reducing the risk of data tampering and provides data in a form that can be used and audited. There's an old saying in the computer industry ``garbage in, garbage out,'' and that applies to data and how you use it. The integrity of the data used as input to the AI model is a necessary criterion in ensuring the value and usefulness of the model. Because it can process and analyze massive quantities of data, AI can use blockchain data to gain valuable insights and detect patterns in how supply chains work and processes behave. Over time, this will generate a valuable source of clean, trusted transactional data that cuts across industries to give us new insights. That includes both structured and unstructured data--everything from Internet of Things (IoT) information to compliance and geospatial data that's stored on a blockchain. AI can use this information to generate valuable insights and detect patterns in near-real time, driving new efficiencies across business operations. For example, IBM Research is working with Everledger, a company that tracks and protects diamonds and other valuables. We're using AI to analyze digital information on one million diamonds Everledger has stored on a blockchain. We can cross-check that data against UN regulations to prevent the sale of conflict diamonds. We can verify time and date stamps. We can certify laser inscriptions in the girdle of the stone. We can perform these analytics directly on the blockchain, without the need to extract the data first. This minimizes opportunities for data tampering and fraud. While this is a specialized application, it shows some of the kinds of data we can collect and analyze at huge scale. We have also partnered with Walmart to use blockchain and AI techniques to ensure food safety. Today's food supply chains are highly complex and involve multiple components, stakeholders, and activities. This complexity makes it difficult to identify sources of contamination, counterfeit substitutions, loss of refrigeration, or food transportation safety issues as products move from their sources to their consumption by consumers. Blockchain supports traceability by tracking the food products from origin to destination and by allowing certification of respective transactions and events along the way. AI- powered technologies are used to analyze this information to help ensure food that can be eaten safely. AI and the Future of Work Artificial intelligence will alter the way people work. This has been true of many new technologies that have benefited human populations over time because they dramatically improved industrial output. They have led to fewer grueling jobs. In the process, new types of jobs have emerged. However, such disruptive improvements have always called for a period of training and adjustment. We need to openly understand and recognize this fact, so that we can create the right conditions to make this transition as successful as possible. As a nation, we need to be prepared to offer the appropriate education and support to manage this change well. There's no question the advent of artificial intelligence will impact jobs. Despite the fear, anxiety, and prediction of massive job loss, history suggests that, even in the face of technological transformation, employment continues to grow and very few occupations disappear. Rather, it is the transformation of occupations that is very likely to be widespread that will impact most workers. Occupations are made up of tasks. It is the tasks that are automated and reorganized where the transformation occurs. Workers will need new skills for the new transformed tasks and occupations. But, it is the tasks that cannot or will not be automated where workers provide the greatest value, commanding higher wages and incomes as a result. Some ``new collar jobs'' will emerge--jobs that require advanced technical skills but do not necessarily require a full undergraduate education. A study by Accenture of more than 1,000 large companies that are already using or testing AI and machine-learning systems identified the emergence of entire categories of new, uniquely human jobs with no precedents. For example, ``trainers'' will be required to teach AI systems how they should perform. They may write scripts for chatbots, helping them to improve their understanding of human communication, or help provide labeled data needed to train the algorithm. They may teach AI systems how to personalize their recommendations, or show compassion when responding to certain problems that require it. Another category of ``explainers'' will be needed to help convey how AI systems have arrived at a decision or a recommendation. They'll monitor the operations of AI systems or perform forensic analyses on algorithms and make corrections to them if they generate an incorrect result. Earlier, I referenced the shortage of qualified cybersecurity professionals. In the future, we'll need far more of them to engage with AI systems, review the recommendations these systems offer and act decisively upon threats. There are actions we must take now to ensure the workforce is prepared to embrace the era of AI and the ways it will augment our economy. To begin, we must address the shortage of workers with the skills needed to make advances in AI, create new solutions and work in partnership with AI systems. We need to match skills education and training with the actual skills that will be required in the emerging age of AI. At IBM, we have an educational model called P-TECH to train new collar workers for a job in technology. P-TECH combines the best of high-school, community college, hands-on-skills training and professional mentoring, and provides public high school students in grades 9-14 a path to post-graduation opportunities in fields aligned with the skills American employers are looking for. Our goal must be to create multiple pathways like this for more people to acquire the skills that will be in demand, as AI use becomes more commonplace. We can use the example of the adoption of software programming as a critical skill that is taught in many high school and colleges. Some colleges require that all students learn how to code since they consider it a necessary skill for success. Students becoming proficient in programming have a wider range of job opportunities. In the future, we may promote and see a similar trend with students gaining understanding of and proficiency in AI techniques such as machine learning. Preparing more U.S. students and workers for success in these well-paying new collar jobs is essential if we want a workforce that is ready to capitalize fully on AI's economic promise. Let me also say that as well-intentioned as it may seem to some, taxing automation will not serve the cause of fostering employment in the new AI economy. It will only penalize technological progress. We should not adopt measures like this one that will harm America's competitiveness. Inevitably, people adapt best by finding higher value in new skills. Technologies that are easiest to integrate and integrate with will be those that improve human productivity. But they should not replace human judgment. IBM Watson was designed from the beginning to work in concert with human expertise. It will only be successful as long as there are people with the right skills to engage with it. Building trust in AI To enjoy the full benefits of AI, we will also need to have confidence in the recommendations, judgments and uses of AI systems. IBM is deeply committed to the responsible and ethical development of AI. Last year, we published one of the first corporate white papers on this subject. The paper, which was intended to help launch a global conversation, centered around the need for safe, ethical, and socially beneficial management of AI systems. Trust in automated systems is not a new concept. We drive cars trusting the brakes will work when the pedal is pressed. We perform laser eye surgery trusting the system to make the right decisions. We have automated systems fly our airplanes trusting they will navigate the air correctly. In these cases, trust comes from confidence that the system will not make a mistake, leveraging system training, exhaustive testing, and experience. We will require similar levels of trust for AI systems, applying these methodologies. In some cases, users of AI systems will need to justify why an AI system produced its recommendation. For example, doctors and clinicians using AI systems to support medical decision-making may be required to provide specific explanations for a diagnosis or course of treatment, both for regulatory and liability reasons. Thus, in these cases, the system will need to provide the reasoning and motivations behind the recommendation, in line with existing regulatory requirements specific to that industry. In the European Union, this will be a requirement for all automated decision-making AI systems as of May 2018. These safeguards can also help to manage the potential for bias in the decision-making process, another important concern with AI. Bias can be introduced both in the datasets that are used to train an AI system and by the algorithms that process that data, and how people interpret and communicate the discerned insights. Our belief is that the data and algorithmic aspects can not only be managed, but also that AI systems themselves can help eliminate many of the biases that already exist in human decision-making models today. At the beginning of this year, IBM issued principles for transparency and trust to guide our development and use of AI systems. In summary, they state the following: We believe AI's purpose is to augment human intelligence We will be transparent about when and where AI is being applied, and about the data and training that went into its recommendations. We believe our clients' data and insights are theirs. We are committed to helping students, workers, and citizens acquire the skills to engage safely, securely, and effectively with cognitive systems, and to do the new kinds of work that will emerge in an AI economy. In the same way that we are committed to the responsible use of AI systems, we are committed to the responsible stewardship of the data they collect. We also believe that government data policies should be fair and equitable and prioritize openness. IBM is actively innovating in this field. We are deriving best practices for how, when, and where AI algorithms should be used. We are creating new AI algorithms that are more explainable and more accurate. We are working on the algorithmic underpinnings of bias and AI, such as creating technologies that can identify and cleanse illegal biases from training datasets. Of course, no single company can guarantee the safe and responsible use of such a pervasive technology. For this reason, IBM is a founding member of the Partnership on AI, a collaboration with other key industry leaders and many scientific and nonprofit organizations. Its goal is to share best practices on AI technologies, advance the public's understanding, and serve as an open platform for discussion and engagement about AI and its influences on people and society. AI has enormous transformative power. Much has been said about its potential to transform sectors and industries. However, AI is also giving us a technological toolkit to address many societal challenges. At IBM we are committed to pioneering new solutions, and showcasing and promoting the opportunities to use AI in social good applications. Three years ago, we launched the AI for Social Good program and have executed a number of AI for Good projects, from using AI to understand patterns of opioid addiction, to prototyping recommendation systems that would aid low-income individuals and help them stay out of poverty, to applying machine learning to understand transmission mechanisms of the Zika virus. Earlier this year, we announced the MIT-IBM Watson AI Lab, a partnership with Massachusetts Institute of Technology (MIT) to carry out fundamental AI research. One of the research areas for the lab is focused on advancing shared prosperity through AI--exploring how AI can deliver economic and societal benefits to a broader range of people, nations and enterprises. Lastly, no discussion of the future of AI would be complete without acknowledging the critical role of government. Public investment and policy support have been the twin pillars of American global technological leadership for the past half-century. We hope and expect the same will be true in the coming age of AI. For this reason, we enthusiastically welcome the interest and support of the United States Senate as this technology continues to evolve. Together, we can ensure that AI serves people at every level of society and advances the common good. Senator Wicker. Thank you very much. Dr. Felten. STATEMENT OF DR. EDWARD W. FELTEN, Ph.D., ROBERT E. KAHN PROFESSOR OF COMPUTER SCIENCE AND PUBLIC AFFAIRS, PRINCETON UNIVERSITY Dr. Felten. Chairman Wicker, Ranking Member Schatz, and members of the Subcommittee, thank you for the opportunity to testify today. Progress in AI has accelerated over the last decade. Machines have met and surpassed human performance on many cognitive tasks, and some longstanding grand challenge problems in AI have been conquered. Recent experience during this time teach us some useful lessons for thinking about AI as a developing technology. First, AI is not a single thing; it's different solutions for different tasks. Success has come in ``narrow AI,'' which applies a toolbox of specific technical approaches to craft solutions for specific applications. There has been a lot less progress on general AI, which tries to create a single, all- purpose artificial brain, like we see in the movies. Second, successful AI doesn't think like a human. If it is an intelligence, it is sort of an alien intelligence. AI and people have different strengths and weaknesses, so teaming up with AI is promising if we can figure out how to work with an intelligence different from our own. And, third, more engineering effort or more data translates into better AI performance. Progress requires a lot of hard work by experts, and that's why our AI workforce is so important. The strategic importance of AI to the United States goes beyond its economic impact to include cybersecurity, intelligence analysis, and military affairs as well. The U.S. is currently the world leader in AI research and applications, but our lead is not insurmountable. Countries around the world are investing heavily in AI, so our industry researchers and workforce need support in their efforts to maintain American leadership in this area. American companies recognized the potential of AI early on and have been investing and moving aggressively to hire top talent. Our lead in research and development is less secure. Federal funding for AI research has been relatively flat. Aggressive hiring by industry has thinned the ranks of the academics who train the next generation of researchers. Industry does a lot of valuable research, but the public research community also plays an important role in basic research and in training young researchers, so investments in policies to support and grow the public research community are important. Policies to enhance access to high-quality education for all American children, especially in computing, lay the foundation for our future workforce. And America has always been a magnet for talent from around the world, and that has to continue if we are to retain our leadership. The many benefits of AI are tempered by some challenges. AI systems may pose safety risks, they may introduce inadvertent bias into decisions, and they may have unforeseen consequences. Much of the criticism of AI has centered on the risk of inadvertent bias, and real-world examples of biased AI are well documented. The good news is that there are technical ways to eliminate bias. Developers can improve datasets to be more representative of the population, they can use algorithms that are more resistant to bias. Promising results on debiasing both data and algorithms are emerging from the research community, and that research should continue to be supported because it points a way to deploying AI more widely with less concern about bias. In considering the risks of AI, it's important to remember that the alternative to relying on AI is to rely on people, and people are also at risk of error and bias. In the long run, AI systems will devise complex data-driven strategies to pursue goals, but people will continue to decide which goals the system should pursue. To better hold AI systems accountable, we need new technologies and new practices to connect AI with the human institutions that will govern it. Regarding regulation, there is no need to create special regulations just for AI at this time. In sectors that are already regulated, the existing regulations are already protecting the public, and regulators need only consider whether and how to adjust the existing regulations to account for changes in practices due to AI. For example, the Department of Transportation, in the previous administration and this one, has been adapting vehicle safety regulation to enable safe deployment of self-driving cars. Government agencies have important roles to play beyond regulation. More expertise, advice, and coordination is needed across the government to help agencies decide how to adapt regulations and use AI in their operations. New structures and new policies to strengthen this expertise would be very beneficial. With good policy choices and the continued hard work and investment of American companies, researchers, and workers, AI can improve the health and welfare of Americans, boost productivity and economic growth, and make us more secure. Americans currently lead the world in AI. We should not step on the brakes; instead, we should reach for the accelerator and the steering wheel. Thank you for the opportunity to testify today. [The prepared statement of Dr. Felten follows:] Prepared Statement of Edward W. Felten Robert E. Kahn Professor of Computer Science and Public Affairs, Princeton University Chairman Wicker, Ranking Member Schatz, and members of the Committee, thank you for inviting me to speak today about how best to realize the benefits of artificial intelligence. Artificial Intelligence (AI) and Machine Learning (ML) Artificial intelligence (AI) and machine learning (ML) have been studied since at least 1950. There has been an unexpected acceleration in technical progress over the last decade, due to three mutually reinforcing factors: the availability of big data sets, which are analyzed by more powerful algorithms, enabled by faster computers. In recent years, machines have met and surpassed human performance on many cognitive tasks, and some longstanding grand challenge problems in AI have been conquered. Industry has recognized the rise of AI as a technical shift as important as the arrival of the Internet or mobile computing. Companies around the world have invested heavily in AI research and development, and leaders of major companies have described adoption of machine learning as a bet-the-company opportunity. The strategic importance of AI/ML to the United States goes beyond its economic impact. These technologies will also profoundly affect the future of security issues such as cybersecurity, intelligence analysis, and military affairs. Fortunately, the United States is currently the world leader in AI/ ML research, development, and applications, in both the corporate and academic spheres. Our national lead is not insurmountable, however. Countries around the world are investing heavily in AI/ML, so our scientists, engineers, and companies need support in their efforts to maintain American leadership. The Nature of AI/ML Today The history of AI teaches some important lessons that are useful in considering policy choices. AI is not a single thing--it is different solutions for different tasks. The greatest progress has been in ``narrow AI,'' which applies a toolbox of specific technical approaches to craft a solution specific to one application or a narrow range of applications. There has been less progress on ``general AI,'' which strives to create a single, all- purpose artificial brain that could address any cognitive challenge and would be as adaptive and flexible as human intelligence. Indeed, there is no clear technical path for achieving general AI, so it appears that for at least the next decade the policy focus should be on the implications of narrow AI. In a world of narrow AI, there will not be a single moment at which machines surpass human intelligence. Instead, machines may surpass human performance at different times for different cognitive tasks; and humans might retain an advantage on some cognitive tasks for a long time. Even if machines surpass humans in the lab for some task, additional time and effort would need to be invested to translate that advance into practical deployment in the economy. Successful AI does not think like a human--if it is an intelligence, it is an alien intelligence. Because AI solutions are task-specific and do not directly mimic the human brain, AI systems tend to ``think'' differently than people. Even when successful, AI systems tend to exhibit a different problem-solving style than humans do. An AI system might handle some extremely complex situations well while failing on cases that seem easy to us. The profound difference between human thinking and AI operation could make human-AI teaming valuable, if the strengths of people and machines can complement each other. At the same time, these differences create challenges in human- AI teaming because the teammates can have trouble understanding each other and predicting their teammates' behavior. On many cognitive tasks, more engineering effort or more data translates into better AI performance. Many AI systems learn from data. Such systems can be improved by re-engineering them to learn more from the available data or by increasing the amount of data available for training. Either way, devoting more effort to engineering and operating an AI system can improve its performance. Machines are generally worse than humans at learning from experience, but a machine with a very large data set has much more ``experience'' from which to learn. Using the narrow AI approaches that have been successful so far, expert AI developers must invest significant effort in applying AI to each specific task. Benefits of AI/ML AI is already creating huge benefits, and its potential will only grow as the technology advances further. For example, AI is a key enabler of precision medicine. AI systems can learn from data about a great many patients, their treatments, and outcomes to enable better choices about how to personalize treatment for the particular needs, history, and genetic makeup of each future patient. AI is also enabling self-driving cars, which will eventually be much safer than human drivers, saving thousands of American lives every year. Self-driving vehicles will improve mobility for elderly and disabled people who cannot drive and will lower the cost and increase the convenience of transporting people and goods. Given the tremendous benefits of AI in these and other areas and the likelihood that the technology will be developed elsewhere even if the United States does not lead in AI, it would be counterproductive to try to stop or substantially slow the development and use of AI. We should not ask the industry and researchers to slam on the brakes. Instead, we should ask them to use the steering wheel to guide the direction of AI development in ways that protect safety, fairness, and accountability. Policies to Support AI Progress America's leadership in AI has been driven by three factors: our companies, our researchers, and our talented workforce. American companies recognized the potential of AI early on and have been investing heavily in AI and moving aggressively to hire top talent. This is the area in which our national leadership in AI seems safest, at least in the short run. In the longer run, however, industry must be able to work with world-leading American researchers and workforce to sustain its advantage. Our lead in research and development is less secure. Federal funding for AI research and development has been relatively flat, even as the importance of the field has dramatically increased. Aggressive hiring by industry has thinned the ranks of the academic researchers and teachers who are needed to train the next generation of leaders. Although industry has carried out and supported a great deal of research, it cannot and does not cover the full spectrum. The public research community plays an important role in basic research, in research areas such as safety and accountability, and in training young researchers, so investments and policies to support and grow that community are a key enabler of continued American leadership. The foundations of the future workforce are laid in our K-12 schools. Policies to enhance access to high-quality education for all American children, especially in computing, can grow the number of students who enter higher education eager and able to pursue studies in technical fields such as AI. The American AI workforce has also been boosted immeasurably over the years by the attractiveness of our universities and industry to the most talented people from around the world. America has been a magnet for talent in AI and other technical fields, and that must continue if we are to retain our leadership. Policies to ensure that America remains an attractive place for foreign-born experts to live, study, work, and start companies are among the most important steps for the future health of our AI enterprise. Risks and Challenges of AI/ML The benefits of AI are tempered by some risks and challenges: AI systems may pose safety risks; they may introduce inadvertent bias into decisions; and they may suffer from the kinds of unforeseen consequences brought on by any novel, complex technology. These are very serious issues that require attention from policymakers, AI developers, and researchers. Much of the criticism of AI/ML systems centers on the risk that adoption of AI/ML will lead inadvertently to biased decisions. There are several ways this could happen. If a system is trained to mimic past human decisions, and those decisions were biased, the system is likely to replicate that bias. If the data used to train a system is derived from one group of people more than another, the result may serve the overrepresented group to the detriment of the underrepresented group. Even with ideal data, statistical artifacts can advantage larger groups to the detriment of smaller ones. Real-world examples of these sorts of biases are well-documented. The solution is not to stop pursuing AI, but rather to take steps to prevent and mitigate bias. Practitioners should work to improve their data, to ensure that datasets are representative of the population and do not rely on past biased decisions. They should also improve their algorithms by developing and using AI systems that are more resistant to bias, so that even if flaws remain in the data, the system can produce results that are more fair. In both areas, data improvement and algorithm improvement, the research community is producing promising early results that will improve the anti-bias toolkit available to practitioners. A robust national AI research effort should include studies of algorithmic bias and how to mitigate it. In considering the risks of bias and accountability in AI, it is important to remember that in most cases the alternative to relying on AI is to rely on human decisions, which are themselves at risk of error, bias, and lack of accountability. In the long run, we will likely rely much more on algorithms to guide decisions, while retaining the human role of determining which goals and criteria should guide each decision. Accountability, Transparency, and Explainability The importance of the decisions now made or assisted by AI/ML systems requires that the systems and their operators are accountable to managers, overseers, regulators, and the public. Yet accountability has proven difficult at times due to the complexity of AI systems and current limitations in the theory underlying AI. Improving practical accountability should be an important goal for the AI community. Transparency is one approach to improve accountability. Disclosing details of a system's code and data can enable outside analysts to study the system and evaluate its behavior and how well the system meets the goals and criteria it is supposed to achieve. Full transparency is often not possible, however. For example, a system's code might include valuable trade secrets that justify withholding aspects of its design, or the data might contain private information about customers or employees that cannot be disclosed. Even where transparency is possible, it is far from perfect as an accountability mechanism. Outside analysts may have limited practical ability to understand or test a system that is highly complex and meant to operate at very large scale. Indeed, even the designers of a system may struggle to understand the nuances of its operation. Computer science theory says that examining a system beforehand cannot hope to reveal everything the system will do when it is exposed to real-world inputs. So transparency, though useful, is far from a complete solution to the accountability problem. Another approach to accountability is inspired by the field of safety engineering. The approach is to state clearly which safety, fairness, or compliance properties a system is designed to provide, as well as the operating conditions under which the system is designed to provide those properties. This is backed up with detailed evidence that the system will have the claimed properties, based on a combination of design reviews, laboratory testing, automated analysis tools, and safety monitoring facilities in place during operation. Rather than revealing everything about how the system works, this approach focuses on specific safety, fairness, or compliance requirements and allows system developers to use the full range of technical tools that exist for ensuring reliable behavior, including the tools that the system developers will already be using internally for quality control. Much needs to be done to make this approach feasible for routine use. Research can develop and test different approaches to proving behavioral properties of systems. Professionals can convene to develop and pilot best practices and standards. The overarching challenge is to understand how to relate the technical process of engineering for reliable operation to the administrative processes of management, oversight, and compliance. Regulation and the Role of Government Agencies There is no need to create special regulations for AI. Where AI is used in sectors or activities that are already regulated, the existing regulations are already protecting the public and regulators need only consider whether and how to adjust the existing regulations to account for changes in practices due to AI. For example, the Department of Transportation (DOT) and National Highway Traffic Safety Administration (NHTSA) have taken useful steps, under the previous and current Administrations, to clarify how existing safety regulations apply to self-driving vehicles and how Federal safety regulations relate to state vehicle laws. These changes will serve to smooth the adoption of self-driving vehicles which, once they are mature and widely adopted, will save many thousands of lives. Similarly, the Federal Aviation Administration (FAA) has been striving to adapt aviation regulations to enable safe, commercial use of unmanned aerial systems (UAS, or ``drones''), which have benefits in many sectors, such as agriculture. The FAA has taken some steps to increase the flexibility to use UAS commercially, but the interagency process on UAS has been moving slowly. Agencies should be urged to work with the FAA to advance this important process. Government agencies have important roles to play beyond regulation. For example, the National Institute of Standards and Technology (NIST) and the Department of Commerce can contribute by setting technical standards, codifying best practices in consultation with the private sector, and convening multi-stakeholder discussions, much as they have done in the area of cybersecurity. All agencies should consider how they might use AI to better accomplish their missions and serve the American people. AI can reduce costs, increase efficiency, and help agencies better target their use of taxpayer dollars and other limited resources. The National Science and Technology Council's subcommittee on Machine Learning and AI can serve as a focal point for interagency coordination and sharing of ideas and best practices. With good policy choices and the continued hard work and investment of American companies, researchers, and workers, AI can improve the health and welfare of Americans, boost productivity and economic growth, and make us more secure. Americans currently lead the world in AI. We should not step on the brakes. Instead, we should reach for the accelerator and the steering wheel. Thank you for the opportunity to testify. I look forward to answering any questions. Senator Wicker. Great. Thank you so much. Let me start with Dr. Gil. Machine learning and artificial intelligence capabilities are accelerating, and I think Americans listening today or maybe insomniacs listening at 2 in the morning 2 weeks from now on C-SPAN, know that we use AI for social media and online search queries and smartphone apps. Ms. Espinel mentioned health diagnosis, and I think she grabbed our attention there. What other industries, Dr. Gil, stand to benefit the most from what we're talking about today? Dr. Gil. I believe actually artificial intelligence is going to touch every profession and every industry, but just to give some concrete examples. Senator Wicker. OK, good. Dr. Gil. From security and cybersecurity, there's a class of category of problems that have to do with responding with low latency. So the nature of the problem that one has to address is too complex, there are too many dimensions of it, and AI can assist a professional detect a threat and be able to assess the proper response. So sometimes it has to do with how much time one has to make a decision, and can you be assisted? In other areas in which, for example, in the health care profession, even though we may have more time in some occasions to perform a diagnosis or select a treatment, just the sheer complexity of the number of documents, or in this case, it could be genomic information that comes into play, goes beyond the expertise that any given person can have. So in that instance, it can assist, for example, in the process of, you know, medical diagnosis, as an example. In agriculture, here we're talking about being able to integrate sensory measurements from soil and weather data, satellite data, to be able to predict what kind of irrigation one may have to deploy to improve the productivity of it, to give another example. So I think that, you know, in every situation where we're trying to integrate knowledge that comes from sometimes measurements from the physical world and also bodies of evidence that we've accumulated through our expertise, combined with our own expertise, we facilitate every worker and professional to be able to make those better decisions. Senator Wicker. OK. Now, Ms. Espinel and to all members of the panel, Ms. Espinel says that our government needs to talk about three changes to policy: open data, more government research, and prioritizing education and workforce development. Also, Dr. Felten says there's no need to create special regulations for AI. Who wants to comment on this? Does anybody want to take issue that we have or is everybody in agreement with all four of these statements? Is there some room for nuances and disagreements? Does anybody want to respond? Yes, Mr. Castro. Mr. Castro. Thanks. So I think one of the most important things to look at in this space is it's about technology adoption. When we're looking at AI, the big question for the United States in terms of competitiveness is, are we going to be the lead in terms of adopting this technology rapidly in our industries before other countries do it in their industries? Because that's going to determine U.S. competitiveness long term. So when we're talking about policies in this space, there are two types. There are the policies that help accelerate deployment and adoption of the technology and our R&D in this space, and then there are the regulations that might slow down adoption or, you know, kind of skew or realign how we do this adoption. So when we're comparing ourselves to Europe, for example, which is also pursuing this, we have to ask the question: Why aren't we doing what Europe is doing to accelerate adoption? And, two, are we having smart regulations that allow us to apply it in our industries kind of better, smarter, and faster than they're doing? Senator Wicker. Do you sort of agree with Dr. Felten, that regulations may actually impede our development of AI? Mr. Castro. I think in most cases, regulation that's focused on AI specifically is probably misguided. If there's a problem there, we need to look broader and say, Why is this problem happening? And is it caused in human situations as well? Senator Wicker. Dr. Felten, have I mischaracterized anything you've said? Dr. Felten. No, you have not. I would agree with Ms. Espinel's points. With regard to regulation, in addition to sectoral regulation, there's an important role for agencies sometimes to create new regulatory structures to allow more activity, as the FAA has been working to do with drones. The FAA has been working to create new rules which allow broader commercial use of drones in the United States. And so although that is a change to regulation, it's one that enables more activity by the commercial sector. Senator Wicker. Thank you very much. Senator Schatz. Senator Schatz. Thank you, Mr. Chairman. Thank you for all of your testimony. The way I see this is there's a competitiveness side of the ledger, which I think is not easy to do, but relatively straightforward. And Ms. Espinel's testimony spells out some of the steps that we can take on a bipartisan basis to make sure that we win that race internationally. That part is, again, not easy, but relatively straightforward morally and as a matter of policy. Where I think it does get difficult is that I think I quibble with you, Mr. Castro, in the sense that I don't think we're--that the endgame here is just that we race as fast as we can in all sectors without regard to consequences and view any regulatory effort as contradictory to our national goals. I think part of what we have to do is recognize that in areas like health care and agriculture, it's a pretty much unalloyed good to have more data, to save lives, to make agriculture more productive, for instance. In defense, it's a little tricky. In criminal justice and policing, it's extremely tricky. And so I don't think anybody in the Senate is talking about European style regulation. I think what we are saying is that this is a complex area that's going to revolutionize society and the way we interact with each other, the way machines interact with each other, and with ourselves, and if we're not careful, we could enshrine some of our thorniest societal problems in datasets. And I'm not as persuaded, Dr. Felten, as maybe you are, that we can program our way out of that. And one of the challenges that I would just maybe ask, if we can start with Dr. Felten and go down the line, I'm worried about diversity in the industry. I think that to the extent that you have software engineers and decisionmakers both at the line level writing the code, but all the way up to project management and the people who are wrestling with some of these moral questions, are mostly white men, and I think that's not a trivial thing because they're not thinking about biases in policing. They may be thinking differently about autonomous weapons. And so I'm wondering how you view--and I don't think this is a place for regulation, but I do think this is a place for us, as a society, to grapple with if this is going to be transformational and change everything, is it fair, is it rational, to have only, or I should say predominantly, white men in charge of setting up these algorithms that most of the rest of society can't even access because it's all proprietary? Dr. Felten. Dr. Felten. Sure. With respect to the question of whether-- of the role of technology in addressing these issues of bias, I think technology has an important role to play, but it can't be the entire solution, as you suggested, Senator. What we need is a combination of institutional oversight and governance with technology. Technology can provide the levers, but we still need institutions that are able to pull those levers to make sure that the results are conducive to what our institutions and our society wants. With respect to the question about diversity in the workforce, this is certainly an issue. The AI workforce is even less diverse than the tech workforce generally. And it's important to take efforts to improve that so we can put our whole team on the field as a nation. And I commend groups like AI for All that are working on that to you. Senator Schatz. Dr. Gil and then Ms. Espinel, we have a minute and 20 seconds left. Dr. Gil. OK. Perhaps I could address the topic of bias associated with these models because bias can be introduced at the level of the dataset, as you properly pointed out, if the data that has been collected, you know, is not representative of the whole population, in this case, it's to make the right assessments. You can introduce then a bias in the---- Senator Schatz. Well, just to be clear, it can be empirically valid, right? I mean, the simplest example is here's where crimes have been committed in the past, right? It turns out Ferguson, right? Load that into the dataset, it's a predictive algorithm, works every time. You go over there, you find more and more crime, and it spins and spins. And then on top of it, enshrining that bias in an algorithm, I'm not sure that--it gets more permanent than it would be if it were up to the individual judgment of sheriffs, VAs, people. Dr. Gil. Yes. So this is a very active field of research in which we're very active that actually has also to do with in the cases in which you may have high degrees of prediction, but you're incorporating protected variables or protected individuals in the case of the assessment that you cannot use because of law, that is a variable that you cannot incorporate. So there are ways actually to perform the data science to be able to achieve, you know, a prediction. I'm not---- Senator Schatz. But if a police department like deals with a vendor, and they say, ``We've got a predictive algorithm, and we can't show you how this predictive algorithm works,'' but in back of that, you know, inside of that black box, they've got census block stuff, they've got all kinds of stuff that you would be not allowed to use in policing, how do we even know? Ms. Espinel, how worried should I be about this? Ms. Espinel. So we talk about bias, but I would actually--I would love to address that as well because it's an issue I'm really passionate about and focused on, and I know we're running low on time here, but you also have sort of led into explainability and accountability, and that's really important, too. So I will try to briefly touch on both and I'm happy to continue this conversation. Senator Schatz. Yes. Ms. Espinel. So in terms of bias, I think there are really two parts of it. So part of it is how the AI systems are trained, how they're built, in essence. And, obviously, as you point out, data can be inaccurate or it can be incomplete, or it can have its own biases that are going to skew the outcomes. And there are a number of things that can be done right now to try to help with that. So part is making sure that data scientists that are building them have the tools and the training to try to counter that. Second, and you already raised this, I think this is another area, or you have another reason why diversity in tech is so important, and I think the more experience and background you have at the table as AI systems are being trained, the more helpful it's going to be to try to avoid that. I think, third, as has been mentioned, there is a lot of research going on in this area, so continuing to support and invest in research that will help lessen the chances of bias in AI is very important. And last I would say, you know, to the extent bias is discovered, obviously companies should be working immediately to try to address that. So I think there are a number of things that can and should be happening now to try to counter that. I think there's another part of this discussion, which we've heard less about, which I think is really important, which is how AI can be used, not trained and built, but how it can be used to try to counter bias and to try to broaden inclusion. And there are a number of really interesting examples here both in terms of hiring practices and in terms of broadening inclusion for people with diseases like--or people with conditions like autism or people that are visually impaired where AI can dramatically transform their ability to interact with society and in workplaces. And so I think having more discussion about how AI can and should be used to try to lessen bias and to try to broaden inclusion is very important. I could talk more. I know we're running low on time. Senator Schatz. I think my time is up. I'll let the other members ask their questions. Senator Wicker. Senator Schatz's time is close to expiring. [Laughter.] Senator Wicker. But we'll take another round. Senator Moran. STATEMENT OF HON. JERRY MORAN, U.S. SENATOR FROM KANSAS Senator Moran. I have four questions, and I'm glad to know that the standard has now increased beyond the 5 minutes. [Laughter.] Senator Moran. Let me quickly try to ask these four questions. First about research and development in the Federal role. A number of us on this Committee are members of the Appropriations Committee. We would think of opportunities to be supportive of this endeavor by funding of NSF, NIH, DoD. What am I missing? Is there something out there that we ought to be paying attention to from an appropriations process that supports Federal research in this regard to AI? Don't pause very long here. Dr. Gil. No, well, in addition to the agencies you listed, I think the DOE also has an important element of it in the intersection of high-performance computing and artificial intelligence and how the computational platform to support historical approaches to be able to do things like modeling of chemical processes and sophisticated approaches to do that to combine it with the more statistical approaches that are being enabled now with artificial intelligence and machine learning. So I think being able to combine those two disciplines in the context of the DOE would be very helpful as well. Senator Moran. Thank you. The significance--the difference between private research, business research, and government research, where do we see the focus? Let me say, where are the most dollars being spent? Is the private sector more engaged than the Federal Government? Dr. Gil. I would say at this point it would be fair to say that in the private sector, certainly in the technology world, AI is the single most important technology in the world today. So--and the levels of investment that we're all making around that is commensurate with that statement. Senator Moran. Yes. Dr. Felten. The private sector investment in AI research and development is much larger than the Federal investment currently. I would agree with the list of agencies that you listed, Senator, and that Dr. Gil listed. And I'd also commend to you the ``National AI Research and Development Strategic Plan'' that was published last October that was put together by the research agencies. Senator Moran. Thank you very much. As we attempt to promote STEM education, in that broad phrase of ``STEM,'' is that sufficient to describe the kind of intellectual and academic excellence that we need in order to develop AI? Is it something more than just promoting science, mathematics, engineering, and research, the traditional kinds of STEM things, as we pursue support of education? Ms. Espinel. Ms. Espinel. So that is very important, and I think trying to ensure that every child in every state, if they want to go into tech, they have the skills to do that, and that's a real viable, realistic career opportunity for them. I think that's critical. I think there are other things that could be very helpful as well. So one of the things that we've been thinking about is trying to modernize vocational training. So for, you know, not just for--important for very young children learning as well, but as young adults are coming out of school, then thinking about where their career path could take them, I think there's a lot that could be done to try to improve those programs as well. Senator Moran. If anyone has suggestions in addition to the short time-frame that I have, that Senator Schatz didn't have, please let me know. We'd like to figure out how we focus our educational support in a way that adds this new dimension or additional dimension to what kids in grade school, middle school, are learning. They're the future. Ms. Espinel. That's fantastic. Thank you. Senator Moran. As a Kansan, I need to ask a question about agriculture. This could be Dr. Gil, Ms. Espinel, Dr. Castro, or, well, really any of you, Dr. Bethel, where is the research taking--who are the leaders in research when it comes to agriculture? Is it the universities or is it the private sector again who is focused on large data and what it can mean to increase productivity and efficiency and better on-the-farm income? Who should I be talking to that's fully engaged in this world? Dr. Gil. Yes. I think there is wonderful work going on across a number of universities, and we can--I can give you some more details of some specific programs that are, you know, very well tailored to this. But certainly in the private sector there is a lot of activity that has had to do with focusing in the instrumentation aspect and the measurement of fields, particularly with satellite data, being able to combine also very unique datasets. As we've aggregated datasets in terms of I alluded before in terms of like soil characteristics, you know, evapotranspiration models that we have, weather data, and be able to combine all these layers of data together to be able to have accurate forecasts and prediction, and to the degree that we have more autonomy also in the agricultural fields to be able to, you know, irrigate with more precision--right?--or be able to use fertilizers with more precision as well, the combination of all of those factors is what is increasing productivity and what it's enabled to do that. Senator Moran. Dr. Felten, my final question. You indicate in your testimony there is no clear technical path for achieving general AI, you have narrow and general. Is what you're telling me is that that's more science fiction, like more of a James Bond movie than where we are today? Dr. Felten. That is the case today. We can't rule out the possibility that general AI may come along far in the future, but from a policymaking standpoint, narrow AI is what we have, and it, I think, should be setting the agenda. We should be alert for the possibility that sometime down the road general AI may come, but it's not close. Senator Moran. Thank you all very much. Thank you, Mr. Chairman. Ms. Espinel. Senator Moran, if I may, you mentioned advances in farming technology, and since you are from Kansas, I just wanted to let you know that we put out a study earlier this year looking at the impact of software across the United States, and Kansas is one of the states we saw the biggest jump in jobs. So over 30 percent growth in software jobs in Kansas, and you're up to nearly 40,000, and part of that is farming technology, but it is other types of software services as well. So Kansas is doing great. Senator Moran. I wouldn't want to forget the aviation world that we live in, too, in Kansas. Thank you very much. Senator Wicker. Senator Peters. Senator Peters. Thank you, Mr. Chairman. Senator Wicker. How's Michigan doing? Senator Peters. Yes, how's Michigan doing? Ms. Espinel. Michigan is doing really, really well. Senator Wicker. But not as good as Kansas, I'm sure. [Laughter.] Ms. Espinel. Well, Kansas did see a huge jump, but Michigan, $13 billion in GDP from software into Michigan. And Michigan is definitely doing better in jobs overall in terms of numbers, maybe not quite as big a jump year to year, but Michigan is a really--is a really strong state, and our---- Senator Wicker. And I guess New Mexico really isn't even in the game. [Laughter.] Ms. Espinel. We actually had an event in Detroit last week through our foundation talking about software and tech, and very focused on the educational system in Michigan and the great things that Michigan is doing to try to advance software and technology in the state, so thank you. Senator Wicker. Senator Peters. STATEMENT OF HON. GARY PETERS, U.S. SENATOR FROM MICHIGAN Senator Peters. Thank you, Chairman Wicker, for bringing up that question so I didn't have to use my time for that. That was very well done. [Laughter.] Senator Peters. And you're right, Michigan is moving very aggressively in this area, and it's primarily driven by self- driving cars and what's happening in that space, which is very exciting, something that I've been intimately involved in over the last few years and months. And we now have some significant legislation moving forward for that. In fact, it's been described to me by folks in the AI space that having self- driving vehicles may be the Moonshot for artificial intelligence, that when AI can pilot a car through a complex city environment with all sorts of things happening all around it, that means AI has developed to the point where it's going to be transformative in every single industry. It's going to be a key barometer of where we are going forward. So we are pleased that that's happening in Michigan. In fact, we had General Mattis, who mentioned the four places for technology in the country, and Michigan was one of those four. So a little different vision than a lot of people here in Washington may have for my great state, so I appreciate the opportunity for that to come up. The question I have--and I'm a believer in all of the wonderful things that you're talking about. I believe this is the most transformative technology in a long, long time, just as it is in the auto industry. It's probably as big as when the first car came off of the assembly line. We know what happened after that, in creating the American middle class, changing everything about our economy. We think the same thing will happen with AI. And so there's incredible promise for it, but I think we also have to be very open to the potential downside to this. And I know some of you have addressed the employment issue, and I want to just talk about that because my experience has been the folks who are big proponents of the technology downplay the employment aspects. Folks who are scared probably overplay the employment aspects. And the truth is going to be somewhere in the middle. And I think one thing it will have an impact on that employment growth and what we've been seeing in the economy recently is we have further concentration of industry and fewer and fewer companies that have larger shares. That has actually suppressed wage growth, it has created a less dynamic environment when it comes to new business formation. I mean, we can go through the economic arguments associated with that. And so, Mr. Castro, you mentioned that whoever comes up-- whatever companies embrace AI will have a significant technological advantage when they do that. We see that in the auto industry. It's why the auto industry is racing to be first, or at least very early, knowing that there is probably going to be fewer car companies once AI is fully implemented as well. So my question to you and other panelists, what sort of implications will AI have for the concentration of business in those companies or those industries, or I should say those individual companies within those industries, that have the resources to be able to utilize this? And will it be more difficult for small businesses? Mr. Castro. Yes, so I think there are two effects. I mean, if you talk to a company like Amazon, it's a--you know, they're using AI more than anyone, and they're growing faster than anyone. And so in some cases, we're going to see--especially when we're talking about global competitiveness, U.S. companies growing because of the technology, and that that growth will outpace the jobs offset. Of course, that won't happen everywhere, and in many cases, what we want to see is more productivity, which means fewer workers in a given space per output. In those cases, and what we've seen historically in this phase, is that the new jobs are not necessarily AI jobs. It's not that, you know, everyone is going to now be, you know, building self-driving cars or designing them, it's that we see more people in other professions, more doctors, you know, classrooms with higher teacher-to-student ratios. The kinds of changes we often say we want to see and can't pay for right now, we can get in the future. We did a really interesting study earlier this year that was looking at occupational change over the last 165 years, and it's actually at the lowest change, so the least disruption, that we've seen in this entire time period. And the reason for that is we get a lot of misperceptions when we see ATMs on the corner, and you think there are fewer of these jobs, when, in reality, you have more banks, you have more prosperity. And so the job losses are usually more visible than the job creation, which is why we have the skewed perception. Senator Peters. But I think we go beyond job losses and actually look at wage differentials and income inequality, and that's probably what I was alluding to, is when you have increased concentration, you have less dynamism in the economy, which I think is consistent with what you just said about the job churn has gone down, it's becoming less dynamic. Many economists believe that's a big reason why we have growing inequality in this country as well. Certainly folks who are able to get these software jobs are going to do extremely well, and God bless them for doing that, but based on trends that we already see of stagnant wages, even though there are increases in productivity, doesn't necessarily translate into everyday wages for everyday folks, that that all could accelerate at a very quick pace that we should be thinking about, and I think it's important for us to be conscious of that impact, it's not just jobs, it's income inequality. Now, Ms. Espinel, if you want to comment on that, please do. Ms. Espinel. I was just going to say briefly that, one, I think you're right, it's something we need to be thinking about. You said, or maybe you were alluding to what Mr. Castro said, in terms of businesses using AI. I guess I would say, you know, we don't think that big businesses or concentrated businesses are the ones that should be using AI. I think our hope is that AI and the technology behind it will be democratized sufficiently so that small businesses will have the ability to use it as well and help them in whatever their business objectives are. So I guess I would say, yes, I would agree. It would be a concern if we saw AI being used primarily by just some large companies, but I personally don't think that will be the future of its deployment, and I know, certainly speaking for our members, they would like for small businesses, for the public sector, for any type of organization that is trying to make decisions and trying themselves to make decisions in a way that is better informed to be able to have the use of the benefits of AI. Senator Peters. Well, I would agree that that's the goal, and we hope we could democratize it, but it hasn't necessarily played out that way. It does take significant amount of capital. And as Mr. Castro mentioned, Amazon is using AI to the greatest extent of any other company, it's growing the fastest, and that's why we have brick-and-mortar retailers that are going out of business all over the country as well. It's great for Amazon, great for AI, great for productivity for Amazon, but it may not be so great for the mom-and-pop store that's there. And I would argue that mom-and-pop store thinks they can't have an AI system like Amazon, that's just simply not realistic for them. And so we need to be thinking. I don't have the answers, but I think we need to be thinking about that or we're going to be facing some significant societal challenges in the future. But thank you for your comments. Senator Wicker. Senator Udall. STATEMENT OF HON. TOM UDALL, U.S. SENATOR FROM NEW MEXICO Senator Udall. Thank you, Mr. Chairman. And thank you to all the witnesses here today. I think this has been very excellent testimony. And obviously there are some very positive aspects to AI. But I wanted to ask about the bots and the software used to hurt consumers. The New York Times recently reported on so-called ``Grinch bots,'' software used to predict the web links of the in-demand toys and other merchandise and to purchase the goods before the general public has access to these items. This practice has caused many of this holiday season's most popular items, such as Fingerlings, Super Nintendo Classic Edition, Barbie Hello Dreamhouse, to be sold out online within seconds. However, one can go to eBay and easily find these same products available for increased costs, sometimes dramatically increased. Have either of your organizations worked with retailers to prevent software bots from taking advantage of Internet deals to jack up the prices of goods? Can you identify other ways to prevent these software bots from using machine learning to become more and more sophisticated? Ms. Espinel. I'm happy to take this as the parent of two young boys that are busy filling out their Christmas lists. So--and I actually think this is an example of an area where AI can be really helpful. So Dr. Gil and others have talked about the use of AI and cybersecurity because AI is really good at looking at large amounts of data, detecting patterns, and then predicting where threats might lie. The Grinch bots I think is an area--AI is also great at fraud detection, and there are a lot of the characteristics of these Grinch bots that are similar. So part of it is, you know, if you see--if you see unusual patterns like huge, large, amounts of purchases that are happening very, very quickly, and these Grinch bots work so they can process transactions in just a few seconds, that is a pattern, and an unusual pattern. And AI is really good at looking at patterns like this, and then alerting the retailers so they can shut down those purchases in much the same way that AI is being used now for credit card fraud detection and can help detect unusual patterns and then, you know, give your bank or you, as the credit card user, the ability to say, ``I didn't approve those, so you need to block my card.'' That--this type of activity I think is actually a great example of how AI could be deployed to try to detect those unusual patterns and then tell the retailers to not process those transactions and shut it down. Senator Udall. Yes, we hope that that happens more because you can see the consumer damage, I think every day, and in a lot of the business coverage. Dr. Bethel, in your testimony, you spoke of the need to access more and more data, including social media and mobile device location history to help refine the uses of artificial data. I'm concerned about the privacy implications that this kind of sharing could have. Could you speak to the ways to obtain relevant data while still protecting consumers' sensitive information? And do others on the panel share my concerns on privacy? Please, Dr. Bethel. Dr. Bethel. I mean, privacy, anytime you're dealing with data, is going to be an issue of concern. And I think we have to be responsible in how we use that data and how we obtain the data. So there does probably need to be some consideration given to how that data is obtained, what that information is used for, and trying to ensure the privacy of the individuals that that information is coming from. It's critical to the success, I think, of AI, and being responsible with that. So there may need to be some regulation in that to protect the general public from harm basically is where regulation I feel is needed. But we need to be aware of how that is happening and try to put in place measures to ensure the safety of that data. Senator Udall. Great. Thank you. Please. Mr. Castro. Thank you. On the question of privacy, there is actually some really interesting where we're seeing, you know, if you look at kind of trends, increased use of AI can actually increase privacy for individuals because it takes personal data out of the hands of humans, which is what generally people are concerned about. There are general privacy concerns when you actually start talking to people, is that, ``I don't want you personally to see my medical record. I don't want this person over here to see my financial data or use it in a harmful way.'' But when they say, you know, ``Are you OK with a computer seeing it?'' suddenly everything is OK. And you see this in a lot of sensitive areas. This is one reason why people like to shop online for certain personal items, they don't want to see a clerk at checkout, but they're absolutely fine with Amazon having that information. So in many cases, what we want to see is how we can have policies that actually shift it so companies can guarantee the data isn't ever touched by a human, it's only touched by a computer, and that is a privacy guarantee. But then we need to make sure we can actually hold companies accountable if they ever allow that data to escape in certain ways. Senator Udall. Yes. Dr. Felten. AI raises the stakes in the privacy debate because it boosts both the uses of your data that you might like as well as the ones that you might not like. And so the importance of protecting data privacy continues and maybe is higher in a world with AI. Dr. Gil. Yes. And I would just say from the perspective of IBM, we take the strong view that our clients' data is their own, and the insights that derive from the application of AI when we help them to do that are their own, too. So we take a very strong view of not using data for other purposes that are not the intended ones. Senator Udall. Great. Ms. Espinel. I'll just say very briefly, I think it's important to distinguish between different types of data and different types of AI. Our BSA companies are at the forefront of protecting privacy, and much of the AI that we built is not built using personal data. So I think it's important to bear in mind. Senator Udall. Thank you very much. I yield back, Mr. Chairman. Senator Wicker. Senator Young. STATEMENT OF HON. TODD YOUNG, U.S. SENATOR FROM INDIANA Senator Young. I thank our panelists for being here today. I want to build on some previous line of questioning about our labor markets and their preparation or lack thereof, as we start to move into an AI-driven economy. What sort of questions should policymakers be asking right now to ensure that we optimize the skills our workforce has to the extent possible to prepare for this, in many ways, exciting and promising new technology? Dr. Felten. Well, I think--I think there's one set of questions with regard to education to make sure--especially in the K-12 system, that kids are getting some basic education in computer science. Things are more in hand, I think, at the university level; it's more a matter of resources there. The more difficult questions relate to adult works and displaced adult workers and what is available in terms of retraining or apprenticeship programs to help them to make sure that those are available and to make sure that those are backed by good data on the effectiveness of programs is important. Ms. Espinel. So I would say I think there are three things to think about. One is, you know, if you look at the situation today, the Department of Labor BLS is saying that by 2020 there's going to be 1.4 million jobs that need a computer science degree, and yet we only have about 400,000 graduates in the pipeline. So we have a gap in our labor pipeline today that definitely needs to be addressed if the United States is going to stay competitive in this area. So that's one thing we need to focus on. Second is in terms of education. I think we need to be rethinking our educational system for young people to ensure that they have access to those skills and the opportunity to acquire them so that if they decide that they want to go into tech--and not every tech job requires a 4-year college degree, and I think that's also something we need to be better about explaining--that they have a realistic ability to do so. But that is going to require some rethink of the educational system that we have now. And I think we need to be modernizing the vocational training programs we have for young people that are coming out of school. And then the third area, which Dr. Felten referred to, in terms of people that are in the workforce now, I think we need to not just be investing in reskilling and retraining programs, but, again, thinking differently about them. I think we need to do a better job of matching the skills that people have with the employment needs that are out there across the country, which is an area where I think there's a lot of work to be done, but a lot of potential. Senator Young. Yes. With respect to this issue of the labor markets and AI, what sort of assumptions are you hearing or reading about that you think either overstate some of the forces that will be unleashed as AI continues to develop and be adopted, perhaps strike you as a bit alarmist, or understate these forces? Dr. Gil. I think---- Senator Young. Yes, sir. Dr. Gil. Oh, sorry. Dr. Felten was, you know, very helpful in providing this journey from the narrow form of artificial intelligence that we have today to a general form of artificial intelligence that ultimately could perform, you know, arbitrary, you know, tasks and domains. We're far away from that. So very often the discussion gets framed in terms of either the policy implications or the implications to the labor market associated with that future form of artificial general intelligence that frankly is decades away at best. So in that sense, when the conversation is framed in that lens, it does come across as alarmist--right?--you know. When we talk about the more narrow form of artificial intelligence that exists today, it's more of a focus of, what domains can it have an impact? Where are proven examples that we can do better on that? And at best, you will perform some narrow specific tasks that can be complemented with human labor. Senator Young. Mr. Castro, do you have some thoughts? Mr. Castro. Yes. I mean, so the number one study that everyone thinks about in terms of jobs is the study that came out of Oxford through research of Frey and Osborne that said 47 percent of U.S. jobs would be eliminated by I think it was 2025 or 2030. That study was--first of all, it's not peer reviewed, and when you look at the data, they use a very kind of, you know, flawed methodology to come up with this estimate. So they list in the back in their Appendix all the different professions that they say will be automated with AI, and it's things like fashion models and barbers, which they tried to walk a robot down the runway in Japan once, and they haven't done it since. So, I mean, you know, kind of realistically, those numbers are very much inflated, and they're not actually tied to what we're seeing in the market today. So, you know, the main thing is I think when you see these studies, and there have been a number of studies that have used their data, it doesn't reflect reality. Senator Young. I'll just close here and indicate, you know, the reason we're focused on this I think is we see some incredible potential here. There are some studies that indicate AI has the potential to increase the rate of economic growth in the U.S. from 2.6 percent to 4.6 percent by 2035. I mean, that's just amazing. That would benefit all Americans. There are some serious policy things we need to wrestle with. I've been partnering with Senator Cantwell, who has a lot of expertise and professional background in the area of technology, and we've developed legislation that would establish a Federal advisory committee to help us better understand some of these issues. So if you have thoughts moving forward about things that a Federal advisory committee should look at as we consider the policy implications and broader market implications of AI, I'd certainly welcome those, and I suspect Senator Cantwell would as well. So thank you, Mr. Chairman. Senator Wicker. What a nice segue to Senator Cantwell. Senator Young. Yes. Senator Wicker Senator Cantwell is now recognized. STATEMENT OF HON. MARIA CANTWELL, U.S. SENATOR FROM WASHINGTON Senator Cantwell. Thank you, Mr. Chairman. And I do look forward to working with Senator Young on this issue and the various aspects of investigation. I wanted to bring up applications because one of the things that I think we should be thinking about is our role as an actual user. And one of the things I'm most interested in is AI's application for cybersecurity. One of our biggest threats obviously we face now is the threat of cybersecurity in all sorts of ways. I've seen some applications by MIT and I think University of Louisville, several entities that are both finding fault in code, basically doing a better job--why wait to find out some guy forgot to do the Apache patch, forgot to do the patch? You know? I mean, one employee at that company cost everybody a lot of money because he didn't put in a patch. AI could help us find errors in code or actually in some of these areas I think predict cyber attacks. So I don't know who on the panel could speak to that, but to me, one of the applications that I hope that we will look at is the government's use of this as it relates to combating cybersecurity. Dr. Felten. There's a huge opportunity there, and it's rapidly becoming a necessity, as the bad guys are adopting AI and automation in their cyber attacks. Government and other institutions need to be using more AI in defense in order to, as you said, Senator, find vulnerabilities before they're exploited in order to be able to react at machine speed when things start to go wrong, and in order to better understand what are the possible implications of the way that systems are set up so we don't get surprised in the way that institutions nowadays too often are by both that a breach occurs and by how bad the consequences are. Dr. Gil. Yes, there are two dimensions are essential in this topic. One of it actually has to do with securing AI itself. The very models that we create to enable these predictions are actually vulnerable to attack, and that in itself, there are many steps that can be taken to be able to secure those models. In fact, you can even extract data from a model that has been created. And the second I mention is the one you alluded to, which is AI itself has to be an integral component to protect against other AI-powered attacks. So in a way, we are going to have AI against AI in the realm of cybersecurity because, you know, some of the bad guys are already using these techniques to be able to attack our networks with the speed and accuracy and adaptability that without the presence of AI, and Ms. Espinel alluded to this already, it will be impossible to defend against. Senator Cantwell. Did you want to comment on that as well? Ms. Espinel. Just I'll just say briefly that IBM is among the BSA members that are focused on this. We have a number of companies that are very focused on cybersecurity and are using AI to try to detect patterns and then predict threats before they happen. So I agree with you, that I think it's an area where AI is already being used, and there is potential for it to be doing even more. Senator Cantwell. Well, I like the notion just in the fact that because there are things called human error, that you can, you know, use this to look for faults in code and weak points and do--because that's what someone else is doing, right? And so the sooner that you can find that yourselves and detect that and create another security layer, the better. Is there any other government application that you think we should be investigating on? Some people say on like health statistics and things of that nature, but I don't know if you-- anybody else on the panel has---- Mr. Castro. We did look last week into government use of AI, and one of the biggest challenges is that we're just not measuring what we're doing within government, that there are lots of opportunities, but there's not a good place where if you're an agency, not even a CIO, but just a team manager basically, who wants to start using, you know, automated calendaring, you know, how do I actually go out and do that quickly? You know, can I quickly procure this? Is there an approved list of, you know, best practices from other agencies? Is there good information sharing? We're starting to do that through GSA, but we're not there yet. So one of the things that this Committee could hopefully help do is to really push agencies to ask--bring the CIOs in here and ask them, ``What are you doing around AI? Have you identified the top three opportunities and are you pursuing them?'' Just like, you know, many agencies were directed to find the high-value datasets, pick three, and get them out there as open data, we can do something similar around AI, pick the top three opportunity targets and pursue that over the next 12 months. Similarly---- Ms. Espinel. I would agree with that, but I think city planning is another area where AI could be really helpful. So like traffic congestion, you know, one of the things that city planners are trying to do is optimize traffic patterns and the changing of traffic lights to try to improve traffic flow, and that's something that AI is really good at. You know, there are a lot of variables, so traffic congestion can seem like a relatively simple question, but if you think of all the variables and traffic patterns, it's actually quite complicated in taking all that data and then giving city planners recommendations for how to optimize traffic flow is something AI can do really well. Senator Cantwell. Thank you. Thank you, Mr. Chairman. Senator Wicker. Dr. Bethel, you are doing practical applications on high-risk law enforcement situations, autonomous cargo transfer, and animal-assisted therapy. Do you receive Federal research dollars for that? And to what extent? Dr. Bethel. The law enforcement application we have submitted to NSF. We have currently one grant under review. The Therabot was funded under NSF funding. And some other projects that we're doing are also funded through the NSF or Department of Defense. Senator Wicker. OK. Now, Senator Schatz mentioned a real concern with regard to law enforcement. What you all are doing, though, at Mississippi State, is scenarios where there is already a threat, and law enforcement needs information about how to respond, how to get inside the building. Is there a child there? Is there something explosive there? Has anyone raised the concerns that he raised, any of your law enforcement people or victims or defendants, raised concerns about data bias? Dr. Bethel. Law enforcement has not mentioned any data bias because when we are using these algorithms, we are looking at more like objects in the scene---- Senator Wicker. Something is already occurring. Dr. Bethel. Something is occurring. We're not trying to use it to target people, we're using it because the scene--some kind of event has occurred, and law enforcement is responding. And so we are trying to provide them as much information as we can prior to them going in so that hopefully they can make better, more safer, decisions when they enter into the environment. For instance, their protocol changes completely if there's a child inside the home that they're going into. They can't use a flash bang, they can't do things they would normally do. So we can, by sending a robot in to be able to see what's happening prior, they can make decisions, so they're probably going to end up saving lives, both civilian and law enforcement lives, because--and they will have better performance because of it. So bias hasn't so far come into our discussions when we've been looking at law enforcement applications, but we are not using it, I think, in the manner in which Senator Schatz has indicated. Senator Wicker. Now, with regard--first of all, on the law enforcement application, how long has Mississippi State been doing this? Dr. Bethel. Six years. I started training with them in 2011. We train monthly. Senator Wicker. OK. AI, according to your testimony, is only as good as the data the system receives. How has your data improved over the 6 years, and could you elaborate on that? Dr. Bethel. So each training we do, we do video recordings, we do sensor recordings, we do different manners of data collection. The more examples of data that we can get and obtain, the better the system is at classifying information, doing the sensor fusion, and incorporating that information to make more informed decisions. So as time has gone on, we've been able to obtain more and more samples of data to be able to use that for the systems. Senator Wicker. OK. Now, most of the panel, if not all the panel, seems to think we lead the world, it's just a question of whether we're going to continue leading the world. Senator Schatz said he doesn't think with regard to national policy we'll take a European approach. Rather than ask him what he means by that, let me ask the panel, are there mistakes that our international competitors, whether in Europe or Asia or Africa or wherever, are making that we need to avoid? Are there overreaches in terms of regulation that we need to avoid? Everybody is trying to get ahead. The testimony is that China is making an important effort, the U.K., Japan, our neighbor to the north, but in terms of things we can avoid, mistakes that other countries have made, does anybody have a suggestion that we need to be mindful of as we work on Senator Young and Senator Cantwell's and Senator Schatz's legislation? Yes, Mr. Castro. Mr. Castro. Yes, so, I mean, the two big things I think some countries are thinking about is the right to explanation that's across the board or right to opt out of automated decisions. So the problem with those two policies is that it minimizes the ability or limits the ability of companies to deploy AI and many commercial applications. So, for example, if you want to be a company, you know, just like before you had companies suddenly become online lenders, if you want to be a company that's an AI lender, streamline your entire business process through AI, you basically can't do that because anyone who got rejected for a loan application, for example, could ask for a human appeal, and you would have to have human support to do that. And the problem with that kind of framework is it basically limits those types of business models, and it's especially problematic when you talk about global competition because it doesn't limit necessarily a foreign competitor from offering a similar service. Now, in certain industries, of course, you have to be licensed in the United States, so there are limits there, but overall, that would limit these kind of opportunities. Ms. Espinel. Can I just say briefly? I think most other governments are still considering what their AI policy environment is going to look like. So many governments are having this discussion right now. I do think that there are governments that are--seem to be skewing toward considering a more regulatory approach, a more heavy-handed regulatory approach, and that raises concerns. And I think one issue I'd raise in particular I think for us, I think any regulatory approach that would compel the disclosure of an algorithm, that would compel companies to hand over source code or algorithms, to a government agency is one that raises a lot of concerns with us. We don't think it's actually going to be an effective way to address policy outcomes, and it raises a lot of competitive concerns to be handing over algorithms or source codes to governments. Senator Wicker. Dr. Gil, on cyber attacks from hostile forces internationally and our defense against that, do I understand you to--well, no, let me rephrase it. Is it conceivable that artificial intelligence will be empowered by a hostile force to make a decision to go forward with a cyber attack without a human being at that end making the decision to pull the trigger? Dr. Gil. Humans will definitely design AI-powered attacks to attack, you know, infrastructure or, you know, an industrial setting or another nation-state---- Senator Wicker. But whether that attack is made today or Valentine's Day next year, an artificial intelligence---- Dr. Gil. System. Senator Wicker.--system might make that decision. Dr. Gil. That's correct. That is possible---- Senator Wicker. How close is that to reality? Dr. Gil.--because, I mean, think about it in the past for-- you could design it with an explicit program model to be able to carry out such attacks, and we would have to, you know, create defense mechanisms against those kinds of attacks. But because the program was well stipulated, you could imagine somebody defending against the attack to be able to interpret what those rules of attack were. Now, the moment that you're employing a more machine- learning-based approach where the type of attack could morph depending on the environment it's detecting, now being able to detect what is the form of the attack that is taking place requires another pattern detection mechanism. So that's why I was referring to--and Ms. Espinel was talking before--about you need AI to defend against AI-powered attacks because it's the only way to make it really adaptive to an adaptive attack. Senator Wicker. OK. Senator Markey, do you mind if we--if we let the two witnesses respond here? And then I'll be generous with your time. Dr. Felten? Dr. Felten. Thank you. What you're talking about in terms of automated cyber attack is something we see already with computer viruses. A virus is software which autonomously spreads itself from place to place, and then does whatever it's programmed to do to cause harm at the places that it infects. So to bring AI into this would just be to have a more sophisticated, more adaptive form of virus. It's not fundamentally a new thing, it is--it's a path that we are already on, or a path that the bad guys are already on. Senator Wicker. And quickly, Ms. Espinel. Ms. Espinel. I still think it's more akin to making a recommendation than making a decision. Of course, you, the person, can decide that whatever recommendation the system gives you, you're going to have it automatically act on, but you still made that decision up front, and I don't think we should be--as a general matter, I don't think we should be abrogating decisionmaking authority, but I think that what you're talking about really is AI making a recommendation, and then the person who designed it deciding whether or not they're going to accept that recommendation automatically or not. Senator Wicker. We're really talking about what's coming at us. Senator Markey. STATEMENT OF HON. EDWARD MARKEY, U.S. SENATOR FROM MASSACHUSETTS Senator Markey. Thank you, Mr. Chairman, very much. I thank you all for being here. The digital revolution's newest innovations--augmented reality, autonomous vehicles, drones--all of these industries use artificial intelligence, which rely heavily on a free and open Internet. Regrettably, these disrupting technologies may be dealt a major blow in their infancy, and that's because in just 2 days, the FCC will vote on a proposal to eliminate net neutrality. Without enforceable net neutrality rules in place, broadband providers, like Comcast, Verizon, AT&T, and Spectrum, could block or slow down the content of innovators and businesses using AI all in an effort to leverage their gatekeeper role to favor their own content and generate additional profits. And what will replace these robust net neutrality protections? Nothing, absolutely nothing is going to replace those rules. Dr. Felten, how would eliminating net neutrality protections impact the deployment and development of innovative technologies that use AI? Dr. Felten. A lot of AI technologies operate within an institution or within a company's data center, and so those technologies would not be much affected by changes in network policy. To the extent that changes in network policy affect the ability of companies to deliver their products to consumers, that would obviously be a policy concern, but in my mind, it is important but also somewhat separate from the issue of development of AI. Senator Markey. So you don't think that the additional cost to the developer, to the innovator, won't have an inhibiting impact upon their ability to go to the capital markets, raise the dough, in order to produce their innovation knowing that there's no guarantee that they can reach the 320 million Americans without paid prioritization or without a threat of throttling or blocking? Dr. Felten. I would say that all pro-innovation policies and all policies that are designed to help small companies enter and compete provide value and are pro-innovation and valuable. As I said, I don't think that net neutrality plays a special role with respect to AI. Senator Markey. But does it---- Dr. Felten. There are other areas of innovation. Senator Markey. Will it play a role in your opinion? Dr. Felten. Yes. I think those decisions do play a role in almost any area of innovation. Senator Markey. OK. And on the question of child privacy, the Child Online Privacy Protection Act of 1998 is still the communications constitution for safeguarding children online, and that's a law I was able to get passed back in 1998. As emerging technologies like AI are deployed, it's important that they honor core American values, including privacy. Dr. Felten, could AI technologies pose a threat to children's privacy? And is there a threat that AI technologies could produce inappropriate content for children? Dr. Felten. This is an issue to pay attention to. AI does raise the stakes on policy--I'm sorry--AI raises the stakes on privacy discussions generally, and that's true with respect to children and others. And, of course, parents are very concerned about what their kids see and what happens, and that's one of the reasons why COPA, for example, requires parental consent before certain uses of data are allowed. Senator Markey. So could relying on AI in children's toys negatively impact kids' ability to develop empathy if we substitute real people with computers that cannot fully understand emotion as humans do, Dr. Felten? Dr. Felten. Well, I think kids are more interested in playing with other kids or using toys as a vehicle for playing with other kids. I'm less worried about kids bonding with something that is not human-like or not companionable. I think kids will reject those on their own. Senator Markey. Yes. Earlier this year I wrote to Mattel with serious concerns about their plan to bring the first all- in-one voice-controlled smart baby monitor to the market. Mattel had planned for the device, Aristotle, to use artificial intelligence to help teach children and respond to their needs. After an outcry of questions, Mattel canceled that product. Dr. Felten, what does that experience expose as to potential negatives of using artificial intelligence with children's devices? Dr. Felten. Well, I think stories like this illustrate that it's important to understand the implications of the technologies that are being deployed. As I would imagine, that parents would be happy to, say, be notified if there is indication that their child is in distress, and AI may help to do that more effectively. But these issues of unintended consequences and safety are paramount, and that's one of the important aspects of clearing the road for responsible deployment of AI, is to make sure these issues are taken care of. Senator Markey. Thank you. Senator Wicker. Senator Cruz. STATEMENT OF HON. TED CRUZ, U.S. SENATOR FROM TEXAS Senator Cruz. Thank you, Mr. Chairman. Thank you to each of the witnesses for coming here this morning to testify. A little over a year ago, the Subcommittee on Science and Space, which I chair, held the first congressional hearing on artificial intelligence. And then, as now, we heard testimony about the extraordinary transformative process that we are engaged in right now and how AI in time can be expected to touch virtually every area of human endeavor, and indeed that this transformation may be of comparable import to the transformation we engaged in, in a prior era in the Industrial Age. Anytime we're seeing dramatic transformations in our economy and our workforce and how we interact with each other, that poses the risk of dislocations, but it also poses policy and government and regulatory challenges for how to interact with the new terrain. In your judgment, what are the biggest barriers right now to developing and expanding AI and its positive impacts on our society and our economy? Ms. Espinel. I'll head off another--so I think one of the biggest barriers is a lack of understanding, a lack of understanding about what AI is, what the actual technology is, and then what it does and what the--you know, what both the intended and unintended consequences are. And so I think, you know, this hearing, the legislation that Senator Schatz is working on, I think trying to increase our collective understanding is critical, it's fundamental. I think there are a number of specific policy issues that would be helpful in terms of eliminating barriers. So, you know, one of those is AI is all about data, and so good data policy in various ways I think is very important. I think investing in research, which we've talked about already, investing both in government research and incentivizing private sector research, is very important. And then I think thinking about jobs and workforce development, both the jobs today, but what will happen tomorrow? And rethinking our educational system and our training and reskilling programs are vitally important. So those are the three specific areas, but I think a greater understanding needs to be part of all of those discussions. Mr. Castro. So I agree that skills are very important. Especially much of this technology is new, and we need people that can basically rapidly be credentialed in how to deploy it. But when we look at the biggest opportunities for AI, it's really in some of the regulated industries. And so I think that's where we head up on challenges because there are two challenges: one, regulators aren't necessarily prepared with how to deal with this; and, two, they don't necessarily have the skill set or capabilities internally within the regulatory system to handle it. And so I think that's something we need to be very focused on, is asking questions---- Senator Cruz. What agencies in the industries in particular? Mr. Castro. So financial regulation, for example. Education is another example, especially when we're talking about using AI in primary education, there are a lot of questions about privacy that get raised. And the financial system it's the questions of, you know, do we have regulation basically to understand the technology? And we'll do things more than just ask, ``Can I see the algorithm?'' but be able to say, ``Can I look at outcomes? Can I actually measure outcomes?'' and then ask questions about, ``OK, is this fair? Is this different than what we have now? Is this moving in the right direction?'' and also have some of that regulatory flexibility. So we need kind of, you know, fewer cops and more sandboxes. Senator Cruz. So one area that has generated fears and concern is general AI, and scientists and innovators ranging from Stephen Hawking to Bill Gates to Elon Musk have raised concerns. Stephen Hawking stated, quote, ``Once humans develop artificial intelligence, it would take off on its own and redesign itself at an ever-increasing rate. Humans, who are limited by slow biological evolution, couldn't compete and would be superseded.'' Elon Musk has referred to it as, quote, ``Summoning the demon.'' How concerned should we be about the prospect of general AI? Or to ask the question differently, in a nod to Terminator, when does Skynet go online? [Laughter.] Dr. Felten. Hopefully never. Senator Cruz. That's the right answer. Dr. Felten. I think there's a lot of debate within the technical community about how likely these sorts of scenarios might be. I think virtually everyone agrees that it would be far in the future. And generally the people who are most involved in AI research and development tend to be the most skeptical about the Skynet or existential risk type of scenarios. In any case, the sorts of risks and concerns that exist now about AI are really baby versions of the ones that we would face with a more sophisticated general AI. And so the tactics we--the policies that make sense now to deal with the issues we face now are the same ones we would use to warm up for a general AI future if it comes. And so from a policy choice standpoint, the possibility of distant general AI seems less important to me. Dr. Gil. Yes, I would completely agree. I think if you ask practitioners in the field when they would envision the possibility of that, I think everybody would say 20-plus years out. And whenever scientists says 20-plus years out is our code word for saying we just don't know. Right? We're nowhere near close to be able to do that. So I do think that while it's an area of very important study, and there are many universities who are now creating a rate of monitoring of what is the progress of AI and the implications that it would have if we eventually reach general artificial intelligence, I do think it would be a mistake to guide our policy decisions at present based on the sort of like long-term hypothetical that we don't even have, even as practitioners, even a credible path to get there. Senator Cruz. Thank you. Senator Wicker. Senator Cortez Masto. STATEMENT OF HON. CATHERINE CORTEZ MASTO, U.S. SENATOR FROM NEVADA Senator Cortez Masto. Thank you, Mr. Chairman. And thank you to the panel members. I'm very excited about the conversation today as well. Obviously, we are standing at the edge of a technological revolution that we must ensure will, and there has been discussion, take our labor force with it. So this is a timely conversation. Workforce development has been a focus of mine since I've entered the Senate, and so has innovation. I've proudly been leading a bipartisan legislation on drone expansion, the use of smart technology in transportation, and in trying to spur the next generation of women to be equal at the forefront of STEM, specifically computer programming, through the introduction of the Code Like a Girl Act. I've worked on this because I've seen the future of these developments through my state in Nevada. And just last week I was visited by the leadership at the Truckee Meadows Community College in Reno, who, in partnership with Panasonic, has developed a curriculum to provide individuals to get the specific training that local jobs are hiring for. These are conversations that constantly are going on in my state. So let me start here, Mr. Felten and Ms. Espinel. The skills gap has been discussed extensively today, and something that obviously is on all of our minds and how we address it. Obviously, we have China investing dramatically in the area of AI, so it begs the question of whether there are investments we can be making at the Federal level to help close any potential skills gaps. I'm curious your thoughts on that. Ms. Espinel. So in terms of the skills gap specifically, we also talked a little bit about Federal support for research funding. But I would say a few, and then others may add to them. I think so one is trying to improve access to computer science education at all states and at very early stages of education. So I think that's one area where the Federal Government could be helpful. I think the second is in rethinking how our vocational training programs work. There are vocational training programs in place, but they could be streamlined and they could be better adapted to the world that we live in today. So I think that's an area where there's a lot that could done at the Federal level. And then the third I would say is that I think there is now, and there will maybe be only an increasing sort of information gap between the skills that people have and employers knowing about those, and that seems like an area where there's a real deficiency now, so therefore real opportunity to try to create programs that not only create-- either create pathways directly into employment or do a better job of matching up skills that people have with the jobs that employers have to offer. Senator Cortez Masto. Thank you. Dr. Felten. There are opportunities to widen the pipeline at all stages starting with K-12, making sure that basic computer science education is available to every child at that level, and more advanced education at the high school level for those who are ready for it. There are opportunities to improve--increase the number of teachers and trainee researchers at the university level, and that's education and research funding specifically in areas of AI and computer science at the university level. And then vocational and adult training and apprenticeship programs also are very important to get people on other career paths, give them an on-ramp into this area. Senator Cortez Masto. Go ahead. Ms. Espinel. Can I just say one more thing? Senator Cortez Masto. Yes. Ms. Espinel. There's a lot that BSA companies are doing in this area, IBM among them. And so I think another area is working with the Federal Government both to scale up programs that are effective now and be more collaborative in this area I think is---- Dr. Gil. Just to touch on just one example of that. On the initiative, on a program, we started a number of years called P-TECH, which is a 9 through 14 educational program that combines both education, hands-on program on vocational, you know, schools. And when they graduate through this--right?--we are talking about the creations of new collar professions that have enough skills to be able to practice and benefit from the advances that are happening in AI without having to go through a full college education, and now that's touching, you know, more than 10,000 students. Senator Cortez Masto. Thank you. No, this is something that's happening in Nevada now. It seems like it's such common sense, but we just don't do it, which is that collaboration and partnership between whether it's a government, private sector, and then our education system or our skills vocational trade, to really say, ``What are the jobs of the future? What are they going to look like?'' work with those employers to develop the curriculum that you're going to need for that skilled workforce so that we can really start educating them now. And that starts I think at a very young age and working with our education system, but also the vocation trades that are out there as well. Not every child is going to go on to get a higher degree, but there are some that are absolutely and rightfully are going to go get that vocation or that trade and that skill that's necessary. So I appreciate the comments today. I am running out of time, so I will submit the rest of my questions to this incredible panel for the record. And I appreciate you being here. Thank you. Senator Wicker. Thank you very much. Senator Blumenthal. STATEMENT OF HON. RICHARD BLUMENTHAL, U.S. SENATOR FROM CONNECTICUT Senator Blumenthal. Thanks, Mr. Chairman. I apologize that I missed some of your testimony so far, but I know that this panel has been very important and enlightening, and I thank you for being here. As you well know, the success of autonomous vehicles is closely linked with the success of AI. An autonomous vehicle can only perform as well as the input it receives. A lot of us know the statistic that 94 percent of the 37,000 deaths on the road each year are attributable to human error--97 percent. The hope is that autonomous vehicles will eliminate that human error. What we really lack is information on the extent, to what extent that is caused by human error will be possibly replaced by computer error. And we know computers are not infallible. I wonder, Dr. Bethel, whether you could talk about some of the tasks in which humans still perform better than computers or an AI system in the context of autonomous vehicles, if you're able. Dr. Bethel. In the context of autonomous vehicles, humans are able to adjust to very rapidly changing, unpredictable environments, things happening in the environment. Our sensor systems and our onboard processing in autonomous vehicles is just not to a point where it can make those kind of adjustments that rapidly currently with current technology to be able to adjust to that. So in those cases where you have an erratic driver who is doing unpredictable behaviors, it's really hard sometimes for the system to be able to detect that and react accordingly. In those cases, a human currently is making better decisions on autonomy--over an autonomous vehicle in that kind of environment. Senator Blumenthal. So if I can just extrapolate from your answer, a computer trying to deal with a drunk driver, either ahead or next to that computer, would have trouble because the drunk driver, by definition, is acting in not only unpredictable, but irrational and sometimes actually self- destructive ways. Dr. Bethel. Right. So it would be much more difficult for a computer to predict that kind of behavior than it would be for a human to slow down and react. I mean, it would react, but it probably is not going to be as effective as a human driver is at this point in the stage of where AI is. Senator Blumenthal. Are there ways to program a computer or create software that deals with those unpredictable situations? Dr. Bethel. To some extent, but I think there's a long way to go on that. Senator Blumenthal. In your testimony, you say, and I'm quoting, ``A current limitation to the advancement of artificial intelligence is the quality and cost effectiveness of sensing capabilities to provide high-quality information or data to the system to make those digital decisions. We have a long way to go in the advancement of artificial intelligence-- ,'' ``We have come a long way in the advancement of artificial intelligence,'' excuse me, ``however, we still have a long way to go.'' Sensing, perceiving, interpreting, surroundings are essential to driving a vehicle. Can you describe some of the limitations in the current computer vision and sensing technologies? Dr. Bethel. I'll probably need to get back to you on some of that because it's not exactly my area of expertise related to computer vision, but there are limitations in the sensing capabilities we currently have. Every sensor has its own limitations, so there's no perfect sensor out there. There are also differences in processing power, so trying to be able to handle large amounts of data coming in from these sensors, and processing that in a timely manner can be a challenge. So that's another area. And computer vision has just mega amounts of data coming in that has to be processed, and so another limitation is the actual processing power to handle that, especially in a timely real-time manner onboard a system. Senator Blumenthal. At what point do you think it would be appropriate to completely remove the human from an autonomous vehicle? Dr. Bethel. Thank you for the question. I--depending on the application, I think there are applications currently where autonomy could be--fully autonomous systems are capable. I think it's not realistic anytime in the near future, and especially in autonomous cars, to be able to say that a fully autonomous car is going to be possible all the time. Senator Blumenthal. Thank you. Thank you very much to the panel. Thank you, Mr. Chairman. Senator Wicker. Senator Schatz. Senator Schatz. Dr. Felten, I just wanted to follow up on the question around autonomous weapons. It seems to me that this is an area that is different than a lot of these other ethical, societal, micro, macro, economic questions; this is about how we engage in warfighting. And so to the degree and extent that some of these algorithms are hackable, to the degree and extent that we have a body of history around how we're supposed to engage our military in an ethical manner, that this is an area where the Federal Government has to make policy. And I'm wondering if you can give us any insight into, in the absence of policymaking, who's making these decisions? Is it defense contractors? Is it individual procurement officers? How is all of this getting decided? Dr. Felten. Well, there does need to be a policy which deals with the very serious issues that you mentioned, both from the standpoint of what our military should be willing to do and what safeguards are needed on the systems that they're using consistent with the need for them to be as effective as possible in battle, and also how we deal with adversaries, who might not be as scrupulous in following the international humanitarian law. And this is a national policy issue that ought to be a matter of policy discussion at the highest levels. And if it's done in a decentralized way, if each contractor, each contracting officer, does one thing, if the State Department goes to international arms control discussions and does their own thing, we get uncoordinated policy and we get a result that doesn't serve the American people well. Senator Schatz. Thank you. Senator Wicker. Well, I would like to thank everyone who has participated. And I think Senator Cortez Masto said this has been an incredible panel, and I would have to agree. Ms. Espinel, you brought information about software and the way a number of states have benefited. Did you have information for all 50 states or just for the states represented by this panel? Ms. Espinel. We have it for all 50 states. Our foundation software data put out a study in September for each of the 50 states. Senator Wicker. OK. Well, then if you would, please enter that into the record. Ms. Espinel. I'd be delighted to. [The information referred to follows:] [GRAPHIC(S) NOT AVAILABLE IN TIFF FORMAT] Senator Wicker. I would like to enter into the record a letter from the Information Technology Industry Council and a letter from the Electronic Privacy Information Center without objection. And that's so ordered. [The information referred to follows:] Information Technology Industry Council Washington, DC, December 11, 2017 Hon. Roger Wicker, Chairman, Hon. Brian Schatz, Ranking Member, Subcommittee on Communications, Technology, Innovation, and the Internet, U.S. Senate Committee on Commerce, Science, and Transportation, Washington, DC. Dear Chairman Wicker and Ranking Member Schatz: In advance of your hearing on ``Digital Decision-Making: The Building Blocks of Machine Learning and Artificial Intelligence,'' I am writing to thank you for interest in and attention to the exciting innovation that is Artificial Intelligence (AI). ITI represents more than 60 of the world's leading information and communications technology (ICT) companies. Our companies are the most dynamic and innovative companies from all corners of the ICT sector, including the development and deployment of AI. I submit this letter on behalf of ITI and its members, and respectfully request that you enter it into the hearing record. Artificial intelligence (AI) technology is an integral part of our daily lives, work, and existence. It's already made an important mark on much of our society and economy, and the exciting part is that we're just seeing the beginning of its benefits. Go to any hospital and medical research center and you will see how doctors and medical providers use AI to save lives. For example, the company Berg uses AI to analyze large amounts of oncological data to create a model of how pancreatic cancer functions, enabling us to develop chemotherapy to which cancer cells are more responsive. Educators across the country use AI to enhance the potential for future generations to grow and learn. Thanks to IBM's Teacher Advisor, a new tool based on its Watson cognitive computing platform, third- grade math teachers can develop personalized lesson plans. Teacher Advisor analyzes education standards, sets targets for skills development, and uses student data to help teachers tailor instructional material for students with varying skill levels. AI also makes day-to-day of life easier--for everyone. Many of our everyday tasks, like making shopping lists and ordering groceries, are streamlined through devices like Alexa. And, through AI technology, researchers at the International Islamic University of Malaysia have developed the Automatic Sign Language Translator (ASLT) that uses machine learning to interpret sign language and convert it into text, easing communications for many. We know AI will revolutionize the way we do business and our overall economy. It's projected AI will add between $7.1 trillion and $13.17 trillion to the global economy by 2025.\1\ --------------------------------------------------------------------------- \1\ https://www.itic.org/resources/AI-Policy-Principles- FullReport2.pdf --------------------------------------------------------------------------- There's no question AI will continue to transform our lives, society, and economy for the better. We understand, however, that there are many questions about this technology and that with transformative innovation, there are going to be points of tension. The tech industry is committed to working with all stakeholders to identify and resolve challenges. In conjunction with the global leaders in AI innovation, ITI recently published AI Policy Principles. These principles are designed to be guidelines for the responsible development and deployment of AI as we develop partnerships with governments, academia, and the public. Our Policy Principles are a conversation catalyst, encouraging all stakeholders, public and private, to collaborate to create smart policies that allow this emerging technology to flourish while addressing the complex issues that arise out of its growth and deployment. Given the reach of AI, we think this kind of partnership and engagement is critical to advance the benefits and responsible growth of AI while also endeavoring to answer the public's questions about the use of this nascent technology. We look forward to working with this Committee, other members of Congress, academia, industry partners, and the public to advance AI responsibly. Thank you, again, for holding the timely and important hearing. Dean Garfield, President and CEO, Information Technology Industry Council (ITI) ______ Electronic Privacy Information Center Washington, DC, December 12, 2017 Hon. John Thune, Chairman, Hon. Bill Nelson, Ranking Member, U.S. Senate Committee on Commerce, Science, and Transportation, Washington, DC. Dear Chairman Thune and Ranking Member Nelson: We write to you regarding the ``Digital Decision-Making: The Building Blocks of Machine Learning and Artificial Intelligence'' hearing.\1\ EPIC is a public interest research center established in 1994 to focus public attention on emerging privacy and civil liberties issues.\2\ EPIC has promoted ``Algorithmic Transparency'' for many years.\3\ --------------------------------------------------------------------------- \1\ Digital Decision-Making: The Building Blocks of Machine Learning and Artificial Intelligence, 115th Cong. (2017), S. Comm. on Commerce, Science, and Transportation, https://www.commerce.senate.gov/ public/index.cfm/hearings?ID=7097E2B0-4A6B-4D92-85C3-D48E100 8C8FD (Dec. 12, 2017). \2\ EPIC, About EPIC, https://epic.org/epic/about.html. \3\ EPIC, Algorithmic Transparency, https://epic.org/algorithmic- transparency/. --------------------------------------------------------------------------- Democratic governance is built on principles of procedural fairness and transparency. And accountability is key to decision making. We must know the basis of decisions, whether right or wrong. But as decisions are automated, and organizations increasingly delegate decisionmaking to techniques they do not fully understand, processes become more opaque and less accountable. It is therefore imperative that algorithmic process be open, provable, and accountable. Arguments that algorithmic transparency is impossible or ``too complex'' are not reassuring. It is becoming increasingly clear that Congress must regulate AI to ensure accountability and transparency: Algorithms are often used to make adverse decisions about people. Algorithms deny people educational opportunities, employment, housing, insurance, and credit.\4\ Many of these decisions are entirely opaque, leaving individuals to wonder whether the decisions were accurate, fair, or even about them. --------------------------------------------------------------------------- \4\ Danielle Keats Citron & Frank Pasquale, The Scored Society: Due Process for Automated Predictions, 89 Wash. L. Rev. 1 (2014). Secret algorithms are deployed in the criminal justice system to assess forensic evidence, determine sentences, to even decide guilt or innocence.\5\ Several states use proprietary commercial systems, not subject to open government laws, to determine guilt or innocence. The Model Penal Code recommends the implementation of recidivism-based actuarial instruments in sentencing guidelines.\6\ But these systems, which defendants have no way to challenge are racially biased, unaccountable, and unreliable for forecasting violent crime.\7\ --------------------------------------------------------------------------- \5\ EPIC v. DOJ (Criminal Justice Algorithms), EPIC, https:// epic.org/foia/doj/criminal-justice-algorithms/; Algorithms in the Criminal Justice System, EPIC, https://epic.org/algorithmic- transparency/crim-justice/. \6\ Model Penal Code: Sentencing Sec. 6B.09 (Am. Law. Inst., Tentative Draft No. 2, 2011). \7\ Julia Angwin et al., Machine Bias, ProPublica (May 23, 2016), https://www.propublica.org/article/machine-bias-risk-assessments-in- criminal-sentencing. Algorithms are used for social control. China's Communist Party is deploying a ``social credit'' system that assigns to each person government-determined favorability rating. ``Infractions such as fare cheating, jaywalking, and violating family-planning rules'' would affect a person's rating.\8\ Low ratings are also assigned to those who frequent disfavored websites or socialize with others who have low ratings. Citizens with low ratings will have trouble getting loans or government services. Citizens with high rating, assigned by the government, receive preferential treatment across a wide range of programs and activities. --------------------------------------------------------------------------- \8\ Josh Chin & Gillian Wong, China's New Tool for Social Control: A Credit Rating for Everything, Wall Street J., Nov. 28, 2016, http:// www.wsj.com/articles/chinas-new-tool-for-social-control-a-credit- rating-for-everything-1480351590 In the United States, U.S. Customs and Border Protection has used secret analytic tools to assign ``risk assessments'' to U.S. travelers.\9\ These risk assessments, assigned by the U.S. Government to U.S. citizens, raise fundamental questions about government accountability, due process, and fairness. They may also be taking us closer to the Chinese system of social control through AI. --------------------------------------------------------------------------- \9\ EPIC v. CBP (Analytical Framework for Intelligence), EPIC, https://epic.org/foia/dhs/cbp/afi/. In a recent consumer complaint to the Federal Trade Commission, EPIC challenged the secret scoring of young athletes.\10\ As EPIC's complaint regarding the Universal Tennis Rating system makes clear, the ``UTR score defines the status of young athletes in all tennis related activity; impacts opportunities for scholarship, education and employment; and may in the future provide the basis for `social scoring' and government rating of citizens.'' \11\ As we explained to the FTC, ``EPIC seeks to ensure that all rating systems concerning individuals are open, transparent and accountable.'' \12\ --------------------------------------------------------------------------- \10\ EPIC, EPIC Asks FTC to Stop System for Secret Scoring of Young Athletes (May 17, 2017), https://epic.org/2017/05/epic-asks-ftc-to- stop-system-f.html; See also Shanya Possess, Privacy Group Challenges Secret Tennis Scoring System, Law360, May 17, 2017, https://www .law360.com/articles/925379; Lexology, EPIC Takes a Swing at Youth Tennis Ratings, June 1, 2017, https://www.lexology.com/library/ detail.aspx?g=604e3321-dfc8-4f46-9afc-abd47c5a5179 \11\ EPIC Complaint to Federal Trade Commission, In re Universal Tennis at 1 (May 17, 2017). \12\ Id. --------------------------------------------------------------------------- In re Universal Tennis, EPIC urged the FTC to (1) Initiate an investigation of the collection, use, and disclosure of children's personal information by Universal Tennis; (2) Halt Universal Tennis's scoring of children without parental consent; (3) Require that Universal Tennis make public the algorithm and other techniques that produce the UTR; (4) Require that Universal Tennis establish formal procedures for rectification of inaccurate, incomplete, and outdated scoring procedures; and (5) Provide such other relief as the Commission finds necessary and appropriate.\13\ --------------------------------------------------------------------------- \13\ Id. at 13. --------------------------------------------------------------------------- ``Algorithmic Transparency'' must be a fundamental principle for consumer protection The phrase has both literal and figurative dimensions. In the literal sense, it is often necessary to determine the precise factors that contribute to a decision. If, for example, a government agency or private company considers a factor such as race, gender, or religion to produce an adverse decision, then the decision- making process should be subject to scrutiny and the relevant factors identified. On October 12, 2016, The White House announced two reports on the impact of Artificial Intelligence on the U.S. economy and related policy concerns. Preparing for the Future of Artificial Intelligence concluded that ``practitioners must ensure that AI-enabled systems are governable; that they are open, transparent, and understandable; that they can work effectively with people; and that their operation will remain consistent with human values and aspirations.'' \14\ --------------------------------------------------------------------------- \14\ Preparing for the Future of Artificial Intelligence, (Oct 2016), Executive Office of the President, National Science and Technology Council, Comm. on Technology, https://obama whitehouse.archives.gov/sites/default/files/whitehouse_files/ microsites/ostp/NSTC/preparing_ for_the_future_of_ai.pdf. --------------------------------------------------------------------------- Some have argued that algorithmic transparency is simply impossible, given the complexity and fluidity of modern processes. But if that is true, there must be some way to recapture the purpose of transparency without simply relying on testing inputs and outputs. We have seen recently that it is almost trivial to design programs that evade testing.\15\ And central to the science and innovation is the provability of results. --------------------------------------------------------------------------- \15\ Jack Ewing, In '06 Slide Show, a Lesson in How VW Could Cheat, N.Y. Times, Apr. 27, 2016, at A1. --------------------------------------------------------------------------- Europeans have long had a right to access ``the logic of the processing'' concerning their personal information.\16\ That principle is reflected in the U.S. in the publication of the FICO score, which for many years remained a black box for consumers, establishing credit worthiness without providing any information about the basis of score.\17\ --------------------------------------------------------------------------- \16\ Directive 95/46/EC--The Data Protection Directive, art 15 (1), 1995, http://www.data protection.ie/docs/EU-Directive-95-46-EC--Chapter-2/93.htm. \17\ Hadley Malcom, Banks Compete on Free Credit Score Offers, USA Today, Jan. 25, 2015, http://www.usatoday.com/story/money/2015/01/25/ banks-free-credit-scores/22011803/. --------------------------------------------------------------------------- The continued deployment of AI-based systems raises profound issues for democratic countries. As Professor Frank Pasquale has said: Black box services are often wondrous to behold, but our black box society has become dangerously unstable, unfair, and unproductive. Neither New York quants nor California engineers can deliver a sound economy or a secure society. Those are the tasks of a citizenry, which can perform its job only as well as it understands the stakes.\18\ --------------------------------------------------------------------------- \18\ Frank Pasquale, The Black Box Society: The Secret Algorithms that Control Money and Information 218 (Harvard University Press 2015). We ask that this Statement from EPIC be entered in the hearing record. We look forward to working with you on these issues of vital importance to the American public. Sincerely, Marc Rotenberg, President, EPIC. Caitriona Fitzgerald, Policy Director, EPIC. Christine Bannan, Policy Fellow. EPIC. Senator Wicker. The hearing record will remain open for 2 weeks. During this time, Senators are asked to submit any questions for the record. Upon receipt, the witnesses are requested to submit their written answers to the Committee as soon as possible. Thank you very much. The hearing is now adjourned. [Whereupon, at 12 p.m., the hearing was adjourned.] A P P E N D I X Response to Written Question Submitted by Hon. Amy Klobuchar to Dr. Cindy L. Bethel Question. Political ads on the Internet are more popular now than ever. In 2016, more than $1.4 billion was spent on digital advertisements and experts project that number will continue to increase. In October, I introduced the Honest Ads Act with Senators Warner and McCain, to help prevent foreign interference in future elections and improve the transparency of online political advertisements. We know that 90 percent of the ads that Russia purchased were issue ads meant to mislead and divide Americans. Increasing transparency and accountability online will benefit consumers and help safeguard future elections. Dr. Bethel, could making more data about political advertisements publicly available help improve the performance of algorithms designed to prevent foreign interference? Answer. More data that is high quality and has varied examples of content may be helpful in improving AI algorithms in general. The data though needs to have features that are detectable to learn from as part of the machine learning process. It is not clear from the political advertisements that have been promoted, as in the previous election, it was not evident what features could be used to learn from to detect involvement by foreign parties of these political advertisements. More data does not always equate to better results. There needs to be sufficient variations in key features that can be detected to be able to develop and test algorithms that will be effective and will have beneficial and meaningful results. ______ Response to Written Question Submitted by Hon. Tom Udall to Dr. Cindy L. Bethel Question. Can you speak of some of the ways that government funded artificial intelligence development is now being used in the private sector? Answer. AI research has been funded for many years through the National Science Foundation, National Institutes of Health, the Department of Defense, USDA, among other agencies. It is possible that there has been government-funded research into artificial intelligence that has moved directly into the private sector for inclusion in product development. Generally, concepts, algorithms, and advancements developed as part of artificial intelligence research, has been integrated into product developments in the private sector but it is typically not a direct transfer from research into the private sector. Currently, the research I am performing with Therabot TM, the robotic therapeutic support companion, funded by the National Science Foundation, developed algorithms associated with artificial intelligence and machine learning that are being used in this application. This is a project that is planned for commercialization and making it available to the public. Mississippi State University has received government funding for projects that include artificial intelligence and have leveraged and transitioned some of the that knowledge into industry-based projects that benefit the private sector. The private sector has been active in funding their own advancements of artificial intelligence and machine learning. Many of the developments and advancements in applications of AI have been developed from private sector research and development groups. There are also cases, where a researcher receives government-funded grants for the development of AI and machine learning, and then may later transition into a private sector position and takes that knowledge with him or her to advance product development that benefits the private sector. There are numerous research developments that have been government funded under SBIR/STTR mechanisms that are joint funding for industry and academic researchers. NIH and USDA have been very active in transitioning the technology developments using their funds into the private sector and commercially available products. NSF and DoD also has programs such as Innovation Corps (I-Corps) that transition developments into commercialized and publicly available products. These have been successful programs available to researchers who have been funded under government grants and contracts. I am not sure of all of the government-funded research to date in artificial intelligence, so I am not sure exactly which projects have benefited or ended up being applied in the private sector. That would be a research project in itself that may be a worthwhile effort. ______ Response to Written Questions Submitted by Hon. Maggie Hassan to Dr. Cindy L. Bethel Question 1. Artificial intelligence, or AI, holds tremendous promise for individuals who experience disabilities. For example, Google and Microsoft have technologies to process language and speech and translate it into a text format to assist individuals who are deaf and hard of hearing. Other technologies will go even further to improve the lives of people with disabilities and I would like to learn more from the panel about what we can expect. What other specific technologies are you aware of in the AI space that will help people who experience disabilities? Answer. There are numerous technologies that use artificial intelligence that are being developed to assist people with different types of disabilities to improve their quality of life. There is the development and use of brain-machine interfaces for operating wheelchairs and other devices through detecting signals sent with intention from the brain. There are exoskeletons that are being developed and prosthetics that learn and detect the signals and impulses from the nervous system to be able to customize how these devices work and enhance the capabilities of the end users. There are in-home assistive robots that are being developed to assist disabled and elderly people in their homes with reminders to take medications, to fetch different items, and to remotely monitor users so they can remain in their homes longer. The Therabot TM robot is being developed as a home therapy support tool to provide comfort and support for people who have experience post-traumatic stress or other types of mental health disorders that can be used to detect and alert clinicians or others when problems are detected. There are continual enhancements to technologies for the hearing and visually impaired users. These are just some of the many examples that use artificial intelligence to improve quality of life for people with different types of disabilities. Question 2. How will manufacturers and developers work to perfect this technology so that it can truly be a reliable tool for these individuals? Answer. These technologies will need to go through extensive testing and user studies/clinical trials to ensure the safety of the developments before they are sold to the public or used by the public. Edge cases need to be tested for events that may not occur frequently but have the possibility of happening. Once these types of technologies are developed and tested, then standards need to be established to ensure ongoing quality of the products for safe use as would be the case of any product used in medical applications or for consumer use. Question 3. What more can Congress to do assist with these efforts? Answer. While research and development is occurring it is important to not establish highly restrictive legislative policies or it will stifle the creativity and development by researchers. Once something is established and has been tested, then it may be necessary to legislate standards of practice for the protection and safety of the public using these items. This would be later in the process. Providing funding to support and an environment supportive of development of these items would allow the U.S. to stay on the top of research developments that use artificial intelligence and machine learning. Legislation should be limited to restrictions and standards for consumer and user safety. Question 4. As we see machine learning and AI increasingly embedded in products and services that we rely on, there are numerous cases of these algorithms falling short of consumer expectations. For example, Google and Facebook both promoted fraudulent news stories in the immediate wake of the Las Vegas Shooting because of their algorithms.\1\ --------------------------------------------------------------------------- \1\ NYT: After Las Vegas Shooting, Fake News Regains Its Megaphone, Kevin Rose, 10/02/2017 https://www.nytimes.com/2017/10/02/business/las- vegas-shooting-fake-news.html --------------------------------------------------------------------------- YouTube Kids is a service designed for children, and marketed as containing videos that are suitable for very young children. In November, YouTube Kids promoted inappropriate content due to algorithms.\2\ While the use of machine learning and AI holds limitless positive potential, at the current point, it faces challenges where we should not risk getting it wrong. --------------------------------------------------------------------------- \2\ NYT: On YouTube Kids, Startling Videos Slip Past Filters, Sapna Maheshwari, 11/04/2017 https://www.nytimes.com/2017/11/04/business/ media/youtube-kids-paw-patrol.html --------------------------------------------------------------------------- Should there be any formal or informal guidelines in place for what tasks are suitable to be done by algorithms, and which are still too important or sensitive to turn over; and what more can be done to ensure better and more accurate algorithms are used as you work to better develop this technology? Answer. In cases, where the algorithms are related to safety or life critical decisions then it may be necessary to have a human in the loop for sanity checks to ensure the best possible decision is made. When it comes to children, this would be case when the system needs to be thoroughly tested with a human involved to ensure the system is working well and there needs to be testing and validation that occurs to ensure those ``edge'' cases or rarer situations are also tested for no matter how unlikely it is for it to occur. Validation and testing should be performed extensively with adults prior to using the system with children and then tested with children with adult supervision. Mistakes can happen, but everything possible needs to be done to prevent issues that could potentially cause harm. In the case of decisions to weaponize autonomous systems there should be a human in the loop when it comes to decisions that impact human lives. There does need to be established standards and benchmarks to assist developers in testing to ensure the safety of a product before it is put in the hands of the public. Question 5. Machine learning and AI hold great promise for assisting us in preventing cybersecurity attacks. According to an IBM survey of Federal IT managers, 90 percent believe that artificial intelligence could help the Federal Government defend against real- world cyber-attacks. 87 percent think AI will improve the efficiency of their cybersecurity workforce.\3\ --------------------------------------------------------------------------- \3\ INFORMATION MANAGEMENT: AI seen as key tool in government's cybersecurity defense, Bob Violino, 11/30/2017 https://www.information- management.com/news/artificial-intelligence-seen-as-key-tool-in- governments-cybersecurity-defense --------------------------------------------------------------------------- While this is promising, the Federal Government currently faces a shortage of qualified cybersecurity employees, and to make matters worse, the pipeline of students studying these topics is not sufficient to meet our needs. A recent GAO report found that Federal agencies have trouble identifying skills gaps, recruiting and retaining qualified staff, and lose out on candidate due to Federal hiring processes. The George Washington University Center for Cyber & Homeland Security recently released a report titled ``Trends in Technology and Digital Security'' which stated: ``Traditional security operations centers are mostly staffed with tier one analysts staring at screens, looking for unusual events or detections of malicious activity. This activity is similar to physical security personnel monitoring video cameras for intruders. It is tedious for humans, but it is a problem really well-suited to machine learning.'' \4\ --------------------------------------------------------------------------- \4\ https://cchs.gwu.edu/sites/cchs.gwu.edu/files/downloads/ Fall%202017%20DT%20symposi um20compendium.pdf What effect will machine learning and AI will have on cybersecurity; and how do you think the Federal Government can best leverage the benefits offered by machine learning and AI to address our cybersecurity workforce shortage? Answer. The use of artificial intelligence and machine learning can definitely help with the tedious task of identifying potential threats to security. The initial evaluation could be performed using computer systems and they are adept at detecting anomalies and maybe even ones that a human may not readily detect. In cases where it is a threat to life or safety of humans involved, then it may need to verified by a cybersecurity trained specialist. Overall, the use of good algorithms and machine learning systems could help fill the gap that has occurred with the lack of trained cybersecurity personnel. If the initial detection work can be performed by the computer systems, then it would require less personnel to verify those findings. Overall, the use of well-developed AI and machine learning systems could be leveraged to address some of the workforce shortage issues associated with cybersecurity professionals. It is also important to better recruit for these programs and to show the benefits of being involved in this type of career. There are limitations to the government hiring practices and payscales, but these can be overcome. It may require changes though to these practices to entice students entering and choosing their career fields to consider careers in these areas. ______ Response to Written Question Submitted by Hon. Tom Udall to Daniel Castro Question. In your testimony, you discussed how regulators will not need to intervene because the private sector will address artificial intelligence's problems--such as bias and discrimination. However, there have been studies that show implicit bias even when artificial intelligence is deployed. For example, a study \1\ about using AI to evaluate resumes found that candidates with names associated with being European American were 50 percent more likely to be offered an interview than candidates with names associated with being African- American. What role should the Federal Government play where there is implicit bias and discrimination--particularly when companies are required to be ``Equal Opportunity Employers''? --------------------------------------------------------------------------- \1\ http://www.sciencemag.org/news/2017/04/even-artificial- intelligence-can-acquire-biases-against-race-and-gender --------------------------------------------------------------------------- Answer. This is an important question. To clarify, regulators will need to continue to intervene to address specific policy goals, such as ensuring non-discrimination in hiring practices. However, policymakers do not necessarily need to create new laws and regulations only for AI to achieve those goals. Existing laws that make these practices illegal still apply, regardless of whether or not a company uses AI to discriminate against a protected class. For example, a company cannot circumvent its obligations in Title VII of the Civil Rights Act to discriminate against a particular race in its hiring practices simply by using an algorithm to review job applicants. There are additional steps policymakers can take to reduce bias. One way to assess bias, whether it be in analog processes or digital algorithms, is to have businesses conduct disparate impact analyses. For example, if a company is using an AI system to screen job applicants, and it has concerns about potential racial bias, it should test this system to assess its accuracy. If government agencies are early adopters of such AI-driven services, they can help identify potential areas of concerns. However, disparate impact analysis is only possible if organizations have data available to them. Moreover, regulators can also identify practices that are known to have a disparate impact, such as using certain criteria for making a credit or housing decision, and discouraging businesses from using those methods.\2\ --------------------------------------------------------------------------- \2\ Travis Korte and Daniel Castro, ``Disparate Impact Analysis is Key to Ensuring Fairness in the Age of the Algorithm,'' Center for Data Innovation (2015), http://datainnovation.org/2015/01/disparate-impact- analysis-is-key-to-ensuring-fairness-in-the-age-of-the-algorithm/. --------------------------------------------------------------------------- In addition, there are likely areas where additional policy is needed to protect workers. For example, Congress should consider passing laws such as the Employment Non-Discrimination Act (ENDA) to ensure that data about sexual orientation and gender identity cannot be used to unfairly harm workers.\3\ These types of laws address specific concerns of vulnerable populations, but they do not apply only to AI. --------------------------------------------------------------------------- \3\ For more on this topic as it relates to data, see: Joshua New and Daniel Castro, ``Accelerating Data Innovation: A Legislative Agenda for Congress,'' Center for Data Innovation (2015), http:// datainnovation.org/2015/05/accelerating-data-innovation-a-legislative- agenda-for-congress/. --------------------------------------------------------------------------- Finally, policymakers should recognize that using AI can often reduce discrimination by limiting the potential for both implicit and explicit human bias. For example, a company that uses AI to screen applicants has the potential to reduce implicit bias of human managers in hiring practices. And while AI systems may not be perfect at the outset, as people identify problems, they will be able to more quickly resolve these issues. The same is not true for strictly human processes, where eliminating bias, is much more difficult. ______ Response to Written Question Submitted by Hon. Gary Peters to Daniel Castro Question. Many have predicted that AI will have a profound effect on the labor market. Most predict that low-wage, routine-based jobs will be under the most pressure for replacement by AI. Meanwhile, recent advancements in technology has led to job creation that will mostly require highly-skilled, highly-educated workers. What evidence have you seen regarding businesses incorporating this labor shift into their business plans? Answer. Some of these predictions are not based on sound analysis. Bureau of Labor Statistics (BLS) projections show that the fastest growing jobs are not in high-skilled occupations. For example, the industry that BLS projects will have the most job growth between 2016- 2026 is the ``food services and drinking places'' industry.\4\ These are not high-wage jobs. Increased use of AI can yield higher rates of automation and hopefully fewer of these low-wages jobs. The way to achieve higher-wage jobs is by increasing productivity. In particular, increasing productivity in low-skill jobs will grow wages in these occupations. --------------------------------------------------------------------------- \4\ See ``Projections of industry employment, 2016-2026,'' Bureau of Labor Statistics, https://www.bls.gov/careeroutlook/2017/article/ projections-industry.htm. --------------------------------------------------------------------------- Some companies have taken steps to address disruption in the workforce. For example, Google and Facebook have made substantial commitments to funding for job retraining programs.\5\ However, overall, U.S. companies are investing less in training now than they were 15 years ago. A knowledge tax credit, where corporates receive a credit for qualified expenditures on worker training, would help address this problem.\6\ --------------------------------------------------------------------------- \5\ ``Google pledges $1 billion to prepare workers for automation,'' Engadget, October 13, 2017, https://www.engadget.com/ 2017/10/13/grow-with-google/. \6\ See Rob Atkinson, ``How a knowledge tax credit could stop decline in corporate training,'' The Hill, http://thehill.com/blogs/ pundits-blog/finance/235018-how-a-knowledge-tax-credit-could-stop- decline-in-corporate. --------------------------------------------------------------------------- ______ Response to Written Questions Submitted by Hon. Maggie Hassan to Daniel Castro Question 1. Artificial intelligence, or AI, holds tremendous promise for individuals who experience disabilities. For example, Google and Microsoft have technologies to process language and speech and translate it into a text format to assist individuals who are deaf and hard of hearing. Other technologies will go even further to improve the lives of people with disabilities and I would like to learn more from the panel about what we can expect. What other specific technologies are you aware of in the AI space that will help people who experience disabilities? Answer. AI will have widespread benefits for people with disabilities. As noted in the question, one of the major areas of impact AI will have is by allowing more people to interface with computer systems using their voice, instead of a keyboard. In particular, the combination of AI with the Internet of Things, will give people with many types of disabilities a better quality of life as they will now be able to control more of the world around them. These functions will allow more people with disabilities to participate in the workforce, go to school, and be more active in their communities. In addition, AI can be used to create smart agents that automate specific tasks, such as scheduling meetings, setting a thermostat, or re-ordering groceries. While these types of actions are conveniences for some people, for people with significant disabilities, they can be empowering and allow individuals significantly more autonomy and independence. Question 2. How will manufacturers and developers work to perfect this technology so that it can truly be a reliable tool for these individuals? Answer. One way to improve is by having industry work more closely with different populations of people with disabilities throughout the design and testing of new products. Working closely with different groups helps developers better anticipate user needs and pursue universal design. Question 3. What more can Congress to do assist with these efforts? Answer. One significant challenge is that the need to design for accessibility for people with disabilities is still underappreciated among technologists. One way to change this is to address this problem at the colleges and universities training the next generation of computer scientists and engineers. For example, Congress could establish NSF-funded Centers of Excellence for Accessible Design to prioritize this skillset and develop more curriculum. In addition, Congress should explore ways to encourage and support more people with disabilities to pursue careers in technology-related fields so they can be involved from the outset in the design and creation of more technologies. Finally, Congress should work to increase access to technology for people with disabilities, including by ensuring that programs designed to close the digital divide, such as PC or Internet access, are updated for newer technologies. Question 4. As we see machine learning and AI increasingly embedded in products and services that we rely on, there are numerous cases of these algorithms falling short of consumer expectations. For example, Google and Facebook both promoted fraudulent news stories in the immediate wake of the Las Vegas Shooting because of their algorithms.\7\ YouTube Kids is a service designed for children, and marketed as containing videos that are suitable for very young children. In November, YouTube Kids promoted inappropriate content due to algorithms.\8\ While the use of machine learning and AI holds limitless positive potential, at the current point, it faces challenges where we should not risk getting it wrong. Should there be any formal or informal guidelines in place for what tasks are suitable to be done by algorithms, and which are still too important or sensitive to turn over; and what more can be done to ensure better and more accurate algorithms are used as you work to better develop this technology? --------------------------------------------------------------------------- \7\ NYT: After Las Vegas Shooting, Fake News Regains Its Megaphone, Kevin Rose, 10/02/2017 https://www.nytimes.com/2017/10/02/business/las- vegas-shooting-fake-news.html \8\ NYT: On YouTube Kids, Startling Videos Slip Past Filters, Sapna Maheshwari, 11/04/2017 https://www.nytimes.com/2017/11/04/business/ media/youtube-kids-paw-patrol.html --------------------------------------------------------------------------- Answer. AI, much like humans, is fallible. There should always be some oversight of AI, just as there should always be some oversight of humans. It is not a problem if AI systems make mistakes, unless these mistakes go undetected. So the key objective, whether a decision is being made by a computer of a human, is whether there is sufficient oversight appropriate to the level of risk for the individuals involved. This will likely be context dependent. This is one reason why it is inappropriate to talk about industry-wide regulation of AI, and much more appropriate to talk about industry-specific regulations of AI. For example, the Department of Transportation may have specific requirements for the types of oversight it wants for autonomous vehicles that looks very different than the type of oversight the Securities and Exchange Commission needs for AI-driven stock trading. Question 5. Machine learning and AI hold great promise for assisting us in preventing cybersecurity attacks. According to an IBM survey of Federal IT managers, 90 percent believe that artificial intelligence could help the Federal Government defend against real- world cyber-attacks. 87 percent think AI will improve the efficiency of their cybersecurity workforce.\9\ While this is promising, the Federal Government currently faces a shortage of qualified cybersecurity employees, and to make matters worse, the pipeline of students studying these topics is not sufficient to meet our needs. A recent GAO report found that Federal agencies have trouble identifying skills gaps, recruiting and retaining qualified staff, and lose out on candidate due to Federal hiring processes. The George Washington University Center for Cyber & Homeland Security recently released a report titled ``Trends in Technology and Digital Security'' which stated: --------------------------------------------------------------------------- \9\ INFORMATION MANAGEMENT: AI seen as key tool in government's cybersecurity defense, Bob Violino, 11/30/2017 https://www.information- management.com/news/artificial-intelligence-seen-as-key-tool-in- governments-cybersecurity-defense ``Traditional security operations centers are mostly staffed with tier one analysts staring at screens, looking for unusual events or detections of malicious activity. This activity is similar to physical security personnel monitoring video cameras for intruders. It is tedious for humans, but it is a problem really well-suited to machine learning.'' \10\ --------------------------------------------------------------------------- \10\ https://cchs.gwu.edu/sites/cchs.gwu.edu/files/downloads/ Fall%202017%20DT%20symposi um%20compendium.pdf What effect will machine learning and AI will have on cybersecurity; and how do you think the Federal Government can best leverage the benefits offered by machine learning and AI to address our cybersecurity workforce shortage? Answer. AI is very good at specific tasks, such as pattern recognition and anomaly detection. This means that it will be useful for identifying attacks in real time, and it will be an especially important line of defense against zero-day attacks (i.e., attacks that use a previously undisclosed vulnerability). AI might also help developers eliminate certain types of vulnerabilities which may be identifiable at the outset, much like a spell-checker or grammar- checker can review documents. However, AI will not be a panacea, as many cybersecurity risks are the result of poor implementation and a lack of adherence to best practices. ______ Response to Written Questions Submitted by Hon. Gary Peters to Victoria Espinel Question 1. I am concerned by recent reports in Nature, The Economist, and Wall Street Journal about large tech firms monopolizing the talent in AI and machine learning. This concentration of talent can lead to several negative outcomes including long-term wage stagnation and income inequality. In your opinion, what steps or incentives might mitigate this concentration, encourage AI-experts to work at small and medium enterprises, or launch their own start-up with the goal of growing a business (rather than having a goal of being bought out by one of the tech giants)? Similarly, what incentives might encourage AI experts to become educators and trainers to help develop the next generation of AI experts? How can the Federal Government compete with the tech giants to attract experts needed to develop and implement AI systems for defense and civil applications? Answer. Artificial intelligence (AI) is a burgeoning field, with market dynamics that are quickly evolving. While the competition for AI expertise is certainly fierce, it is important to remember that the economic benefits of AI will be spread throughout the economy. By helping people make better data-driven decisions, AI is stimulating growth in every industry sector. It is helping to optimize manufacturing, improve supply chains, secure networks, and enhance products and services. The history of the technology industry suggests that innovation will continue to emerge from enterprises of all sizes. Indeed, BSA's membership is a testament to just how fiercely competitive the enterprise technology landscape is. Datastax, DocuSign, Salesforce, Splunk and Workday are just a few of the young companies that have disrupted the industry over the past 10 years and contributed to a wave of innovation that has made the U.S. software industry the envy of the world. Moreover, despite intense competition for AI expertise, small and medium-sized firms continue to play an incredibly important role in driving AI innovation. In fact, a recent study found that there are currently more than 2,000 AI startups that have raised almost $30 billion in funding.\1\ --------------------------------------------------------------------------- \1\ See Vala Afshar, AI Is Tranformational Technology and Major Sector Disruptor, Huffington Post (Dec. 5, 2017), https:// www.huffingtonpost.com/entry/ai-is-transformational-technology-and- major-sector_us_5a259dbfe4b05072e8b56b6e. --------------------------------------------------------------------------- Because AI will be a huge driver of the global economy in the years ahead, it is vital that we examine the important issues that you have raised to ensure that the United States remains the global hub for AI innovation. There are three specific ways in which the government can increase the talent pool and attract that talent to the government for defense and civil applications. First, the government should increase its commitment to STEM education so that the United States prepares more men and women for AI-related careers. As part of this effort, government agencies can explore partnerships with academic institutions to provide internships to students studying in this field, which could pique their interests in pursuing careers in public service, and develop opportunities for academic researchers to share their technical expertise with government agencies. Second, the government should increase its funding for AI research, providing targeted investments into the ``high risk, high reward'' areas of basic research that are typically underfunded by the private sector.\2\ Third, the government should be ambitious in its goals. The greater the vision for how AI will improve government services and capabilities, the better it will do in attracting talent. --------------------------------------------------------------------------- \2\ See Jason Furman, Is this Time Different? The Opportunities and Challenges of Artificial Intelligence, AI Now: The Social and Economic Implications of Artificial Intelligence in the Near Term (July 7, 2016), available at https://goo.gl/pzFDYw (``In 2015, American businesses devoted almost 1.8 percent of GDP to research and development, the highest share on record. But government investments in R&D have fallen steadily as a share of the economy since the 1960s. While business investment is critical, it is not sufficient. Basic research discoveries often have great social value because of their broad applicability, but there tends to be underinvestment in basic research by private firms because it is difficult for a private firm to appropriate the gains from such research. In fact, while the private sector accounts for roughly two-thirds of all spending on R&D, it is important to keep in mind that it largely invests in applied research while the Federal Government provides 60 percent of the funding for basic research.''). Question 2. Many have predicted that AI will have a profound effect on the labor market. Most predict that low-wage, routine-based jobs will be under the most pressure for replacement by AI. Meanwhile, recent advancements in technology has led to job creation that will mostly require highly-skilled, highly-educated workers. What evidence have you seen regarding businesses incorporating this labor shift into their business plans? Answer. The benefits of AI will be widespread, likely enhancing operations in every industry. As a result, AI also will likely create shifts in the labor market across the economy. The precise impact of AI on employment is uncertain. However, it is clear that AI will create new opportunities within existing jobs and new roles that require skills that the current workforce does not yet have. As a result, many BSA companies have launched initiatives to train employees, youth, and military veterans to help meet the demands of the future labor market. BSA would like to work with Congress to ensure we have the right programs and resources in place for the jobs of the future. We would be happy to come in and discuss with you the initiatives of the software industry that address this important issue. ______ Response to Written Questions Submitted by Hon. Maggie Hassan to Victoria Espinel Question 1. Artificial intelligence, or AI, holds tremendous promise for individuals who experience disabilities. For example, Google and Microsoft have technologies to process language and speech and translate it into a text format to assist individuals who are deaf and hard of hearing. Other technologies will go even further to improve the lives of people with disabilities and I would like to learn more from the panel about what we can expect. What other specific technologies are you aware of in the AI space that will help people who experience disabilities? Answer. There are numerous ways in which AI is being used to improve the lives of people who experience disabilities. Below, I highlight a few examples. Visual impairment--Microsoft recently released an intelligent camera app that uses a smartphone's built-in camera functionality to describe to low-vision individuals the objects that are round them. See Microsoft, Seeing AI, https:// www.microsoft.com/en-us/seeing-ai/. The app opens up new possibilities for the visually impaired to navigate the world with more independence. Autism--IBM researchers are using AI to develop tools that will help people with cognitive and intellectual disabilities, such as autism, by breaking down complex sentences and phrases to help them better understand normal speech and communicate more effectively. See https://www-03.ibm.com/able/content- clarifier.html. Accessible public transportation--As part of a public- private partnership, an innovative project is underway that aims to help disabled people and those with special needs access public transportation by providing real-time information through an Internet of Things system that helps them find the right track, platform, train, and place to board, and alerts them when to disembark. See David Louie, Artificial Intelligence Research Is Helping the Disabled Use Public Transportation, (July 12, 2017), http://abc7news.com/ technology/ai-being-used-to-help-disabled-using-public- transportation/2210112/. Mobility Impairments--Microsoft's Windows 10 operating system introduced Eye Control, a built-in eye tracking feature that enables people with motor neurone disease and other mobility impairments to navigate their computers. See Tas Bindi, Microsoft Using AI to Empower Living With Disabilities, Zdnet (Nov. 15, 2017), http://www.zdnet.com/article/microsoft- using-ai-to-empower-people-living-with-disabilities/. Alzheimer's--Researchers in Italy and Canada have developed machine-learning algorithms to help identify patients that are at risk of developing Alzheimer's. In early tests, the technology has identified changes in the brain that lead to Alzheimer's almost a decade before clinical symptoms would appear. See Daisy Yuhas, Doctors Have Trouble Diagnosing Alzheimer's. AI Doesn't, NBC News (Oct. 30, 2017), https:// www.nbcnews.com/mach/science/doctors-have-trouble-diagnosing- alzheimer-s-ai-doesn-t-ncna815561. As researchers continue to apply AI to new settings, the myriad ways in which AI is used to enhance the lives of people with disabilities will only increase. Question 2. How will manufacturers and developers work to perfect this technology so that it can truly be a reliable tool for these individuals? Answer. BSA members that design and offer AI products and services have strong incentives to ensure that the technology is reliable, as they understand that building trust and confidence in AI systems is integral to successful and widespread deployment of AI services. There are a number of strategies that companies already employ to accomplish this objective. For example, one key priority is ensuring access to vast, robust, and representative data sets. Because AI technologies process and learn from data inputs, ensuring sufficient quantity and quality of data used to train AI systems is very important to enhancing the reliability of these services. In addition, another key step companies take to enhance reliability is testing their AI systems to ensure that they operate as intended, and making appropriate adjustments where they identify errors. Companies also recognize the need to protect AI systems from cyberattacks and are investing heavily in the development of advanced security tools. As companies continue to seek to expand the capabilities of AI technologies, investment in research and development will continue to be important to unleash the full potential of innovation, strengthen cybersecurity, and enhance the overall reliability of AI systems. Question 3. What more can Congress to do assist with these efforts? Answer. Congress can play a very important role in facilitating the deployment of AI services that help people with disabilities by ensuring, more broadly, that the U.S. maintains a flexible policy framework that spurs innovation in AI. Specifically, as I highlighted in my testimony, I think Congress can assist with these efforts in three key ways. First, Congress should pass the OPEN Government Data Act, which recognizes that government- generated data is a national resource that can serve as a powerful engine for creating new jobs and a catalyst for economic growth, and that it is incredibly valuable in fostering innovation in AI and other data-driven services. Second, Congress should support efforts to promote digital trade and facilitate data flows. In a global economy, real-time access to data around the world has become increasingly critical for AI and other digital services to function. As a result, Congress should support modernizing trade initiatives, such as NAFTA, that seek to facilitate digital trade and limit inappropriate restrictions on cross-border data transfers. Third, Congress should promote U.S. investment in AI research, education, and workforce development to ensure that the U.S. remains globally competitive. Strategic investment in education and workforce development can help ensure that the next generation and our current workforce are prepared for the jobs of the future. In addition, promoting public sector and incentivizing private sector research will be essential to unlocking additional capabilities that AI can provide. Question 4. As we see machine learning and AI increasingly embedded in products and services that we rely on, there are numerous cases of these algorithms falling short of consumer expectations. For example, Google and Facebook both promoted fraudulent news stories in the immediate wake of the Las Vegas Shooting because of their algorithms.\3\ YouTube Kids is a service designed for children, and marketed as containing videos that are suitable for very young children. In November, YouTube Kids promoted inappropriate content due to algorithms.\4\ While the use of machine learning and AI holds limitless positive potential, at the current point, it faces challenges where we should not risk getting it wrong. Should there be any formal or informal guidelines in place for what tasks are suitable to be done by algorithms, and which are still too important or sensitive to turn over; and what more can be done to ensure better and more accurate algorithms are used as you work to better develop this technology? --------------------------------------------------------------------------- \3\ NYT: After Las Vegas Shooting, Fake News Regains Its Megaphone, Kevin Rose, 10/02/2017 https://www.nytimes.com/2017/10/02/business/las- vegas-shooting-fake-news.html \4\ NYT: On YouTube Kids, Startling Videos Slip Past Filters, Sapna Maheshwari, 11/04/2017 https://www.nytimes.com/2017/11/04/business/ media/youtube-kids-paw-patrol.html --------------------------------------------------------------------------- Answer. Because AI is ultimately a technology that is intended to help people and organizations make better uses of data, I would be reluctant to prescribe any bright line rules about when its use may or may not be appropriate. However, it is important for companies that develop AI systems, and their customers, to consider the unique risks and potential unintended consequences that can arise when AI is deployed in particular settings. While AI is an invaluable tool for making sense of large quantities of data, there are settings where the intuition of subject matter experts will remain important. For instance, while AI systems certainly have an important role to play in helping to diagnose patients, they are a resource for a medical professional to consider in making a diagnosis or prescribing treatment, they should not be replacing a doctor's judgment. Question 5. Machine learning and AI hold great promise for assisting us in preventing cybersecurity attacks. According to an IBM survey of Federal IT managers, 90 percent believe that artificial intelligence could help the Federal Government defend against real- world cyber-attacks. 87 percent think AI will improve the efficiency of their cybersecurity workforce.\5\ --------------------------------------------------------------------------- \5\ INFORMATION MANAGEMENT: AI seen as key tool in government's cybersecurity defense, Bob Violino, 11/30/2017 https://www.information- management.com/news/artificial-intelligence-seen-as-key-tool-in- governments-cybersecurity-defense --------------------------------------------------------------------------- While this is promising, the Federal Government currently faces a shortage of qualified cybersecurity employees, and to make matters worse, the pipeline of students studying these topics is not sufficient to meet our needs. A recent GAO report found that Federal agencies have trouble identifying skills gaps, recruiting and retaining qualified staff, and lose out on candidate due to Federal hiring processes. The George Washington University Center for Cyber & Homeland Security recently released a report titled ``Trends in Technology and Digital Security'' which stated: ``Traditional security operations centers are mostly staffed with tier one analysts staring at screens, looking for unusual events or detections of malicious activity. This activity is similar to physical security personnel monitoring video cameras for intruders. It is tedious for humans, but it is a problem really well-suited to machine learning.'' \6\ --------------------------------------------------------------------------- \6\ https://cchs.gwu.edu/sites/cchs.gwu.edu/files/downloads/ Fall%202017%20DT%20symposi um%20compendium.pdf What effect will machine learning and AI will have on cybersecurity; and how do you think the Federal Government can best leverage the benefits offered by machine learning and AI to address our cybersecurity workforce shortage? Answer. AI tools are revolutionizing network security, helping analysts parse through hundreds of thousands of security incidents per day to weed out false positives and identify threats that warrant further attention by network administrators. By automating responses to routine incidents and enabling security professionals to focus on truly significant threats, AI-enabled cyber tools are helping enterprises stay ahead of their malicious adversaries. For instance, AI has helped an enterprise security operations center ``reduce the time to remediate spearphishing attacks from there hours per incident to less than two minutes per incident.'' \7\ Importantly, AI is also helping to train the next generation of security analysts, teaching them to more quickly identify threats that need to be escalated through the chain of command.\8\ Greater deployment of AI is therefore a critical factor for addressing the cyber workforce shortage, which experts now estimate will climb 1.8 million positions by 2022. --------------------------------------------------------------------------- \7\ See Robert Lemos, AI Is Changing SecOps: What Security Analysts Need to Know, TechBeacon (Dec. 19, 2017), https://techbeacon.com/ai- changing-secops-what-security-analysts-need-know. \8\ Id. --------------------------------------------------------------------------- However, AI alone will not solve the cyber workforce shortage. It is therefore incumbent on governments and industry to work collaboratively to grow the pipeline of cyber talent. To that end, BSA recently launched a new cybersecurity agenda \9\ that highlights four pathways for developing a 21st century cybersecurity workforce: --------------------------------------------------------------------------- \9\ BSA/The Software Alliance, A Cybersecurity Agenda for the Connected Age, available at www.bsa.org//media/Files/Policy/ BSA_2017CybersecurityAgenda.pdf. Increase access to computer science education: Expand cybersecurity education for K-12 as well as in undergraduate computer science programs, increase scholarships, and --------------------------------------------------------------------------- incentivize minority students. Promote alternative paths to cybersecurity careers: Launch careers through apprenticeship programs, community colleges, cybersecurity ``boot camps,'' and government or military service. Modernize training for mid-career professionals: Reform Trade Adjustment Assistance, and update other mid-career re- training programs, to provide American workers with high-demand cybersecurity and IT skills as digitalization transforms the global economy. Improve the exchange of cybersecurity professionals between the government and private sector: Enable private sector experts to join the government for periodic or short-term assignments. ______ Response to Written Question Submitted by Hon. Amy Klobuchar to Dr. Dario Gil, Ph.D. Question. While I was at the hearing there was significant discussion about the future security applications for machine learning and artificial intelligence. As the Ranking Member on the Rules Committee, I am working with Senators Lankford, Harris and Graham on a bill to upgrade our election equipment to protect against cyber- attacks. The Department of Homeland Security recently confirmed that hackers targeted 21 states' election systems in the run-up to the 2016 election. As we prepare for 2018 and beyond, we must ensure that our election systems are secure, both from a hardware and a software perspective because election security is national security. Dr. Gil, can artificial intelligence and machine learning be used to identify and prevent cyber-attacks? Answer. The power of AI, like most machine learning techniques, lies in identifying broader trends, building models of what normal and expected behavior is and flagging anomalies. There has been tremendous value shown by deploying AI in the field of security, and we use it in a plethora of use cases: in, applications to flag systems and networks, to monitor for anomalies and raise alerts when these behaviors change. Such anomalous behavior may indicate an attack (or a benign error). AI has also been leveraged for generalizations to make it easier to identify new instances of known attacks. It can be used to learn about malware and exploits of vulnerabilities and can use that to detect new infections and intrusions better than rule-based systems. AI and AI- based techniques can also help in hardening security protections to make it more difficult for attackers to successfully exploit a system. For example, automating tests (fuzzing) to probe for vulnerabilities using reinforcement learning can be more efficient than an exhaustive scan. It should be borne in mind that while there is no panacea for security, AI is a very powerful tool that can be employed to increase security when combined with a standard suite of best practices. ______ Response to Written Questions Submitted by Hon. Tom Udall to Dr. Dario Gil, Ph.D. Question 1. As you are aware, New Mexico is home to two national laboratories--Sandia and Los Alamos. Can you speak of any partnership you have with the national laboratories? Answer. We have had, and continue to have, a number of partnerships with Los Alamos and Sandia National Labs. At this time, we cannot comment on individual projects, but we value the joint work we have with the labs and all of our external partners. Question 2. Can you speak of some of the ways that government funded artificial intelligence development is now being used in the private sector? Answer. Government-funded AI development is currently being used by IBM in the following ways: U.S. Army Research Labs funded the development of technologies for rule-based systems and policy management. The technology developed is being used now by IBM to improve and support the ability of employees in a company to configure computer software to efficiently execute high-volume, highly transactional process functions, boosting capabilities and saving time and money. U.S. Army Research Labs funded the development of AI enabled algorithms for analyzing and understanding properties of large numbers of moving objects. The technology developed has been used to create commercial cloud-based services such as The Weather Company LongShip service to predict the influence of weather impacts on traffic. It also has been incorporated into commercial software products such as DB/2. Looking more broadly, there are a variety of ways in which government funded research can be used by the private sector. These include: Government funded technology has been used to create dual- purpose technologies--those which serve the needs of the private sector as well as the needs of the government agency which sponsored the work. One expression of this is when the private sector produces a COTS (Custom Off the Shelf) product which allows the government to meet their specific requirement at a lower overall cost. Government funded research (primarily in basic sciences) has a research horizon which is typically longer than the private sector. As a result, government funded research has been used to create technology at the earlier stage. There are many instances where government support has been used for technologies at the TRL (Technology Readiness Level) of 1-4, and the successful ones from this level of exploration have led to commercialized products at TRL level 5 and above. Government funded technology has been used to produce open source software, which private sector companies develop further and use to create new offerings. In many cases, the open source software is used by academics and the public at large for knowledge creation. Government funded alliances have driven collaboration between private sector researchers, academics and government researchers. This cross-fertilization of researchers working towards a common goal has been beneficial to all parties-- including employment opportunities for students, infusion of new ideas in industry activities, improvements in government, and formation of lasting collaborations. ______ Response to Written Question Submitted by Hon. Gary Peters to Dr. Dario Gil, Ph.D. Question. A major challenge AI and machine learning developers need to address is the ability to ensure prolonged safety, security, and fairness of the systems. This is especially true of systems designed to work in complex environments that may be difficult to replicate in training and testing, or systems that are designed for significant learning after deployment. Dr. Gil, you testified that IBM is looking to build trust in AI by following a set of principles to guide your development and use of AI systems. Would you please provide more detail about how these principles are being implemented? How will these principles prevent a system designed to learn after deployment from developing unacceptable behavior over time? Answer. The currently available AI products, such as factory robots, personal digital assistants, and healthcare decision support systems, are designed to perform one narrow task, such as assemble a product, provide a weather forecast or make a purchase order, or help a radiologist interpret an X-ray. When these technologies learn after deployment, they do so in the context of that narrow task, and do not have the ability to learn other tasks on their own, even similar ones. The kind of AI systems that can acquire new skills and perform new cognitive and reasoning tasks autonomously are actively being researched. This effort does not include only the development of underlying learning and reasoning capabilities; the AI research community is actively pursuing the capabilities that would ensure the safety, security and fairness of these systems. To start with, the principles of safe design are applied to a wide variety of engineered systems, such as trains, safety breaks, industrial plants, flight autopilot systems, and robotic laser surgery. Some of these principles apply directly to the design of AI systems, some will be adapted, and new ones will have to be defined. For example, it is possible to constrain the space of outcomes or actions a robot can perform, to ensure that it does not accidentally come into contact with human workers and cause injury. Similarly, robots in complex environments that encounter completely new situations could be designed to require human intervention. Another direction is to embed principles of safe and ethical behavior in the AI reasoning mechanisms, so that they can distinguish between right and wrong actions. With respect to the fairness of the AI systems, we are currently pursuing a range of efforts aimed at developing and embedding in our services and offerings techniques for bias detection, certification, and mitigation. For example, we have developed algorithms that can de- bias training data so that any AI system that learns from such data does not discriminate against protected groups (e.g., those defined by race or gender). We also are working on using blockchain to ensure the integrity of an AI system by making sure that it is secure, auditable and used as intended. We also are developing capabilities to enhance the explainability and interpretability of AI systems, so that unacceptable behaviors can be easily discovered and removed. IBM has established the following principles for the artificial intelligence/cognitive era: Purpose: The purpose of AI and cognitive systems developed and applied by the IBM company is to augment human intelligence. Our technology, products, services and policies will be designed to enhance and extend human capability, expertise and potential. Our position is based not only on principle but also on science. Cognitive systems will not realistically attain consciousness or independent agency. Rather, they will increasingly be embedded in the processes, systems, products and services by which business and society function--all of which will and should remain within human control. Transparency: For cognitive systems to fulfill their world-changing potential, it is vital that people have confidence in their recommendations, judgments and uses. Therefore, the IBM company will make clear: When and for what purposes AI is being applied in the cognitive solutions we develop and deploy. The expertise that informs the insights of cognitive solutions, as well as the methods used to train those systems and solutions. The principle that clients own their own business models and intellectual property and that they can use AI and cognitive systems to enhance the advantages they have built. We will work with our clients to protect their data and insights, and will encourage our clients, partners and industry colleagues to adopt similar practices. Skills: The economic and societal benefits of this new era will not be realized if the human side of the equation is not supported. This is uniquely important with cognitive technology, which augments human intelligence and expertise and works collaboratively with humans. Therefore, the IBM company will work to help students, workers and citizens acquire the skills and knowledge to engage safely, securely and effectively in a relationship with cognitive systems, and to perform the new kinds of work and jobs that will emerge in a cognitive economy. Data: Since AI is heavily based on data, IBM has developed a framework of best practices for data stewardship \1\ that ensures great care and responsibility in data ownership, storage, security, and privacy. IBM abides by these practices and, as a result, serves as a data steward providing transparent and secure services. For example, we write client agreements with full transparency and will not use client data unless they agree to such use. We will limit that use to the specific purposes clearly described in the agreement. IBM does not put `backdoors' in its products for any government agency, nor do we provide source code or encryption keys to any government agency for the purpose of accessing client data. --------------------------------------------------------------------------- \1\ IBM: Data Responsibility@IBM, Khttps://www.ibm.com/blogs/ policy/wp-content/uploads/2017/10/IBM_DataResponsibility-A4_WEB.pdf --------------------------------------------------------------------------- We are working on a range of efforts aimed at developing and embedding in our services and offerings techniques for bias detection, certification, and mitigation. For example, we are working on improving the accuracy of directly interpretable decision-support algorithms, such as decision trees and rule sets, as well as enhancing the interpretability of deep learning neural net models. Moreover, as we develop innovative AI systems, we are guided by the principles of safety engineering. Some of these principles could be directly applied to the design of AI systems, some will be adapted, and new ones will have to be defined. For example, robots in complex environments that encounter completely new situations could be designed to require human intervention. Another direction is to embed principles of safe and ethical behavior in their reasoning mechanisms, so they can distinguish between right and wrong actions. Finally, we are working to develop AI systems that act according to human values that are relevant for the scenarios and communities in which such systems will be deployed. This means constraining the learning, reasoning and optimization machinery inside AI systems to behavioral constraints. These constraints will ensure that the actions of the AI system comply with values and guidelines that humans define as appropriate for the specific use case and application. Such behavioral constraints should be learned offline (i.e., by training the system with data or via simulation), modified only by humans, and given a higher priority compared to online policies (outcomes that an AI system learns post-deployment, which are based on reinforcement learning or other machine learning approaches aimed at reward maximization and optimization). ______ Response to Written Questions Submitted by Hon. Maggie Hassan to Dr. Dario Gil, Ph.D. Question 1. Artificial intelligence, or AI, holds tremendous promise for individuals who experience disabilities. For example, Google and Microsoft have technologies to process language and speech and translate it into a text format to assist individuals who are deaf and hard of hearing. Other technologies will go even further to improve the lives of people with disabilities and I would like to learn more from the panel about what we can expect. What other specific technologies are you aware of in the AI space that will help people who experience disabilities? Answer. AI technologies are enabling many exciting new assistive functions by enhancing machines' ability to see, hear, interpret complex signals, and operate in the real world through action and dialog. Essential building blocks for these new capabilities are machine vision, speech to text, text to speech, natural language understanding and generation, emotion recognition, and machine learning to interpret sensor data. For example, with AI vision, it is now becoming possible to describe an image, a local environment, or a video to a person with visual impairment. Further, these technologies will soon support wearable assistants that can recognize people, objects, landmarks and obstacles in the environment, and guide a person safely to an unfamiliar destination. AI speech to text capabilities, coupled with natural language understanding, enable a quadriplegic individual to control their environment through speech commands, providing a new level of autonomy. Machine learning techniques can translate brain and nerve signals into commands for prosthetic limbs and convey a sense of touch. AI natural language understanding and generation enable communication of knowledge in the form most easily understood by an individual, whether that means generating a description of a graph for a blind person, reading text aloud for a person with dyslexia, simplifying a complex document for a person with an intellectual disability, or, one day, translating between spoken languages and sign languages used by people who are deaf and hard of hearing. AI embedded in autonomous vehicles, intelligent wheelchairs and interactive assistance robots will provide physical independence and assistance for many. For example, IBM, the CTA (Consumer Technology Association) Foundation, and Local Motors are exploring applications of Watson technologies to developing the world's most accessible self- driving vehicle, able to adapt its communication and personalize the overall experience to suit each passenger's unique needs. Machine learning on sensor data from instrumented environments can support an older adult in living independently and safely at home by learning their normal patterns of behavior and providing assistance and alerts. Just as importantly, AI technologies will benefit people with disabilities by analyzing websites and applications, finding and fixing accessibility problems in a more automated way than was previously possible. Question 2. How will manufacturers and developers work to perfect this technology so that it can truly be a reliable tool for these individuals? Answer. Perfecting AI technologies will take significant experimentation and depends on the availability of data. For these core technologies to be reliable, applications for people with disabilities should be explored as early as possible and used to drive requirements. This includes consulting with, and testing by, people with disabilities in realistic environments. At IBM, we are exploring the potential of AI technologies to support people with disabilities, and older adults through several initiatives and collaborations. IBM researchers are exploring how Watson's language-processing software could help people with cognitive disabilities by simplifying text, how older adults' patterns of activity can be learned, how a blind person can navigate and find objects effectively using machine vision, and how AI can enable our accessibility test tools to move from pointing out problems to actively suggesting solutions. Secondly, it is essential that people with disabilities are represented adequately in training data, to prevent new forms of discrimination from emerging. For example, an authentication system based on voice recognition should be able to recognize people with dysarthric speech. Manufacturers and developers applying AI technologies should incorporate mechanisms to recognize and gracefully handle exceptions, falling back on human judgment for cases that are outside their training. Question 3. What more can Congress to do assist with these efforts? Answer. For AI to deliver the promised economic and societal benefits to a broader range of people, including people with disabilities, both policy support and public investment from the U.S. Government are critical. Given the great diversity of human abilities, it is a challenge for manufacturers and developers to ensure diversity in training data. For example, speech recognition training data should ideally include people who stutter. Government investment in initiatives to make diverse data broadly available would accelerate our ability to make AI technology more inclusive, and to apply AI techniques to new accessibility problems. Government support for controlled studies with people with disabilities will also accelerate the inclusion of people with disabilities. Access to other forms of data is also critical. An indoor-outdoor navigation system for blind people relies on public outdoor maps, but indoor maps are privately owned. A centralized mechanism to share such maps for accessibility purposes would remove a practical barrier to the widespread use of such systems. AI vision techniques depend on the use of images or video to describe people and objects to people with visual impairment. Government leadership is needed to address privacy concerns of individuals and copyright concerns of organizations over the use of images of their faces or products for accessibility purposes. There is a copyright exception for converting books into braille, and a similar solution could be effective here. Secondly, policy support can help to counter the danger of new forms of discrimination. For example, reinstating the Department of Justice rulemaking on accessibility guidelines for public websites would emphasize the importance of accessibility, and spur efforts by industry to include people with disabilities in development. Most of the examples in question 1 describe ways that AI technologies can assist people with sensory or physical impairments. There is a need to foster standards and policies to address the needs of people with cognitive disabilities, which will encourage application of AI technologies to these challenges. Question 4. As we see machine learning and AI increasingly embedded in products and services that we rely on, there are numerous cases of these algorithms falling short of consumer expectations. For example, Google and Facebook both promoted fraudulent news stories in the immediate wake of the Las Vegas Shooting because of their algorithms.\2\ YouTube Kids is a service designed for children, and marketed as containing videos that are suitable for very young children. In November, YouTube Kids promoted inappropriate content due to algorithms.\3\ While the use of machine learning and AI holds limitless positive potential, at the current point, it faces challenges where we should not risk getting it wrong. Should there be any formal or informal guidelines in place for what tasks are suitable to be done by algorithms, and which are still too important or sensitive to turn over; and what more can be done to ensure better and more accurate algorithms are used as you work to better develop this technology? --------------------------------------------------------------------------- \2\ NYT: After Las Vegas Shooting, Fake News Regains Its Megaphone, Kevin Rose, 10/02/2017 https://www.nytimes.com/2017/10/02/business/las- vegas-shooting-fake-news.html \3\ NYT: On YouTube Kids, Startling Videos Slip Past Filters, Sapna Maheshwari, 11/04/2017 https://www.nytimes.com/2017/11/04/business/ media/youtube-kids-paw-patrol.html --------------------------------------------------------------------------- Answer. AI is already more capable than humans in narrow domains, some of which involve delicate decision making. Humanity is not threatened by them, but many people could be affected by their decisions. Examples are autonomous online trading agents, media and news services, and soon autonomous cars. Even though AI algorithms are usually evaluated based on their accuracy, that is, their ability to produce correct results, this is only one component of a bigger picture. We need to be able to assess the impact of their decisions in the narrow domains where they will function. To understand the suitability of an AI system with respect to performing a specific task, one must consider not only their accuracy, but also the context, the possible errors, and the consequences on the impacted communities. Furthermore, the assessment of risk should be carried out with respect to both the risk of ``doing it'' and the risk of ``not doing it'', as in many fields we already know the consequences of wrong decisions made by humans. For example, melanoma detection from skin images is a task that AI algorithms can perform at high levels of accuracy. Even though there is still a possibility of error, it is beneficial to deploy such systems in healthcare decision support, in a way that would augment human decision-making process. On the other hand, let us consider automated trading systems. A bad decision in these systems may be (and has been) a financial disaster for many people. That will also be the case for self-driving cars. Some of their decisions will be critical and possibly affect lives. Because sectors like finance and transportation can carry large risks, protections have always been in place through existing regulations. These existing protections are properly designed to provide consumer protection even with the advent of new technologies like AI. Finally, we believe that in many applications, rather than considering only fully autonomous AI solutions, the most effective approach is to build AI systems that support humans and work with them in performing a task. For example, in a breast cancer detection study, it has been shown that doctors and AI working together achieve a higher degree of accuracy than just doctors or AI separately. Question 5. Machine learning and AI hold great promise for assisting us in preventing cybersecurity attacks. According to an IBM survey of Federal IT managers, 90 percent believe that artificial intelligence could help the Federal Government defend against real- world cyber-attacks. 87 percent think AI will improve the efficiency of their cybersecurity workforce.\4\ --------------------------------------------------------------------------- \4\ INFORMATION MANAGEMENT: AI seen as key tool in government's cybersecurity defense, Bob Violino, 11/30/2017 https://www.information- management.com/news/artificial-intelligence-seen-as-key-tool-in- governments-cybersecurity-defense --------------------------------------------------------------------------- While this is promising, the Federal Government currently faces a shortage of qualified cybersecurity employees, and to make matters worse, the pipeline of students studying these topics is not sufficient to meet our needs. A recent GAO report found that Federal agencies have trouble identifying skills gaps, recruiting and retaining qualified staff, and lose out on candidate due to Federal hiring processes. The George Washington University Center for Cyber & Homeland Security recently released a report titled ``Trends in Technology and Digital Security'' which stated: ``Traditional security operations centers are mostly staffed with tier one analysts staring at screens, looking for unusual events or detections of malicious activity. This activity is similar to physical security personnel monitoring video cameras for intruders. It is tedious for humans, but it is a problem really well-suited to machine learning.'' \5\ --------------------------------------------------------------------------- \5\ https://cchs.gwu.edu/sites/cchs.gwu.edu/files/downloads/ Fall%202017%20DT%20symposi um%20compendium.pdf What effect will machine learning and AI will have on cybersecurity; and how do you think the Federal Government can best leverage the benefits offered by machine learning and AI to address our cybersecurity workforce shortage? Answer. AI and machine learning will be a disruptive force in the field of cybersecurity by providing the potential for aiding both in the defense and protection of critical infrastructure, leveling the playing field between large nation states and smaller niche players. From a defensive standpoint, AI has shown promise in automating defenses, such as probing systems for weaknesses, including software vulnerabilities and configuration errors. Penetration testing and bug finding tools have benefited tremendously from AI techniques in improving their efficiency to more quickly evaluate systems for weaknesses and increase coverage of the evaluated space. Security monitoring tools have also benefited greatly from AI and will continue to do so as AI systems improve. Automation can be leveraged to process suspicious alerts and events that warrant investigation, performing many of the rote tasks typically performed by low-level analysts. These automated tools will provide an analyst a more complete picture of the events unfolding, highlight meaningful information and context, triage, and allow the analyst to provide a higher-level response. This can allow security analysts to investigate far more alerts than are currently possible, and hopefully make fewer errors in how those alerts are processed. Security operations can be conducted at machine-scale as opposed to human-scale. The Federal Government can learn from the experience in research, industry and academia in leveraging AI to develop and deploy the next generation of AI-powered defenses that will be necessary to protect the Nation's critical infrastructure. This requires significant leadership and outreach on behalf of the Government to industry and academia on the following fronts: Declare AI Leadership in Cyber Security as a national research and development priority. Evolve and develop the Nation's cybersecurity strategy to address the AI-powered threats to critical infrastructure with AI-powered defenses. Initiate U.S. Government programs, through various policy and funding agencies (e.g., OSTP, DARPA, IARPA, NSF, NIST etc.) to fund and sponsor leading edge research in areas of intersection between AI and security Set policies and standards for procurement of next generation security controls by the U.S. Government. ______ Response to Written Question Submitted by Hon. Amy Klobuchar to Dr. Edward W. Felten, Ph.D. Question. Political ads on the Internet are more popular now than ever. In 2016, more than $1.4 billion was spent on digital advertisements and experts project that number will continue to increase. In October, I introduced the Honest Ads Act with Senators Warner and McCain, to help prevent foreign interference in future elections and improve the transparency of online political advertisements. We know that 90 percent of the ads that Russia purchased were issue ads meant to mislead and divide Americans. Increasing transparency and accountability online will benefit consumers and help safeguard future elections. Dr. Felten, can machine learning be used to help identify issue ads and stop misinformation from spreading online? Answer. Yes, machine learning can be useful in several ways. Machine learning can help to classify the nature or topic of ads, to distinguish issue ads from others and to characterize the issue being addressed by an ad. Machine learning can be useful in in determining the source of an ad, including in identifying when a single source is trying to disguise itself as a set of separate, independent sources. More broadly, machine learning can be helpful in identifying misinformation and disinformation campaigns, and in targeting countermeasures to maximize the impact on a harmful campaign while minimizing collateral damage. Three caveats are in order, however. First, more research will be necessary to take full advantage of these opportunities. That research is best done using realistic datasets derived from platforms' experience with past disinformation campaigns. Second, machine learning methods will necessarily be less than perfectly accurate. Not only will they fail to spot some disinformation campaigns, they will also sometimes misclassify content or a user as malicious when they are in fact benign. Appropriate use of machine learning in this setting will require both a careful technical evaluation of the likelihood of errors, and a policy approach that recognizes the harm that might be done by errors. Finally, machine learning systems for detecting anomalies depend on a variety of data sources and signals, and the success of machine learning depends on the characteristics of those sources and signals. Real-world data is sometimes erroneous and often incomplete, in ways that could frustrate the use of machine learning for this application or render it less accurate. Where data signals derive from the votes or clicks of users, the resulting system may be subject to gaming or manipulation, so such signals should be used with caution, especially in systems that aim to limit disinformation. ______ Response to Written Question Submitted by Hon. Tom Udall to Dr. Edward W. Felten, Ph.D. Question. In your testimony, you discussed how adoption of artificial intelligence can inadvertently lead to biased decisions. What specific steps should the Federal Government and other users take to improve the data and ensure that datasets minimize societal bias-- especially with regard to vulnerable populations? Answer. The results of a machine learning system can only be as accurate as the dataset on which the system was trained. If a community is underrepresented in the dataset, relative to its representation in the population, then that community is likely to be poorly served by the system, as the system will not put enough weight on the characteristics of the underrepresented community. Practitioners should take care to ensure that datasets are representative of the population, to the extent possible. Where this is not possible, deficiencies in the dataset should be noted carefully, and steps should be taken to mitigate the deficiencies. For example, it is sometimes possible to correct for a group's underrepresentation in a data analysis or machine learning procedure by putting greater weight on data points that represent that group. Additional statistical methods exist that can counteract the effect of non-representative datasets. Another common source of error or bias in machine learning occurs when a system is tasked with learning from examples of past decisions made by people. If those past decisions were biased, the machine is likely to learn to replicate that bias. Whenever a system is trained based on past human decisions, care should be taken to consider the social and historical context of those past decisions and to look for indications of bias in the system's output, and anti-bias techniques should be used in designing or training the system if possible. In addition to technical measures in the design and use of AI systems, the possibilities of bias--whether that AI will introduce bias or that AI will open opportunities to measure and counteract human bias--should be taken into account in making policy decisions. Consulting with technical experts, and including technical expertise in the policymaking conversation, are important steps toward good policy in this area. ______ Response to Written Question Submitted by Hon. Gary Peters to Dr. Edward W. Felten, Ph.D. Question. I am concerned by recent reports in Nature, The Economist, and Wall Street Journal about large tech firms monopolizing the talent in AI and machine learning. This concentration of talent can lead to several negative outcomes including long-term wage stagnation and income inequality. In your opinion, what steps or incentives might mitigate this concentration, encourage AI-experts to work at small and medium enterprises, or launch their own start-up with the goal of growing a business (rather than having a goal of being bought out by one of the tech giants)? Similarly, what incentives might encourage AI experts to become educators and trainers to help develop the next generation of AI experts? How can the Federal Government compete with the tech giants to attract experts needed to develop and implement AI systems for defense and civil applications? Answer. One approach to making smaller companies attractive to top AI talent is to adopt pro-competition policies generally. Because AI often relies on large datasets, and those datasets are more likely to be held by large companies, there may be a natural tendency toward concentration in AI-focused industry sectors. Public policy can help to ensure that smaller companies can be viable competitors. The Federal Government can also provide some large, high quality datasets that may be useful to individuals and companies of all sizes. At present, the demand for highly skilled AI experts exceeds the supply, leading to a scarcity of those experts in all but the best- funded environments. In the long run, steps to increase the education and training of AI professionals are the most important means to strengthen our national talent base and broaden the availability of expertise. The talent pipeline can be widened at every stage. At the K-12 level, access to a good computer science course should be available to every student. A bipartisan coalition of states and nonprofit actors is working toward this goal. At the university and graduate level, access to education is limited by the number of trained faculty available to teach advanced AI and machine learning courses. It is difficult to underestimate the importance of supporting a large and robust public research community. This ensures that access to the latest knowledge and techniques in AI is available to the public and not limited to a few companies' researchers. It widens the talent pipeline because AI research funding enables faculty hiring in AI, which increases the national capacity to train AI leaders. Federally- funded research projects serve as the main training ground for the next generation of research leaders. ______ Response to Written Questions Submitted by Hon. Maggie Hassan to Dr. Edward W. Felten, Ph.D. Question 1. Artificial intelligence, or AI, holds tremendous promise for individuals who experience disabilities. For example, Google and Microsoft have technologies to process language and speech and translate it into a text format to assist individuals who are deaf and hard of hearing. Other technologies will go even further to improve the lives of people with disabilities and I would like to learn more from the panel about what we can expect. What other specific technologies are you aware of in the AI space that will help people who experience disabilities? Answer. There are many examples, of which I will highlight three here. First, self-driving vehicles will improve mobility and lower the cost of transportation for people who are unable to drive. These vehicles will have major safety benefits in the long run, and they are already starting to benefit people with disabilities. Maintaining policies to encourage safety-conscious testing and deployment of self- driving vehicles will benefit all Americans, and especially those with disabilities. Second, computer vision and image interpretation systems have the potential to help those with visual disabilities process information about their surroundings. These systems are demonstrating an increasing capacity to identify specific objects and people in complex scenes, and to model and predict what might happen next, such as warning of potential dangers. Third, AI can help to identify barriers to accessibility. For example, Project Sidewalk at the University of Maryland combines crowdsourced data collection with AI techniques to build a database of curb, ramp, and sidewalk locations, and analyze it to identify accessibility problems. This can help city planners and property owners recognize accessibility failures. Question 2. How will manufacturers and developers work to perfect this technology so that it can truly be a reliable tool for these individuals? Answer. As with any new, complex technology, careful testing is needed to understand the implications of using a system. Such testing must be done in a realistic environment and must involve the community of potential users. The best design practices are user-centered, meaning that the potential user community for a product is involved throughout the design process, from initial concept exploration through final testing. This is especially important if the designer might experience the world differently than the user community. Question 3. What more can Congress to do assist with these efforts? Answer. Three significant things that Congress can do are (1) provide funding for research on applications of AI for use by people with disabilities; (2) work with agencies to ensure they are giving proper attention to these issues and the interests of people with disabilities; and (3) highlight the need for work in this area and highlight the successes of those already working in the area. Question 4. [citations omitted] As we see machine learning and AI increasingly embedded in products and services that we rely on, there are numerous cases of these algorithms falling short of consumer expectations. For example, Google and Facebook both promoted fraudulent news stories in the immediate wake of the Las Vegas Shooting because of their algorithms. YouTube Kids is a service designed for children, and marketed as containing videos that are suitable for very young children. In November, YouTube Kids promoted inappropriate content due to algorithms. While the use of machine learning and AI holds limitless positive potential, at the current point, it faces challenges where we should not risk getting it wrong. Should there be any formal or informal guidelines in place for what tasks are suitable to be done by algorithms, and which are still too important or sensitive to turn over; and what more can be done to ensure better and more accurate algorithms are used as you work to better develop this technology? Answer. In considering a switch from human to AI based decision making, we should not demand perfection of the AI system. The alternative to AI is often to rely on human judgment, which is also prone to bias and mistakes. Instead of demanding perfection of the AI system, an organization needs to understand the potential consequences of adopting the AI system well enough to conclude with justified confidence that switching to an automated system is an improvement and on balance the effects are benign. There are also considerations of scale. AI can operate at larger scale (that is, on larger number amounts of data or more decisions per second) than any human organization can hope to achieve. As a result, in some cases the choice in designing a function is not between using AI and using a human, but rather between using AI and not providing the function at all. The tremendous value provided by many of today's information, communication, and publishing tools relies at least in part on the use of AI. That said, the potential risks of AI-based systems must be considered and addressed. It is too early to establish formal guidelines, because not enough is known about how best to address these problems. Informal guidelines are needed, and the industry and other stakeholders should be encouraged to develop them collaboratively. Multi-stakeholder groups such as the Partnership on AI may be useful venues for these discussions. The guidelines, best practices, and technical tools to address these problems will evolve with time. Switching a process based on human decision-making to one based on AI can have unpredictable consequences, so experimentation is needed in a safe environment to adequately understand the implications of such a change before it is made. Organizations can be more transparent by publishing information about what kinds of testing and analysis were done in preparation for the introduction of AI into an existing process, thereby enabling stakeholders to understand why a change was made and query the organization if concerns remain. In many cases, an organization will have valid reasons, such as trade secrets or user privacy, to refrain from publishing the full details of how a system works. This does not preclude the organization from publishing information about how it tested the system and evaluated the pros and cons of adopting it. Question 5. [citations omitted] Machine learning and AI hold great promise for assisting us in preventing cybersecurity attacks. According to an IBM survey of Federal IT managers, 90 percent believe that artificial intelligence could help the Federal Government defend against real-world cyber-attacks. 87 percent think AI will improve the efficiency of their cybersecurity workforce. While this is promising, the Federal Government currently faces a shortage of qualified cybersecurity employees, and to make matters worse, the pipeline of students studying these topics is not sufficient to meet our needs. A recent GAO report found that Federal agencies have trouble identifying skills gaps, recruiting and retaining qualified staff, and lose out on candidate due to Federal hiring processes. The George Washington University Center for Cyber & Homeland Security recently released a report titled ``Trends in Technology and Digital Security'' which stated: ``Traditional security operations centers are mostly staffed with tier one analysts staring at screens, looking for unusual events or detections of malicious activity. This activity is similar to physical security personnel monitoring video cameras for intruders. It is tedious for humans, but it is a problem really well-suited to machine learning.'' What effect will machine learning and AI will have on cybersecurity; and how do you think the Federal Government can best leverage the benefits offered by machine learning and AI to address our cybersecurity workforce shortage? Answer. As the GWU report suggests, cybersecurity tasks of a routine nature can be automated, thereby reducing the need for human operators do to lower-level work. This can free up workers to concentrate on higher-level tasks requiring more skill and judgment, and can help to mitigate the Federal Government's cybersecurity personnel shortage. Notwithstanding these opportunities, the Federal Government will continue to face challenges in recruiting and retaining the best technology talent. Many proposals exist to address these challenges by improving hiring authorities, pay scales, and working conditions for Federal technology workers, and by instituting or expanding training and scholarship-for-service programs. Question 6. Mr. Felten, as we heard you recently served as the deputy U.S. Chief Technology Officer at the White Office of Science and Technology Policy. One of the projects you worked on in that office was a major report on Artificial Intelligence. That report was one of the many important projects taken on by the Office of Science and Technology Policy in recent years. And it's extremely disappointing that President Trump has failed to nominate leaders for that office, now more than ten months into his presidency. That's the longest a president has gone without a science advisor since the Office of Science and Technology Policy was established in law in 1976. I've led two letters to President Trump urging him to nominate well-qualified experts to lead this office, but so far we have seen nothing from this administration. As a former leader at the Office of Science and Technology Policy, could you please explain why this office is important, and what kinds of qualities you look for in good nominees for this office? Answer. OSTP's importance derives from the central role of science and technology in nearly every area of policy. In major policy areas such as national security and defense, transportation, education, and the economy, technology is critical to the most important challenges and opportunities. Making the best decisions in these areas requires input from and dialog with the technical community. Congress assigned that role within the White House to OSTP. AI is just one of the areas of interest for OSTP, but it connects to many important policy questions. What should DoD's policy be on autonomous weapons systems, and what position should the United States take in international talks about such weapons? How will AI-driven automation affect the job market, and how can American schoolchildren and adults be educated and trained for the future workplace? What needs to be done to improve highway safety as automated vehicles become practical? How can American farmers, journalists, and businesses be freed to use drones, while strengthening our defenses against potential terrorist uses of the same technology? How will changes in information technology affect the mission of the Intelligence Community, and what kinds of people and capabilities will the IC need in the future? How will cybersecurity concerns affect all of these goals? Each of these questions can be better answered with the help of technical advisors who have deep domain knowledge, connections to the relevant technical communities, and a seat at the policy table. A successful OSTP Director will be a trusted advisor to the President and the President's senior advisors, a liaison to departments and agencies on science and technology issues, and an ambassador to scientific and technical communities in the United States and around the world. A candidate for OSTP Director should be a highly respected member of the scientific/technical community, with a reputation for technical knowledge and policy judgment. The candidate should be able to work successfully across disciplines, acquiring knowledge and providing advice across many subject areas with appropriate staff support. They should be able to work successfully within the unusual administrative and legal environment of the White House, and they should be able to recruit, motivate, and lead a team of highly-skilled domain experts and policy advisors. Because the subject matter of science and technology is so extensive, and the United States is blessed with leading experts in so many specialties, no one person can hope to have the knowledge, experience, and connections needed to provide advice in all technical areas. A successful OSTP Director will recruit a team of topic-area advisors who can provide context and guidance in specific areas and can expand OSTP's ``surface area'' in coordinating with agencies, outside experts, and the public. [all]
MEMBERNAME | BIOGUIDEID | GPOID | CHAMBER | PARTY | ROLE | STATE | CONGRESS | AUTHORITYID |
---|---|---|---|---|---|---|---|---|
Wicker, Roger F. | W000437 | 8263 | S | R | COMMMEMBER | MS | 115 | 1226 |
Blunt, Roy | B000575 | 8313 | S | R | COMMMEMBER | MO | 115 | 1464 |
Moran, Jerry | M000934 | 8307 | S | R | COMMMEMBER | KS | 115 | 1507 |
Thune, John | T000250 | 8257 | S | R | COMMMEMBER | SD | 115 | 1534 |
Baldwin, Tammy | B001230 | 8215 | S | D | COMMMEMBER | WI | 115 | 1558 |
Udall, Tom | U000039 | 8260 | S | D | COMMMEMBER | NM | 115 | 1567 |
Capito, Shelley Moore | C001047 | 8223 | S | R | COMMMEMBER | WV | 115 | 1676 |
Capito, Shelley Moore | C001047 | 8223 | S | R | COMMMEMBER | WV | 115 | 1676 |
Cantwell, Maria | C000127 | 8288 | S | D | COMMMEMBER | WA | 115 | 172 |
Klobuchar, Amy | K000367 | 8249 | S | D | COMMMEMBER | MN | 115 | 1826 |
Heller, Dean | H001041 | 8060 | S | R | COMMMEMBER | NV | 115 | 1863 |
Peters, Gary C. | P000595 | 7994 | S | D | COMMMEMBER | MI | 115 | 1929 |
Gardner, Cory | G000562 | 7862 | S | R | COMMMEMBER | CO | 115 | 1998 |
Young, Todd | Y000064 | 7948 | S | R | COMMMEMBER | IN | 115 | 2019 |
Blumenthal, Richard | B001277 | 8332 | S | D | COMMMEMBER | CT | 115 | 2076 |
Lee, Mike | L000577 | 8303 | S | R | COMMMEMBER | UT | 115 | 2080 |
Johnson, Ron | J000293 | 8355 | S | R | COMMMEMBER | WI | 115 | 2086 |
Duckworth, Tammy | D000622 | S | D | COMMMEMBER | IL | 115 | 2123 | |
Schatz, Brian | S001194 | S | D | COMMMEMBER | HI | 115 | 2173 | |
Cruz, Ted | C001098 | S | R | COMMMEMBER | TX | 115 | 2175 | |
Fischer, Deb | F000463 | S | R | COMMMEMBER | NE | 115 | 2179 | |
Booker, Cory A. | B001288 | S | D | COMMMEMBER | NJ | 115 | 2194 | |
Sullivan, Dan | S001198 | S | R | COMMMEMBER | AK | 115 | 2290 | |
Cortez Masto, Catherine | C001113 | S | D | COMMMEMBER | NV | 115 | 2299 | |
Hassan, Margaret Wood | H001076 | S | D | COMMMEMBER | NH | 115 | 2302 | |
Inhofe, James M. | I000024 | 8322 | S | R | COMMMEMBER | OK | 115 | 583 |
Markey, Edward J. | M000133 | 7972 | S | D | COMMMEMBER | MA | 115 | 735 |
Nelson, Bill | N000032 | 8236 | S | D | COMMMEMBER | FL | 115 | 859 |
Disclaimer:
Please refer to the About page for more information.