Chairwoman Mace, Ranking Member Connolly, and members of the Subcommittee:

Thank you for holding this hearing on “White House Overreach on AI” and for the invitation to appear before you. My name is Adam Thierer and I am a senior fellow at the R Street Institute, where I focus on emerging technology issues. I also recently served as a commissioner on the U.S. Chamber of Commerce’s Commission on Artificial Intelligence Competitiveness, Inclusion, and Innovation.[1]

My message today boils down to three main points:

  1. First, the United States needs a national pro-innovation policy framework for artificial intelligence (AI) that builds on the winning bipartisan vision Congress crafted for the internet and digital technology in the 1990s.
  1. Second, if we fail to get this AI policy framework right, it will have profound consequences for America’s global competitive advantage and geopolitical security.
  1. Third, the Biden administration’s recent AI Executive Order is no substitute for this needed national policy framework and, worse yet, it could undermine that objective.

I will briefly elaborate on each point.

The U.S. Vision Bore Rich Fruit

First, it is important to recall the foundational principles behind the bipartisan national framework for digital commerce that Congress and the Clinton administration crafted a quarter century ago.[2] Freedom to innovate was made America’s policy default. Lawmakers rejected the inefficient old regulatory models of the analog era, which constrained entrepreneurialism and competition. We allowed new digital technologies to be “born free” and to flourish without excessive micromanagement, and then used ongoing multistakeholder efforts and flexible regulatory responses to address concerns.[3]

Europe took the opposite path of heavy-handed technocratic mandates and has “regulate[d] its way to last place,” as a recent Wall Street Journal headline observed.[4] In fact, 18 of the 25 largest digital companies in the world today are U.S.-based, while it is difficult to name any from Europe.[5] While some people have concerns about large technology companies today, we should agree that it is better that these firms are primarily based in the United States instead of China, Europe, or any other country or continent.[6]

Good Policy Creates a Strong Technology Base

Further, there is a second point about the connection between AI policy and broader national objectives. A strong digital technology base is an important source of strength and prosperity, so it is essential that our nation not shoot itself in the foot as the next great technological race gets underway with China and the rest of the world.[7]

Consider this scenario. When OpenAI launched ChatGPT in late 2022, it quickly became the most rapidly adopted digital technology in history, and competing AI services from U.S. developers followed quickly.[8] Had a Chinese operator launched a major generative AI model first, it would have been a “Sputnik moment” for America. Luckily, it is instead foreign nations who are today left scratching their heads and wondering how America once again raced ahead of them on digital technology.

Wise policy choices not only strengthen our economy and provide better products and jobs, but also bolster national security and allow our values to shape information technology platforms and markets globally.[9]

We need a national AI policy that is flexible and pro-innovation to make sure our firms, workers, and values continue to lead the world in this fashion.[10]

AI Policy by Executive Fiat is Bad Policy

This brings me to the Biden administration’s October executive order.[11] This wide-ranging 100-plus page directive has been praised by some as a logical response to congressional inaction on AI, but many others have rightly noted that it stretches executive authority over emerging technology well beyond statutory limits and raises the danger of overregulation.[12]

For example, the order flips the Defense Production Act on its head and converts a 1950s law meant to encourage production, into an expansive regulatory edict intended to curtail some forms of algorithmic innovation.[13] Twenty state attorneys general recently filed a letter with the Department of Commerce noting how the order, “is about regulating technological development, not about encouraging the production of anything,” and also objecting to its effort, “to centralize governmental control over an emerging technology being developed by the private sector.”[14]

The order also contains open-ended language about taking steps to combat “algorithmic discrimination,” and pushes the Federal Trade Commission to get more aggressive in policing the AI marketplace. These steps open the door to a new regulatory regime for AI without any express authority from Congress.[15]

While other provisions of the order are more reasonable, Congress still needs to reassert itself to ensure that administrative overreach is curtailed and that agencies adhere to the Constitution and their congressionally designated powers.[16]

The AI Framework America Needs

Instead of arbitrary executive mandates, Congress needs to craft an AI policy vision that:

  1. is rooted in a flexible, risk-based framework that relies more on ongoing multistakeholder negotiations and evolutionary standards that are more closely matched to rapidly changing algorithmic technologies;[17]
  2. builds on existing government powers on a sectoral basis instead of trying to develop an entirely new regulatory superstructure for AI;[18]
  3. preempts state and local government AI laws that create a confusing patchwork of conflicting mandates;[19] and, most importantly,
  4. gives algorithmic entrepreneurs a green light and avoids treating new AI services as “guilty until proven innocent,” as the executive order does.[20]

In sum, our nation must create a positive innovation culture and avoid trapping our AI innovators in a regulatory cage if we hope to prosper economically and ensure a safer, more secure technological base.[21]

It is essential that we strike the right policy balance as we face serious competition from China and other nations who are looking to counter America’s early lead in computational systems and data-driven digital technologies.

Thank you for holding the hearing and for your consideration of my views. I look forward to any questions you may have.

See pages 5-8 for enclosure of“Artificial Intelligence Task Force: 10 Principles to Guide AI Policy,” R Street Institute,Feb. 21, 2024. https://www.rstreet.org/commentary/artificial-intelligence-task-force-10-principles-to-guide-ai-policy.

Artificial Intelligence Task Force: 10 Principles to Guide AI Policy

By Adam Thierer

Feb. 21, 2024

On Feb. 20, House Speaker Mike Johnson and Democratic Leader Hakeem Jeffries announced the creation of a new bipartisan Task Force on artificial intelligence (AI) that will be chaired by Rep. Jay Obernolte (R-Calif.) and Rep. Ted Lieu (D-Calif.). This task force “will seek to produce a comprehensive report that will include guiding principles, forward-looking recommendations and bipartisan policy proposals developed in consultation with committees of jurisdiction.”

Although other AI task forces have already been established, this latest one represents a sensible step forward for Congress, and the bipartisan nature of the new task force makes it even more important. The choice of Reps. Obernolte and Lieu to co-chair the task force is sensible because they have shown thoughtful leadership on AI policy issues and have been regular speakers on the issue at hearings and events.

It remains unclear whether this Congress can craft any AI policy with the legislative clock ticking fast during this election year. Nonetheless, Congress needs to create a governance framework for AI that both ensures that citizens can enjoy the enormous benefits associated with advanced algorithmic technologies and preserves the United States’ position as a world leader in digital technology as China and other nations race to catch up.

Toward that end, here are 10 principles the new congressional AI task force should consider to help craft a sensible AI policy framework for America.

1) Reiterate that the freedom to innovate remains the default for U.S. digital policy. U.S. policy should give entrepreneurs a green light to experiment with bold, new ideas without having to seek permission to innovate. What separates the United States from China, the European Union, and other governments is that our technology policy creates a positive innovation culture, not an innovation cage, for digital entrepreneurialism. We must keep it that way if we are to once again lead in the next great technological revolution. The new AI task force should build on the Clinton administration’s 1997 Framework for Global Electronic Commerce, which charted a principled, pro-innovation vision like this and inspired a generation of visionaries to become digital technology leaders in many different global technology sectors.

Recommended reading: Getting AI Innovation Culture Right

2) Ensure that AI policy remains rooted in a flexible, risk-based framework that relies on ongoing multistakeholder negotiations and evolutionary standards. Our approach must be able to closely match rapidly changing algorithmic technologies. The National Telecommunications and Information Administration and other agencies have brought together diverse stakeholders repeatedly to hammer out solutions to complicated digital technology problems, and the National Institute of Standards and Technology has taken important steps through its AI Risk Management Framework to help developers craft standards and solutions to address potential algorithmic risks in a flexible fashion. This is the agile and iterative governance model that America used for internet policy issues over the past quarter century. AI policy should continue to encourage private developers to work with other stakeholders to refine best practices and ethical guidelines for algorithmic technologies, without imposing heavy-handed government mandates preemptively. 

Recommended reading: Flexible, Pro-Innovation Governance Strategies for Artificial Intelligence

3) Target regulation by using a sectoral approach that breaks down AI policy into smaller, more manageable chunks. As its first order of business, the new AI task force should inventory the extensive regulatory capacity that already exists rather than attempt to develop an entirely new regulatory superstructure for AI. Former National Economic Director Bill Whyman, who spent his career at the intersection of finance, emerging technologies, and government policy, notes that, while the United States “is not likely to pass a broad national AI law over the next few years,” it is important to understand that, “no broad national law does not mean no regulation.” Indeed, this is already how digital policy works. America does not have a Federal Computer Commission for computing or the internet but instead relies on the wide variety of laws, regulations, and agencies that existed long before digital technologies came along. The federal government has 436 federal departments, and many of them are already aggressively investigating how to oversee AI developments in their areas. In fact, some of them may already be over-regulating AI technologies. If this new AI task force wants to help Congress get something done in the short term, it should recommend that federal policymakers break AI policy down into its smaller subcomponents and then prioritize among them. For example, it would make more sense to address data privacy and driverless car issues in separate legislation, and plenty of broad, bipartisan support exists for measures in those fields. Efforts to incorporate those issues and everything else in one mega-bill are a recipe for legislative failure.

Recommended reading: Artificial Intelligence Legislative Outlook: Fall 2023 Update

4) When considering AI policies, focus on the outputs of algorithmic systems instead of the inputs into them. Regardless of how AI policy takes shape, it is crucial that lawmakers make it clear that the focus of regulation will be on algorithmic outputs or outcomes, not inputs or processes. Rep. Obernolte has repeatedly stressed this point in recent talks and essays when explaining why it is essential that policymakers avoid AI mandates “that stifle innovation by focusing on mechanisms instead of on outcomes.” In other words, policy should focus on how AI technologies perform and whether they do so in a generally safe manner. Too much of the AI policy discussion today is instead focused on hypothetical, worst-case scenarios pulled straight from the dystopian plots of science fiction stories. Those fear-based narratives then prompt calls for preemptive regulation of computational processes and treat AI innovations as guilty until proven innocent. That is the wrong standard for AI policy. As Rep. Obernolte correctly recommends, we should focus on real-world outputs and outcomes and then judge them accordingly.

Recommended reading: The Most Important Principle for AI Regulation

5) Ensure that AI is being defined in a sensible and consistent fashion, and avoid efforts that require all algorithmic systems to be “explainable.” Perhaps the greatest of all policy challenges relating to AI is simply defining the term. “There is no single universally accepted definition of AI, but rather differing definitions and taxonomies,” a U.S. Government Accountability Office report noted in 2018. There may be no escaping this definitional dilemma, but lawmakers must do their best to ensure that definitions are as clear and consistent as possible. Because AI concerns and regulations will vary widely by sector and context, the easiest way to avoid definitional confusion is to keep AI policy focused on targeted, sectoral approaches. It is equally important that lawmakers not demand that all AI systems be perfectly “explainable” in terms of how they operate. Algorithmic models are highly complex, and it would be virtually impossible to explain everything that went into creating them or explaining how they reasoned their way to certain conclusions. Again, this is why it is important to focus on real-world outcomes instead of underlying processes.

Recommended reading: Comments of the R Street Institute to the National Telecommunications and Information Administration (NTIA) on AI Accountability Policy

6) Examine how to preempt state and local government AI regulations that would impede the development of a robust national marketplace in algorithmic systems. While Congress is moving slowly on AI policy, many states and localities are aggressively advancing new AI proposals. Some parochial AI regulation will obviously be enacted, but a confusing patchwork of inconsistent state and local mandates would encumber algorithmic innovation across the nation. While Congress cannot completely preempt all state and local AI regulatory activity, it can set some guidelines on what types of regulation might impinge upon interstate commerce or speech. Again, this is why breaking AI policy into smaller chunks makes even more sense; in some areas (like autonomous vehicles, drones, or intellectual property) preemption may be easier than others (such as insurance markets, policing issues, or education policy).

Recommended reading: State and local meddling threatens to undermine the AI revolution

7) Ensure continued liability protections for online speech and commerce by preserving Section 230 of the Telecommunications Act of 1996. This provision has played an essential role in protecting digital speech and commerce and has allowed the internet to flourish without fear of frivolous lawsuits being filed at every juncture. It remains unclear whether Section 230 as written applies to generative AI systems, but a good case can be made that it should. Unfortunately, some lawmakers have been trying to gut Section 230—or at least make sure it does not extend to next-generation digital systems. The new AI task force should carefully consider how such a move would undermine algorithmic innovation by unleashing a flood of litigation against algorithmic innovators. Optimally, Section 230 protections would be extended to cover new AI systems, but, at a minimum, lawmakers must ensure that the law is preserved for existing speech and commerce.

Recommended reading: Without Section 230 Protections, Generative AI Innovation Will Be Decimated

8) Provide more resources for anti-fraud efforts or criminal enforcement involving algorithmic capabilities where needed. The most important AI enforcement tools will likely be existing consumer protections, anti-fraud laws, and other mechanisms for the recall authority possessed by some federal regulatory agencies. In most cases, these agencies already have the right tools to address algorithmic problems that could develop, but they might need additional resources or training to address them more effectively. The new AI task force can identify where those gaps might exist.

9) Ensure an open door for global AI firms, talent, and investment. Online commerce exploded in the United States because the nation opened its doors to skilled immigrants and global investors who were eager to come here to enjoy the benefits of vibrant markets, world-class higher education institutions, and research facilities. In essence, the United States attracted the world’s best and brightest away from other nations by simply providing boundless opportunities through the general freedom to innovate. America needs to double down on that winning approach. The Biden administration’s recent AI executive order included some sensible steps to improve skilled immigration policy, and the new AI task force should explore how to follow through on efforts to attract even more talent and investment to our shores.

10) Carefully evaluate the strategic importance of AI systems and their role in enhancing America’s national security and geopolitical standing. It is now widely accepted that AI has important ramifications not only for global competitiveness, but also for national security and cybersecurity. An earlier task force, the National Security Commission on Artificial Intelligence, produced a major 2021 report on these issues and concluded that “America is not prepared to defend or compete in the AI era.” The new AI task force should explore how a strong digital technology base is an important source of both national prosperity and security. It is essential that the United States be a leader in AI to counter China and other nations attempting to overtake U.S. innovation in next-generation computational systems. It is especially important for American companies to once again lead this technological revolution to ensure that our values can shape information technology platforms and markets going forward. This is why our nation must get AI policy right and not shoot itself in the foot as the next great technological race gets underway with China and the rest of the world.

Recommended reading: Existential Risks & Global Governance Issues around AI & Robotics


[1] Commission on Artificial Intelligence Competitiveness, Inclusion, and Innovation, Commission on Artificial Intelligence Competitiveness, Inclusion, and Innovation: Report and Recommendations, U.S. Chamber of Commerce, March 2023. https://www.uschamber.com/assets/documents/CTEC_AICommission2023_Report_v6.pdf.

[2] “The Framework for Global Electronic Commerce,” The White House, July 1997. https://clintonwhitehouse4.archives.gov/WH/New/Commerce.

[3] Adam Thierer, “Getting AI Innovation Culture Right,” R Street Policy Study No. 281 (March 30, 2023). https://www.rstreet.org/research/getting-ai-innovation-culture-right.

[4] Greg Ip, “Europe Regulates Its Way to Last Place,” The Wall Street Journal, Jan. 31, 2024. https://www.wsj.com/economy/europe-regulates-its-way-to-last-place-2a03c21d.

[5] “Largest tech companies by market cap,” Companies Market Cap, last accessed Feb. 1, 2024. https://companiesmarketcap.com/tech/largest-tech-companies-by-market-cap.

[6] Daniel Gouré, “The New Arsenal of Democracy: The U.S. Commercial High-Tech Industry’s Role In Countering The China Threat,” Lexington Institute, June 24, 2022. https://www.lexingtoninstitute.org/the-new-arsenal-of-democracy-the-u-s-commercial-high-tech-industrys-role-in-countering-the-china-threat.

[7] James Pethokoukis, “What’s Really at Stake If We Get AI Regulation Wrong,” Faster, Please! Oct. 30, 2023. https://www.aei.org/articles/whats-really-at-stake-if-we-get-ai-regulation-wrong.

[8] Andrew Tarantola, “How OpenAI’s ChatGPT has changed the world in just a year,” Engadget, Nov. 30, 2023. https://www.engadget.com/how-openais-chatgpt-has-changed-the-world-in-just-a-year-140050053.html.

[9] Loren B. Thompson, “Why U.S. National Security Requires A Robust, Innovative Technology Sector,” Lexington Institute, Oct. 8, 2020. https://www.lexingtoninstitute.org/why-u-s-national-security-requires-a-robust-innovative-technology-sector.

[10] Adam Thierer, “Flexible, Pro-Innovation Governance Strategies for Artificial Intelligence,” R Street Policy Study No. 283 (April 20, 2023). https://www.rstreet.org/research/flexible-pro-innovation-governance-strategies-for-artificial-intelligence.

[11] Adam Thierer, “White House Executive Order Threatens to Put AI in a Regulatory Cage,” R Street Institute, Oct. 30, 2023. https://www.rstreet.org/commentary/white-house-executive-order-threatens-to-put-ai-in-regulatory-cage.

[12]  Mohar Chatterjee and Brendan Bordelon, “The campaign to take down the Biden AI executive order,” Politico, Jan. 26, 2024. https://www.politico.com/news/2024/01/25/conservatives-prepare-attack-on-bidens-ai-order-00137935.

[13] “AFP Foundation Files Lawsuit Over President Biden’s AI Overreach,” Americans for Prosperity, Jan. 26, 2024. https://americansforprosperity.org/afp-foundation-files-lawsuit-over-president-bidens-ai-overreach/amp. Will Rinehart, “Unpacking the Executive Order on AI,” The Exformation Newsletter, Nov. 9, 2023.  https://exformation.williamrinehart.com/p/unpacking-the-executive-order-on.

[14] “Any regulation must comport with the Constitution including only authorized executive action, as well as protecting against government censorship,” the letter also notes. See Utah Office of the Attorney General, “AG Reyes Leads 20-State Letter Asking AI to be Regulated by the People, Not Politics,” State of Utah, Feb. 5, 2024. https://attorneygeneral.utah.gov/ag-reyes-leads-20-state-letter-asking-ai-to-be-regulated-by-the-people-not-politics.

[15] Thierer. https://www.rstreet.org/commentary/white-house-executive-order-threatens-to-put-ai-in-regulatory-cage.

[16] If nothing else, we should remember than an executive order is simply a statement of direction by a sitting president to his agencies and it can be changed with the stroke of a pen by a different president. This is all the more reason why the legislative branch should establish a more stable policy framework.

[17] Rep. Jay Obernolte, “The Role of Congress in Regulating Artificial Intelligence,” The Ripon Forum 57:3 (June 2023). https://riponsociety.org/article/the-role-of-congress-in-regulating-artificial-intelligence. Adam Thierer, “The Most Important Principle for AI Regulation,” R Street Institute, June 21, 2023. https://www.rstreet.org/commentary/the-most-important-principle-for-ai-regulation.

[18] Many agencies are already addressing algorithmic systems that touch their fields using existing regulatory tools like product recall authority, unfair and deceptive practices law, and civil rights laws. Our government already has 2.2 million civilian workers spread across 436 federal departments. We should give those existing agencies and processes a chance to work before adding another federal bureaucracy or burdensome licensing regime to the mix. The courts and our common law system can also address algorithmic risks that are unforeseeable but develop over time. See Adam Thierer and Neil Chilson, “The Problem with AI Licensing & an ‘FDA for Algorithms,’” The Federalist Society, June 5, 2023. https://fedsoc.org/commentary/fedsoc-blog/the-problem-with-ai-licensing-an-fda-for-algorithms.

[19] Adam Thierer, “State and local meddling threatens to undermine the AI revolution,” The Hill, Jan. 21, 2024. https://thehill.com/opinion/4420144-state-and-local-meddling-threatens-to-undermine-the-ai-revolution.

[20] Steven Sinofsky, “Regulating AI by Executive Order is the Real AI Risk,” Hardcore Software by Steven Sinofsky, Nov. 1, 2023. https://hardcoresoftware.learningbyshipping.com/p/211-regulating-ai-by-executive-order.

[21]  Adam Thierer, “A balanced AI governance vision for America,” The Hill, April 16, 2023. https://thehill.com/opinion/congress-blog/3953916-a-balanced-ai-governance-vision-for-america.