
In a move to regulate the expanding use of artificial intelligence (AI) in federal agencies, the White House has announced stringent measures aimed at safeguarding Americans’ rights and ensuring safety.
The directive, issued by the Office of Management and Budget (OMB) on Thursday, mandates federal agencies to adopt concrete safeguards by December 1, as reported by Reuters.
Under the new guidelines, agencies utilizing AI technologies are obligated to monitor, assess, and test the impacts of AI on the public. Additionally, efforts must be made to mitigate the risks of algorithmic discrimination while providing transparent insights into the government’s AI usage. This entails conducting thorough risk assessments and establishing operational and governance metrics to ensure accountability and transparency.
President Joe Biden had previously signed an executive order in October, invoking the Defense Production Act to compel developers of AI systems posing risks to national security, economy, public health, or safety to share safety test results with the U.S. government prior to public release.
Related: White House Pushes for Pro-Small Business AI Policy
The White House emphasized that the implementation of these safeguards is crucial, particularly in instances where AI deployment could impact Americans’ rights or safety. Detailed public disclosures regarding the usage of AI by the government will be made to ensure transparency and accountability.
Notable provisions include the ability for air travelers to opt-out from Transportation Security Administration (TSA) facial recognition screenings without delay and the requirement for human oversight in federal healthcare systems where AI supports diagnostic decisions.
Generative AI, which has raised both excitement and concerns, particularly regarding job displacement and potential societal upheavals, is also addressed in the directive. Government agencies are now mandated to release inventories of AI use cases, report metrics on AI usage, and disclose government-owned AI code, models, and data, provided they do not pose significant risks.
The Biden administration underscored the ongoing utilization of AI across various federal agencies. For instance, the Federal Emergency Management Agency (FEMA) employs AI to assess structural hurricane damage, while the Centers for Disease Control and Prevention (CDC) utilizes AI for disease spread prediction and opioid use detection. Additionally, the Federal Aviation Administration (FAA) leverages AI to enhance air traffic management in major metropolitan areas, ultimately improving travel efficiency.
Source: Reuters
Featured News
House Judiciary Committee Examines Antitrust Issues in Medical Residency Market
Mar 17, 2025 by
CPI
Belgian Competition Authority Accuses Roche of Anticompetitive Practices in Cancer Drug Market
Mar 17, 2025 by
CPI
PepsiCo to Acquire Prebiotic Soda Brand Poppi for $1.95 Billion
Mar 17, 2025 by
CPI
Tech Companies Face Tightened Online Safety Rules in the UK, Starting Monday
Mar 17, 2025 by
CPI
UK Government to Tighten Merger Scrutiny in New Plan by Chancellor Reeves
Mar 17, 2025 by
CPI
Antitrust Mix by CPI
Antitrust Chronicle® – Self-Preferencing
Feb 26, 2025 by
CPI
Platform Self-Preferencing: Focusing the Policy Debate
Feb 26, 2025 by
Michael Katz
Weaponized Opacity: Self-Preferencing in Digital Audience Measurement
Feb 26, 2025 by
Thomas Hoppner & Philipp Westerhoff
Self-Preferencing: An Economic Literature-Based Assessment Advocating a Case-By-Case Approach and Compliance Requirements
Feb 26, 2025 by
Patrice Bougette & Frederic Marty
Self-Preferencing in Adjacent Markets
Feb 26, 2025 by
Muxin Li