Close Menu
Technology Mag

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    A Group of Young Cybercriminals Poses the ‘Most Imminent Threat’ of Cyberattacks Right Now

    July 3, 2025

    Apple’s colorful Watch Solo Loop bands are up to 70 percent off now

    July 3, 2025

    How to Protest Safely in the Age of Surveillance

    July 3, 2025
    Facebook X (Twitter) Instagram
    Subscribe
    Technology Mag
    Facebook X (Twitter) Instagram YouTube
    • Home
    • News
    • Business
    • Games
    • Gear
    • Reviews
    • Science
    • Security
    • Trending
    • Press Release
    Technology Mag
    Home » America’s Big AI Safety Plan Faces a Budget Crunch
    Business

    America’s Big AI Safety Plan Faces a Budget Crunch

    News RoomBy News RoomDecember 26, 20233 Mins Read
    Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Email

    The lawmakers’ letter also claims that NIST is being rushed to define standards even though research into testing AI systems is at an early stage. As a result there is “significant disagreement” among AI experts over how to work on or even measure and define safety issues with the technology, it states. “The current state of the AI safety research field creates challenges for NIST as it navigates its leadership role on the issue,” the letter claims.

    NIST spokesperson Jennifer Huergo confirmed that the agency had received the letter and said that it “will respond through the appropriate channels.”

    NIST is making some moves that would increase transparency, including issuing a request for information on December 19, soliciting input from outside experts and companies on standards for evaluating and red-teaming AI models. It is unclear if this was a response to the letter sent by the members of Congress.

    The concerns raised by lawmakers are shared by some AI experts who have spent years developing ways to probe AI systems. “As a nonpartisan scientific body, NIST is the best hope to cut through the hype and speculation around AI risk,” says Rumman Chowdhury, a data scientist and CEO of Parity Consulting who specializes in testing AI models for bias and other problems. “But in order to do their job well, they need more than mandates and well wishes.”

    Yacine Jernite, machine learning and society lead at Hugging Face, a company that supports open source AI projects, says big tech has far more resources than the agency given a key role in implementing the White House’s ambitious AI plan. “NIST has done amazing work on helping manage the risks of AI, but the pressure to come up with immediate solutions for long-term problems makes their mission extremely difficult,” Jernite says. “They have significantly fewer resources than the companies developing the most visible AI systems.”

    Margaret Mitchell, chief ethics scientist at Hugging Face, says the growing secrecy around commercial AI models makes measurement more challenging for an organization like NIST. “We can’t improve what we can’t measure,” she says.

    The White House executive order calls for NIST to perform several tasks, including establishing a new Artificial Intelligence Safety Institute to support the development of safe AI. In April, a UK taskforce focused on AI safety was announced. It will receive $126 million in seed funding.

    The executive order gave NIST an aggressive deadline for coming up with, among other things, guidelines for evaluating AI models, principles for “red-teaming” (adversarially testing) models, developing a plan to get US-allied nations to agree to NIST standards, and coming up with a plan for “advancing responsible global technical standards for AI development.”

    Although it isn’t clear how NIST is engaging with big tech companies, discussions on NIST’s risk management framework, which took place prior to the announcement of the executive order, involved Microsoft; Anthropic, a startup formed by ex-OpenAI employees that is building cutting-edge AI models; Partnership on AI, which represents big tech companies; and the Future of Life Institute, a nonprofit dedicated to existential risk, among others.

    “As a quantitative social scientist, I’m both loving and hating that people realize that the power is in measurement,” Chowdhury says.

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp Reddit Email
    Previous ArticleOur Favorite Photobooks of 2023
    Next Article Apple is now banned from selling its latest Apple Watches in the US

    Related Posts

    For Today’s Business Traveler, It’s All About Work-Life Integration

    July 3, 2025

    Affluent Travelers Are Ditching Business Class for Business Jets

    July 2, 2025

    Airplane Wi-Fi Is Now … Good?

    July 2, 2025

    Business Travel Is Evolving Faster Than Ever. We’ll Help You Navigate It

    July 2, 2025

    Airport Lounges Are Sexy Again—if You Can Get In

    July 2, 2025

    Business Class Ain’t What It Used to Be. Don’t Tell First Class

    July 2, 2025
    Our Picks

    Apple’s colorful Watch Solo Loop bands are up to 70 percent off now

    July 3, 2025

    How to Protest Safely in the Age of Surveillance

    July 3, 2025

    E Ink is turning the laptop touchpad into an e-reader for AI apps

    July 3, 2025

    US Supreme Court Upholds Texas Porn ID Law

    July 3, 2025
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo
    Don't Miss
    Science

    A European Startup’s Spacecraft Made It to Orbit. Now It’s Lost at Sea

    By News RoomJuly 3, 2025

    A European company that seeks to develop orbital spacecraft for cargo, and eventually humans, took…

    ‘Persona 5: The Phantom X’ Brings the Series to Your Phone—and It’s Shockingly Good

    July 3, 2025

    Paramount Plus slashes prices to $2 for two months

    July 3, 2025

    Whoop MG review: a big whoop for a small crowd

    July 3, 2025
    Facebook X (Twitter) Instagram Pinterest
    • Privacy Policy
    • Terms of use
    • Advertise
    • Contact
    © 2025 Technology Mag. All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.