Close Menu
Technology Mag

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot
    The First Radio Signal From Comet 3I/Atlas Ends the Debate About Its Nature

    The First Radio Signal From Comet 3I/Atlas Ends the Debate About Its Nature

    November 19, 2025
    Nvidia says its AI GPUs are sold out, grows data center business by B in a single quarter

    Nvidia says its AI GPUs are sold out, grows data center business by $10B in a single quarter

    November 19, 2025
    OpenAI’s Fidji Simo Plans to Make ChatGPT Way More Useful—and Have You Pay For It

    OpenAI’s Fidji Simo Plans to Make ChatGPT Way More Useful—and Have You Pay For It

    November 19, 2025
    Facebook X (Twitter) Instagram
    Subscribe
    Technology Mag
    Facebook X (Twitter) Instagram YouTube
    • Home
    • News
    • Business
    • Games
    • Gear
    • Reviews
    • Science
    • Security
    • Trending
    • Press Release
    Technology Mag
    Home » The US Government Wants You—Yes, You—to Hunt Down Generative AI Flaws
    Security

    The US Government Wants You—Yes, You—to Hunt Down Generative AI Flaws

    News RoomBy News RoomAugust 24, 20243 Mins Read
    Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Email
    The US Government Wants You—Yes, You—to Hunt Down Generative AI Flaws

    At the 2023 Defcon hacker conference in Las Vegas, prominent AI tech companies partnered with algorithmic integrity and transparency groups to sic thousands of attendees on generative AI platforms and find weaknesses in these critical systems. This “red-teaming” exercise, which also had support from the US government, took a step in opening these increasingly influential yet opaque systems to scrutiny. Now, the ethical AI and algorithmic assessment nonprofit Humane Intelligence is taking this model one step further. On Wednesday, the group announced a call for participation with the US National Institute of Standards and Technology, inviting any US resident to participate in the qualifying round of a nationwide red-teaming effort to evaluate AI office productivity software.

    The qualifier will take place online and is open to both developers and anyone in the general public as part of NIST’s AI challenges, known as Assessing Risks and Impacts of AI, or ARIA. Participants who pass through the qualifying round will take part in an in-person red-teaming event at the end of October at the Conference on Applied Machine Learning in Information Security (CAMLIS) in Virginia. The goal is to expand capabilities for conducting rigorous testing of the security, resilience, and ethics of generative AI technologies.

    “The average person utilizing one of these models doesn’t really have the ability to determine whether or not the model is fit for purpose,” says Theo Skeadas, chief of staff at Humane Intelligence. “So we want to democratize the ability to conduct evaluations and make sure everyone using these models can assess for themselves whether or not the model is meeting their needs.”

    The final event at CAMLIS will split the participants into a red team trying to attack the AI systems and a blue team working on defense. Participants will use the AI 600-1 profile, part of NIST’s AI risk management framework, as a rubric for measuring whether the red team is able to produce outcomes that violate the systems’ expected behavior.

    “NIST’s ARIA is drawing on structured user feedback to understand real-world applications of AI models,” says Humane Intelligence founder Rumman Chowdhury, who is also a contractor in NIST’s Office of Emerging Technologies and a member of the US Department of Homeland Security AI safety and security board. “The ARIA team is mostly experts on sociotechnical test and evaluation, and [is] using that background as a way of evolving the field toward rigorous scientific evaluation of generative AI.”

    Chowdhury and Skeadas say the NIST partnership is just one of a series of AI red team collaborations that Humane Intelligence will announce in the coming weeks with US government agencies, international governments, and NGOs. The effort aims to make it much more common for the companies and organizations that develop what are now black-box algorithms to offer transparency and accountability through mechanisms like “bias bounty challenges,” where individuals can be rewarded for finding problems and inequities in AI models.

    “The community should be broader than programmers,” Skeadas says. “Policymakers, journalists, civil society, and nontechnical people should all be involved in the process of testing and evaluating of these systems. And we need to make sure that less represented groups like individuals who speak minority languages or are from nonmajority cultures and perspectives are able to participate in this process.”

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp Reddit Email
    Previous ArticleMicrosoft to host CrowdStrike and others to discuss Windows security changes
    Next Article The Covid-19 Summer Wave Is So Big, the FDA Might Release New Vaccines Early

    Related Posts

    This Is the Platform Google Claims Is Behind a ‘Staggering’ Scam Text Operation

    This Is the Platform Google Claims Is Behind a ‘Staggering’ Scam Text Operation

    November 19, 2025
    DOJ Issued Seizure Warrant to Starlink Over Satellite Internet Systems Used at Scam Compound

    DOJ Issued Seizure Warrant to Starlink Over Satellite Internet Systems Used at Scam Compound

    November 19, 2025
    DHS Kept Chicago Police Records for Months in Violation of Domestic Espionage Rules

    DHS Kept Chicago Police Records for Months in Violation of Domestic Espionage Rules

    November 18, 2025
    Scam Ads Are Flooding Social Media. These Former Meta Staffers Have a Plan

    Scam Ads Are Flooding Social Media. These Former Meta Staffers Have a Plan

    November 15, 2025
    The Government Shutdown Is a Ticking Cybersecurity Time Bomb

    The Government Shutdown Is a Ticking Cybersecurity Time Bomb

    November 14, 2025
    Zohran Mamdani Just Inherited the NYPD Surveillance State

    Zohran Mamdani Just Inherited the NYPD Surveillance State

    November 12, 2025
    Our Picks
    Nvidia says its AI GPUs are sold out, grows data center business by B in a single quarter

    Nvidia says its AI GPUs are sold out, grows data center business by $10B in a single quarter

    November 19, 2025
    OpenAI’s Fidji Simo Plans to Make ChatGPT Way More Useful—and Have You Pay For It

    OpenAI’s Fidji Simo Plans to Make ChatGPT Way More Useful—and Have You Pay For It

    November 19, 2025
    Netflix signs a three year deal to stream MLB live events and games

    Netflix signs a three year deal to stream MLB live events and games

    November 19, 2025
    This Is the Platform Google Claims Is Behind a ‘Staggering’ Scam Text Operation

    This Is the Platform Google Claims Is Behind a ‘Staggering’ Scam Text Operation

    November 19, 2025
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo
    Don't Miss
    Google’s new Scholar Labs search uses AI to find relevant studies News

    Google’s new Scholar Labs search uses AI to find relevant studies

    By News RoomNovember 19, 2025

    Google has announced it’s testing a new AI-powered search tool, Scholar Labs, that’s designed to…

    For Black Friday, get the M4 MacBook Air that’s only 0 more than the slower M1 model

    For Black Friday, get the M4 MacBook Air that’s only $150 more than the slower M1 model

    November 19, 2025
    Meta Claims Downloaded Porn at Center of AI Lawsuit Was for ‘Personal Use’

    Meta Claims Downloaded Porn at Center of AI Lawsuit Was for ‘Personal Use’

    November 19, 2025
    You Won’t Be Able to Offload Your Holiday Shopping to AI Agents Anytime Soon

    You Won’t Be Able to Offload Your Holiday Shopping to AI Agents Anytime Soon

    November 19, 2025
    Facebook X (Twitter) Instagram Pinterest
    • Privacy Policy
    • Terms of use
    • Advertise
    • Contact
    © 2025 Technology Mag. All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.