Close Menu
Technology Mag

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot
    We visited the heart of America’s chipmaking boom

    We visited the heart of America’s chipmaking boom

    November 21, 2025
    Mark Zuckerberg Opened an Illegal School at His Palo Alto Compound. His Neighbors Revolted

    Mark Zuckerberg Opened an Illegal School at His Palo Alto Compound. His Neighbors Revolted

    November 21, 2025
    A Major Leak Spills a Chinese Hacking Contractor’s Tools and Targets

    A Major Leak Spills a Chinese Hacking Contractor’s Tools and Targets

    November 21, 2025
    Facebook X (Twitter) Instagram
    Subscribe
    Technology Mag
    Facebook X (Twitter) Instagram YouTube
    • Home
    • News
    • Business
    • Games
    • Gear
    • Reviews
    • Science
    • Security
    • Trending
    • Press Release
    Technology Mag
    Home » OpenAI Touts New AI Safety Research. Critics Say It’s a Good Step, but Not Enough
    Business

    OpenAI Touts New AI Safety Research. Critics Say It’s a Good Step, but Not Enough

    News RoomBy News RoomJuly 18, 20244 Mins Read
    Facebook Twitter Pinterest LinkedIn Reddit WhatsApp Email
    OpenAI Touts New AI Safety Research. Critics Say It’s a Good Step, but Not Enough

    OpenAI has faced opprobrium in recent months from those who suggest it may be rushing too quickly and recklessly to develop more powerful artificial intelligence. The company appears intent on showing it takes AI safety seriously. Today it showcased research that it says could help researchers scrutinize AI models even as they become more capable and useful.

    The new technique is one of several ideas related to AI safety that the company has touted in recent weeks. It involves having two AI models engage in a conversation that forces the more powerful one to be more transparent, or “legible,” with its reasoning so that humans can understand what it’s up to.

    “This is core to the mission of building an [artificial general intelligence] that is both safe and beneficial,” Yining Chen, a researcher at OpenAI involved with the work, tells WIRED.

    So far, the work has been tested on an AI model designed to solve simple math problems. The OpenAI researchers asked the AI model to explain its reasoning as it answered questions or solved problems. A second model is trained to detect whether the answers are correct or not, and the researchers found that having the two models engage in a back and forth encouraged the math-solving one to be more forthright and transparent with its reasoning.

    OpenAI is publicly releasing a paper detailing the approach. “It’s part of the long-term safety research plan,” says Jan Hendrik Kirchner, another OpenAI researcher involved with the work. “We hope that other researchers can follow up, and maybe try other algorithms as well.”

    Transparency and explainability are key concerns for AI researchers working to build more powerful systems. Large language models will sometimes offer up reasonable explanations for how they came to a conclusion, but a key concern is that future models may become more opaque or even deceptive in the explanations they provide—perhaps pursuing an undesirable goal while lying about it.

    The research revealed today is part of a broader effort to understand how large language models that are at the core of programs like ChatGPT operate. It is one of a number of techniques that could help make more powerful AI models more transparent and therefore safer. OpenAI and other companies are exploring more mechanistic ways of peering inside the workings of large language models, too.

    OpenAI has revealed more of its work on AI safety in recent weeks following criticism of its approach. In May, WIRED learned that a team of researchers dedicated to studying long-term AI risk had been disbanded. This came shortly after the departure of cofounder and key technical leader Ilya Sutskever, who was one of the board members who briefly ousted CEO Sam Altman last November.

    OpenAI was founded on the promise that it would make AI both more transparent to scrutiny and safer. After the runaway success of ChatGPT and more intense competition from well-backed rivals, some people have accused the company of prioritizing splashy advances and market share over safety.

    Daniel Kokotajlo, a researcher who left OpenAI and signed an open letter criticizing the company’s approach to AI safety, says the new work is important, but incremental, and that it does not change the fact that companies building the technology need more oversight. “​The situation we are in remains unchanged,” he says. “Opaque, unaccountable, unregulated corporations racing each other to build artificial superintelligence, with basically no plan for how to control it.”

    Another source with knowledge of OpenAI’s inner workings, who asked not to be named because they were not authorized to speak publicly, says that outside oversight of AI companies is also needed. “The question is whether they’re serious about the kinds of processes and governance mechanisms you need to prioritize societal benefit over profit,” the source says. “Not whether they let any of their researchers do some safety stuff.”

    Share. Facebook Twitter Pinterest LinkedIn WhatsApp Reddit Email
    Previous ArticleThe Boys season 4 finale gets a disclaimer following Trump rally shooting
    Next Article Hackers Claim to Have Leaked 1.1 TB of Disney Slack Messages

    Related Posts

    Mark Zuckerberg Opened an Illegal School at His Palo Alto Compound. His Neighbors Revolted

    Mark Zuckerberg Opened an Illegal School at His Palo Alto Compound. His Neighbors Revolted

    November 21, 2025
    In Alex Karp’s World, Palantir Is the Underdog

    In Alex Karp’s World, Palantir Is the Underdog

    November 20, 2025
    Gemini 3 Is Here—and Google Says It Will Make Search Smarter

    Gemini 3 Is Here—and Google Says It Will Make Search Smarter

    November 19, 2025
    The 4 Things You Need for a Tech Bubble

    The 4 Things You Need for a Tech Bubble

    November 19, 2025
    OpenAI’s Fidji Simo Plans to Make ChatGPT Way More Useful—and Have You Pay For It

    OpenAI’s Fidji Simo Plans to Make ChatGPT Way More Useful—and Have You Pay For It

    November 19, 2025
    Meta Claims Downloaded Porn at Center of AI Lawsuit Was for ‘Personal Use’

    Meta Claims Downloaded Porn at Center of AI Lawsuit Was for ‘Personal Use’

    November 19, 2025
    Our Picks
    Mark Zuckerberg Opened an Illegal School at His Palo Alto Compound. His Neighbors Revolted

    Mark Zuckerberg Opened an Illegal School at His Palo Alto Compound. His Neighbors Revolted

    November 21, 2025
    A Major Leak Spills a Chinese Hacking Contractor’s Tools and Targets

    A Major Leak Spills a Chinese Hacking Contractor’s Tools and Targets

    November 21, 2025
    Google’s Nano Banana Pro generates excellent conspiracy fuel

    Google’s Nano Banana Pro generates excellent conspiracy fuel

    November 21, 2025
    Gemini 3 is almost as good as Google says it is

    Gemini 3 is almost as good as Google says it is

    November 20, 2025
    • Facebook
    • Twitter
    • Pinterest
    • Instagram
    • YouTube
    • Vimeo
    Don't Miss
    Feds charge four with illegally smuggling Nvidia AI chips to China News

    Feds charge four with illegally smuggling Nvidia AI chips to China

    By News RoomNovember 20, 2025

    Federal prosecutors have charged four people with illegally smuggling Nvidia GPUs and HP supercomputers with…

    Meta’s Hyperscape is ready to turn your real living room into a VR hangout

    Meta’s Hyperscape is ready to turn your real living room into a VR hangout

    November 20, 2025
    Microsoft’s AI-powered copy and paste can now use on-device AI

    Microsoft’s AI-powered copy and paste can now use on-device AI

    November 20, 2025
    A Simple WhatsApp Security Flaw Exposed 3.5 Billion Phone Numbers

    A Simple WhatsApp Security Flaw Exposed 3.5 Billion Phone Numbers

    November 20, 2025
    Facebook X (Twitter) Instagram Pinterest
    • Privacy Policy
    • Terms of use
    • Advertise
    • Contact
    © 2025 Technology Mag. All Rights Reserved.

    Type above and press Enter to search. Press Esc to cancel.