OpenAI’s GPT-4 Demonstrates Limited Utility In Bioweapon Development: Study

    Date:

    Loading…

    Loading…

    The latest artificial intelligence offering from startup ChatGPT-parent OpenAI, known as GPT-4, has shown a minimal propensity to aid in the development of biological threats, a new study reveals.

    What Happened: OpenAI carried out tests to gauge the probable dangers of its GPT-4 being exploited to assist in the production of biological weapons. 

    This study was initiated in response to apprehensions voiced by legislators and tech industry leaders over potential misuse of AI technologies for harmful purposes, reported Bloomberg. 

    Last October, an executive order signed by President Joe Biden mandated the Department of Energy to ensure AI technologies do not pose chemical, biological, or nuclear threats. In response, OpenAI assembled a “preparedness” team to tackle potential risks associated with AI.

    See Also: Microsoft CEO Satya Nadella Doesn’t Know How He’d Work ‘If You Took Away’ This AI Tool — Hint: It’s Not ChatGPT

    As part of the team’s inaugural study, unveiled on Wednesday, OpenAI’s researchers engaged with 50 biology experts and 50 college-level biology students who were divided into two groups. 

    One group used a special version of GPT-4 and the internet for tasks related to making a biological threat, while the other group used only the internet.

    Despite a slight uptick in accuracy and completeness for the group using the AI model, the researchers concluded that GPT-4 only marginally improves the acquisition of information for creating biological threats.

    Alexander Madry, who heads the preparedness team, stated that this study is part of a series aimed at understanding potential abuses of OpenAI’s technology. Other ongoing studies are exploring AI’s potential use in creating cybersecurity threats and influencing personal beliefs.

    Why It Matters: The increasing concern over the potential misuse of AI tools was evident when the Biden administration considered regulating AI tools like ChatGPT over fears of potential harm and discrimination.

    OpenAI CEO Sam Altman’s subsequent launch of a “preparedness challenge” in October 2023, was a clear response to growing fears over AI’s potential misuse.

    Concerns over AI’s dual-use potential were further amplified with the release of GPT-4, the latest AI engine behind ChatGPT, which experts warned could make it easier for malicious actors to build malware programs and create new cybersecurity threats. 

    Photo Courtesy: Shutterstock.com

    Check out more of Benzinga’s Consumer Tech coverage by following this link.

    Read Next: The Hidden ChatGPT Trick: Being Nice Can Give Surprisingly Better Results

    Disclaimer: This content was partially produced with the help of Benzinga Neuro and was reviewed and published by Benzinga editors.

    Loading…

    Loading…

    Go Source

    Chart

    Sign up for Breaking Alerts

    Share post:

    Popular

    More like this
    Related

    Monthly Expiration to Traders: “Remember Me?”

    Your Privacy When you visit any website it may use...

    Why is the Gold Rally Leaving Silver Behind?

    Your Privacy When you visit any website it may use...

    What is the US national debt’s risk to investments?

    Key takeaways Debt not a disaster While the US national debt...