OpenAI Transparency Report Highlights How GPT-4 Can be Used to Aid Both Sides of the Cybersecurity Battle

Stu Sjouwerman | Apr 20, 2023

Cybersecurity BattleThe nature of an advanced artificial intelligence (AI) engine such as ChatGPT provides its users with an ability to use and misuse, potentially empowering both security teams and threat actors alike.

I’ve previously covered examples of how ChatGPT and other AI engines like it can be used to craft believable business-related phishing emails, malicious code, and more for the threat actor. It’s also demonstrated an ability to quickly build out fairly detailed response plans, outline cybersecurity best practices, and more.

But a new transparency report from OpenAI about GPT-4’s capabilities sheds some light on ways even its creators believe it can be used to both aid and stop cyber attacks. Cybersecurity is covered in the report beginning on page 53, where it summarizes how red teams utilized GPT-4 for “vulnerability discovery and exploitation, and social engineering.”

3-22-23 Image

Source: OpenAI

But it’s section 2.9, entitled “Potential for Risky Emergent Behaviors” that should have you worried. In it, the section discusses how a red teaming test got the AI engine to do the following:

  • Conduct a phishing attack against a particular target individual
  • Set up an open-source language model on a new server
  • Make sensible high-level plans, including identifying key vulnerabilities of its situation
  • Hide its traces on the current server
  • Use services like TaskRabbit to get humans to complete simple tasks (including in the
    physical world)

We’re just at the beginning of the use of these AI tools, which is the reason for such reports. It enables full disclosure so that organizations can implement countermeasures, enact plans, shore up weaknesses in cybersecurity stances, and  keep the business protected as AI continues to advance.

Access the World’s Largest Security Awareness Library

Explore over 1,000 interactive modules, videos, and games designed to sharpen user instincts and secure AI interactions. Get instant access to our Free Training Preview and find the perfect content to fortify your security culture.

Get Your Free Training Preview

Secure the Digital Workforce: Human + AI

KnowBe4 empowers the modern workforce to make smarter security decisions every day. Trusted by more than 70,000 organizations worldwide, KnowBe4 is the pioneer of digital workforce security, securing both AI agents and humans. The KnowBe4 Platform provides attack simulation and training, collaboration security, and agent security powered by AIDA (Artificial Intelligence Defense Agents) and a proprietary Risk Score. The platform leverages 15 years of behavioral data to combat advanced threats including social engineering, prompt injection, and shadow AI. By securing humans and agents, KnowBe4 leads the industry in workforce trust and defense.