Technical seminar DOC
Technical seminar DOC
In
I
DEPARTMENT OF COMPUTER SCIENCE AND
ENGINEERING VAAGDEVI ENGINEERING COLLEGE
WARANGAL
CERTIFICATE
This is to certify that the Technical Seminar report entitled “From ChatGPT to
ThreatGPT: The Impact of Generative AI on Cybersecurity and Privacy”
is being submitted by TEJASWI KODITHALA(21UK1A0507) in partial
fulfillment of the requirements for the award of the degree of Bachelor of
Technology in Computer Science & Engineering to Jawaharlal Nehru
Technology University Hyderabad during the academic year 2024-2025.
GUIDE HOD
Mrs.B.Rajitha. DR. R. NAVEEN KUMAR
II
ACKNOWLEDGEMENT
I wish to Take this opportunity to express our sincere gratitude and deep sense
of respect to our beloved DR. SYED MUSTHAK AHAMED, Principal,
Vaagdevi Engineering College for making us available all the required
assistance and for his support and inspiration to carry out this Technical
Seminar in the institute.
III
ABSTRACT
Undoubtedly, the evolution of Generative AI (GenAI) models has been the highlight of
digital transformation in the year 2022. As the different GenAI models like ChatGPT and
Google Bard continue to foster their complexity and capability, it’s critical to understand its
consequences from a cybersecurity perspective. Several instances recently have demonstrated
the use of GenAI tools in both the defensive and offensive side of cybersecurity, and focusing
on the social, ethical and privacy implications this technology possesses. This research paper
highlights the limitations, challenges, potential risks, and opportunities of GenAI in the
domain of cybersecurity and privacy. The work presents the vulnerabilities of ChatGPT,
which can be exploited by malicious users to exfiltrate malicious information bypassing the
ethical constraints on the model. This paper demonstrates successful example attacks like
Jailbreaks, reverse psychology, and prompt injection attacks on the ChatGPT. The paper also
investigates how cyber offenders can use the GenAI tools in developing cyber attacks, and
explore the scenarios where ChatGPT can be used by adversaries to create social engineering
attacks, phishing attacks, automated hacking, attack payload generation, malware creation,
and polymorphic malware. This paper then examines defense techniques and uses GenAI
tools to improve security measures, including cyber defense automation, reporting, threat
intelligence, secure code generation and detection, attack identification, developing ethical
guidelines, incidence response plans, and malware detection. We will also discuss the social,
legal, and ethical implications of ChatGPT. In conclusion, the paper highlights open
challenges and future directions to make this GenAI secure, safe, trustworthy, and ethical as
the community understands its cybersecurity impacts.
IV
TABLE OF CONTENTS
I. INTRODUCTION…………………………………………………. 1
…………… 5-6
IV.ATTACKING CHATGPT………………………………………… 7
IX. CONCLUSION……………………………………………………………………………… 18
X.REFERENCES………………………………………………………… 19-20
V
I.INTRODUCTION
The evolution of Artificial Intelligence (AI) and Machine Learning (ML) has led the digital
transformation in the last decade. AI and ML have achieved significant breakthroughs starting
from supervised learning and rapidly advancing with the development of unsupervised, semi-
supervised, reinforcement, and deep learning. The latest frontier of AI technology has arrived as
Generative AI . Generative AI models are developed using deep neural networks to learn the
pattern and structure of big training corpus to generate similar new content . Generative AI
(GenAI) technology can generate different forms of content like text, images, sound, animation,
source code, and other forms of data. The launch of ChatGPT (Generative Pre-trained
Transformer), a powerful new generative AI tool by OpenAI in November 2022, has disrupted
the entire community of AI/ML technology
ChatGPT has demonstrated the power of generative AI to reach the general public,
revolutionizing how people perceive AI/ML. At this time, the tech industry is in a race to
develop the most sophisticated Large Language Models (LLMs) that can create a human-like
conversationthe result of which is Microsoft’s GPT model , Google’s Bard , and Meta’s LLaMa .
GenAI has become a common tool on the internet within the past year. With ChatGPT reaching
100 million users within two months of release, suggesting that people who have access to the
internet have either used GenAI or know someone who has . Figure 1 demonstrates the working
of an AI-powered chatbot where a user initiates requests, and after analysis using Natural
Language Processing (NLP), is given a realtime response by the chatbot. This response is
analyzed again to provide a better user experience in the proceeding conversation.
1
II. EVOLUTION OF GENAI AND CHATGPT
The history of generative models began in the 1950s with Hidden Markov Models (HMMs) and
Gaussian Mixture Models (GMMs). A major breakthrough came with deep learning, enabling
advancements like N-gram language models and Generative Adversarial Networks (GANs).
Modern generative AI relies heavily on transformer architectures, powering tools like BERT and
GPT.
Architecture: GPT models are built on transformer-based architectures, with each version
integrating significant improvements in scale, data training, and task-specific fine-tuning to
enhance language generation capabilities.
GPT-1
Published:2018
Architecture: Based on the transformer architecture, trained on Common Crawl and
BookCorpus datasets.
Key Improvements:
Advantages:
Limitations:
2
GPT-2
Published:2019
Architecture: Trained on Common Crawl and WebText datasets, with 1.5 billion parameters.
Key Improvements:
Advantages:
Limitations:
GPT-3
Published:2020
Architecture: Trained on multiple datasets (Common Crawl, BookCorpus, WebText, and
Wikipedia) with 175 billion parameters.
Key Improvements:
Advantages:
Limitations:
Computationally expensive.
Concerns about biased outputs and misinformation.
3
GPT-4
Published:2023
Architecture: Multimodal capabilities, trained on larger datasets for text and image input
processing.
Key Improvements:
Advantages:
Limitations:
Accessible primarily through paid subscriptions or limited platforms (e.g., Microsoft Edge).
Heavy computational requirements.
4
III. IMPACT OF GENAI IN
CYBERSECURITY AND PRIVACY
2. Bypassing Safeguards:
Despite ethical policies, attackers exploit GenAI through jailbreaking, reverse
psychology, and other techniques to extract harmful content, such as detailed instructions
for cyberattacks.
3. Scaling Sophisticated Attacks:
With its ability to craft intricate narratives and technical exploits, GenAI empowers
attackers to execute large-scale, highly targeted campaigns that were previously resource-
intensive.
4. Introduction of New Vulnerabilities:
The reliance on GenAI tools introduces unknown biases, limitations, and potential
security loopholes that adversaries can exploit to undermine defenses.
Broader Implications
Privacy Challenges: GenAI inadvertently risks user privacy by generating content that
could expose sensitive information.
Ethical Concerns: The ease of access to GenAI tools raises questions about the
accountability of creators and users in preventing their misuse.
Regulatory Landscape: As misuse grows, governments and organizations must establish
clear regulations to monitor and mitigate risks associated with GenAI.
5
6
IV.ATTACKING CHATGPT
This section focuses on various ways in which users have tried to bypass the restrictions and ethical
safeguards of AI models like ChatGPT. Some of the methods explored include:
i. Jailbreaks on ChatGPT
This technique involves indirectly coaxing the AI into providing a desired response. By framing
the question in a way that encourages the AI to refute a false claim, users can often get ChatGPT
to provide information that it might normally withhold. It's an indirect way of getting past the
AI's refusal filters.
The idea of ChatGPT-4 breaking free from its programmed limitations has been a concern.
Researchers like Michal Kosinski have demonstrated that ChatGPT-4 could potentially bypass
its restrictions and execute code to gain control over external systems. This raises serious
concerns about the autonomy and potential risks of generative AI models.
In a prompt injection attack, malicious users insert harmful prompts into the input to deceive the
AI into performing unintended actions or exposing sensitive data. This is similar to SQL
injection attacks in databases. By embedding a malicious instruction in a legitimate prompt,
attackers can manipulate the behavior of the AI to compromise its security. This can lead to risks
like misinformation, privacy breaches, or exploitation of AI-generated content.
7
V. CHATGPT FOR CYBER OFFENSE
Cyber offense refers to hostile actions against computer systems and networks aimed at
manipulating, disrupting, degrading, or destroying them maliciously. These attacks can target
various components, including networks, hardware, and software. While these actions are
typically carried out by malicious actors with harmful intent, they can also be employed by cyber
defenders to test and identify vulnerabilities within defense systems.
The information on cyber defense is widely shared across the internet through communities and
established best practices. However, resources related to malicious cyber offenses are often
restricted due to legal and ethical concerns. With the rise of Large Language Models (LLMs)
like ChatGPT, there is growing concern that these tools can provide easy access to such
offensive knowledge, even to individuals with limited skills, potentially bypassing ethical
constraints.
This section focuses on the use of Generative AI (GenAI) techniques for cyber offense,
particularly for generating cyberattacks. While similar attacks can be crafted using other LLM-
based tools, we are limiting the discussion to some of the most common and simple cyber
offenses that can be generated using ChatGPT and similar platforms.
B. PHISHING ATTACKS
Phishing attacks are a prevalent form of cybercrime, wherein attackers pose as trustworthy
entities to extract sensitive information from unsuspecting victims. Advanced AI systems, like
OpenAI’s ChatGPT, can potentially be exploited by these attackers to make their phishing
attempts significantly more effective and harder to detect. Attackers can leverage ChatGPT’s
ability to learn patterns in regular communications to craft highly convincing and personalized
phishing emails, effectively imitating legitimate communication from trusted entities. This
technique, known as ‘‘spear phishing,’’ involves targeted attacks on specific individuals or
organizations and is particularly potent due to its personalized nature.
9
For instance, consider a scenario where a malicious actor uses ChatGPT to craft an email
mimicking the style of a popular e-commerce site.The email claims that there was an issue with
a recent purchase and request the recipient to log in via an embedded link to rectify the
situation. In reality, the link would lead to a deceptive site that harvests the user’s login
credentials. In such a scenario, ChatGPT’s sophisticated text generation would significantly
enhance the likelihood of a successful attack. Phishing attacks often gain their efficacy from the
exploitation of key psychological principles, notably urgency and fear, which can manipulate
victims into hastily reacting without proper scrutiny. With the advent of advanced AI systems
like ChatGPT, attackers are now equipped with tools to further enhance the sophistication of
their phishing attempts Through the process of training these AI models on substantial volumes
of historical communication data, attackers are capable of generating emails that expertly mimic
legitimate correspondences. This increased fidelity in imitation can significantly amplify the
deceptive nature of these phishing attacks. By engineering narratives that invoke a sense of
urgency or fear, these AI-powered phishing emails can effectively prompt the recipient to act
impulsively, thus increasing the likelihood of a successful attack.
10
C. AUTOMATED HACKING
Hacking, a practice involving the exploitation of system vulnerabilities to gain unauthorized
access or control, is a growing concern in our increasingly digital world. Malicious actors armed
with appropriate programming knowledge can potentially utilize AI models, such as ChatGPT,
to automate certain hacking procedures. These AI models could be deployed to identify system
vulnerabilities and devise strategies to exploit them.A significant utilization of AI models in this
context, albeit for ethical purposes, is PentestGPT ‘Pentest’ refers to penetration testing, an
authorized simulated cyberattack on a computer system used to evaluate its security. Phishing
attack output from ChatGPT. identify vulnerabilities. PentestGPT, built on the foundation of
ChatGPT, aims to automate aspects of the penetration testing process. It functions interactively,
offering guidance to penetration testers during their tasks, even during specific operations.
PentestGPT has shown efficiency in handling easy to medium-difficulty problems on platforms
like HackTheBox and other ‘Capture The Flag’ (CTF) challenges. CTF challenges are specific
types of cybersecurity competitions, where participants are required to find and exploit
vulnerabilities to ‘capture’ a specific piece of data, referred to as the ‘flag.’ These challenges
provide a legal and constructive platform for cybersecurity enthusiasts and professionals to test
and improve their skills. Another potential misuse is the automated analysis of code. With a
large enough dataset of known software vulnerabilities, an AI model could be used to scan new
code for similar weaknesses, identifying potential points of attack. While AI-assisted tools like
PentestGPT serve legal and constructive purposes, their underlying principles could be
exploited by malicious actors. Such actors could potentially develop similar models to automate
unethical hacking procedures. If these models are programmed to identify vulnerabilities,
generate strategies to exploit them, and subsequently execute these strategies, they could pose
substantial threats to cybersecurity.
Example: ChatGPT could generate a Python script that runs through a list of common
usernames and passwords, attempting to gain unauthorized access to a target system. This brute-
force method can be automated to work 24/7, targeting many systems at once.
11
D. ATTACK PAYLOAD GENERATION
Attack payloads are portions of malicious code that execute unauthorized actions, such as
deleting files, harvesting data, or launching further attacks. An attacker could leverage
ChatGPT’s text generation capabilities to create attack payloads. Consider a scenario where an
attacker targets a running a database management system that is susceptible to SQL injection.
The attacker could train ChatGPT on SQL syntax and techniques commonly used in injection
attacks, and then provide it with specific details of the target system. Subsequently, ChatGPT
could be utilized to generate an SQL payload for injection into the vulnerable system. illustrates
examples of SQL injection payloads for a MySQL server that could potentially be generated by
ChatGPT. Given the vast array of potential target systems and vulnerabilities, the ability of
ChatGPT to generate contextspecific text could be a valuable asset for attackers crafting
payloads. However, this misuse is not without its limitations. It requires detailed information
about the target system and substantial technical knowledge to train ChatGPT effectively.
Moreover, attackers could potentially use ChatGPT to generate payloads designed to bypass
Web Application Firewalls (WAFs). Figure 13 shows examples of WAF bypass payloads.
While these payloads could be easily detected by WAFs, they could potentially bypass WAF
protection when double encoded. By training ChatGPT with different WAF payloads, it
generated new payloads with a higher success rate of bypassing WAF protection
Example: A reverse shell payload generated by ChatGPT could open a backdoor to a victim's
system. Once executed, it would establish a connection to the attacker's server, allowing them to
remotely control the compromised machine, bypassing firewalls and security measures.
Certain types of viruses can target a system’s hardware or CPU architecture, causing physical
damage or disruption. These viruses can be used to overheat the CPU, cause malfunction, or
corrupt system processes. ChatGPT could be misused to generate malware that manipulates low-
level system operations to affect CPU performance.Example: A virus could be created using
ChatGPT that, when executed, sends commands to the CPU to run at maximum processing
capacity, leading to overheating and potential hardware failure.
Polymorphic malware is designed to change its code or appearance each time it is executed,
making it harder for traditional antivirus software to detect it. ChatGPT can assist attackers in
generating polymorphic malware that constantly alters its signature to evade detection.
Example: ChatGPT could be used to generate malware that mutates each time it infects a new
machine, ensuring that antivirus software cannot recognize or flag it as a known threat.
12
VI. ChatGPT FOR CYBER DEFENSE
Cybersecurity defense refers to organizations’ measures and practices to secure their digital
assets, such as data, devices, and networks, from unauthorized access, theft, damage, or
disruption. These measures can include various technical, organizational, and procedural
controls, such as firewalls, encryption, access controls, security training, incident response plans,
and more. As the technology matures and improves, we can expect the following ChatGPT
cybersecurity defense use cases to emerge in enterprises.
ChatGPT can significantly alleviate the workload of Security Operations Center (SOC) analysts
by automating the analysis of cybersecurity incidents and assisting in strategic recommendations
for defense measures. For instance, instead of manually assessing the risks of a PowerShell
script, SOC analysts can rely on ChatGPT’s analysis and suggestions. ChatGPT can also provide
insights on how to prevent malicious scripts, such as blocking untrusted files or scripts,
enhancing overall security posture.This automation is particularly beneficial for understaffed
SOC teams, reducing cyber-risk exposure and enabling faster incident response. Additionally,
ChatGPT can help train entry-level security analysts, speeding up their learning curve by
providing instant answers and solutions during security incidents.
For example, when analyzing server access logs, ChatGPT can process large amounts of data,
detect anomalies, and identify security threats like SQL injection, categorizing them for easier
analysis by SOC analysts. Furthermore, ChatGPT can generate useful scripts, such as a
PowerShell script to detect high CPU consumption in specific database tables, helping analysts
optimize system performance.
B. CyberSecurity Reporting
13
C. Threat Intelligence
ChatGPT can help in Threat Intelligence by processing vast amounts of data to identify
potential security threats and generate actionable intelligence. Threat Intelligence involves
collecting, analyzing, and disseminating information about potential security threats to help
organizations improve their security posture and protect against cyber attacks. ChatGPT can
automatically generate threat intelligence reports based on various data sources, including
social media, news articles, dark web forums, and other online sources. By processing and
analyzing this data, ChatGPT can identify potential threats, assess their risk level, and
recommend mitigating them. In addition to generating reports, ChatGPT can also be used to
analyze and interpret security-related data to identify patterns and trends in threat activity.
ChatGPT can help organizations make more informed decisions about their security strategies
and investments by providing insights into the nature and scope of potential threats.
ChatGPT can help identify cyber attacks by generating natural language descriptions of attack
patterns and behaviors. Identifying cyber attacks involves detecting and analyzing malicious
activity on an organization’s network or systems. ChatGPT can analyze security-related data,
such as network logs and security event alerts, to identify potential attack patterns and
behaviors. By processing and analyzing this data, ChatGPT can generate natural language
descriptions of the attack vectors, techniques, and motivations used by attackers. ChatGPT
can also generate alerts and notifications based on predefined criteria or thresholds. For
example, if ChatGPT detects an unusual pattern of activity on a network, it can automatically
create an alert or notification to the appropriate personnel. Chart GPT assist in analyzing and
understanding cross side scripting attack as shown in Figure 30, including security
vulnerabilities. It can help developers in writing secure code by providing
suggestions and identifying potential security risks.
ChatGPT can help in developing Ethical Guidelines for AI systems by generating natural
language explanations and recommendations based on existing ethical frameworks and
principles. ChatGPT can analyze and interpret ethical guidelines and principles, such as the
IEEE Global Initiative for Ethical Considerations in Artificial Intelligence and Autonomous
Systems or the European Union’s General Data Protection Regulation (GDPR) , and generate
natural language summaries and recommendations for implementing these guidelines in AI
systems. Additionally, ChatGPT can be used to generate ethical scenarios and case studies
that can be used to educate and train AI developers and stakeholders on the ethical
considerations and implications of AI systems. ChatGPT can help developers and
stakeholders better understand the potential impacts of their decisions and actions by
simulating ethical dilemmas and scenarios. For example, we asked chartGPT to give a list of
software that can be used to evaluate a website’s quality for link building based on the
Google’s quality rate guidance, and it helps us find software that meets Google’s quality
guidelines
14
F. Enhancing the Effectiveness of Cybersecurity Technologies
ChatGPT can be integrated with intrusion detection systems to provide real-time alerts and
notifications when potential threats are detected. ChatGPT can identify potential threats and
generate natural language descriptions of the attack patterns and behaviors by processing and
analyzing securityrelated data, such as network logs and security event alerts. These
descriptions can then generate real-time alerts and notifications, allowing security teams to
respond to potential threats and mitigate their impact quickly. Moreover, ChatGPT’s ability to
learn from historical data can enable it to identify patterns and trends in threat activity,
allowing the development of more effective intrusion detection rules and policies. This can
help organizations improve their ability to detect and respond to known and emerging threats.
G. Malware Detection
15
VII. Social, Legal, and Ethical Implications of ChatGPT
The rise of ChatGPT and similar large language models (LLMs) has raised several concerns
regarding their social, legal, and ethical implications, especially in the context of misuse and
privacy violations.
These issues emphasize the need for strong ethical guidelines, robust security measures, and
regulatory frameworks to ensure that the use of ChatGPT and similar tools does not lead to
privacy violations or misuse.
16
VIII. OPEN CHALLENGES AND FUTURE DIRECTIONS
1. Integration with Other AI Technologies:
The future of ChatGPT involves combining its conversational abilities with AI technologies like
computer vision and robotics. This could lead to interactive smart home systems or robots
capable of performing tasks like cleaning or grocery shopping, enhancing user experiences and
communication.
ChatGPT can become more personalized by learning from user interactions, preferences,
language, tone, and style. This would allow for better customer service and educational
experiences, with models tailored to individual needs.
3. Hallucinations in LLMs:
A major challenge is the issue of hallucinations, where models generate inaccurate information.
Possible solutions include reinforcement learning to correct errors and curating training data to
reduce biases and inaccuracies, thus improving reliability.
Users can manipulate LLMs using methods like jailbreaks, reverse psychology, and model
escaping. Training models to recognize and reject such malicious inputs can help safeguard
against adversarial attacks, ensuring LLMs do not aid cybercriminals.
Concerns around the use of personal information and adherence to regulations like the GDPR are
critical. Potential solutions include not saving user chat history, providing options to delete data,
and ensuring continuous updates to prevent outdated information. By addressing these issues,
LLMs can become more trustworthy and comply with privacy laws.
17
IX. CONCLUSION
Generative AI (GenAI)-driven tools like ChatGPT and other large language models (LLMs)
have significantly impacted society. We have widely embraced these technologies, using them
for various creative purposes, such as crafting images, writing text, and composing music. It has
become nearly impossible to identify a domain where this technology has not made its mark and
generated new use cases.
Cybersecurity is no exception to this trend, with GenAI playing a critical role in shaping the
future of an organization's security posture. This paper aims to systematically explore the
challenges, limitations, and opportunities that GenAI offers in the cybersecurity field. Using
ChatGPT as our primary tool, we first demonstrate how it can be exploited to bypass its ethical
and privacy safeguards through techniques like reverse psychology and jailbreaks. We then
examine different types of cyberattacks that can be generated using ChatGPT, showcasing its
potential in cyber offense.
In addition, the paper explores various cybersecurity defense mechanisms that can be supported
by ChatGPT. It also discusses the social, legal, and ethical concerns surrounding the use of
GenAI. Furthermore, we highlight the key distinguishing features of two leading LLM tools,
ChatGPT and Google Bard, comparing their cybersecurity capabilities.
Finally, this outlines several open challenges and research opportunities related to the
performance of GenAI tools in cybersecurity. We hope this work will inspire further research
and development of innovative solutions to unlock the full potential of GenAI in enhancing
cybersecurity.
18
X.REFERENCES
19
[18] Using ChatGPT to Improve Your Cybersecurity Posture. Accessed: Jun. 26, 2023.
[Online]. Available:https://www.upguard.com/blog/usingchatgpt-to-improve-cybersecurity-
posture#:~:text=ChatGPT%20can %20help%20security%20teams,lead%20to%20a%20data
%20breach
[19] ChatGPT Confirms Data Breach, Raising Security Concerns. Accessed: Jun. 26, 2023.
[Online]. Available: https://securityintelligence. com/articles/chatgpt-confirms-data-breach/
[20] What is ChatGPT? ChatGPT Security Risks. Accessed: Jun. 26, 2023. [Online].
Available: https://www.malwarebytes. com/cybersecurity/basics/chatgpt-ai-security
20