OpenAI's ChatGPT chatbot has gained considerable attention and praise in the tech and mobile user community in recent months. However, there is a legitimate concern if ChatGPT is used as a tool for network attacks through software containing malicious code. This could result in severe damage.
So, how does the ChatGPT story unfold in the creation of this malicious code? Let's explore it with Mytour Supermarket in the following article!

Malicious actors exploiting ChatGPT can generate code to compromise user data
How can ChatGPT generate dangerous malicious code?
Recently, security expert Aaron Mulgrew of Forcepoint shared his method of creating a 'zero-day' malware using the assistance of ChatGPT.
Although this application has implemented protective measures against anyone attempting to request the creation of malicious code, Mulgrew, with his extensive experience, found a loophole. He utilized this chatbot to generate distinct lines of malicious code, functioning based on customizable features.
Security expert crafts malicious software
Specifically, this malicious software has been disguised by Mulgrew as a screen protector application for phones, be it Samsung or PC. Upon activation by the user, the application will automatically initiate. Subsequently, it will sift through files (such as images, Word documents, and PDFs) to extract valuable data.
After customizing individual functions requested with ChatGPT's prior assistance, Mulgrew created a virtually undetectable data-stealing file. This sophisticated malware is so cunning that it can bypass all supported antivirus operating systems he tested.

Cybersecurity expert develops malicious software leveraging ChatGPT
Hidden Risks of Concern
Noteworthy is the Zero-day software using steganography to break down stolen data into smaller parts and conceal them within images on the computer. These images are then uploaded to a Google Drive folder, an intricately designed detection-avoidance process.

The potential threat of ChatGPT may be further intensified by hackers
Not stopping there, Mulgrew fine-tuned and added his code to counter virus detection using simple 'PROMPT' queries on ChatGPT to inquire about the safety level of usage. He tested this application on VirusTotal without any detection of harmful warnings.
However, the entire process of testing to create Zero-day was conducted by Mulgrew on his personal computer, and there have been no attacks from the malicious code generated by ChatGPT. Mulgrew mentioned lacking experience in complex programming, but it seems ChatGPT's protective measures are still not strong enough to resist his tests. If exploited by hackers, this could be highly alarming.
Recently, there have been reports about ChatGPT potentially creating undetectable malicious code on phones like the Samsung Galaxy S23 Ultra 5G 256GB or computers. It is hoped that in the near future, OpenAI will need to enhance ChatGPT further. Stay tuned to Mytour Supermarket for more updates!
