Close Menu
Best in TechnologyBest in Technology
  • News
  • Phones
  • Laptops
  • Gadgets
  • Gaming
  • AI
  • Tips
  • More
    • Web Stories
    • Global
    • Press Release

Subscribe to Updates

Get the latest tech news and updates directly to your inbox.

What's On

This Is Why Tesla’s Robotaxi Launch Needed Human Babysitters

4 July 2025

Vivo X Fold 5, Vivo X200 FE Price in India Leaked Ahead of Debut on July 14

4 July 2025

The EU Proposes New Rules to Govern the European Space Race

4 July 2025
Facebook X (Twitter) Instagram
Just In
  • This Is Why Tesla’s Robotaxi Launch Needed Human Babysitters
  • Vivo X Fold 5, Vivo X200 FE Price in India Leaked Ahead of Debut on July 14
  • The EU Proposes New Rules to Govern the European Space Race
  • Android 16’s Live Updates to Show Active Navigation, Ongoing Phone Calls, and More on Lock Screen
  • Google Pixel 6a Owners Eligible for $100 Cash or $150 Store Credit Under Battery Performance Programme
  • Tecno Pova 7 5G – Price in India, Specifications (4th July 2025)
  • Tecno Pova 7 5G, Pova 7 Pro 5G Launched in India With MediaTek Dimensity 7300 Ultimate SoC
  • Tecno Pova 7 Pro 5G – Price in India, Specifications (4th July 2025)
Facebook X (Twitter) Instagram Pinterest Vimeo
Best in TechnologyBest in Technology
  • News
  • Phones
  • Laptops
  • Gadgets
  • Gaming
  • AI
  • Tips
  • More
    • Web Stories
    • Global
    • Press Release
Subscribe
Best in TechnologyBest in Technology
Home » This Prompt Can Make an AI Chatbot Identify and Extract Personal Details From Your Chats
News

This Prompt Can Make an AI Chatbot Identify and Extract Personal Details From Your Chats

News RoomBy News Room18 October 20243 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
Share
Facebook Twitter LinkedIn Pinterest Email

The researchers say that if the attack were carried out in the real world, people could be socially engineered into believing the unintelligible prompt might do something useful, such as improve their CV. The researchers point to numerous websites that provide people with prompts they can use. They tested the attack by uploading a CV to conversations with chatbots, and it was able to return the personal information contained within the file.

Earlence Fernandes, an assistant professor at UCSD who was involved in the work, says the attack approach is fairly complicated as the obfuscated prompt needs to identify personal information, form a working URL, apply Markdown syntax, and not give away to the user that it is behaving nefariously. Fernandes likens the attack to malware, citing its ability to perform functions and behavior in ways the user might not intend.

“Normally you could write a lot of computer code to do this in traditional malware,” Fernandes says. “But here I think the cool thing is all of that can be embodied in this relatively short gibberish prompt.”

A spokesperson for Mistral AI says the company welcomes security researchers helping it to make its products safer for users. “Following this feedback, Mistral AI promptly implemented the proper remediation to fix the situation,” the spokesperson says. The company treated the issue as one with “medium severity,” and its fix blocks the Markdown renderer from operating and being able to call an external URL through this process, meaning external image loading isn’t possible.

Fernandes believes Mistral AI’s update is likely one of the first times an adversarial prompt example has led to an LLM product being fixed, rather than the attack being stopped by filtering out the prompt. However, he says, limiting the capabilities of LLM agents could be “counterproductive” in the long run.

Meanwhile, a statement from the creators of ChatGLM says the company has security measures in place to help with user privacy. “Our model is secure, and we have always placed a high priority on model security and privacy protection,” the statement says. “By open-sourcing our model, we aim to leverage the power of the open-source community to better inspect and scrutinize all aspects of these models’ capabilities, including their security.”

A “High-Risk Activity”

Dan McInerney, the lead threat researcher at security company Protect AI, says the Imprompter paper “releases an algorithm for automatically creating prompts that can be used in prompt injection to do various exploitations, like PII exfiltration, image misclassification, or malicious use of tools the LLM agent can access.” While many of the attack types within the research may be similar to previous methods, McInerney says, the algorithm ties them together. “This is more along the lines of improving automated LLM attacks than undiscovered threat surfaces in them.”

However, he adds that as LLM agents become more commonly used and people give them more authority to take actions on their behalf, the scope for attacks against them increases. “Releasing an LLM agent that accepts arbitrary user input should be considered a high-risk activity that requires significant and creative security testing prior to deployment,” McInerney says.

For companies, that means understanding the ways an AI agent can interact with data and how they can be abused. But for individual people, similarly to common security advice, you should consider just how much information you’re providing to any AI application or company, and if using any prompts from the internet, be cautious of where they come from.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleNASA’s Perseverance rover shares update during tricky Mars climb
Next Article Windows 11 to let you use your phone as a webcam

Related Articles

News

This Is Why Tesla’s Robotaxi Launch Needed Human Babysitters

4 July 2025
News

The EU Proposes New Rules to Govern the European Space Race

4 July 2025
News

The Person in Charge of Testing Tech for US Spies Has Resigned

4 July 2025
News

Trump’s Defiance of TikTok Ban Prompted Immunity Promises to 10 Tech Companies

4 July 2025
News

The 61 Best Early Amazon Prime Day Deals

3 July 2025
News

A Game Called Date Everything Literally Lets You Date Everything—Except People

3 July 2025
Demo
Top Articles

ChatGPT o1 vs. o1-mini vs. 4o: Which should you use?

15 December 2024101 Views

Costco partners with Electric Era to bring back EV charging in the U.S.

28 October 202495 Views

Oppo Reno 14, Reno 14 Pro India Launch Timeline and Colourways Leaked

27 May 202581 Views

Subscribe to Updates

Get the latest tech news and updates directly to your inbox.

Latest News
Laptops

Tecno Pova 7 5G – Price in India, Specifications (4th July 2025)

News Room4 July 2025
Phones

Tecno Pova 7 5G, Pova 7 Pro 5G Launched in India With MediaTek Dimensity 7300 Ultimate SoC

News Room4 July 2025
Laptops

Tecno Pova 7 Pro 5G – Price in India, Specifications (4th July 2025)

News Room4 July 2025
Most Popular

The Spectacular Burnout of a Solar Panel Salesman

13 January 2025124 Views

ChatGPT o1 vs. o1-mini vs. 4o: Which should you use?

15 December 2024101 Views

Costco partners with Electric Era to bring back EV charging in the U.S.

28 October 202495 Views
Our Picks

Android 16’s Live Updates to Show Active Navigation, Ongoing Phone Calls, and More on Lock Screen

4 July 2025

Google Pixel 6a Owners Eligible for $100 Cash or $150 Store Credit Under Battery Performance Programme

4 July 2025

Tecno Pova 7 5G – Price in India, Specifications (4th July 2025)

4 July 2025

Subscribe to Updates

Get the latest tech news and updates directly to your inbox.

Facebook X (Twitter) Instagram Pinterest
  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact Us
© 2025 Best in Technology. All Rights Reserved.

Type above and press Enter to search. Press Esc to cancel.