mastouille.fr est l'un des nombreux serveurs Mastodon indépendants que vous pouvez utiliser pour participer au fédiverse.
Mastouille est une instance Mastodon durable, ouverte, et hébergée en France.

Administré par :

Statistiques du serveur :

597
comptes actifs

#promptinjection

0 message0 participant0 message aujourd’hui

"Nikkei Asia has found that research papers from at least 14 different academic institutions in eight countries contain hidden text that instructs any AI model summarizing the work to focus on flattering comments.

Nikkei looked at English language preprints – manuscripts that have yet to receive formal peer review – on ArXiv, an online distribution platform for academic work. The publication found 17 academic papers that contain text styled to be invisible – presented as a white font on a white background or with extremely tiny fonts – that would nonetheless be ingested and processed by an AI model scanning the page."

#PromptInjection
#AcademicsBehavingBadly

theregister.com/2025/07/07/sch

The Register · Scholars sneaking phrases into papers to fool AI reviewersPar Thomas Claburn

⚠️ 𝗛𝗮𝘃𝗲 𝘆𝗼𝘂 𝗵𝗲𝗮𝗿𝗱 𝗮𝗯𝗼𝘂𝘁 𝗘𝗰𝗵𝗼𝗟𝗲𝗮𝗸? Researchers showed a single email could silently pull data from Microsoft Copilot—the first documented zero-click attack on an AI agent.

Last week, we shared a new paper dropped outlining six guardrail patterns to stop exactly this class of exploit.

Worth pairing the real-world bug with the proposed fixes. Links on the replies.
#PromptInjection #AIDesign #FOSS #Cybersecurity

"Aim Labs reported CVE-2025-32711 against Microsoft 365 Copilot back in January, and the fix is now rolled out.

This is an extended variant of the prompt injection exfiltration attacks we've seen in a dozen different products already: an attacker gets malicious instructions into an LLM system which cause it to access private data and then embed that in the URL of a Markdown link, hence stealing that data (to the attacker's own logging server) when that link is clicked.

The lethal trifecta strikes again! Any time a system combines access to private data with exposure to malicious tokens and an exfiltration vector you're going to see the same exact security issue.

In this case the first step is an "XPIA Bypass" - XPIA is the acronym Microsoft use for prompt injection (cross/indirect prompt injection attack). Copilot apparently has classifiers for these, but unsurprisingly these can easily be defeated:"

simonwillison.net/2025/Jun/11/

Simon Willison’s WeblogBreaking down ‘EchoLeak’, the First Zero-Click AI Vulnerability Enabling Data Exfiltration from Microsoft 365 CopilotAim Labs reported CVE-2025-32711 against Microsoft 365 Copilot back in January, and the fix is now rolled out. This is an extended variant of the prompt injection exfiltration attacks we've …

AI-powered features are the new attack surface! Check out our new blog in which LMG Security’s Senior Penetration Tester Emily Gosney @baybedoll shares real-world strategies for testing AI-driven web apps against the latest prompt injection threats.

From content smuggling to prompt splitting, attackers are using natural language to manipulate AI systems. Learn the top techniques—and why your web app pen test must include prompt injection testing to defend against today’s AI-driven threats.

Read now: lmgsecurity.com/are-your-ai-ba

LMG SecurityAre Your AI-Backed Web Apps Secure? Why Prompt Injection Testing Belongs in Every Web App Pen Test | LMG SecurityDiscover how prompt injection testing reveals hidden vulnerabilities in AI-enabled web apps. Learn real-world attack examples, risks, and why your pen test must include LLM-specific assessments.

New #Blog: #Perplexity #AI is susceptible to #promptinjection by scraped content (and some other issues)

I had a quick play around with Perplexity after reading @robb's excellent post on their behaviour.

I fell down a bit of a rabbit hole and this post is the result.

Includes:

- Prompt injections from arbitrary pages
- Reproducing Robb's findings
- Messing with images on their results page
- Catching them grabbing images
- Testing prompt robustness

bentasker.co.uk/posts/blog/sec

www.bentasker.co.uk · Perplexity.AI Is Susceptible to Prompt Injection From Arbitrary Pages After reading an excellent blog post last week, followed by some news stories, I decided to have a play around with Perplexity.AI to see whether I could reproduce what was reported. However, I got a l

Okay, new anti-AI scraping prompt injection attack incoming!

I'm still testing, but currently ChatGPT3.5 just gets stuck in a loop of making ASCII art penises after reading my profile. It also causes it to make ASCII art dicks for every other profile it tries to scrape afterwards too 🤣🤣

#AI#LLM#GPT

💬 GitHub Copilot Chat: From Prompt Injection to Data Exfiltration · Embrace The Red

「 GitHub Copilot Chat is a VS Code Extension that allows a user to chat with source code, refactor code, get info about terminal output, or general help about VS Code, and things along those lines 」

embracethered.com/blog/posts/2

Embrace The Red GitHub Copilot Chat: From Prompt Injection to Data Exfiltration · Embrace The Red

There are quite a few LLM pen testing tools out there, breaking the boundaries of what models are supposed to do by employing prompt injection and jail breaking techniques.
With Microsoft releasing #PyRIT and getting a lot of visibility for it, we wanted to highlight some of the other tools for the community:

- garak
github.com/leondz/garak

- HouYi
github.com/LLMSecurity/HouYi

- JailbreakingLLMs
github.com/patrickrchao/Jailbr

- llm-attacks
github.com/llm-attacks/llm-att

- PromptInject
github.com/agencyenterprise/Pr

- LLM-Canary
github.com/LLM-Canary/LLM-Cana

- And now, of course, PyRIT
github.com/Azure/PyRIT

With thanks to Idan Gelbourt and Simo Jaanus for the research!

More AI security related posts and research will be published in the future from Knostic even while we’re still in stealth. Follow us to stay in the loop.