News

Not a very smart home: crims could hijack smart-home boiler, open and close powered windows and more. Now fixed ...
A prompt injection attack using calendar invites can be used for real-world effects, like turning off lights, opening window ...
OpenAI's ChatGPT can easily be coaxed into leaking your personal data — with just a single "poisoned" document. As Wired ...
For likely the first time ever, security researchers have shown how AI can be hacked to create real world havoc, allowing ...
Researchers demonstrated a way to hack Google Home devices via Gemini. Keeping your devices up-to-date on security patches is ...
Security researchers found a weakness in OpenAI’s Connectors, which let you hook up ChatGPT to other services, that allowed ...
Researchers used a calendar invite to make Gemini control lights, windows, and more in a real-world smart home hack.
Once they're in, a hacker can use Gemini to start Zoom calls, send spam, read browser content, and delete calendar events.
The hack, laid out in a paper titled “Invitation Is All You Need!”, the researchers lay out 14 different ways they were able ...
The promptware attack begins with a calendar appointment containing a description that is actually a set of malicious ...
This Wired article shows how an indirect prompt injection attack against a Gemini-powered AI assistant could cause the bot to ...
Anywhere a user can put stuff is prone to injection flaws. Tip: Always validate and sanitize anything users can send. It’s ...