Cybercriminals are tricking AI into leaking your data, executing code, and sending you to malicious sites. Here's how.
Researchers say a prompt injection bug in Google's Antigravity AI coding tool could have let attackers run commands, despite ...
Antigravity Strict Mode bypass disclosed Jan 7, 2026, patched Feb 28, enables arbitrary code execution via fd -X flag.