🎉 [Gate 30 Million Milestone] Share Your Gate Moment & Win Exclusive Gifts!
Gate has surpassed 30M users worldwide — not just a number, but a journey we've built together.
Remember the thrill of opening your first account, or the Gate merch that’s been part of your daily life?
📸 Join the #MyGateMoment# campaign!
Share your story on Gate Square, and embrace the next 30 million together!
✅ How to Participate:
1️⃣ Post a photo or video with Gate elements
2️⃣ Add #MyGateMoment# and share your story, wishes, or thoughts
3️⃣ Share your post on Twitter (X) — top 10 views will get extra rewards!
👉
Google AI Agent Uncovers Critical SQLite Flaw Before Exploitation
HomeNews* Google used its AI-powered framework to spot a major security flaw in the open-source SQLite database before it was widely exploited.
Google described this security flaw as critical, noting that threat actors were aware of it and could have exploited it. “Through the combination of threat intelligence and Big Sleep, Google was able to actually predict that a vulnerability was imminently going to be used and we were able to cut it off beforehand,” said Kent Walker, President of Global Affairs at Google and Alphabet, in an official statement. He also said, “We believe this is the first time an AI agent has been used to directly foil efforts to exploit a vulnerability in the wild.”
Last year, Big Sleep also detected a separate SQLite vulnerability—a stack buffer underflow—that could have led to crashes or attackers running arbitrary code. In response to these incidents, Google released a white paper that recommends clear human controls and strict operational boundaries for AI agents.
Google says traditional software security controls are not enough, as they don’t provide the needed context for AI agents. At the same time, security based only on AI’s judgment does not provide strong guarantees because of weaknesses like prompt injection. To tackle this, Google uses a multi-layered, “defense-in-depth” approach that blends traditional safeguards and AI-driven defenses. These layers aim to reduce risks from attacks, even if the agent’s internal process is manipulated by threats or unexpected input.
Previous Articles: