What does it say when the man selling the technology tells you to be skeptical of it? In an exclusive BBC interview, Alphabet CEO Sundar Pichai urged people not to �blindly trust� AI outputs, acknowledging that even state-of-the-art models are prone to errors? The timing matters: Google is rolling out Gemini deeper into Search with an �AI Mode,� while racing rivals for market share?
Useful, powerful�and fallible
Pichai�s message is blunt: large language models hallucinate? �People have to learn to use these tools for what they�re good at, and not blindly trust everything they say,� he said, adding that Google will be �bold and responsible� in deployment? He noted the need for a �rich information ecosystem� beyond chatbots�think traditional search, official sources, and human experts�which dovetails with BBC testing this year that found major chatbots introduced significant inaccuracies when summarizing news?
That caution sits alongside a confidence play? Google is integrating Gemini more tightly into Search to give users an �expert-like� conversational experience? It�s a bet that better orchestration will make AI more useful without promising infallibility?
AI is already in the cyber fight�on both sides
If you need a reason not to over-trust AI, look at how attackers are using it? Anthropic recently reported a state-linked group used its tools to automate 80�90% of a cyber-espionage operation across about 30 targets�reconnaissance, vulnerability discovery, exploitation, and data exfiltration included? Anthropic banned related accounts and warned defenders to assume a �fundamental change� and start using AI for SOC automation and incident response?
But the story isn�t one-sided? Independent researchers pushed back on the �90% autonomous� claim, noting that success rates were low and that the attackers leaned heavily on standard open-source tools? Dan Tentler of Phobos Group questioned whether models reliably deliver such attack assistance given common guardrails and hallucinations? In other words: AI can accelerate the boring middle of an intrusion�but it still makes things up, and humans still drive strategy?
Defensive pressure is rising fast
Meanwhile, the background noise of the internet is getting louder? Microsoft says it absorbed a record 15?7 Tbit/s distributed denial-of-service (DDoS) attack in October, driven by a Mirai-style botnet abusing compromised home routers and cameras? DDoS floods overwhelm systems with traffic; even when blocked, they force costly capacity planning and mitigation spending?
On the endpoint side, Dell warned that its ControlVault3 credential storage�used to secure passwords and biometrics�contained multiple vulnerabilities, including buffer overflows and a hardcoded password in earlier versions? Patching to version 6?2?36?47 is urgent for enterprises; a compromised credential vault is a jackpot for attackers?
Market exuberance meets operational reality
Pichai also acknowledged the elephant in the boardroom: a potential AI bubble? �No company is going to be immune,� he told the BBC, likening today�s exuberance to the dot-com era�excess investment now, enduring impact later? That doesn�t negate the value of AI; it reframes near-term expectations? For leaders, the risk is misallocating capital to shiny demos while underfunding the unglamorous controls�data governance, red-team testing, patching, and incident preparedness�that keep the lights on?
What leadership looks like now
If you�re rolling AI into customer service, finance, or code generation, Pichai�s warning is operational guidance:
- Layered verification: Pair generative outputs with retrieval from vetted sources and enforce human-in-the-loop sign-off for high-stakes actions?
- Abuse-resistant design: Expect prompt-bypass attempts? Test models against jailbreaking and tool orchestration abuse, and log/alert on unusual chaining behavior?
- Security parity: Budget for DDoS mitigation, identity hardening, and rapid patch management�especially for credential stores and endpoint agents?
- Reality checks: Track measurable ROI (defect rates, ticket resolution times, revenue uplift) and treat vendor claims about autonomy with healthy skepticism?
Pichai also underscored an image-authenticity initiative: open-sourcing tech to detect whether an image was AI-generated? That�s a helpful building block, but as Anthropic�s episode shows, the attack surface is bigger than content authenticity? The agenda now is end-to-end resilience: identity, data, network, and model safety?
The takeaway: AI can be transformative and untrustworthy at the same time? Businesses that thrive will treat it like a powerful intern�fast, tireless, occasionally wrong�and build systems and culture that assume both speed and error? Skepticism isn�t a brake on innovation; it�s how you ship responsibly?

