San Francisco: A growing divide is emerging within Google as employees raise concerns over being required to use internal artificial intelligence tools they do not fully trust. The issue highlights tensions between the company’s aggressive AI adoption strategy and the practical challenges faced by its engineering teams.

Reports suggest that while leadership is encouraging widespread use of in-house AI systems, not all employees are convinced the tools are ready to meet their needs—particularly when better-performing alternatives exist but remain restricted.

Unequal access to AI tools fuels frustration

At the centre of the issue is limited access to advanced external AI tools. Teams within Google DeepMind have reportedly been allowed to use Claude, a popular AI coding assistant developed by Anthropic.

However, most other teams across Google are expected to rely on the company’s own AI model, Gemini. This uneven access has led to frustration among engineers who believe internal tools are still catching up in terms of coding performance and efficiency.

According to reports, this has created a paradox where employees are pushed to integrate AI deeply into their workflows but feel they are not being given the most effective tools available.

AI usage tied to performance metrics

The pressure to adopt AI is not merely informal. In some divisions, employees are reportedly being evaluated based on their use of AI tools. Engineers are expected not only to incorporate AI into coding tasks but also to build systems that enhance productivity through automation.

This marks a shift in how performance is measured at Google, a company traditionally known for prioritising engineering excellence and autonomy. Some employees view this change as premature, given their concerns about the reliability of current internal tools.

Google’s “dogfooding” strategy under scrutiny

Google’s preference for its own tools is rooted in its long-standing “dogfooding” philosophy—encouraging employees to use in-house products to refine them before public release. The company also cites technical reasons, including the complexity of integrating third-party tools into its customised infrastructure.

While this approach can accelerate product improvement, some employees argue it is limiting their efficiency and flexibility in day-to-day work.

Comparisons with competitors intensify debate

The contrast becomes more pronounced when compared to rivals like Meta, where employees reportedly have greater freedom to use external AI tools, including Claude. This openness has led some Googlers to question whether stricter internal policies may be slowing innovation.

The debate recently gained public attention after comments from Steve Yegge, who claimed that AI adoption within Google was lagging behind expectations. His remarks drew a sharp rebuttal from Demis Hassabis, who dismissed the claims as “completely false” and “pure clickbait.”

Despite the denial, the discussion continued, with Yegge later stating that some Google employees had privately supported aspects of his observations, particularly around uneven access to tools.

Internal tensions over future access

One of the more contentious developments reportedly involved a proposal to remove access to Claude altogether—even from DeepMind teams. The suggestion was met with strong resistance, with some engineers reportedly threatening to leave if such restrictions were enforced.

This reaction underscores the depth of concern among employees, not just about access to specific tools, but about broader questions of trust, efficiency, and the direction of Google’s AI strategy.

Conclusion

Google’s internal AI push reflects a broader industry shift towards automation and AI-driven productivity. However, the situation also highlights the challenges of balancing innovation with employee trust and practical usability.

As the company continues to invest heavily in AI, how it addresses these internal concerns could play a key role in shaping both its workplace culture and its competitive position in the rapidly evolving AI landscape.