In this Acquired episode, hosts explore how Google has evolved into a major player in artificial intelligence, tracing the company's AI journey from Larry Page's early vision through to its current position. The discussion covers Google's strategic moves in AI development, including its recruitment of key researchers, the acquisition of DeepMind, and the development of custom AI hardware through Tensor Processing Units (TPUs).
The episode examines Google's response to recent AI developments, particularly following ChatGPT's emergence, which led to significant organizational changes and new initiatives like Bard and Gemini. The hosts analyze how Google's extensive infrastructure and data resources provide advantages in AI development, while also considering how AI assistants could affect the company's traditional search-based revenue model.

Sign up for Shortform to access the whole episode summary along with additional materials like counterarguments and context.
The history of AI at Google was shaped by Larry Page's early vision, influenced by his father's pioneering work in AI during the 1960s and 70s. During the 2000s, Google strategically concentrated AI talent by hiring prominent researchers like Ilya Sutskever, Jeff Hinton, and Sebastian Thrun. This concentration of expertise led to groundbreaking developments in language models and, in 2014, the strategic acquisition of DeepMind for $550 million, which significantly boosted Google's AI capabilities.
Google's Brain team revolutionized AI with the development of the Transformer architecture in 2017, which dramatically improved machine translation and laid the foundation for modern large language models. To support these advancing technologies, Google engineer Jonathan Ross initiated the development of Tensor Processing Units (TPUs), custom chips designed specifically for neural network operations. These TPUs gave Google a significant hardware advantage, reducing reliance on external providers and enabling efficient scaling through Google Cloud.
Following ChatGPT's unprecedented success, Google CEO Sundar Pichai issued a "code red" in December 2022, leading to the rapid development of Bard and a major reorganization. Under this restructuring, Google merged its Brain and DeepMind teams under Demis Hassabis's leadership and launched the Gemini initiative, aiming to standardize AI development across Google's products and services.
According to Ben Gilbert and David Rosenthal, Google's extensive infrastructure provides a unique advantage in AI development. The company's $50 billion cloud infrastructure and in-house TPU production allow for cost-effective AI deployment at scale. However, they note that AI assistants could potentially disrupt Google's traditional search-based ad revenue model. In response, Google is balancing its core search business while investing in new AI products, leveraging its vast data resources from services like Gmail, Maps, and Android to create personalized AI experiences.
1-Page Summary
Discussions on a podcast reveal the deep history of AI at Google, driven by the vision of its founders and a strategic concentration of AI talent, laying the foundations for significant breakthroughs in the field.
Larry Page, one of Google's co-founders, envisioned Google as fundamentally an artificial intelligence company from the very beginning. His vision was influenced by his father, who was a computer science professor specializing in AI at the University of Michigan. Page's father had conducted research in the field of AI during its first wave in the 1960s and 70s when it was not a popular computer science field. This had a profound impact on Larry's outlook.
David Rosenthal and Ben Gilbert discuss how Google concentrated AI talent by employing almost every notable person in AI during the 2000s. This includes hiring luminaries such as Ilya Sutskever, Jeff Hinton, and Alex Krizhevsky, not to mention other prominent figures like Dario Amodei, Andrej Karpathy, Andrew Ng, Sebastian Thrun, and Noam Shazir. These hires found fertile ground at Google to advance their groundbreaking work.
Larry Page hired Sebastian Thurn, then the head of Stanford’s AI Lab, to work on machine learning applications. Thurn suggested that Google should bring in AI academics part-time to work on various projects, which the founders agreed to. George Herrick, an engineer with a machine learning PhD from the University of Michigan, advanced a notion that would foreshadow the development of large language models. Jeff Hinton, a machine learning professor from the University of Toronto, was brought in by Thrun to give a tech talk at Google about new work on neural networks.
It was mentioned that Google engineers developed groundbreaking language models with applications across Google’s products. These models demonstrate the capability to understand and respond to human language. Jeff Dean is noted to have used the "Fill" language model to implement AdSense in a week, contributing to billions in new revenue and expanded ad inventory. By the mid-2000s, such models were using significant data center resources due to their wide applications, including predicting search queries and improving ad quality scores.
The podcast describes a series of events ...
The Historical Foundations of AI at Google
Google's relentless pursuit of innovation has led to the creation and refinement of AI technologies that have become foundational in the field.
Google's development of the Transformer architecture represents a paradigm shift in the AI field. Before the advent of the Transformer, Google Translate had relied on recurrent neural networks, which were limited by short context windows and did not scale efficiently. In 2017, the Google Brain team published a paper describing the Transformer model, which leveraged attention mechanisms to achieve better parallelization and longer context retention. This mechanism mimics how human translators work, understanding the entire context before translating. Due to its elegance, energy efficiency, and scalability, the Transformer model marked the beginning of a modern AI era.
The Transformer architecture was not only a milestone for Google Translate but also set the stage for large language models such as BERT, and later influenced models developed by other organizations, like OpenAI's ChatGPT. Recognizing its potential, Google Brain planned to extend the Transformer architecture beyond text to other modalities, such as images and audio. The Transformer model continued Google's legacy in language models and significantly reduced Google Translate's error rate by 60% when it replaced LSTMs in 2016.
Anticipating the need for high-volume matrix multiplications required by neural networks, Google embarked on the development of the Tensor Processing Unit (TPU). Initiated by Google engineer Jonathan Ross and materialized in a formal project, the custom ASIC was designed to support neural network operations efficiently and at scale.
The TPUs provided Google with a significant hardware edge. Initially kept top secret, TPUs were fundamental in powering projects like AlphaGo. Over time, TPUs evolved, with versions significantly increasing in efficiency. Google's decision to create TPUs internali ...
Key AI Technologies and Breakthroughs Developed at Google
In the face of burgeoning AI technologies from competitors like OpenAI, Google has taken critical steps to advance its AI development and maintain its position in the market.
Since the release of OpenAI's ChatGPT, Google has recognized the serious potential of large language models to disrupt traditional search paradigms. The rapid growth of ChatGPT shows its capability to reshape the way users engage with search engines, challenging Google's core business. Google, aware of the progress within its AI sector, including raw chatbot models like Lambda, responded to the tangible threat by significantly shifting its strategy.
After the launch of ChatGPT, which witnessed unprecedented adoption rates, reaching 1 million users within a week and 100 million by the end of January 2023, Google realized that AI was transitioning from a sustaining innovation to a disruptive one. As such, it posed an existential threat. Microsoft's investment of another $10 billion in OpenAI and its subsequent 100 million registered users for ChatGPT further emphasized the immediacy of the threat.
In response, Sundar Pichai, CEO of Google, issued a "code red" in December 2022, urging the company to expedite the development of native AI products. This urgency led to the release of Bard, which rebranded the Lambda model with a new chatbot interface, following the success of ChatGPT with Bing. However, the initial public release of Bard was met with criticism over its performance.
Sundar Pichai initiated a significant reorganization within Google, directing the merger of Google Brain and DeepMind into a unified team. This decision was aimed at consolidating Google's AI efforts and streamlining capabilities across the company. Demis Hassabis of DeepMind was appointed as the CEO of this integrated AI division, suggesting a concerted effort to centralize Google's AI initiatives.
Following the reorganization, a new high-performance AI system initiative called G ...
Google's Response to External AI Developments and Competition
As artificial intelligence (AI) continues to evolve, Google is uniquely positioned to benefit from its vast infrastructure and to navigate the challenges AI may pose to its traditional business models.
Ben Gilbert and David Rosenthal unpack how Google’s well-established infrastructure could give the company a significant edge in the competitive AI landscape.
Google's switch from CPUs to GPUs, orchestrated by Alex Krashevsky, Jeff Hinton, and Ilya Sutskever, led to a transformative $130 million purchase of 40,000 GPUs from Nvidia. This investment, approved by Larry Page, shows Google's understanding of deep learning's strategic importance for its future. Their massive infrastructure has integrated a profound number of GPUs for AI training and inference, providing a likely competitive advantage.
Gilbert emphasizes that Google's unique cloud infrastructure, with revenues of $50 billion, allows the company to spread the substantial costs of model training across its diverse user services. By producing their own Tensor Processing Units (TPUs), and managing all infrastructure in-house, Google achieves a more favorable cost structure for deploying AI at scale.
Running an AI data center incurs substantial costs, primarily from chips and depreciation, but Google stands out as a low-cost provider, with access to low markup hardware. Gilbert mentions that the model training costs are amortized across every Google search, allowing Google to process 10 trillion to nearly 1 quadrillion inference tokens within a surprisingly short period. This vast scale distributes the hardware and training expenses over a significant amount of value creation.
David Rosenthal notes that Google’s ownership of personalized data from products like Gmail, Maps, Docs, Chrome, and Android allows it to create highly personalized AI products. This edge comes from the fact that Google can use this data in ways that other companies cannot, given they lack access to such abundant and individualized information.
Despite Google’s strategic preparations for AI, there is a recognized potential disruption to its traditional revenue streams, specifically those tied to web search and advertising.
The rise of AI-driven products, like OpenAI's ChatGPT, suggests a potential disruption to Google’s search-based ad revenue. Gilbert and Rosenthal discuss how AI advancements, particularly AI assistants, might reduce the volume of traditional web searches. High-value queries, such as those related to trip pl ...
Business and Strategic Implications of AI for Google
Download the Shortform Chrome extension for your browser
