How Amazon Is Making Custom Chips To Catch Up In Generative A.I. Race

CNBC
12 Aug 202315:42

TLDRAmazon's Austin, Texas chip lab is developing custom microchips to compete with industry giants like Intel and Nvidia, aiming to save costs and enhance performance. These chips power large language models for the AI boom. Amazon Web Services (AWS), the world's largest cloud computing provider, has seen its profit margins decrease but still accounts for a significant portion of Amazon's overall profit. AWS is investing in custom silicon and a range of developer tools to remain at the forefront of AI advancements. Despite the competition, Amazon is confident in its ability to supply the capacity needed for AI applications, leveraging its extensive customer base and robust infrastructure.

Takeaways

  • 🏢 Amazon has a chip lab in Austin, Texas, where they design custom microchips to compete with established chip giants like Intel and Nvidia.
  • 🔧 The custom chips are aimed at powering data centers and AI applications, particularly in the wake of the AI boom sparked by ChatGPT.
  • 💡 Amazon Web Services (AWS) is the world's largest cloud computing provider and a major contributor to Amazon's overall profitability.
  • 📉 Despite a year-over-year decrease in Q2 operating income, AWS still accounts for a significant portion of Amazon's total operating profit.
  • 🚀 AWS has a history of innovation in the chip industry, having acquired Annapurna Labs and produced its own ARM-based server chip, Graviton.
  • 🧠 Amazon's AI-focused chips, Trainium and Inferentia, target the training and inference stages of machine learning, respectively.
  • 🌐 While Nvidia remains the leader in AI chip technology, Amazon is not the only company venturing into custom silicon; Apple and Google have also developed their own chips.
  • 🤖 Amazon has been leveraging AI and machine learning technologies since the late 1990s, with applications ranging from recommendation engines to robotics in fulfillment centers.
  • 🛠️ AWS offers a broad range of AI services and tools, including Bedrock for accessing large language models and SageMaker for machine learning development.
  • 📈 AWS has a significant customer base utilizing its machine learning services, and the company is focused on helping these customers expand into AI-driven solutions.
  • 🔒 As generative AI advances, companies like Amazon are also focusing on ensuring responsible and secure use of these technologies, including compliance with data protection regulations.

Q & A

  • Where is Amazon's chip lab located?

    -Amazon's chip lab is located in Austin, Texas.

  • What are the two types of microchips designed in Amazon's lab?

    -Amazon's lab designs microchips to power data centers and AI boom applications.

  • How does Amazon's custom microchips benefit the company?

    -Amazon's custom microchips help the company save money and boost performance, as Amazon is one of the biggest buyers of data center chips in the world.

  • What is the significance of the chips that CEO Adam Selipsky referred to in the CNBC interview?

    -The chips that CEO Adam Selipsky mentioned are significant because they power large language models and are part of the AI boom.

  • What is Amazon Web Services (AWS) and its role in Amazon's profitability?

    -Amazon Web Services (AWS) is the world's biggest cloud computing provider and the most profitable arm of Amazon, accounting for 70% of its overall Q2 operating profit.

  • How does AWS plan to leverage its customer's data stored on its platform?

    -AWS plans to leverage its customers' data stored on its platform to help them customize the models that they're using to power their generative AI applications.

  • What is the current status of AWS's profit margins?

    -AWS's profit margins have historically been far higher than those at Google Cloud, but those margins have been narrowing and its growth is slowing down compared to before.

  • What is Amazon's strategy in response to the generative AI boom?

    -Amazon's strategy is to focus on providing a broad AI infrastructure for machine learning with dozens of services and custom silicon, and to offer tools like Bedrock to enhance software using generative AI.

  • What are Trainium and Inferentia chips, and how do they contribute to AWS's AI offerings?

    -Trainium and Inferentia are AWS's own AI-focused chips. Trainium, launched in 2021, helps customers train their own machine learning models with improved price-performance. Inferentia, launched in 2019, is designed for low-cost, high-throughput, low-latency machine learning inference.

  • How does AWS's approach to generative AI differ from Microsoft and Google?

    -AWS focuses on providing tools and a broad AI infrastructure rather than building a direct competitor to ChatGPT. It offers access to multiple providers' models through its Bedrock service, allowing customers to choose the right tool for the job.

  • What are some of the AI services and products offered by AWS?

    -AWS offers a variety of AI services and products, including Amazon SageMaker, AWS HealthScribe, CodeWhisperer, and a $100 million generative AI innovation center.

  • How is AWS addressing concerns about the responsible and secure use of generative AI?

    -AWS ensures that any model used through its Bedrock service runs in an isolated, virtual private cloud environment with encryption and the same AWS access controls. The company is also involved in initiatives to establish regulatory guardrails for AI.

Outlines

00:00

🏢 Amazon's Custom Chip Lab in Texas

The video script begins with a scene set in an unmarked office building in Austin, Texas, where Amazon's chip lab is located. The lab employs a handful of workers who design two types of microchips intended for data centers and AI applications. Unlike chips from well-known companies like Nvidia and AMD, Amazon's chips are part of their strategy to save costs and enhance performance, given their massive demand for data center chips. The lab's activities include rigorous testing of the chips they produce. The narrative then shifts to the broader context of Amazon Web Services (AWS), the world's leading cloud computing provider and a highly profitable part of Amazon. Despite a year-over-year decrease in operating income, AWS remains a significant contributor to Amazon's overall profits. This financial strength positions AWS well to invest in custom silicon and a suite of developer tools, potentially positioning Amazon at the heart of the AI boom. The script also mentions that AWS customers are utilizing large amounts of data stored on AWS to refine their AI models, and that while competitors like Microsoft and Google have made significant investments in AI, Amazon is playing catch-up in the generative AI space.

05:05

🤖 Advancements in AI Chips and Training

This paragraph discusses Amazon's AI chip offerings, including Trainium and Inferentia, which were introduced to the market in 2021 and 2019, respectively. Trainium is designed to assist customers in training their own machine learning and generative AI models, offering significant improvements in price-performance compared to other training methods on AWS. However, Nvidia's GPUs are still the leading choice for training due to their extensive software ecosystem. The paragraph also touches on the fact that other tech giants like Apple and Google have developed their own custom silicon, and that Microsoft is reportedly working on an AI chip in partnership with AMD. Amazon's long history with machine learning and its current focus on providing a broad AI infrastructure, including numerous services and tools like Bedrock, are also highlighted. The narrative emphasizes Amazon's strategy of offering various state-of-the-art models from multiple providers to meet different customer needs.

10:05

💡 New AI Services and Initiatives by Amazon

The script describes several of Amazon's new AI services, including AWS HealthScribe, which assists doctors in drafting patient summaries, and CodeWhisperer, a tool that generates code recommendations from natural language prompts. The paragraph also mentions SageMaker, Amazon's machine learning hub, and how it has been utilized by companies like Autodesk to design lighter aircraft components. In addition, the script covers Amazon's $100 million investment in a generative AI innovation center, aimed at helping customers understand and implement generative AI solutions. The narrative also explores the reasons why companies might choose AWS over competitors like Google and Microsoft, especially those already familiar with AWS's services and infrastructure. The paragraph concludes with a discussion on the rapid development of generative AI applications and the need for responsible and secure usage of these tools.

15:06

🏁 The Race for Generative AI and Chips

The final paragraph of the script likens the competition among tech giants in the generative AI space to a long race, emphasizing that the initial steps are less important than the finish line. It suggests that the true measure of success will be determined by the long-term outcomes and innovations in AI and chip technology. The paragraph also touches on the importance of addressing regulatory and national security concerns as the development of AI applications and chips continues to accelerate.

Mindmap

Keywords

💡microchips

Microchips, as discussed in the video, are tiny electronic components that serve as the building blocks of modern computing and communication devices. They are designed to process and store information, enabling the operation of complex systems such as data centers and AI technologies. In the context of the video, Amazon is developing custom microchips to power data centers and AI applications, aiming to compete with established chip manufacturers like Nvidia and Intel.

💡AI boom

The AI boom refers to the rapid growth and increased interest in artificial intelligence technologies, applications, and industries. This surge is driven by advancements in machine learning, big data, and computing power, which have made AI more accessible and effective for various tasks. In the video, the AI boom is a key factor that has prompted Amazon to design its own microchips to meet the demand for powerful and efficient AI processing capabilities.

💡data centers

Data centers are large facilities that house computer servers and associated components for storing, processing, and distributing large amounts of data. They are the backbone of cloud computing and the internet, providing the infrastructure necessary for the operation of online services, applications, and data-intensive tasks like AI. In the context of the video, Amazon is designing microchips specifically to enhance the performance and efficiency of data centers, which are crucial for running complex AI algorithms and handling massive datasets.

💡custom silicon

Custom silicon refers to the process of designing and manufacturing specialized microchips tailored to specific needs or functions. This customization allows for optimized performance, efficiency, and cost savings compared to using standard, off-the-shelf chips. In the video, Amazon's investment in custom silicon is a strategic move to enhance its competitiveness in the AI and cloud computing markets, as well as to reduce reliance on third-party chip suppliers.

💡Amazon Web Services (AWS)

Amazon Web Services, or AWS, is a subsidiary of Amazon that provides on-demand cloud computing platforms and APIs to individuals, companies, and governments. It is the most profitable part of Amazon and plays a significant role in the global cloud computing market. In the video, AWS's involvement in chip design and AI tools underscores its commitment to maintaining its market leadership by offering a comprehensive suite of services and technologies for its customers.

💡Inferentia

Inferentia is one of the AI-focused chips developed by Amazon. It is designed for machine learning inference, which is the process of making predictions based on trained models. Inferentia aims to provide high throughput and low latency, making it an efficient choice for handling the rapid response requirements of generative AI applications. The chip's performance is highlighted as being significantly better than other options available on AWS for inference tasks.

💡Trainium

Trainium is Amazon's AI chip designed for training machine learning models. It is engineered to offer significant improvements in price-performance ratio compared to other training methods on AWS. Trainium is part of Amazon's broader strategy to provide specialized hardware for different stages of AI processing, aiming to capture a larger share of the AI market and enhance its cloud services.

💡Nvidia

Nvidia is a major technology company known for its graphics processing units (GPUs), which have become essential for AI and machine learning applications due to their parallel processing capabilities. Nvidia's GPUs are considered the industry standard for training complex AI models, and the company has built a robust software ecosystem around its products over the years. In the video, Nvidia's dominance in the AI chip market is emphasized, with Amazon even launching new AI acceleration hardware powered by Nvidia's H100s.

💡generative AI

Generative AI refers to artificial intelligence systems that can create new content, such as text, images, or audio, based on patterns learned from existing data. This type of AI is particularly relevant for applications like chatbots, content creation, and automated decision-making. The video discusses the competition among tech giants to develop and deploy generative AI tools and infrastructure, with Amazon aiming to catch up with its own offerings like Inferentia and Trainium chips.

💡Bedrock

Bedrock is a cloud service announced by Amazon that aims to help developers enhance their software using generative AI. It provides access to large language models (LLMs) from various providers, including Amazon's own Titan models, enabling customers to integrate AI capabilities into their applications without having to build their own models from scratch. This service is part of Amazon's broader AI infrastructure and strategy to support its customers in leveraging AI technologies.

💡machine learning services

Machine learning services are cloud-based platforms and tools that enable developers and data scientists to build, train, and deploy machine learning models. These services often provide a range of algorithms, data processing capabilities, and infrastructure to facilitate the development and deployment of AI applications. In the video, Amazon's machine learning services, such as SageMaker, are highlighted as a key part of its AI strategy, offering customers over 20 different services to support their AI initiatives.

Highlights

Amazon's Austin, Texas chip lab is designing custom microchips to compete with giants like Intel and Nvidia.

The custom chips are intended to save money and boost performance for Amazon, which is one of the largest buyers of data center chips globally.

CEO Adam Selipsky stated that Amazon is better positioned than any other company to supply the capacity needed for AI boom.

Amazon Web Services (AWS) is the world's largest cloud computing provider and the most profitable part of Amazon.

AWS's custom silicon initiative is funded by its significant operating income, which supports the development of custom chips and developer tools.

AWS customers have large amounts of data stored on AWS, which can help customize models for generative AI applications.

Despite slowing growth and narrowing profit margins, AWS continues to invest in custom chips and AI technologies.

Amazon's acquisition of Annapurna Labs in 2015 accelerated its entry into the chip business.

AWS designs chips in multiple locations, including Silicon Valley, Canada, and Israel, and partners with manufacturers like TSMC for production.

AWS's Nitro is a specialized hardware chip that has been widely adopted across AWS servers.

The Graviton chip, launched by AWS, competes with x86 CPUs from AMD and Intel and is now in its third generation.

Trainium and Inferentia are Amazon's AI-focused chips, designed for training and inference in machine learning models.

AWS has launched new AI acceleration hardware powered by Nvidia GPUs, improving performance and reducing costs.

Amazon is working on a broader AI infrastructure, offering a range of services beyond just chips.

Amazon's AI products, including Titan and Bedrock, are being used by major customers like Philips, 3M, Old Mutual, and HSBC.

AWS has established a $100 million generative AI innovation center to help customers with their AI initiatives.

Amazon is focusing on tools rather than directly competing with ChatGPT, offering a platform that integrates various AI models.

AWS has over 100,000 customers using machine learning services, many of whom have standardized on SageMaker for building custom models.

The development of generative AI applications and the custom chips to power them is a long-term race, with various companies adopting different strategies.