Google CEO: The model processes over 16 billion tokens per minute

robot
Abstract generation in progress

On the evening of April 22 Beijing time, the Google Cloud Next conference opened, and Google CEO Sundar Pichai posted a message stating that by directly using APIs, Google’s models process over 16 billion tokens per minute, compared to 10 billion tokens last quarter. He predicts that by 2026, more than half of Google’s total investment in machine learning computing will be dedicated to cloud services. At the conference, Google will also launch the eighth-generation TPU, along with two chips for intelligent agents—TPU 8t and TPU 8i. The blog mentioned that they are designed to handle demanding AI workloads and adapt to the continuously evolving model architectures. (First Financial)

View Original
This page may contain third-party content, which is provided for information purposes only (not representations/warranties) and should not be considered as an endorsement of its views by Gate, nor as financial or professional advice. See Disclaimer for details.
  • Reward
  • Comment
  • Repost
  • Share
Comment
Add a comment
Add a comment
No comments
  • Pin