April 22, 2022, 1:04 p.m. | Sascha Kirch

Towards Data Science - Medium towardsdatascience.com

Understand your Hardware to Optimize your Software

Cloud TPUv3 POD by Google Cloud under (CC BY 4.0)

In this post I will show you the basic principles of tensor processing units (TPUs) from a hardware perspective and show you step-by-step how you can perform accelerated distributed training on a TPU using TensorFlow to train your own models.

Outline

  1. Introduction
    1.1. Tensor Processing Units
    1.2. Distribution Strategies
  2. Implementing Distributed Training on TPU with TensorFlow
    2.1. Hardware Detection
    2.2. Distributed Dataset
    2.3. …

distributed distributed systems google python tensorflow tpu training

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

C003549 Data Analyst (NS) - MON 13 May

@ EMW, Inc. | Braine-l'Alleud, Wallonia, Belgium

Marketing Decision Scientist

@ Meta | Menlo Park, CA | New York City