This page explains Agent Platform's PyTorch integration and provides resources that show you how to use PyTorch on Agent Platform. Agent Platform's PyTorch integration makes it easier for you to train, deploy, and orchestrate PyTorch models in production.
Run code in notebooks
Agent Platform provides two options for running your code in notebooks, Colab Enterprise and Agent Platform Workbench. To learn more about these options, see choose a notebook solution.
Prebuilt containers for training
Agent Platform provides prebuilt Docker container images for model training. These containers are organized by machine learning frameworks and framework versions and include common dependencies that you might want to use in your training code. To learn about which PyTorch versions have prebuilt training containers and how to train models with a prebuilt training container, see Prebuilt containers for custom training.
Prebuilt containers for serving inferences
Agent Platform provides prebuilt Docker container images for serving both batch and online inferences. These containers are organized by machine learning frameworks and framework versions and include common dependencies that you might want to use in your inference code. To learn about which PyTorch versions have prebuilt inference containers and how to serve models with a prebuilt inference container, see Prebuilt containers for custom training.
Distributed training
You can run distributed training of PyTorch models on Agent Platform. For multi-worker training, you can use Reduction Server to optimize performance even further for all-reduce collective operations. To learn more about distributed training on Agent Platform, see Distributed training.
Resources for using PyTorch on Agent Platform
To learn more and start using PyTorch in Agent Platform, see the following resources:
- How to train and tune PyTorch models on Agent Platform: Learn how to use Agent Platform Training to build and train a sentiment text classification model using PyTorch and Agent Platform Hyperparameter Tuning to tune hyperparameters of PyTorch models.
- How to deploy PyTorch models on Agent Platform: Walk through the deployment of a Pytorch model using TorchServe as a custom container, by deploying the model artifacts to a Agent Platform Inference service.
- Orchestrating PyTorch ML Workflows on Agent Platform Pipelines: See how to build and orchestrate ML pipelines for training and deploying PyTorch models on Google Cloud Agent Platform using Agent Platform Pipelines.
- Scalable ML Workflows using PyTorch on Kubeflow Pipelines and Vertex Pipelines: Take a look at examples of PyTorch-based ML workflows on OSS Kubeflow Pipelines, (part of the Kubeflow project) and Agent Platform Pipelines. We share new PyTorch built-in components added to the Kubeflow Pipelines.
- Serving PyTorch image models with prebuilt containers on Vertex AI: This notebook deploys a PyTorch image classification model on Agent Platform using prebuilt PyTorch serving images.
What's next
- Tutorial: Use Agent Platform to train a PyTorch image
classification model in one of Agent Platform's prebuilt container environments
by using the Google Cloud console.
To follow step-by-step guidance for this task directly in the Google Cloud console, click Guide me: