Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

What is Vertex AI?

When you run a machine learning experiment on a raw virtual machine (VM), you are responsible for provisioning the machine, managing Python environments, monitoring your training jobs, and remembering to shut everything down when you are done. This works fine for one or two experiments, but it becomes a real burden as your workload scales.

Vertex AI is Google Cloud’s fully managed, end-to-end ML platform. It abstracts away the infrastructure so you can focus entirely on your models and data. Instead of thinking about virtual machines and infrastructure setup, you work with higher-level concepts like jobs, datasets, experiments, and pipelines.

Core Capabilities of Vertex AI

Vertex AI is not a single tool. It is a suite of services that covers the entire ML workflow:

How Vertex AI Fits Into GCP?

Vertex AI does not work in isolation. It is deeply integrated with the rest of GCP:

Vertex AI vs Raw VMs on GCP

Both raw virtual machines (VMs) and Vertex AI can run the same experiments, but they differ in operational overhead and management:

FeatureRaw VMVertex AI Custom Training JobVertex AI Workbench
ManagementManualFully managedSemi-managed
Idle costsYes, until deletedNo, auto-terminatesYes, until stopped
AccessSSH / TerminalCLI / Console logsBrowser (Jupyter)
Setup complexityLowMediumLow

🔑 Key Takeaways

🚀 What Next?

In the next two hands-on sessions, we will run the same Gemma 1B fine-tuning experiment on Vertex AI in two different ways, focusing on the developer experience.

These sessions are designed to help you understand the workflow of working within Vertex AI.


📚 References & Further Reading