Principal GCP Data Architect | Location: Remote
• 16–20 years of overall IT experience.
• 5–6 years of recent, hands-on experience with GCP.
• Proven experience working on Google Data projects.
• Valid and recent GCP certification.
• Prior involvement in Google-funded projects.
• Must possess a Data Readiness Placement (DRP) ID with a verified score of 50 or above.
Job Description
We are seeking a hands-on
• *Principal GCP Data Architect
to lead the design, and implementation of large-scale data ecosystems.
The ideal candidate is an architect who has led massive cloud migrations, built self-serve data platforms from the ground up, and maintains a rigorous focus on data governance and security.
Technical Requirements
• Cloud Expertise:
Expert-level mastery of the GCP Data Stack:
• Storage & Warehouse:
BigQuery (including BigLake and Omni), Google Cloud Storage.
• Processing:
Dataflow (Apache Beam), Dataproc (Spark/Hadoop), Cloud Composer (Airflow).
• Messaging:
Pub/Sub and Confluent/Kafka integration.
• Analytics & AI:
Looker, Vertex AI, and BigQuery ML.
• Certifications:
Must hold an active
• *GCP Professional Data Engineer
certification.
• DRP ID Performance:
A
• *Data Readiness Placement (DRP) ID
with a verified score of
• *50+
is highly preferable, demonstrating a high level of technical proficiency and architectural maturity.
• Modern Data Stack:
Deep experience with dbt, Airflow, and containerization (GKE/Kubernetes).
Experience & Qualifications
• Years of Experience:
18–20 years in Data Engineering, Data Warehousing, and Business Intelligence, with at least 6+ years focused specifically on GCP.
• Migration Track Record:
Proven experience leading at least two enterprise-scale migrations (PB-scale) to the cloud.
• Leadership:
Demonstrated experience leading large, cross-functional engineering teams in an Agile/DevOps environment.
• Education:
Bachelor’s or Master’s degree in Computer Science, Information Systems, or a related technical field.
Roles & Responsibilities
• Strategic Architecture:
Define the long-term roadmap for data platforms on GCP, ensuring alignment with global business goals and "North Star" data strategies.
• Self-Serve Platform Design:
Architect and oversee the build-out of modern, self-serve data architectures that empower decentralized teams while maintaining central standards.
• Large-Scale Migrations:
Lead complex end-to-end migration from AWS to GCP.
• Data Governance & Security:
Design and enforce robust data security frameworks including IAM, VPC Service Controls, Data Masking/Encryption, and automated governance using
• *Dataplex
.
• Infrastructure as Code (IaC):
Drive automation using Terraform or Pulumi to ensure repeatable, scalable, and version-controlled infrastructure.
• Executive Stakeholder Management:
Act as a trusted advisor to the VP of Data, articulating the ROI of data initiatives and managing technical risk.