Minimum 7 years of experience in data architecture, database design, and data engineering
Strong expertise in Google Cloud Platform (GCP), including Dataplex, BigQuery, Dataflow (Apache Beam), and other GCP-native data tools
Experience with Apache-based data pipelining tools (e.g., Beam, Airflow, Kafka, Spark) for scalable data processing and orchestration
Expertise in data modeling (conceptual, logical, and physical) for structured and semi-structured data
Solid knowledge of ETL/ELT processes, data transformation, and integration techniques in cloud environments
Strong understanding of data governance, metadata management, and data security within GCP
Excellent communication and presentation skills — ability to engage with clients, present technical solutions, and translate complex data concepts into clear business insights
Proven ability to collaborate with cross-functional teams, including engineers, analysts, and business stakeholders, to ensure data integrity and accessibility
Your responsibilities
Design, implement, and optimize data solutions in GCP for cloud-native data architecture
Leverage Dataplex for data governance, cataloging, and lifecycle management
Build and manage Apache-based data pipelines (Beam, Airflow, Kafka) to ensure efficient and scalable data processing
Develop and maintain ETL/ELT workflows, focusing on cloud-based and streaming architectures
Define and enforce data governance and compliance best practices across platforms
Collaborate with engineering and analytics teams to ensure data availability, reliability, and performance
Provide expertise in big data processing and enterprise-scale analytics solutions on GCP
Stay updated with emerging data technologies and recommend improvements to existing data architecture