Position: GCP Big Data Engineer
Locations: Bangalore, Pune, Mumbai, Gurgaon (Hybrid)
Experience: 4–15 years
Budget: ₹20–35 LPA
Notice Period: Immediate to 30 days
Job Description:
We are looking for an experienced GCP Big Data Engineer with strong expertise in building and managing large-scale data processing systems on Google Cloud Platform.
The ideal candidate should have hands-on experience with BigQuery, Dataflow, Dataproc, Pub/Sub, Cloud Storage, and related GCP data services.
Key Responsibilities:
- Design, develop, and maintain scalable data pipelines and architectures on GCP.
- Optimize and manage data ingestion, transformation, and storage solutions.
- Collaborate with data scientists and analysts to ensure efficient data access.
- Implement data governance, security, and performance optimization best practices.
- Work with cross-functional teams to deliver end-to-end data solutions.
Technical Skills:
- Expertise in GCP Data Services (BigQuery, Dataflow, Dataproc, Pub/Sub, Cloud Composer, etc.)
- Proficiency in Python, SQL, and Spark
- Experience with ETL design, data modeling, and data warehousing
- Knowledge of CI/CD pipelines and version control systems (Git)
- Understanding of data security, compliance, and access control