We are seeking skilled and motivated Spark & Databricks Developers to join our dynamic team for a long-term project.
The ideal candidate will have strong hands-on experience in Apache Spark, Databricks, and GitHub-based development workflows.
Key Responsibilities:
- Design, develop, and optimize big data pipelines using Apache Spark.
- Build and maintain scalable data solutions on Databricks.
- Collaborate with cross-functional teams for data integration and transformation.
- Manage version control and code collaboration using GitHub.
- Ensure data quality, performance tuning, and job optimization.
- Participate in code reviews, testing, and documentation activities.
Must-Have Skills:
- 5–8 years of experience in Data Engineering or related roles
- Strong hands-on expertise in Apache Spark (Batch & Streaming)
- Proficiency in Databricks for developing and managing data workflows
- Experience with GitHub (version control, pull requests, branching strategies)
- Good understanding of Data Lake and Data Warehouse architectures
- Strong SQL and Python scripting skills
- In-depth knowledge of Python programming
Good-to-Have Skills:
- Experience with Azure Data Lake, AWS S3, or GCP BigQuery
- Familiarity with Delta Lake and Databricks SQL
- Exposure to CI/CD pipelines and DevOps practices
- Experience with ETL tools or data modeling
- Understanding of data governance, security, and performance tuning best practices