Role Overview 
We are seeking a skilled and motivated Data Engineer with 5–8 years of experience in building scalable data pipelines using Python, PySpark, and AWS services.
The ideal candidate will have hands-on expertise in big data processing, orchestration using AWS Step Functions, and serverless computing with AWS Lambda.
Familiarity with DynamoDB and deployment of ETL programs in AWS is essential.
Key Responsibilities 
- Design, develop, and maintain robust data pipelines using Python and PySpark 
 - Handle large-scale data processing and transformation using AWS services 
 - Implement orchestration workflows using AWS Step Functions 
 - Develop and manage serverless components using AWS Lambda 
 - Deploy and monitor ETL programs in AWS environments 
 - Configure and optimize DynamoDB for data storage and retrieval 
 - Collaborate with cross-functional teams to understand data requirements and deliver scalable solutions 
 - Ensure data quality, integrity, and security across all stages of the pipeline 
 
Required Skills & Qualifications 
-  5–8 years of experience in data engineering or related field 
 -  Strong proficiency in Python and PySpark 
 -  Solid understanding of AWS services including S3, Lambda, Step Functions, Glue, and DynamoDB  Experience deploying and managing ETL workflows in AWS 
 -  Familiarity with NoSQL databases, especially DynamoDB 
 -  Knowledge of CI/CD practices and infrastructure-as-code tools (e.g., CloudFormation, Terraform) is a plus 
 -  Excellent problem-solving and communication skills 
 -  Ability to work independently in a remote setup
 
What We Offer
- Fully remote work environment 
 - Opportunity to work on cutting-edge data engineering projects 
 - Collaborative and inclusive team culture 
 - Competitive compensation and benefits