Role Overview   
We are seeking a skilled and motivated Data Engineer with 5–8 years of experience in building scalable data pipelines using Python, PySpark, and AWS services.
The ideal candidate will have hands-on expertise in big data processing, orchestration using AWS Step Functions, and serverless computing with AWS Lambda.
Familiarity with DynamoDB and deployment of ETL programs in AWS is essential.
Key Responsibilities   
- Design, develop, and maintain robust data pipelines using Python and PySpark  
- Handle large-scale data processing and transformation using AWS services  
- Implement orchestration workflows using AWS Step Functions  
- Develop and manage serverless components using AWS Lambda  
- Deploy and monitor ETL programs in AWS environments  
- Configure and optimize DynamoDB for data storage and retrieval  
- Collaborate with cross-functional teams to understand data requirements and deliver scalable solutions  
- Ensure data quality, integrity, and security across all stages of the pipeline  
Required Skills & Qualifications   
-  5–8 years of experience in data engineering or related field  
-  Strong proficiency in Python and PySpark  
-  Solid understanding of AWS services including S3, Lambda, Step Functions, Glue, and DynamoDB  Experience deploying and managing ETL workflows in AWS  
-  Familiarity with NoSQL databases, especially DynamoDB  
-  Knowledge of CI/CD practices and infrastructure-as-code tools (e.g., CloudFormation, Terraform) is a plus  
-  Excellent problem-solving and communication skills  
-  Ability to work independently in a remote setup 
What We Offer 
- Fully remote work environment  
- Opportunity to work on cutting-edge data engineering projects  
- Collaborative and inclusive team culture  
- Competitive compensation and benefits