Experience : 5 to 8 years 
Job description:  Python API + Data Engineer 
We are looking for a self-motivated Python Data Engineer  with strong skills in backend/API development and large-scale data processing , who can work independently and thrive in a fast-paced, agile environment.
Key Responsibilities - Design, develop, and optimize data pipelines  using PySpark, AWS Glue, and EMR .
 
 
- Build and maintain Python backend/API services  (FastAPI, async preferred).
 
 
- Support ML/MLOps workflows  leveraging AWS SageMaker  and other cloud-native tools.
 
 
- Write clean, testable, and scalable code with unit testing (pytest)  and CI/CD practices.
 
 
- Work with AWS services  (S3, Athena, Glue, EMR, SageMaker) for data engineering and model deployment.
 
 
- Collaborate cross-functionally to provide production support, troubleshooting, and performance optimization .
 
 
- Use modern DevOps tools  (Git, Docker, Kubernetes, Terraform) to automate deployments and ensure reliability.
 
 
- Communicate effectively with global teams and adapt quickly to changing requirements.
 
 
Required Skills & Experience- Strong expertise in Python programming .
 
 
- Hands-on experience with PySpark  and large-scale data processing.
 
 
- Proven experience in Python API/backend development  (FastAPI, REST/async).
 
 
- Proficiency in AWS cloud services : Glue, EMR, SageMaker, S3, Athena.
 
 
- Exposure to ML/MLOps workflows  and production-level deployment.
 
 
- Solid knowledge of SQL/PostgreSQL  and database query optimization.
 
 
- Familiarity with unit testing (pytest), Git, Docker, Kubernetes, Terraform .
 
 
- Strong problem-solving, production support, and communication skills.
 
 
Work Environment - Remote-friendly, with expectation of independent ownership and accountability.
 
 
- Timing: 2 PM - 11 PM IST ( 4 hours overlap with EST)