Experience  : 5 to 8 years  
  
Job description:   Python API + Data Engineer  
  
We are looking for a self-motivated Python Data Engineer   with strong skills in backend/API development and large-scale data processing  , who can work independently and thrive in a fast-paced, agile environment.
 
  
Key Responsibilities  - Design, develop, and optimize data pipelines   using PySpark, AWS Glue, and EMR  .
 
 
- Build and maintain Python backend/API services   (FastAPI, async preferred).
 
 
- Support ML/MLOps workflows   leveraging AWS SageMaker   and other cloud-native tools.
 
 
- Write clean, testable, and scalable code with unit testing (pytest)   and CI/CD practices.
 
 
- Work with AWS services   (S3, Athena, Glue, EMR, SageMaker) for data engineering and model deployment.
 
 
- Collaborate cross-functionally to provide production support, troubleshooting, and performance optimization  .
 
 
- Use modern DevOps tools   (Git, Docker, Kubernetes, Terraform) to automate deployments and ensure reliability.
 
 
- Communicate effectively with global teams and adapt quickly to changing requirements.
 
 
Required Skills & Experience- Strong expertise in Python programming  .
 
 
- Hands-on experience with PySpark   and large-scale data processing.
 
 
- Proven experience in Python API/backend development   (FastAPI, REST/async).
 
 
- Proficiency in AWS cloud services  : Glue, EMR, SageMaker, S3, Athena.
 
 
- Exposure to ML/MLOps workflows   and production-level deployment.
 
 
- Solid knowledge of SQL/PostgreSQL   and database query optimization.
 
 
- Familiarity with unit testing (pytest), Git, Docker, Kubernetes, Terraform  .
 
 
- Strong problem-solving, production support, and communication skills.
 
 
  
Work Environment  - Remote-friendly, with expectation of independent ownership and accountability.
 
 
- Timing: 2 PM - 11 PM IST ( 4 hours overlap with EST)