Experience : 5 to 8 years
Job description: Python API + Data Engineer
We are looking for a self-motivated Python Data Engineer with strong skills in backend/API development and large-scale data processing , who can work independently and thrive in a fast-paced, agile environment.
Key Responsibilities - Design, develop, and optimize data pipelines using PySpark, AWS Glue, and EMR .
- Build and maintain Python backend/API services (FastAPI, async preferred).
- Support ML/MLOps workflows leveraging AWS SageMaker and other cloud-native tools.
- Write clean, testable, and scalable code with unit testing (pytest) and CI/CD practices.
- Work with AWS services (S3, Athena, Glue, EMR, SageMaker) for data engineering and model deployment.
- Collaborate cross-functionally to provide production support, troubleshooting, and performance optimization .
- Use modern DevOps tools (Git, Docker, Kubernetes, Terraform) to automate deployments and ensure reliability.
- Communicate effectively with global teams and adapt quickly to changing requirements.
Required Skills & Experience - Strong expertise in Python programming .
- Hands-on experience with PySpark and large-scale data processing.
- Proven experience in Python API/backend development (FastAPI, REST/async).
- Proficiency in AWS cloud services : Glue, EMR, SageMaker, S3, Athena.
- Exposure to ML/MLOps workflows and production-level deployment.
- Solid knowledge of SQL/PostgreSQL and database query optimization.
- Familiarity with unit testing (pytest), Git, Docker, Kubernetes, Terraform .
- Strong problem-solving, production support, and communication skills.
Work Environment - Remote-friendly, with expectation of independent ownership and accountability.
- Timing: 2 PM - 11 PM IST ( 4 hours overlap with EST)