Job Role: -AWS Data Engineer
Job Location: -Pune/Hyderabad/Chennai/Mysuru/Bhubaneswar/Mangalore/Trivandrum/Chandigarh/Jaipur/Nagpur/Indore/Gurgaon
Experience: - 7+ Years
Job Roles & Responsibilities: -
- Design, develop, and maintain data pipelines and assets on AWS.
- Optimize and refactor legacy PySpark / Spark SQL code for performance and maintainability.
- Implement unit testing / TDD to ensure robust, bug-free code.
- Debug and resolve complex performance, concurrency, and logic issues.
- Manage code versioning and repositories (Git, JFrog Artifactory).
- Leverage AWS services (S3, EC2, Lambda, Redshift, CloudFormation) for scalable data solutions.
Job Skills & Requirements: -
- 7+ years hands-on experience in Python, PySpark, Boto3, and related frameworks/libraries.
- Proven expertise in Spark SQL & PySpark optimization.
- Strong knowledge of AWS architecture (S3, EC2, Lambda, Redshift, CloudFormation).
- Experience in code refactorization for clean, maintainable solutions.
- Familiarity with Git, JFrog Artifactory, and modern CI/CD practices.
- Strong debugging and problem-solving skills.
- Solid understanding of unit testing and TDD methodologies.