Job Role: -AWS Data Engineer
 
 Job Location: - Pune/Hyderabad/Chennai/Mysuru/Bhubaneswar/Mangalore/Trivandrum/Chandigarh/Jaipur/Nagpur/Indore/Gurgaon
 
 Experience: - 7+ Years
 
 Job Roles & Responsibilities: -
 
  - Design, build, and optimize data pipelines and scalable data assets.
  - Develop and maintain high-performance code using PySpark/Python with best practices.
  - Perform code optimization using Spark SQL and PySpark for efficiency.
  - Implement code refactoring to modernize legacy codebases for readability and maintainability.
  - Write and maintain unit tests (TDD) to ensure code reliability and reduce bugs.
  - Debug and resolve complex issues including performance, concurrency, and logic flaws.
  - Work with AWS services (S3, EC2, Lambda, Redshift, CloudFormation) to architect and deliver solutions.
  - Manage version control and artifact repositories with Git and JFrog Artifactory.
  
 
 Job Skills & Requirements: -
 
  - 7+ years of strong hands-on experience in PySpark, Python, Boto3, and Python frameworks/libraries.
  - Expertise in Spark SQL & PySpark optimization.
  - Solid understanding of AWS architecture and cloud-native solutions.
  - Proficiency in Git repositories & JFrog Artifactory for code versioning.
  - Strong experience in refactoring, debugging, and optimizing legacy and new code.
  - Knowledge of TDD/Unit testing to ensure robust solutions.