Job description
 
                         Ready to shape the future of work?
 At Genpact, we don’t just adapt to change—we drive it.
AI and digital innovation are redefining industries, and we’re leading the charge.
Genpact’s AI Gigafactory, our industry-first accelerator, is an example of how we’re scaling advanced technology solutions to help global enterprises work smarter, grow faster, and transform at scale.
From large-scale models to agentic AI, our breakthrough solutions tackle companies’ most complex challenges.
 
 If you thrive in a fast-moving, tech-driven environment, love solving real-world problems, and want to be part of a team that’s shaping the future, this is your moment.
 
 Genpact (NYSE: G) is anadvanced technology services and solutions company that deliverslastingvalue for leading enterprisesglobally.Through ourdeep business knowledge, operational excellence, and cutting-edge solutions – we help companies across industries get ahead and stay ahead.Powered by curiosity, courage, and innovation,our teamsimplementdata, technology, and AItocreate tomorrow, today.Get to know us atgenpact.comand onLinkedIn,X,YouTube, andFacebook.
 Inviting applications for the role of Lead Consultant-Data Engineer, AWS+Python, Spark, Kafka for ETL! 
 Responsibilities 
  Develop, deploy, and manage ETL pipelines using AWS services, Python, Spark, and Kafka.
    Integrate structured and unstructured data from various data sources into data lakes and data warehouses.
    Design and deploy scalable, highly available, and fault-tolerant AWS data processes using AWS data services (Glue, Lambda, Step, Redshift) 
    Monitor and optimize the performance of cloud resources to ensure efficient utilization and cost-effectiveness.
 
    Implement and maintain security measures to protect data and systems within the AWS environment, including IAM policies, security groups, and encryption mechanisms.
 
    Migrate the application data from legacy databases to Cloud based solutions (Redshift, DynamoDB, etc) for high availability with low cost 
    Develop application programs using Big Data technologies like Apache Hadoop, Apache Spark, etc with appropriate cloud-based services like Amazon AWS, etc.
    Build data pipelines by building ETL processes (Extract-Transform-Load) 
    Implement backup, disaster recovery, and business continuity strategies for cloud-based applications and data.
 
    Responsible for analysing business and functional requirements which involves a review of existing system configurations and operating methodologies as well as understanding evolving business needs 
    Analyse requirements/User stories at the business meetings and strategize the impact of requirements on different platforms/applications, convert the business requirements into technical requirements 
    Participating in design reviews to provide input on functional requirements, product designs, schedules and/or potential problems 
    Understand current application infrastructure and suggest Cloud based solutions which reduces operational cost, requires minimal maintenance but provides high availability with improved security 
    Perform unit testing on the modified software to ensure that the new functionality is working as expected while existing functionalities continue to work in the same way 
    Coordinate with release management, other supporting teams to deploy changes in production environment 
   Qualifications we seek in you! 
 Minimum Qualifications 
  Experience in designing, implementing data pipelines, build data applications, data migration on AWS 
    Strong experience of implementing data lake using AWS services like Glue, Lambda, Step, Redshift 
    Experience of Databricks will be added advantage 
    Strong experience in Python and SQL 
     Proven expertise in AWS services such as S3, Lambda, Glue, EMR, and Redshift.
     Advanced programming skills in Python for data processing and automation.
     Hands-on experience with Apache Spark for large-scale data processing.
     Experience with Apache Kafka for real-time data streaming and event processing.
    Proficiency in SQL for data querying and transformation.
    Strong understanding of security principles and best practices for cloud-based environments.
 
    Experience with monitoring tools and implementing proactive measures to ensure system availability and performance.
 
    Excellent problem-solving skills and ability to troubleshoot complex issues in a distributed, cloud-based environment.
 
    Strong communication and collaboration skills to work effectively with cross-functional teams.
 
   Preferred Qualifications/ Skills 
  Master’s Degree-Computer Science, Electronics, Electrical.
    AWS Data Engineering & Cloud certifications, Databricks certifications 
    Experience with multiple data integration technologies and cloud platforms 
    Knowledge of Change & Incident Management process 
   Why join Genpact?
  Be a transformation leader – Work at the cutting edge of AI, automation, and digital innovation  
    Make an impact – Drive change for global enterprises and solve business challenges that matter  
    Accelerate your career – Get hands-on experience, mentorship, and continuous learning opportunities  
    Work with the best – Join 140,000+ bold thinkers and problem-solvers who push boundaries every day  
    Thrive in a values-driven culture – Our courage, curiosity, and incisiveness - built on a foundation of integrity and inclusion - allow your ideas to fuel progress  
   Come join the tech shapers and growth makers at Genpact and take your career in the only direction that matters: Up.  
 Let’s build tomorrow together.
 
 
                    
                    Required Skill Profession
 
                     
                    
                    Other General