Job Description
<p><p><b>Description : </b></p><br/><p>Job Title : ETL/DataStage SQL Engineer</p><br/><p><b>Key Responsibilities : </b></p><br/><p>- Act as a critical member of the Colocation Team and Site Strategy, supporting both on-premises and cloud-based data platforms to ensure seamless data operations.</p><br/><p>- Design, develop, and maintain robust ETL workflows and data pipelines using DataStage and other ETL tools, ensuring high data quality, consistency, and reliability.</p><br/><p>- Support, monitor, and optimize on-prem Teradata environments while planning, migrating, and integrating data with modern cloud platforms such as Snowflake and Databricks.</p><br/><p>- Analyze, optimize, and troubleshoot complex SQL queries, ensuring optimal database performance and minimal latency in data processing.</p><br/><p>- Implement, maintain, and improve workflow orchestration using Airflow to automate data pipelines efficiently.</p><br/><p>- Integrate modern DevOps best practices into ETL development, including version control via GitHub, automated CI/CD pipelines, and leveraging GitHub Copilot for enhanced coding efficiency.</p><br/><p>- Collaborate closely with business analysts, data architects, and technical teams to design and deliver scalable, reliable, and reusable data solutions that meet evolving business needs.</p><br/><p>- Explore and leverage AI/ML frameworks and tools to automate repetitive tasks, optimize processes, and enhance data engineering efficiency.</p><br/><p>- Document technical workflows, ETL processes, and best practices to ensure knowledge transfer and maintain operational continuity.</p><br/><p>- Participate actively in a hybrid work model, spending a minimum of three days per week onsite at HCL Hyderabad or Noida offices, while maintaining remote collaboration for flexibility.</p><br/><p><b>Required Skills & Experience : </b></p><br/><p>- Strong hands-on experience with UNIX/Linux environments, including shell scripting and command-line tools.</p><br/><p>- Expertise in ETL design and development, with hands-on experience in DataStage or equivalent ETL platforms.</p><br/><p>- Solid understanding of Data Warehousing concepts, dimensional modeling, and SQL query optimization.</p><br/><p>- Proficiency in Snowflake, Databricks, and Teradata environments, with experience in data migration and integration.</p><br/><p>- Advanced programming skills in Python and PySpark for data manipulation, automation, and pipeline optimization.</p><br/><p>- Experience in workflow orchestration using Airflow, including scheduling, monitoring, and error handling.</p><br/><p>- Familiarity with GitHub Copilot, modern coding practices, and DevOps workflows to improve development efficiency.</p><br/><p>- Awareness of AI/ML concepts and practical ability to leverage automation and data-driven solutions.</p><br/><p>- Strong analytical, problem-solving, and communication skills, with the ability to work collaboratively in a fast-paced environment.</p><br/><p>- (Nice to have) Prior experience in the Healthcare domain, including HIPAA-compliant data handling and reporting.</p><br/></p> (ref:hirist.tech)