Job Title:  Data Integration Engineer 
Location:  (Add Location or mention “Remote/Hybrid”) 
 Company:  LumenData 
Role Overview  
We are seeking a Data Integration Engineer  with expertise in building and orchestrating data pipelines using Apache Airflow  to integrate data from diverse sources into Snowflake .
 The ideal candidate will have strong experience with JDBC  and API-based integrations (REST/JSON) , hands-on proficiency with Postman , and solid skills in SQL encryption/decryption , Python development , and pipeline monitoring .
Key Responsibilities  
- Design, develop, and maintain Airflow DAGs  to orchestrate end-to-end data workflows.
 
 
- Integrate structured and unstructured data from multiple systems into Snowflake  using JDBC connectors , APIs , and flat-file ingestion .
 
 
- Use Postman  and other tools to test, validate, and automate API integrations.
 
 
- Implement SQL encryption/decryption  techniques to protect sensitive datasets.
 
 
- Perform data quality checks , including row-level validation, hash-based reconciliation, and exception handling.
 
 
- Develop transformation logic using Python  and SQL , ensuring performance, scalability, and maintainability.
 
 
- Implement detailed logging , monitoring , and alerting  to ensure pipeline reliability and compliance.
 
 
- Collaborate with stakeholders to understand requirements and deliver scalable, production-ready solutions .
 
 
Required Skills & Experience  
- Strong proficiency in Apache Airflow  for workflow orchestration.
 
 
- Hands-on experience with Snowflake  as a data warehouse.
 
 
- Proven ability to integrate data via JDBC drivers , REST APIs , and Postman-tested endpoints .
 
 
- Advanced knowledge of SQL , including encryption/decryption techniques .
 
 
- Strong programming skills in Python  for ETL/ELT development .
 
 
- Experience with logging , monitoring , and data observability  practices.