Know ATS Score
CV/Résumé Score
  • Expertini Resume Scoring: Our Semantic Matching Algorithm evaluates your CV/Résumé before you apply for this job role: Data Engineer ETL/PySpark.
India Jobs Expertini

Urgent! Data Engineer - ETL/PySpark Job Opening In Pune – Now Hiring Confidential

Data Engineer ETL/PySpark



Job description

<p></p><p><b>Description :</b><br/><br/></p><p>Experience : 6 to 10 Years<br/><br/>Location : Any Xebia Location (Bangalore, Chennai, Pune, Bhopal, Hyderabad, Jaipur, Gurgaon)<br/><br/>Contract Type : TPC 6 Months (Extendable)<br/><br/>Interview Platform : Flow Career (Ensure weekend availability for interviews)<br/><br/><b>About the Role :</b><br/><br/></p><p>We are seeking an experienced Data Engineer with deep expertise in Databricks, Python, SQL, and Postgres.

The ideal candidate will have practical experience working with Vector Databases (pgvector, Qdrant, Pinecone, etc.) and exposure to Generative AI use cases such as Retrieval-Augmented Generation (RAG) pipelines and embedding-based search.<br/><br/>You will collaborate with cross-functional teams to design, build, and optimize scalable data pipelines and contribute to innovative AI-driven data solutions on Azure.<br/><br/><b>Key Responsibilities :</b><br/><br/></p><p>- Design and develop robust ETL/ELT pipelines using Databricks, PySpark, and Azure Data Factory.<br/><br/>- Implement scalable data processing workflows integrating Delta Lake and Azure Data Lake Storage Gen2.<br/><br/>- Build, optimize, and maintain Postgres databases (schema design, indexing, performance tuning).<br/><br/>- Develop and manage RAG pipelines and vector search capabilities using pgvector, Qdrant, or Pinecone.<br/><br/>- Work on embedding generation and integration with Azure OpenAI or equivalent Generative AI services.<br/><br/>- Write and optimize complex SQL queries for analytics and data modeling.<br/><br/>- Automate deployments and manage version control using CI/CD and Git-based workflows.<br/><br/>- Collaborate with data scientists, ML engineers, and business teams to deliver high-quality data products.<br/><br/><b>Required Skills :</b><br/><br/></p><p>- Python (PySpark, pandas, API integration)<br/><br/>- Databricks (notebooks, Delta Lake, workflows)<br/><br/>- SQL (data modeling, optimization, analytics)<br/><br/>- Postgres (schema design, indexing, query tuning)<br/><br/>- Azure Services Data Factory, Data Lake Storage Gen2, Synapse Analytics<br/><br/>- Vector Databases pgvector, Qdrant, Pinecone (any)<br/><br/>- Generative AI Exposure RAG pipelines, embeddings, integration with Azure OpenAI<br/><br/>- Version Control & DevOps Git, CI/CD pipelines<br/><br/><b>Additional Information :</b><br/><br/></p><p>- Immediate joiners preferred.<br/><br/>- Interviews will be conducted via Flow Career please ensure weekend availability for interviews.</p><br/><p></p> (ref:hirist.tech)


Required Skill Profession

Computer Occupations



Your Complete Job Search Toolkit

✨ Smart • Intelligent • Private • Secure

Start Using Our Tools

Join thousands of professionals who've advanced their careers with our platform

Rate or Report This Job
If you feel this job is inaccurate or spam kindly report to us using below form.
Please Note: This is NOT a job application form.


    Unlock Your Data Engineer Potential: Insight & Career Growth Guide