- Design, build, and maintain scalable and efficient data pipelines and ETL processes.
- Develop and optimize Clickhouse databases for high-performance analytics.
- Create RESTful APIs using FastAPI to expose data services.
- Work with Kubernetes for container orchestration and deployment of data services.
- Write complex SQL queries to extract, transform, and analyze data from PostgreSQL and Clickhouse.
- Collaborate with data scientists, analysts, and backend teams to support data needs and ensure data quality.
- Monitor, troubleshoot, and improve performance of data infrastructure.
- Strong experience in Clickhouse - data modeling, query optimization, performance tuning.
- Expertise in SQL - including complex joins, window functions, and optimization.
- Proficient in Python, especially for data processing (Pandas, NumPy) and scripting.
- Experience with FastAPI for creating lightweight APIs and microservices.
- Hands-on experience with PostgreSQL - schema design, indexing, and performance.
- Solid knowledge of Kubernetes managing containers, deployments, and scaling.
- Understanding of software engineering best practices (CI/CD, version control, testing).
- Experience with cloud platforms like AWS, GCP, or Azure.
- Knowledge of data warehousing and distributed data systems.
- Familiarity with Docker, Helm, and monitoring tools like Prometheus/Grafana.
Skills Required
Python, Aws, Gcp, Postgresql, Kubernetes