- Detailed Job Description Skill Set.
- Author transformations, optimizations, and manipulations of datasets to meet the needs of the use case and its outcomes.
- Monitor and curate the most important datasets in one or multiple projects, adhering to the central Program Team's guidelines.
- Contribute back to Ontology according to defined contribution process.
- Perform modeling, machine learning, classification, and regressions to enrich and visualize data in service of decision-making workflows.
- Understand the concepts of data warehousing, including ETL (Extract, Transform, Load) processes and data modeling is essential.
- Expertise on PySpark, Spark SQL, Data frames, SQL Server etc.
Mandatory Skills.
- PySpark/Spark SQL, Data frames, SQL Server, Bigdata, NumPy, Pandas.
- Machine Learning, Data Science.
- Monitor and curate the most important datasets in one or multiple projects, adhering to the central Program Team's guidelines.
- Understand the concepts of data warehousing, including ETL (Extract, Transform, Load) processes and data modeling is essential.
Skills Required
Sql Server, Bigdata, Numpy, Pandas, Sql