You will be part of the data delivery team and will have the opportunity to develop a deep understanding of the domain/function.
You will design and drive the work plan for the optimization/automation and standardization of the processes incorporating best practices to achieve efficiency gains.
You will run data engineering pipelines, link raw client data with data model, conduct data assessment, perform data quality checks, and transform data using ETL tools.
You will perform data transformations, modeling, and validation activities, as well as configure applications to the client context. You will also develop scripts to validate, transform, and load raw data using programming languages such as Python and / or PySpark.
In this role, you will determine database structural requirements by analyzing client operations, applications, and programming.
You will develop cross-site relationships to enhance idea generation, and manage stakeholders.
Lastly, you will collaborate with the team to support ongoing business processes by delivering high-quality end products on-time and perform quality checks wherever required.
Bachelor’s degree in Engineering or Computer Science; Master’s degree is a plus
3+ years of professional work experience with a reputed analytics firm
Expertise in handling large amount of data through Python or PySpark
Conduct data assessment, perform data quality checks and transform data using SQL and ETL tools
Experience of deploying ETL / data pipelines and workflows in cloud technologies and architecture such as Azure and Amazon Web Services will be valued
Comfort with data modelling principles (e.g. database structure, entity relationships, UID etc.) and software development principles (e.g. modularization, testing, refactoring, etc.)
A thoughtful and comfortable communicator (verbal and written) with the ability to facilitate discussions and conduct training
Strong problem-solving, requirement gathering, and leading.