Leading life insurance and financial services companyabout the job.
- Create and implement a data strategy using Azure Data tools for advanced data processing and analytics.
- Lead the development of data pipelines with a focus on optimizing Databricks components and cluster performance.
- Set up data integration configurations for Extract, Transform, Load (ETL) processes aligned with target data architecture.
- Understand and configure enterprise data models, including CDM and departmental data marts, using hive Metastore and Unity Catalog.
- Design and maintain robust ETL/ELT processes using Azure Data Factory and Azure Databricks for seamless data integration.
- Enhance data platform resilience through access control frameworks and best practice education for users.
- Implement data governance practices using Databricks Unity Catalog and Azure Purview to ensure compliance and data quality.
- Continuously monitor and optimize the cost and performance of Databricks environments, including data partitioning and indexing.
- Collaborate with cross-functional teams to understand data needs and communicate technical specifications effectively.
- Implement security best practices for data management and ensure compliance with data privacy regulations and internal policies.
skills & experiences required.
- University/College graduate in Computing, IT, or a related discipline.
- Minimum 5 years of relevant experience, with at least 3 years of hands-on experience using Databricks.
- In-depth knowledge of Azure Databricks for data engineering, including proficiency in Apache Spark, PySpark, and Delta Lake.
- Familiarity with Databricks components such as Workspace, Run-time, Clusters, workflows, DLT, functions, hive Metastore, SQL Warehouse, Delta sharing, and Unity Catalog.
- Knowledge of the insurance industry is preferred.
- Experience in ETL/ELT and data integration with an understanding of enterprise data models like CDM and departmental data marts.
- Proficiency in using Azure Data Factory (ADF) to build complex data pipelines and integrate data from various sources.
- Strong experience in SQL, Python, and PySpark for building and optimizing data workflows.
- Knowledge of Jenkins DevOps for implementing CI/CD pipelines and managing code repositories.
- Familiarity with Azure Purview, Azure Key Vault, Azure Active Directory, and RBAC for managing security and compliance in data platforms.
Please feel free to drop me a message or share your updated resume to đź“©Cheryl.lau@randstad.com.hk