Data Engineer - Enterprise Data Warehouse
Michael Page
- Hong Kong Island, Hong Kong
- Permanent
- Full-time
- Supporting the business plan through the creation and meticulous maintenance of optimal data models and pipelines
- Enhance efficiencies for different teams, promoting automated processes which replace manual labour, and enhancing data delivery systems. This pivotal transformation permits team members to channel their focus on tasks that directly add value
- Liaise with business and other stakeholders to elicit and clarify data requirements, including assessing implementation feasibility and delivery
- Perform tasks in accordance with quality procedures or architecture requirements as applicable.
- Build and maintain advanced analytics platforms, tools, and products
- Build and maintain data pipelines to capture, clean, and democratise information
- Lead, design and build tactical data architectures and databases to support internal product development
- Lead, design and build highly complex solutions in collaboration with data scientist, full stack developers and subject matter experts
- Produce high quality code using internal guidelines, frameworks, documentation to scale and maintain the internal technology stack
- Bachelor's degree in a relevant IT, Software, Computer Science or Engineering discipline.
- Agile Certification desirable
- At least 2+ years of experience as a Data Engineer
- Experience building and optimising 'big data' data pipelines, ETLs, architectures and data sets
- Experience in managing, maintaining, and improving quality in data pipelines
- Strong knowledge in building processes supporting data transformation, data structures, metadata, dependency, and workload management
- Knowledge in manipulating, processing, and extracting value from large, disconnected datasets
- Strong experience in SQL applied on different platforms
- Experience in Python, Pyspark, RedPanda, Dagster, Kubernetes (Developer), Dockers
- Demonstrated experience in using big data & cloud development such as databricks, datafactory, Kafka, Spark, Hadoop and microservices