- Data Science
- Shanghai - China
About the role
As a Data Engineer you will part of a pioneering team building the analytics platform for the Greater China region and will be responsible for developing, industrialising, and optimising Dyson's big data platform running on AWS. You will ingest new data sources, write data pipelines as code, and transform, enrich and publish data using the most efficient methods.
Working with data from across Dyson’s global data estate, you will understand the best way to serve up data at scale to a global audience of analysts. You will work closely with data architects, data scientists and data product managers on the team to ensure that we are building an integrated, performant solutions.
Ideally you will have a Software Engineering mind-set, be able to leverage CI/CD and apply critical thinking to the work you undertake. The role would suit candidates looking to make the move from working with traditional big data stacks such as Spark and Hadoop to using cloud native technologies (DataFlow, Redshift, Docker/Kubernetes, Pub/Sub, Cloud Functions). Candidates who also have strong software development skills and wishing to make the leap to working with Data at scale will also be considered.
Designing and building end to end Data Engineering solutions on the AWS Platform (Redshift)
Being a proactive member of DevOps / Agile scrum driven team; always looking for ways to tune and optimise all aspects of work delivered on the platform
Aligning work to both core development standards and architectural principles
Person specification / Core Competencies:
Resilient and comfortable with high pace change.
Detailed understanding of the AWS data landscape
Strong programming skills in languages such as Python/Java/Scala including building, testing and releasing code into production
Strong SQL skills and experience working with relational/columnar databases (e.g. Redshift, SQL Server, Postgres, Oracle, Presto, Hive, BigQuery)
Knowledge of data modelling techniques and integration patterns
Practical experience writing data analytic pipelines
Experience integrating/interfacing with REST APIs / Web Services
Experience handling data securely
Experience with DevOps software delivery and CI/CD processes
A willingness to learn and find solutions to complex problems
Familiarity with commercial data sources used in the Greater China Region
We are following the government guidelines regarding COVID19. At this time all interviews will be conducted via video or telephone. We’re taking these precautionary measures to protect both our employee and candidate wellbeing. Our Talent Acquisition team will work with you and provide further information as appropriate.